diff options
| author | YurenHao0426 <blackhao0426@gmail.com> | 2026-01-27 09:57:37 -0600 |
|---|---|---|
| committer | YurenHao0426 <blackhao0426@gmail.com> | 2026-01-27 09:57:37 -0600 |
| commit | dc801c07cf38b0c495686463e6ca6f871a64440e (patch) | |
| tree | 599f03114775921dbc472403c701f4a3a8ea188a /collaborativeagents/slurm/logs/run_collab_baselines_14355966.err | |
| parent | e43b3f8aa36c198b95c1e46bea2eaf3893b13dc3 (diff) | |
Add collaborativeagents module and update gitignore
- Add collaborativeagents subproject with adapters, agents, and evaluation modules
- Update .gitignore to exclude large binary files (.whl, .tar), wandb logs, and results
Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
Diffstat (limited to 'collaborativeagents/slurm/logs/run_collab_baselines_14355966.err')
| -rw-r--r-- | collaborativeagents/slurm/logs/run_collab_baselines_14355966.err | 14466 |
1 files changed, 14466 insertions, 0 deletions
diff --git a/collaborativeagents/slurm/logs/run_collab_baselines_14355966.err b/collaborativeagents/slurm/logs/run_collab_baselines_14355966.err new file mode 100644 index 0000000..6d3191e --- /dev/null +++ b/collaborativeagents/slurm/logs/run_collab_baselines_14355966.err @@ -0,0 +1,14466 @@ +2025-12-25 10:37:47,003 - INFO - Loaded dataset: math-500 +2025-12-25 10:37:47,143 - INFO - Loaded 100 profiles from ../data/complex_profiles_v2/profiles_100.jsonl +2025-12-25 10:37:47,144 - INFO - Running method: contextual +2025-12-25 10:37:51,592 - INFO - Profile 1/20 +2025-12-25 10:37:53,048 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,062 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,063 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,066 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,067 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,069 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,070 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,073 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,074 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,076 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,077 - INFO - Profile 2/20 +2025-12-25 10:37:53,077 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,080 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,081 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,083 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,084 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,087 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,088 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,090 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,091 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,094 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,094 - INFO - Profile 3/20 +2025-12-25 10:37:53,094 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,097 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,098 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,101 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,101 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,104 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,105 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,107 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,108 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,111 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,111 - INFO - Profile 4/20 +2025-12-25 10:37:53,112 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,114 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,115 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,118 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,118 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,121 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,122 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,124 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,125 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,128 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,128 - INFO - Profile 5/20 +2025-12-25 10:37:53,129 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,131 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,132 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,135 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,136 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,138 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,139 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,142 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,142 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,145 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,145 - INFO - Profile 6/20 +2025-12-25 10:37:53,146 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,149 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,149 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,152 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,153 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,156 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,156 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,159 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,160 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,162 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,163 - INFO - Profile 7/20 +2025-12-25 10:37:53,163 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,166 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,167 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,169 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,170 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,173 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,174 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,176 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,177 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,180 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,180 - INFO - Profile 8/20 +2025-12-25 10:37:53,181 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,184 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,184 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,187 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,188 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,190 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,191 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,194 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,195 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,198 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,198 - INFO - Profile 9/20 +2025-12-25 10:37:53,198 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,201 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,202 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,205 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,206 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,208 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,209 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,212 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,213 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,215 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,215 - INFO - Profile 10/20 +2025-12-25 10:37:53,216 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,219 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,220 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,222 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,223 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,226 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,227 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,230 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,231 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,233 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,233 - INFO - Profile 11/20 +2025-12-25 10:37:53,234 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,237 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,238 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,240 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,241 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,244 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,245 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,247 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,248 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,251 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,251 - INFO - Profile 12/20 +2025-12-25 10:37:53,252 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,255 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,255 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,258 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,259 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,262 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,263 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,265 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,266 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,269 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,269 - INFO - Profile 13/20 +2025-12-25 10:37:53,270 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,273 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,273 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,276 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,277 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,280 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,281 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,283 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,284 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,287 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,287 - INFO - Profile 14/20 +2025-12-25 10:37:53,288 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,291 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,291 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,294 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,295 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,298 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,299 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,301 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,302 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,305 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,305 - INFO - Profile 15/20 +2025-12-25 10:37:53,306 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,309 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,310 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,312 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,313 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,316 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,317 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,320 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,321 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,323 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,323 - INFO - Profile 16/20 +2025-12-25 10:37:53,324 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,327 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,328 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,331 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,332 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,334 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,335 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,338 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,339 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,341 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,342 - INFO - Profile 17/20 +2025-12-25 10:37:53,343 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,345 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,346 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,349 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,350 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,352 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,354 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,356 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,357 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,360 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,360 - INFO - Profile 18/20 +2025-12-25 10:37:53,361 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,364 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,365 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,367 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,368 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,371 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,372 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,375 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,376 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,378 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,379 - INFO - Profile 19/20 +2025-12-25 10:37:53,380 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,382 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,383 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,386 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,387 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,390 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,391 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,393 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,394 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,397 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,397 - INFO - Profile 20/20 +2025-12-25 10:37:53,398 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,401 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,402 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,405 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,406 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,408 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,409 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,412 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,413 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,416 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/contextual_adapter.py", line 93, in generate_response + response = self._agent.generate_collaborator_response(context) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,422 - INFO - Running method: reflection +2025-12-25 10:37:53,427 - INFO - Profile 1/20 +2025-12-25 10:37:53,428 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,431 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,432 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,435 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,435 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,438 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,439 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,441 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,442 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,445 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,445 - INFO - Profile 2/20 +2025-12-25 10:37:53,445 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,448 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,449 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,451 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,452 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,455 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,455 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,458 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,459 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,461 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,462 - INFO - Profile 3/20 +2025-12-25 10:37:53,462 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,465 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,466 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,468 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,469 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,472 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,472 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,475 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,476 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,478 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,478 - INFO - Profile 4/20 +2025-12-25 10:37:53,479 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,482 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,482 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,485 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,486 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,488 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,489 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,492 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,493 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,495 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,495 - INFO - Profile 5/20 +2025-12-25 10:37:53,496 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,499 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,499 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,502 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,503 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,505 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,506 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,509 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,509 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,513 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,513 - INFO - Profile 6/20 +2025-12-25 10:37:53,514 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,517 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,517 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,520 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,521 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,523 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,524 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,527 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,527 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,530 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,530 - INFO - Profile 7/20 +2025-12-25 10:37:53,531 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,534 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,534 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,537 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,538 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,540 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,541 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,544 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,544 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,547 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,547 - INFO - Profile 8/20 +2025-12-25 10:37:53,548 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,550 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,551 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,554 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,554 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,557 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,558 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,560 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,561 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,564 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,564 - INFO - Profile 9/20 +2025-12-25 10:37:53,565 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,567 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,568 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,571 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,571 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,574 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,575 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,577 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,578 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,581 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,581 - INFO - Profile 10/20 +2025-12-25 10:37:53,581 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,584 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,585 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,587 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,588 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,591 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,591 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,594 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,595 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,597 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,598 - INFO - Profile 11/20 +2025-12-25 10:37:53,598 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,601 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,602 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,604 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,605 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,608 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,608 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,611 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,612 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,614 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,614 - INFO - Profile 12/20 +2025-12-25 10:37:53,615 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,618 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,618 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,621 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,622 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,624 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,625 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,628 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,628 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,631 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,631 - INFO - Profile 13/20 +2025-12-25 10:37:53,632 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,635 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,635 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,638 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,639 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,641 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,642 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,645 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,645 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,648 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,648 - INFO - Profile 14/20 +2025-12-25 10:37:53,649 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,651 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,652 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,655 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,655 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,658 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,659 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,661 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,662 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,665 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,665 - INFO - Profile 15/20 +2025-12-25 10:37:53,666 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,668 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,669 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,672 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,672 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,675 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,676 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,678 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,679 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,682 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,682 - INFO - Profile 16/20 +2025-12-25 10:37:53,682 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,685 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,686 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,688 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,689 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,692 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,692 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,695 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,696 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,698 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,699 - INFO - Profile 17/20 +2025-12-25 10:37:53,699 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,702 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,703 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,705 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,706 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,709 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,709 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,712 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,713 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,715 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,715 - INFO - Profile 18/20 +2025-12-25 10:37:53,716 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,719 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,719 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,722 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,723 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,725 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,726 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,729 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,729 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,732 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,732 - INFO - Profile 19/20 +2025-12-25 10:37:53,733 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,735 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,736 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,739 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,739 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,742 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,743 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,746 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,746 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,749 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,749 - INFO - Profile 20/20 +2025-12-25 10:37:53,750 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,752 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,753 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,756 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,756 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,759 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,760 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,762 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,763 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,766 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 93, in generate_collaborator_response + response = self.completion(messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,785 - INFO - Running method: reflection_grpo +2025-12-25 10:37:53,791 - INFO - Profile 1/20 +2025-12-25 10:37:53,791 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,795 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,795 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,798 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,799 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,801 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,802 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,805 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,805 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,808 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,808 - INFO - Profile 2/20 +2025-12-25 10:37:53,809 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,812 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,812 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,815 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,816 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,818 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,819 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,822 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,823 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,825 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,825 - INFO - Profile 3/20 +2025-12-25 10:37:53,826 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,829 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,829 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,832 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,833 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,835 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,836 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,839 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,840 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,842 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,842 - INFO - Profile 4/20 +2025-12-25 10:37:53,843 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,846 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,846 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,849 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,850 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,853 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,853 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,856 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,857 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,859 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,859 - INFO - Profile 5/20 +2025-12-25 10:37:53,860 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,863 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,864 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,866 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,867 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,870 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,870 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,873 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,874 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,876 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,876 - INFO - Profile 6/20 +2025-12-25 10:37:53,877 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,880 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,881 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,883 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,884 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,887 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,887 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,890 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,891 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,893 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,893 - INFO - Profile 7/20 +2025-12-25 10:37:53,894 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,897 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,897 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,900 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,901 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,903 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,904 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,907 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,908 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,910 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,910 - INFO - Profile 8/20 +2025-12-25 10:37:53,911 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,914 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,914 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,917 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,918 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,920 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,921 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,924 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,925 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,927 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,927 - INFO - Profile 9/20 +2025-12-25 10:37:53,928 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,931 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,931 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,934 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,935 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,937 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,938 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,941 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,942 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,944 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,944 - INFO - Profile 10/20 +2025-12-25 10:37:53,945 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,948 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,949 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,951 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,952 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,955 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,955 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,958 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,959 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,961 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,961 - INFO - Profile 11/20 +2025-12-25 10:37:53,962 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,965 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,966 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,968 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,969 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,972 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,972 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,975 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,976 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,978 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,979 - INFO - Profile 12/20 +2025-12-25 10:37:53,979 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,982 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,983 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,985 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,986 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,989 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,989 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,992 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,993 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,995 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:53,995 - INFO - Profile 13/20 +2025-12-25 10:37:53,996 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:53,999 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,000 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,002 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,003 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,006 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,006 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,009 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,010 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,012 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,012 - INFO - Profile 14/20 +2025-12-25 10:37:54,013 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,016 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,017 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,019 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,020 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,023 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,023 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,026 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,027 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,029 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,030 - INFO - Profile 15/20 +2025-12-25 10:37:54,030 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,033 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,034 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,036 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,037 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,040 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,040 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,043 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,044 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,047 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,047 - INFO - Profile 16/20 +2025-12-25 10:37:54,047 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,050 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,051 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,053 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,054 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,057 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,057 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,060 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,061 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,064 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,064 - INFO - Profile 17/20 +2025-12-25 10:37:54,064 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,067 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,068 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,070 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,071 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,074 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,074 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,077 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,078 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,080 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,081 - INFO - Profile 18/20 +2025-12-25 10:37:54,081 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,084 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,085 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,087 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,088 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,091 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,091 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,094 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,095 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,097 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,098 - INFO - Profile 19/20 +2025-12-25 10:37:54,098 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,101 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,102 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,104 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,105 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,108 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,108 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,111 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,112 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,114 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,115 - INFO - Profile 20/20 +2025-12-25 10:37:54,115 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,118 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,119 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,121 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,122 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,125 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,125 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,128 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,129 - ERROR - Error in session: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' +2025-12-25 10:37:54,131 - ERROR - Full traceback: +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1274, in wrapper + result = original_function(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4080, in completion + raise exception_type( + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 1243, in completion + model, custom_llm_provider, dynamic_api_key, api_base = get_llm_provider( + ^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 455, in get_llm_provider + raise e + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/litellm_core_utils/get_llm_provider_logic.py", line 432, in get_llm_provider + raise litellm.exceptions.BadRequestError( # type: ignore +litellm.exceptions.BadRequestError: litellm.BadRequestError: LLM Provider NOT provided. Pass in the LLM provider you are trying to call. You passed model=meta-llama/Llama-3.3-70B-Instruct + Pass model as E.g. For 'Huggingface' inference endpoints pass in `completion(model='huggingface/starcoder',..)` Learn more: https://docs.litellm.ai/docs/providers + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4094, in completion_with_retries + import tenacity +ModuleNotFoundError: No module named 'tenacity' + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/scripts/run_experiments.py", line 201, in run_single_session + response = agent_adapter.generate_response(query, turns[:-1]) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/adapters/reflection_grpo_adapter.py", line 97, in generate_response + response = self._agent.generate_collaborator_response(self._conversation_history) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 88, in generate_collaborator_response + conversation = self.add_scaffolding_to_conversation(conversation) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 71, in add_scaffolding_to_conversation + scaffolding_response = self.completion(scaffolding_messages) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/collaborativeagents/collaborativeagents/agents/collaborator_agent.py", line 50, in completion + response = llm_completion(model=self.model_name, messages=messages, num_retries=self.num_retries, **self.kwargs).choices[0].message.content + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/utils.py", line 1385, in wrapper + return litellm.completion_with_retries(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/litellm/main.py", line 4096, in completion_with_retries + raise Exception( +Exception: tenacity import failed please run `pip install tenacity`. ErrorNo module named 'tenacity' + +2025-12-25 10:37:54,141 - INFO - Report saved to ../results/collab_baselines_20251225_103724/20251225_103746/report.md |
