summaryrefslogtreecommitdiff
path: root/src/personalization/feedback/local_llm_reward.py
diff options
context:
space:
mode:
Diffstat (limited to 'src/personalization/feedback/local_llm_reward.py')
-rw-r--r--src/personalization/feedback/local_llm_reward.py36
1 files changed, 32 insertions, 4 deletions
diff --git a/src/personalization/feedback/local_llm_reward.py b/src/personalization/feedback/local_llm_reward.py
index 9837ff0..70bbeb8 100644
--- a/src/personalization/feedback/local_llm_reward.py
+++ b/src/personalization/feedback/local_llm_reward.py
@@ -307,11 +307,39 @@ class LocalLLMRewardClient:
This is the main entry point for batch reward estimation.
"""
- return asyncio.run(self.judge_batch_async(samples))
-
- def judge(self, sample: TurnSample) -> RewardResult:
+ try:
+ loop = asyncio.get_running_loop()
+ except RuntimeError:
+ loop = None
+
+ if loop is not None:
+ # Already in an event loop - create a new thread to run the coroutine
+ import concurrent.futures
+ with concurrent.futures.ThreadPoolExecutor() as executor:
+ future = executor.submit(asyncio.run, self.judge_batch_async(samples))
+ return future.result()
+ else:
+ return asyncio.run(self.judge_batch_async(samples))
+
+ async def judge(self, sample: TurnSample) -> RewardResult:
+ """Judge a single turn (async interface for compatibility with LLMRewardClient)."""
+ return await self.judge_async(sample)
+
+ def judge_sync(self, sample: TurnSample) -> RewardResult:
"""Judge a single turn (sync wrapper)."""
- return asyncio.run(self.judge_async(sample))
+ try:
+ loop = asyncio.get_running_loop()
+ except RuntimeError:
+ loop = None
+
+ if loop is not None:
+ # Already in an event loop - create a new thread to run the coroutine
+ import concurrent.futures
+ with concurrent.futures.ThreadPoolExecutor() as executor:
+ future = executor.submit(asyncio.run, self.judge_async(sample))
+ return future.result()
+ else:
+ return asyncio.run(self.judge_async(sample))
# --- Convenience Functions ---