diff options
Diffstat (limited to 'src/personalization/feedback/local_llm_reward.py')
| -rw-r--r-- | src/personalization/feedback/local_llm_reward.py | 36 |
1 files changed, 32 insertions, 4 deletions
diff --git a/src/personalization/feedback/local_llm_reward.py b/src/personalization/feedback/local_llm_reward.py index 9837ff0..70bbeb8 100644 --- a/src/personalization/feedback/local_llm_reward.py +++ b/src/personalization/feedback/local_llm_reward.py @@ -307,11 +307,39 @@ class LocalLLMRewardClient: This is the main entry point for batch reward estimation. """ - return asyncio.run(self.judge_batch_async(samples)) - - def judge(self, sample: TurnSample) -> RewardResult: + try: + loop = asyncio.get_running_loop() + except RuntimeError: + loop = None + + if loop is not None: + # Already in an event loop - create a new thread to run the coroutine + import concurrent.futures + with concurrent.futures.ThreadPoolExecutor() as executor: + future = executor.submit(asyncio.run, self.judge_batch_async(samples)) + return future.result() + else: + return asyncio.run(self.judge_batch_async(samples)) + + async def judge(self, sample: TurnSample) -> RewardResult: + """Judge a single turn (async interface for compatibility with LLMRewardClient).""" + return await self.judge_async(sample) + + def judge_sync(self, sample: TurnSample) -> RewardResult: """Judge a single turn (sync wrapper).""" - return asyncio.run(self.judge_async(sample)) + try: + loop = asyncio.get_running_loop() + except RuntimeError: + loop = None + + if loop is not None: + # Already in an event loop - create a new thread to run the coroutine + import concurrent.futures + with concurrent.futures.ThreadPoolExecutor() as executor: + future = executor.submit(asyncio.run, self.judge_async(sample)) + return future.result() + else: + return asyncio.run(self.judge_async(sample)) # --- Convenience Functions --- |
