diff options
Diffstat (limited to 'kg_rag/prompt_based_generation/GPT/run_true_false_generation.py')
| -rw-r--r-- | kg_rag/prompt_based_generation/GPT/run_true_false_generation.py | 33 |
1 files changed, 33 insertions, 0 deletions
diff --git a/kg_rag/prompt_based_generation/GPT/run_true_false_generation.py b/kg_rag/prompt_based_generation/GPT/run_true_false_generation.py new file mode 100644 index 0000000..0d248db --- /dev/null +++ b/kg_rag/prompt_based_generation/GPT/run_true_false_generation.py @@ -0,0 +1,33 @@ +from kg_rag.utility import * +import sys + + +CHAT_MODEL_ID = sys.argv[1] + +QUESTION_PATH = config_data["TRUE_FALSE_PATH"] +SYSTEM_PROMPT = system_prompts["TRUE_FALSE_QUESTION_PROMPT_BASED"] +SAVE_PATH = config_data["SAVE_RESULTS_PATH"] +TEMPERATURE = config_data["LLM_TEMPERATURE"] + +CHAT_DEPLOYMENT_ID = CHAT_MODEL_ID + +save_name = "_".join(CHAT_MODEL_ID.split("-"))+"_prompt_based_one_hop_true_false_binary_response.csv" + + +def main(): + start_time = time.time() + question_df = pd.read_csv(QUESTION_PATH) + answer_list = [] + for index, row in question_df.iterrows(): + question = "Question: "+ row["text"] + output = get_GPT_response(question, SYSTEM_PROMPT, CHAT_MODEL_ID, CHAT_DEPLOYMENT_ID, temperature=TEMPERATURE) + answer_list.append((row["text"], row["label"], output)) + answer_df = pd.DataFrame(answer_list, columns=["question", "label", "llm_answer"]) + answer_df.to_csv(os.path.join(SAVE_PATH, save_name), index=False, header=True) + print("Completed in {} min".format((time.time()-start_time)/60)) + + + +if __name__ == "__main__": + main() + |
