From f4cb192ebed9bdff951c3c5b4ea21bbdccfc70e0 Mon Sep 17 00:00:00 2001 From: Yufan Song <33971064+yufansong@users.noreply.github.com> Date: Mon, 10 Jun 2024 23:55:33 +0800 Subject: [PATCH] Fix llm key leaks bug (#2376) * fix bug * fix bug * add --- evaluation/EDA/run_infer.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/evaluation/EDA/run_infer.py b/evaluation/EDA/run_infer.py index 984d307e84..6c2aed78a1 100644 --- a/evaluation/EDA/run_infer.py +++ b/evaluation/EDA/run_infer.py @@ -65,7 +65,9 @@ AGENT_CLS_TO_INST_SUFFIX = { } -def process_instance(instance, agent_class, metadata, reset_logger: bool = True): +def process_instance( + instance, agent_class, metadata, openai_api_key, reset_logger: bool = True +): # Setup the logger properly, so you can run multi-processing to parallelize the evaluation eval_output_dir = metadata['eval_output_dir'] if reset_logger: @@ -107,7 +109,7 @@ def process_instance(instance, agent_class, metadata, reset_logger: bool = True) answerer_model=metadata['answerer_model'], guesser_model=None, num_turns=metadata['max_iterations'], - openai_api_key=metadata['openai_api'], + openai_api_key=openai_api_key, guesser_kargs=guesser_kargs, ) @@ -234,7 +236,6 @@ if __name__ == '__main__': 'data_split': args.data_split, 'answerer_model': args.answerer_model, 'agent_class': agent_class, - 'openai_api': args.OPENAI_API_KEY, 'model_name': model_name, 'max_iterations': max_iterations, 'eval_output_dir': eval_output_dir, @@ -317,6 +318,7 @@ if __name__ == '__main__': instance, agent_class, metadata, + args.OPENAI_API_KEY, reset_logger=bool(num_workers > 1), ) future.add_done_callback(update_progress)