diff --git a/examples/babyagi.py b/examples/babyagi.py index d21630da0..a745c93ee 100644 --- a/examples/babyagi.py +++ b/examples/babyagi.py @@ -134,6 +134,8 @@ def one_cycle(objective: str, task_list, next_task_id: int): ) new_tasks = model(prompt) + new_tasks = create_tasks_fmt(new_tasks) + for task in new_tasks: next_task_id += 1 task_list.append({"task_id": next_task_id, "task_name": task}) @@ -143,6 +145,8 @@ def one_cycle(objective: str, task_list, next_task_id: int): ) prioritized_tasks = model(prompt) + prioritized_tasks = prioritize_tasks_fmt(prioritized_tasks) + return task, result, prioritized_tasks, next_task_id diff --git a/examples/meta_prompting.py b/examples/meta_prompting.py index dcfdad42a..b41843db7 100644 --- a/examples/meta_prompting.py +++ b/examples/meta_prompting.py @@ -22,7 +22,7 @@ def solve(question): Let's solve this problem by splitting it into steps. """ - complete = models.text_completion.openai(model_name) + complete = models.text_completion.openai(model_name, max_tokens=500) prompt = solve(question) answer = complete(prompt) @@ -43,12 +43,12 @@ def determine_goal(question): def solve(memory): """{{memory}}. Let's begin.""" - complete = models.text_completion.openai(model_name) + complete = models.text_completion.openai(model_name, max_tokens=500) prompt = determine_goal(question) answer = complete(prompt, stop_at=["."]) prompt = solve(prompt + answer) - answer = complete(prompt, stop_at=["."]) + answer = complete(prompt) completed = prompt + answer return completed @@ -76,14 +76,14 @@ def find_expert(question): @text.prompt def get_answer(question, expert, memory): """ - {{memory}} + {{memory}}". I am ready to ask my question. "{{expert}}" I say, {{question}} """ complete_expert = models.text_completion.openai(model_name) - complete_answer = models.text_completion.openai(model_name) + complete_answer = models.text_completion.openai(model_name, max_tokens=500) prompt = find_expert(question) expert = complete_expert(prompt, stop_at=['"']) @@ -111,7 +111,7 @@ def get_answer(expert, memory): """ model_expert = models.text_completion.openai(model_name) - model_answer = models.text_completion.openai(model_name) + model_answer = models.text_completion.openai(model_name, max_tokens=500) prompt = find_expert(question) expert = model_expert(prompt, stop_at=["\n", "."]) @@ -157,7 +157,9 @@ def run_example(model_fn, question, model_name): meaning_q = "What is the meaning of life?" run_example(split_into_steps, math_q, args.model) - run_example(split_into_steps, sat_q, args.model) + run_example( + split_into_steps, sat_q, args.model + ) # gpt>3.5 usually gets this one right run_example(fill_in_the_blanks, sat_q, args.model) run_example(ask_an_expert, alignment_q, args.model) run_example(ask_an_expert_simple, meaning_q, args.model)