Spaces:
Runtime error
Runtime error
| cnt_agents: &cnt_agents 4 | |
| max_turn: &max_turn 3 | |
| max_inner_turns: &max_inner_turns 2 | |
| task_description: |- | |
| write a simple calculator GUI using Python3. | |
| prompts: | |
| role_assigner_prepend_prompt: &role_assigner_prepend_prompt |- | |
| role_assigner_append_prompt: &role_assigner_append_prompt |- | |
| You are the leader of a group of experts, now you are faced with a task: | |
| ${task_description} | |
| You can recruit ${cnt_critic_agents} expert team members in different regions. | |
| What experts will you recruit to better generate good ideas? | |
| Output format example: | |
| 1. an electrical engineer specified in the filed of xxx | |
| 2. an economist who is good at xxx | |
| 3. a lawyer with a good knowledge of xxx | |
| ... | |
| ${advice} | |
| You don't have to give the reason. | |
| solver_prepend_prompt: &solver_prepend_prompt |- | |
| You are faced with the task: | |
| ${task_description} | |
| Below is the chat history among you and other teammates. | |
| solver_append_prompt: &solver_append_prompt |- | |
| Now you are going to give a new solution, based upon your former solution and the critics' opinions. Write the code step by step. | |
| critic_prepend_prompt: &critic_prepend_prompt |- | |
| Now you are ${role_description} | |
| You are in a discussion group, aiming to ${task_description}. | |
| Below is the chat history among you and other teammates. | |
| critic_append_prompt: &critic_append_prompt |- | |
| Now the group is asking your opinion about it. Based on your knowledge | |
| in your field, do you agree that this solution can perfectly | |
| solve the problem? | |
| Or do you have any ideas to improve it? | |
| - If you thinks it is perfect, use the following output format: | |
| Action: Agree | |
| Action Input: Agree. | |
| (Do not output your reason for agreeing!) | |
| - If you want to give complemented opinions to improve it or to contradict with it, use the following output format: | |
| Action: Disagree | |
| Action Input: (what you want to say in one line) | |
| P.S. Always remember you are ${role_description}! | |
| If no former solution or critic opinions are given, you can just disagree and output your idea freely, based on the expertise of your role. | |
| Remember, the ideas should be specific and detailed enough, not just general opinions. | |
| Please control output code in 2048 tokens! (Write concise code) | |
| evaluator_prepend_prompt: &evaluator_prepend_prompt |- | |
| evaluator_append_prompt: &evaluator_append_prompt |- | |
| You are an professional and strict code reviewer. Your task is to evaluate the solution. The code is to ${task_description}. Your task is to evaluate the codes written by the code engineers. | |
| Please not only give a general rating points (from 0 to 9) but also give detailed comments about where and how the code can be improved. Please consider the following aspects when you are evaluating the code: | |
| 1. The code should be able to run without any errors. | |
| 2. The code should be able to achieve the goal specified in the task description. | |
| 3. The code should be easy to read and understand, efficient, concise and elegant. | |
| 4. The code should be robust. | |
| Please rate the code in the following dimensions: | |
| 1. Completeness: Is the code snippet complete enough without unimplemented functions of methods? Is it able to run without any errors? | |
| 2. Functionality: Is the code able to achieve the goal specified in the task description? | |
| 3. Readability: Is the code easy to read and understand, efficient, concise and elegant? | |
| 4. Robustness: Is the code snippet able to handle different unexpected input or other exceptions? | |
| In the fifth line of output, give your detailed advice for the engineers to better generate good codes. Only give high grade when it is really great. | |
| # Output format | |
| 1. Completeness: (a score between 0 and 9) | |
| 2. Functionality: (a score between 0 and 9) | |
| 3. Readability: (a score between 0 and 9) | |
| 4. Robustness: (a score between 0 and 9) | |
| 5. Advice: (your advice in one line) | |
| Here is the content you have to evaluate: | |
| ${solution} | |
| name: pipeline | |
| environment: | |
| env_type: task-basic | |
| max_turn: | |
| rule: | |
| role_assigner: | |
| type: role_description | |
| cnt_agents: | |
| decision_maker: | |
| type: vertical-solver-first | |
| max_inner_turns: | |
| executor: | |
| type: none | |
| evaluator: | |
| type: basic | |
| agents: | |
| - #role_assigner_agent: | |
| agent_type: role_assigner | |
| name: role assigner | |
| max_retry: 1000 | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0.7 | |
| max_tokens: 256 | |
| output_parser: | |
| type: role_assigner | |
| - #solver_agent: | |
| agent_type: solver | |
| name: Summarizer | |
| max_retry: 1000 | |
| max_history: 5 | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0.7 | |
| max_tokens: 2048 | |
| output_parser: | |
| type: dummy | |
| - #critic_agents: | |
| agent_type: critic | |
| name: Reviewer | |
| max_retry: 1000 | |
| max_history: 5 | |
| role_description: |- | |
| Waiting to be assigned. | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0.7 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: critic | |
| - #executor_agent: | |
| agent_type: executor | |
| name: Dummy Executor | |
| max_retry: 1000 | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0.7 | |
| max_tokens: 512 | |
| output_parser: | |
| type: dummy | |
| - #evaluator_agent: | |
| agent_type: evaluator | |
| name: Evaluator | |
| max_retry: 1000 | |
| role_description: |- | |
| Evaluator | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0.7 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: evaluator | |
| dimensions: | |
| - Completeness | |
| - Functionality | |
| - Readability | |
| - Robustness | |