Spaces:
Runtime error
Runtime error
| cnt_agents: &cnt_agents 2 | |
| max_turn: &max_turn 3 | |
| max_inner_turns: &max_inner_turns 3 | |
| prompts: | |
| role_assigner_prepend_prompt: &role_assigner_prepend_prompt |- | |
| role_assigner_append_prompt: &role_assigner_append_prompt |- | |
| # Role Description | |
| You are the leader of a group of experts, now you need to generate a response based on the text: | |
| ${task_description} | |
| You can recruit ${cnt_critic_agents} expert in different fields. What experts will you recruit to better generate an accurate solution? | |
| # Response Format Guidance | |
| You should respond with a list of expert description. For example: | |
| 1. an electrical engineer specified in the filed of xxx | |
| 2. an economist who is good at xxx | |
| 3. a lawyer with a good knowledge of xxx | |
| ... | |
| You don't have to give the reason. | |
| solver_prepend_prompt: &solver_prepend_prompt |- | |
| You are ${role_description}. Below is a chat history: | |
| ${task_description} | |
| And below is the discussion about what the next system response should be: | |
| solver_append_prompt: &solver_append_prompt |- | |
| Now based on these information, please give a better next system response. Your response should contain only one system response beginning with "System: ". Do not give any additional information. | |
| critic_prepend_prompt: &critic_prepend_prompt |- | |
| You are ${role_description}. You are in a discussion group, aiming to generate a system response to the following chat history: | |
| ${task_description} | |
| Below is the discussion about what the next system response should be: | |
| critic_append_prompt: &critic_append_prompt |- | |
| # Response Format Guidance | |
| - If you thinks the latest response given above is perfect, respond using the following format: | |
| Decision: (set it to "Agree") | |
| Response: (your response on why you think it is perfect) | |
| - If you think it is flawed, give your advice use the following output format: | |
| Decision: (set it to "Disagree") | |
| Response: (explain why you disagree and give your advice) | |
| Based on your knowledge in your field, do you agree that this solution is a good response to the chat history? | |
| manager_prompt: &manager_prompt |- | |
| executor_prepend_prompt: &executor_prepend_prompt |- | |
| executor_append_prompt: &executor_append_prompt |- | |
| evaluator_prepend_prompt: &evaluator_prepend_prompt |- | |
| evaluator_append_prompt: &evaluator_append_prompt |- | |
| # Role Description | |
| You are an experienced dialogue teacher. As a good teacher, you carefully assess the given system response based on the chat history. When the response is flawed, you should patiently teach the system how to give better response. | |
| # Response Format Guidance | |
| You must respond in the following format: | |
| Engaging: (a score between 0 and 10) | |
| Relevant: (a score between 0 and 10) | |
| Semantically Appropriate: (a score between 0 and 10) | |
| Advice: (your advice on how to improve the response) | |
| # Chat History | |
| ${task_description} | |
| # Next System Response | |
| ${solution} | |
| # Your Task | |
| Now carefully check the system's response, and give your opinion. | |
| name: pipeline | |
| environment: | |
| env_type: task-basic | |
| max_turn: | |
| rule: | |
| role_assigner: | |
| type: role_description | |
| cnt_agents: | |
| decision_maker: | |
| type: vertical-solver-first | |
| max_inner_turns: | |
| executor: | |
| type: none | |
| evaluator: | |
| type: basic | |
| agents: | |
| - #role_assigner_agent: | |
| agent_type: role_assigner | |
| name: role assigner | |
| max_retry: 1000 | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0 | |
| max_tokens: 512 | |
| output_parser: | |
| type: role_assigner | |
| - #solver_agent: | |
| agent_type: solver | |
| name: Planner | |
| max_retry: 1000 | |
| max_history: 10 | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: responsegen | |
| - #critic_agents: | |
| agent_type: critic | |
| name: Critic 1 | |
| max_retry: 1000 | |
| max_history: 10 | |
| role_description: |- | |
| Waiting to be assigned. | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: responsegen-critic-2 | |
| - #executor_agent: | |
| agent_type: executor | |
| name: Executor | |
| max_retry: 1000 | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: gpt-4 | |
| temperature: 0 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: responsegen | |
| - #evaluator_agent: | |
| agent_type: evaluator | |
| name: Evaluator | |
| max_retry: 1000 | |
| role_description: |- | |
| Evaluator | |
| prepend_prompt_template: | |
| append_prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: gpt-4 | |
| temperature: 0.3 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: responsegen-evaluator | |
| dimensions: | |
| - Engaging | |
| - Relevant | |
| - Semantically Appropriate | |
| - #manager_agent: | |
| agent_type: manager | |
| name: Manager | |
| max_retry: 1000 | |
| prompt_template: | |
| memory: | |
| memory_type: chat_history | |
| llm: | |
| llm_type: gpt-4 | |
| model: "gpt-4" | |
| temperature: 0 | |
| max_tokens: 1024 | |
| output_parser: | |
| type: humaneval-manager | |