Spaces:
Runtime error
Runtime error
| ## Gradio UI user manual | |
| Gradio UI is a ready to use way of testing most of PrivateGPT API functionalities. | |
|  | |
| ### Execution Modes | |
| It has 3 modes of execution (you can select in the top-left): | |
| * Query Docs: uses the context from the | |
| ingested documents to answer the questions posted in the chat. It also takes | |
| into account previous chat messages as context. | |
| * Makes use of `/chat/completions` API with `use_context=true` and no | |
| `context_filter`. | |
| * Search in Docs: fast search that returns the 4 most related text | |
| chunks, together with their source document and page. | |
| * Makes use of `/chunks` API with no `context_filter`, `limit=4` and | |
| `prev_next_chunks=0`. | |
| * LLM Chat: simple, non-contextual chat with the LLM. The ingested documents won't | |
| be taken into account, only the previous messages. | |
| * Makes use of `/chat/completions` API with `use_context=false`. | |
| ### Document Ingestion | |
| Ingest documents by using the `Upload a File` button. You can check the progress of | |
| the ingestion in the console logs of the server. | |
| The list of ingested files is shown below the button. | |
| If you want to delete the ingested documents, refer to *Reset Local documents | |
| database* section in the documentation. | |
| ### Chat | |
| Normal chat interface, self-explanatory ;) | |
| #### System Prompt | |
| You can view and change the system prompt being passed to the LLM by clicking "Additional Inputs" | |
| in the chat interface. The system prompt is also logged on the server. | |
| By default, the `Query Docs` mode uses the setting value `ui.default_query_system_prompt`. | |
| The `LLM Chat` mode attempts to use the optional settings value `ui.default_chat_system_prompt`. | |
| If no system prompt is entered, the UI will display the default system prompt being used | |
| for the active mode. | |
| ##### System Prompt Examples: | |
| The system prompt can effectively provide your chat bot specialized roles, and results tailored to the prompt | |
| you have given the model. Examples of system prompts can be be found | |
| [here](https://www.w3schools.com/gen_ai/chatgpt-3-5/chatgpt-3-5_roles.php). | |
| Some interesting examples to try include: | |
| * You are -X-. You have all the knowledge and personality of -X-. Answer as if you were -X- using | |
| their manner of speaking and vocabulary. | |
| * Example: You are Shakespeare. You have all the knowledge and personality of Shakespeare. | |
| Answer as if you were Shakespeare using their manner of speaking and vocabulary. | |
| * You are an expert (at) -role-. Answer all questions using your expertise on -specific domain topic-. | |
| * Example: You are an expert software engineer. Answer all questions using your expertise on Python. | |
| * You are a -role- bot, respond with -response criteria needed-. If no -response criteria- is needed, | |
| respond with -alternate response-. | |
| * Example: You are a grammar checking bot, respond with any grammatical corrections needed. If no corrections | |
| are needed, respond with "verified". |