Spaces:
Sleeping
Sleeping
| [tokens] | |
| ; Enter one/all of your API key here. | |
| ; E.g., OPENAI_API_KEY = sk-xxxxxxx | |
| OPENAI_API_KEY = sk-proj-2JwvyIn7WoKlkbjPOYVWT3BlbkFJnGAk65YAzvPH6cEVQXmr | |
| ANTHROPIC_API_KEY = xxxxx | |
| TOGETHER_API_KEY = xxxxx | |
| ; if you use Meta-Llama models, you may need Huggingface token to access. | |
| HUGGINGFACE_TOKEN = xxxxx | |
| VERSION = 1.0.1 | |
| [directory] | |
| ; Directory for source files. | |
| DOCS_DIR = ./data | |
| ; Directory to store embeddings and Langchain documents. | |
| DB_DIR = ./database_store | |
| LOCAL_MODEL_DIR = ./models | |
| ; The below parameters are optional to modify: | |
| ; -------------------------------------------- | |
| [parameters] | |
| ; Model name schema: Model Provider|Model Name|Model File. Model File is only valid for GGUF format, set None for other format. | |
| ; For example: | |
| ; OpenAI|gpt-3.5-turbo|None | |
| ; OpenAI|gpt-4|None | |
| ; Anthropic|claude-2.0|None | |
| ; Together|togethercomputer/llama-2-70b-chat|None | |
| ; HuggingFace|TheBloke/Llama-2-70B-chat-GGUF|llama-2-70b-chat.q4_K_M.gguf | |
| ; HuggingFace|meta-llama/Llama-2-70b-chat-hf|None | |
| ; The full Together.AI model list can be found in the end of this file; We currently only support quantized gguf and the full huggingface local LLMs. | |
| MODEL_NAME = OpenAI|gpt-4-1106-preview|None | |
| ; LLM temperature | |
| TEMPURATURE = 0 | |
| ; Maximum tokens for storing chat history. | |
| MAX_CHAT_HISTORY = 800 | |
| ; Maximum tokens for LLM context for retrieved information. | |
| MAX_LLM_CONTEXT = 1200 | |
| ; Maximum tokens for LLM generation. | |
| MAX_LLM_GENERATION = 1000 | |
| ; Supported embeddings: openAIEmbeddings and hkunlpInstructorLarge. | |
| EMBEDDING_NAME = openAIEmbeddings | |
| ; This is dependent on your GPU type. | |
| N_GPU_LAYERS = 100 | |
| ; this is depend on your GPU and CPU ram when using open source LLMs. | |
| N_BATCH = 512 | |
| ; The base (small) chunk size for first stage document retrieval. | |
| BASE_CHUNK_SIZE = 100 | |
| ; Set to 0 for no overlap. | |
| CHUNK_OVERLAP = 0 | |
| ; The final retrieval (medium) chunk size will be BASE_CHUNK_SIZE * CHUNK_SCALE. | |
| CHUNK_SCALE = 3 | |
| WINDOW_STEPS = 3 | |
| ; The # tokens of window chunk will be BASE_CHUNK_SIZE * WINDOW_SCALE. | |
| WINDOW_SCALE = 18 | |
| ; Ratio of BM25 retriever to Chroma Vectorstore retriever. | |
| RETRIEVER_WEIGHTS = 0.5, 0.5 | |
| ; Number of retrieved chunks will range from FIRST_RETRIEVAL_K to 2*FIRST_RETRIEVAL_K due to the ensemble retriever. | |
| FIRST_RETRIEVAL_K = 3 | |
| ; Number of retrieved chunks will range from SECOND_RETRIEVAL_K to 2*SECOND_RETRIEVAL_K due to the ensemble retriever. | |
| SECOND_RETRIEVAL_K = 3 | |
| ; Number of windows (large chunks) for the third retriever. | |
| NUM_WINDOWS = 2 | |
| ; (The third retrieval gets the final chunks passed to the LLM QA chain. The 'k' value is dynamic (based on MAX_LLM_CONTEXT), depending on the number of rephrased questions and retrieved documents.) | |
| [logging] | |
| ; If you do not want to enable logging, set enabled to False. | |
| enabled = True | |
| level = INFO | |
| filename = IncarnaMind.log | |
| format = %(asctime)s [%(levelname)s] %(name)s: %(message)s | |
| ; Together.AI supported models: | |
| ; 0 Austism/chronos-hermes-13b | |
| ; 1 EleutherAI/pythia-12b-v0 | |
| ; 2 EleutherAI/pythia-1b-v0 | |
| ; 3 EleutherAI/pythia-2.8b-v0 | |
| ; 4 EleutherAI/pythia-6.9b | |
| ; 5 Gryphe/MythoMax-L2-13b | |
| ; 6 HuggingFaceH4/starchat-alpha | |
| ; 7 NousResearch/Nous-Hermes-13b | |
| ; 8 NousResearch/Nous-Hermes-Llama2-13b | |
| ; 9 NumbersStation/nsql-llama-2-7B | |
| ; 10 OpenAssistant/llama2-70b-oasst-sft-v10 | |
| ; 11 OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5 | |
| ; 12 OpenAssistant/stablelm-7b-sft-v7-epoch-3 | |
| ; 13 Phind/Phind-CodeLlama-34B-Python-v1 | |
| ; 14 Phind/Phind-CodeLlama-34B-v2 | |
| ; 15 SG161222/Realistic_Vision_V3.0_VAE | |
| ; 16 WizardLM/WizardCoder-15B-V1.0 | |
| ; 17 WizardLM/WizardCoder-Python-34B-V1.0 | |
| ; 18 WizardLM/WizardLM-70B-V1.0 | |
| ; 19 bigcode/starcoder | |
| ; 20 databricks/dolly-v2-12b | |
| ; 21 databricks/dolly-v2-3b | |
| ; 22 databricks/dolly-v2-7b | |
| ; 23 defog/sqlcoder | |
| ; 24 garage-bAInd/Platypus2-70B-instruct | |
| ; 25 huggyllama/llama-13b | |
| ; 26 huggyllama/llama-30b | |
| ; 27 huggyllama/llama-65b | |
| ; 28 huggyllama/llama-7b | |
| ; 29 lmsys/fastchat-t5-3b-v1.0 | |
| ; 30 lmsys/vicuna-13b-v1.3 | |
| ; 31 lmsys/vicuna-13b-v1.5-16k | |
| ; 32 lmsys/vicuna-13b-v1.5 | |
| ; 33 lmsys/vicuna-7b-v1.3 | |
| ; 34 prompthero/openjourney | |
| ; 35 runwayml/stable-diffusion-v1-5 | |
| ; 36 stabilityai/stable-diffusion-2-1 | |
| ; 37 stabilityai/stable-diffusion-xl-base-1.0 | |
| ; 38 togethercomputer/CodeLlama-13b-Instruct | |
| ; 39 togethercomputer/CodeLlama-13b-Python | |
| ; 40 togethercomputer/CodeLlama-13b | |
| ; 41 togethercomputer/CodeLlama-34b-Instruct | |
| ; 42 togethercomputer/CodeLlama-34b-Python | |
| ; 43 togethercomputer/CodeLlama-34b | |
| ; 44 togethercomputer/CodeLlama-7b-Instruct | |
| ; 45 togethercomputer/CodeLlama-7b-Python | |
| ; 46 togethercomputer/CodeLlama-7b | |
| ; 47 togethercomputer/GPT-JT-6B-v1 | |
| ; 48 togethercomputer/GPT-JT-Moderation-6B | |
| ; 49 togethercomputer/GPT-NeoXT-Chat-Base-20B | |
| ; 50 togethercomputer/Koala-13B | |
| ; 51 togethercomputer/LLaMA-2-7B-32K | |
| ; 52 togethercomputer/Llama-2-7B-32K-Instruct | |
| ; 53 togethercomputer/Pythia-Chat-Base-7B-v0.16 | |
| ; 54 togethercomputer/Qwen-7B-Chat | |
| ; 55 togethercomputer/Qwen-7B | |
| ; 56 togethercomputer/RedPajama-INCITE-7B-Base | |
| ; 57 togethercomputer/RedPajama-INCITE-7B-Chat | |
| ; 58 togethercomputer/RedPajama-INCITE-7B-Instruct | |
| ; 59 togethercomputer/RedPajama-INCITE-Base-3B-v1 | |
| ; 60 togethercomputer/RedPajama-INCITE-Chat-3B-v1 | |
| ; 61 togethercomputer/RedPajama-INCITE-Instruct-3B-v1 | |
| ; 62 togethercomputer/alpaca-7b | |
| ; 63 togethercomputer/codegen2-16B | |
| ; 64 togethercomputer/codegen2-7B | |
| ; 65 togethercomputer/falcon-40b-instruct | |
| ; 66 togethercomputer/falcon-40b | |
| ; 67 togethercomputer/falcon-7b-instruct | |
| ; 68 togethercomputer/falcon-7b | |
| ; 69 togethercomputer/guanaco-13b | |
| ; 70 togethercomputer/guanaco-33b | |
| ; 71 togethercomputer/guanaco-65b | |
| ; 72 togethercomputer/guanaco-7b | |
| ; 73 togethercomputer/llama-2-13b-chat | |
| ; 74 togethercomputer/llama-2-13b | |
| ; 75 togethercomputer/llama-2-70b-chat | |
| ; 76 togethercomputer/llama-2-70b | |
| ; 77 togethercomputer/llama-2-7b-chat | |
| ; 78 togethercomputer/llama-2-7b | |
| ; 79 togethercomputer/mpt-30b-chat | |
| ; 80 togethercomputer/mpt-30b-instruct | |
| ; 81 togethercomputer/mpt-30b | |
| ; 82 togethercomputer/mpt-7b-chat | |
| ; 83 togethercomputer/mpt-7b | |
| ; 84 togethercomputer/replit-code-v1-3b | |
| ; 85 upstage/SOLAR-0-70b-16bit | |
| ; 86 wavymulder/Analog-Diffusion |