Last active
May 16, 2023 00:06
-
-
Save pancutan/8d95dbe52bf3173077a40695d1271b14 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
~/privateGPT $ lsb_release -a | |
LSB Version: n/a | |
Distributor ID: ManjaroLinux | |
Description: Manjaro Linux | |
Release: 22.1.1 | |
Codename: Talos | |
~/privateGPT $ python --version | |
Python 3.10.10 | |
~/privateGPT $ cat .env | |
PERSIST_DIRECTORY=db | |
LLAMA_EMBEDDINGS_MODEL=/mnt/mochila-linux/PrivateGPT_models/ggml-model-q4_0.bin | |
MODEL_TYPE=GPT4All | |
MODEL_PATH=/mnt/mochila-linux/PrivateGPT_models/ggml-gpt4all-j-v1.3-groovy.bin | |
MODEL_N_CTX=1000 | |
~/privateGPT $ python privateGPT.py | |
llama.cpp: loading model from /mnt/mochila-linux/PrivateGPT_models/ggml-model-q4_0.bin | |
llama.cpp: can't use mmap because tensors are not aligned; convert to new format to avoid this | |
llama_model_load_internal: format = 'ggml' (old version with low tokenizer quality and no mmap support) | |
llama_model_load_internal: n_vocab = 32000 | |
llama_model_load_internal: n_ctx = 1000 | |
llama_model_load_internal: n_embd = 4096 | |
llama_model_load_internal: n_mult = 256 | |
llama_model_load_internal: n_head = 32 | |
llama_model_load_internal: n_layer = 32 | |
llama_model_load_internal: n_rot = 128 | |
llama_model_load_internal: ftype = 2 (mostly Q4_0) | |
llama_model_load_internal: n_ff = 11008 | |
llama_model_load_internal: n_parts = 1 | |
llama_model_load_internal: model size = 7B | |
llama_model_load_internal: ggml ctx size = 4113748.20 KB | |
llama_model_load_internal: mem required = 5809.33 MB (+ 2052.00 MB per state) | |
................................................................................................... | |
. | |
llama_init_from_file: kv self size = 1000.00 MB | |
AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 | | |
Using embedded DuckDB with persistence: data will be stored in: db | |
gptj_model_load: loading model from '/mnt/mochila-linux/PrivateGPT_models/ggml-gpt4all-j-v1.3-groovy.bin' - please wait ... | |
gptj_model_load: n_vocab = 50400 | |
gptj_model_load: n_ctx = 2048 | |
gptj_model_load: n_embd = 4096 | |
gptj_model_load: n_head = 16 | |
gptj_model_load: n_layer = 28 | |
gptj_model_load: n_rot = 64 | |
gptj_model_load: f16 = 2 | |
gptj_model_load: ggml ctx size = 4505.45 MB | |
gptj_model_load: memory_size = 896.00 MB, n_mem = 57344 | |
gptj_model_load: ................................... done | |
gptj_model_load: model size = 3609.38 MB / num tensors = 285 | |
Enter a query: could you show me a hello world example in brainfuck? | |
llama_print_timings: load time = 15942.12 ms | |
llama_print_timings: sample time = 0.00 ms / 1 runs ( 0.00 ms per run) | |
llama_print_timings: prompt eval time = 16909.54 ms / 14 tokens ( 1207.82 ms per token) | |
llama_print_timings: eval time = 0.00 ms / 1 runs ( 0.00 ms per run) | |
llama_print_timings: total time = 16966.38 ms | |
Traceback (most recent call last): | |
File "/home/s/privateGPT/privateGPT.py", line 57, in <module> | |
main() | |
File "/home/s/privateGPT/privateGPT.py", line 42, in main | |
res = qa(query) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/base.py", line 140, in __call__ | |
raise e | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/base.py", line 134, in __call__ | |
self._call(inputs, run_manager=run_manager) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/retrieval_qa/base.py", line 119, in _call | |
docs = self._get_docs(question) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/retrieval_qa/base.py", line 181, in _get_docs | |
return self.retriever.get_relevant_documents(question) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/base.py", line 366, in get_relevant_documents | |
docs = self.vectorstore.similarity_search(query, **self.search_kwargs) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 181, in similarity_search | |
docs_and_scores = self.similarity_search_with_score(query, k, filter=filter) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 228, in similarity_search_with_score | |
results = self.__query_collection( | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/utils.py", line 50, in wrapper | |
return func(*args, **kwargs) | |
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 120, in __query_collection | |
return self._collection.query( | |
File "/home/s/.local/lib/python3.10/site-packages/chromadb/api/models/Collection.py", line 219, in query | |
return self._client._query( | |
File "/home/s/.local/lib/python3.10/site-packages/chromadb/api/local.py", line 408, in _query | |
uuids, distances = self._db.get_nearest_neighbors( | |
File "/home/s/.local/lib/python3.10/site-packages/chromadb/db/clickhouse.py", line 583, in get_nearest_neighbors | |
uuids, distances = index.get_nearest_neighbors(embeddings, n_results, ids) | |
File "/home/s/.local/lib/python3.10/site-packages/chromadb/db/index/hnswlib.py", line 230, in get_nearest_neighbors | |
raise NoIndexException( | |
chromadb.errors.NoIndexException: Index not found, please create an instance before querying |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment