Skip to content

Instantly share code, notes, and snippets.

@pancutan
Last active May 16, 2023 00:06
Show Gist options
  • Save pancutan/8d95dbe52bf3173077a40695d1271b14 to your computer and use it in GitHub Desktop.
Save pancutan/8d95dbe52bf3173077a40695d1271b14 to your computer and use it in GitHub Desktop.
~/privateGPT $ lsb_release -a
LSB Version: n/a
Distributor ID: ManjaroLinux
Description: Manjaro Linux
Release: 22.1.1
Codename: Talos
~/privateGPT $ python --version
Python 3.10.10
~/privateGPT $ cat .env
PERSIST_DIRECTORY=db
LLAMA_EMBEDDINGS_MODEL=/mnt/mochila-linux/PrivateGPT_models/ggml-model-q4_0.bin
MODEL_TYPE=GPT4All
MODEL_PATH=/mnt/mochila-linux/PrivateGPT_models/ggml-gpt4all-j-v1.3-groovy.bin
MODEL_N_CTX=1000
~/privateGPT $ python privateGPT.py
llama.cpp: loading model from /mnt/mochila-linux/PrivateGPT_models/ggml-model-q4_0.bin
llama.cpp: can't use mmap because tensors are not aligned; convert to new format to avoid this
llama_model_load_internal: format = 'ggml' (old version with low tokenizer quality and no mmap support)
llama_model_load_internal: n_vocab = 32000
llama_model_load_internal: n_ctx = 1000
llama_model_load_internal: n_embd = 4096
llama_model_load_internal: n_mult = 256
llama_model_load_internal: n_head = 32
llama_model_load_internal: n_layer = 32
llama_model_load_internal: n_rot = 128
llama_model_load_internal: ftype = 2 (mostly Q4_0)
llama_model_load_internal: n_ff = 11008
llama_model_load_internal: n_parts = 1
llama_model_load_internal: model size = 7B
llama_model_load_internal: ggml ctx size = 4113748.20 KB
llama_model_load_internal: mem required = 5809.33 MB (+ 2052.00 MB per state)
...................................................................................................
.
llama_init_from_file: kv self size = 1000.00 MB
AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 |
Using embedded DuckDB with persistence: data will be stored in: db
gptj_model_load: loading model from '/mnt/mochila-linux/PrivateGPT_models/ggml-gpt4all-j-v1.3-groovy.bin' - please wait ...
gptj_model_load: n_vocab = 50400
gptj_model_load: n_ctx = 2048
gptj_model_load: n_embd = 4096
gptj_model_load: n_head = 16
gptj_model_load: n_layer = 28
gptj_model_load: n_rot = 64
gptj_model_load: f16 = 2
gptj_model_load: ggml ctx size = 4505.45 MB
gptj_model_load: memory_size = 896.00 MB, n_mem = 57344
gptj_model_load: ................................... done
gptj_model_load: model size = 3609.38 MB / num tensors = 285
Enter a query: could you show me a hello world example in brainfuck?
llama_print_timings: load time = 15942.12 ms
llama_print_timings: sample time = 0.00 ms / 1 runs ( 0.00 ms per run)
llama_print_timings: prompt eval time = 16909.54 ms / 14 tokens ( 1207.82 ms per token)
llama_print_timings: eval time = 0.00 ms / 1 runs ( 0.00 ms per run)
llama_print_timings: total time = 16966.38 ms
Traceback (most recent call last):
File "/home/s/privateGPT/privateGPT.py", line 57, in <module>
main()
File "/home/s/privateGPT/privateGPT.py", line 42, in main
res = qa(query)
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/base.py", line 140, in __call__
raise e
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/base.py", line 134, in __call__
self._call(inputs, run_manager=run_manager)
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/retrieval_qa/base.py", line 119, in _call
docs = self._get_docs(question)
File "/home/s/.local/lib/python3.10/site-packages/langchain/chains/retrieval_qa/base.py", line 181, in _get_docs
return self.retriever.get_relevant_documents(question)
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/base.py", line 366, in get_relevant_documents
docs = self.vectorstore.similarity_search(query, **self.search_kwargs)
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 181, in similarity_search
docs_and_scores = self.similarity_search_with_score(query, k, filter=filter)
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 228, in similarity_search_with_score
results = self.__query_collection(
File "/home/s/.local/lib/python3.10/site-packages/langchain/utils.py", line 50, in wrapper
return func(*args, **kwargs)
File "/home/s/.local/lib/python3.10/site-packages/langchain/vectorstores/chroma.py", line 120, in __query_collection
return self._collection.query(
File "/home/s/.local/lib/python3.10/site-packages/chromadb/api/models/Collection.py", line 219, in query
return self._client._query(
File "/home/s/.local/lib/python3.10/site-packages/chromadb/api/local.py", line 408, in _query
uuids, distances = self._db.get_nearest_neighbors(
File "/home/s/.local/lib/python3.10/site-packages/chromadb/db/clickhouse.py", line 583, in get_nearest_neighbors
uuids, distances = index.get_nearest_neighbors(embeddings, n_results, ids)
File "/home/s/.local/lib/python3.10/site-packages/chromadb/db/index/hnswlib.py", line 230, in get_nearest_neighbors
raise NoIndexException(
chromadb.errors.NoIndexException: Index not found, please create an instance before querying
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment