-
-
Save joshuacox/dc9ed633e28260f16af0ec02c69c0c74 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
go mod edit -replace github.com/go-skynet/go-llama.cpp=/build/go-llama | |
go mod edit -replace github.com/nomic-ai/gpt4all/gpt4all-bindings/golang=/build/gpt4all/gpt4all-bindings/golang | |
go mod edit -replace github.com/go-skynet/go-ggml-transformers.cpp=/build/go-ggml-transformers | |
go mod edit -replace github.com/donomii/go-rwkv.cpp=/build/go-rwkv | |
go mod edit -replace github.com/ggerganov/whisper.cpp=/build/whisper.cpp | |
go mod edit -replace github.com/go-skynet/go-bert.cpp=/build/go-bert | |
go mod edit -replace github.com/go-skynet/bloomz.cpp=/build/bloomz | |
go mod edit -replace github.com/mudler/go-stable-diffusion=/build/go-stable-diffusion | |
go mod edit -replace github.com/mudler/go-piper=/build/go-piper | |
go mod edit -replace github.com/mudler/go-ggllm.cpp=/build/go-ggllm | |
go mod download | |
touch prepare | |
mkdir -p backend-assets/grpc | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/langchain-huggingface ./cmd/grpc/langchain-huggingface/ | |
make -C go-ggml-transformers BUILD_TYPE=cublas libtransformers.a | |
make[1]: Entering directory '/build/go-ggml-transformers' | |
I libtransformers build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -I -O3 -DNDEBUG -std=c11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wno-unused-function -pthread -march=native -mtune=native | |
I CXXFLAGS: -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native | |
I LDFLAGS: | |
I CMAKE_ARGS: -DGGML_CUBLAS=ON | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native starcoder.cpp -o starcoder.o -c | |
In file included from [01m[Kstarcoder.cpp:19[m[K: | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:[m[K In function '[01m[Kint main_starcoder(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:799:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
799 | for (int i = 0; [01;35m[Ki < embd_inp.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:821:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
821 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:837:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
837 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:859:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
859 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/starcoder/main.cpp:861:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
861 | if ([01;35m[Kembd.size() >= params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:[m[K In function '[01m[Kint starcoder_predict(void*, void*, char*)[m[K': | |
[01m[Kstarcoder.cpp:80:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
80 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:96:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
96 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:118:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
118 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:120:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
120 | if ([01;35m[Kembd.size() >= params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:36:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
36 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kstarcoder.cpp:47:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
47 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native falcon.cpp -o falcon.o -c | |
In file included from [01m[Kfalcon.cpp:19[m[K: | |
[01m[Kimplementations/falcon.cpp:[m[K In function '[01m[Kbool falcon_model_load(const string&, falcon_model&, gpt_vocab&)[m[K': | |
[01m[Kimplementations/falcon.cpp:187:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
187 | [01;35m[K.[m[Kmem_size = ctx_size, | |
| [01;35m[K^[m[K | |
[01m[Kimplementations/falcon.cpp:188:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
188 | [01;35m[K.[m[Kmem_buffer = NULL, | |
| [01;35m[K^[m[K | |
[01m[Kimplementations/falcon.cpp:189:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
189 | [01;35m[K.[m[Kno_alloc = false, | |
| [01;35m[K^[m[K | |
In file included from [01m[Kfalcon.cpp:19[m[K: | |
[01m[Kimplementations/falcon.cpp:[m[K In function '[01m[Kbool falcon_eval(const falcon_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[Kimplementations/falcon.cpp:410:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
410 | [01;35m[K.[m[Kmem_size = buf_size, | |
| [01;35m[K^[m[K | |
[01m[Kimplementations/falcon.cpp:411:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
411 | [01;35m[K.[m[Kmem_buffer = buf, | |
| [01;35m[K^[m[K | |
[01m[Kimplementations/falcon.cpp:412:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
412 | [01;35m[K.[m[Kno_alloc = false, | |
| [01;35m[K^[m[K | |
[01m[Kfalcon.cpp:[m[K In function '[01m[Kint falcon_predict(void*, void*, char*)[m[K': | |
[01m[Kfalcon.cpp:69:34:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
69 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon.cpp:85:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
85 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon.cpp:107:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
107 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon.cpp:109:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
109 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon.cpp:36:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
36 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon.cpp:48:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
48 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native gptj.cpp -o gptj.o -c | |
In file included from [01m[Kgptj.cpp:19[m[K: | |
[01m[Kggml.cpp/examples/gpt-j/main.cpp:[m[K In function '[01m[Kint main_gptj(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/gpt-j/main.cpp:674:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
674 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-j/main.cpp:690:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
690 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-j/main.cpp:712:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
712 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-j/main.cpp:714:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
714 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:[m[K In function '[01m[Kint gptj_predict(void*, void*, char*)[m[K': | |
[01m[Kgptj.cpp:72:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
72 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:88:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
88 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:110:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
110 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:112:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
112 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:36:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
36 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kgptj.cpp:48:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
48 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native mpt.cpp -o mpt.o -c | |
In file included from [01m[Kmpt.cpp:19[m[K: | |
[01m[Kggml.cpp/examples/mpt/main.cpp:[m[K In function '[01m[Kbool mpt_model_load(const string&, mpt_model&, gpt_vocab&)[m[K': | |
[01m[Kggml.cpp/examples/mpt/main.cpp:246:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::__cxx11::basic_string<wchar_t>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
246 | for (int w = 0; [01;35m[Kw < word_multibytes.size()[m[K; w++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kmpt.cpp:[m[K In function '[01m[Kint mpt_predict(void*, void*, char*)[m[K': | |
[01m[Kmpt.cpp:37:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
37 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kmpt.cpp:49:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
49 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
mkdir build | |
cd build && cmake ../ggml.cpp -DGGML_CUBLAS=ON && make VERBOSE=1 ggml && cp -rf src/CMakeFiles/ggml.dir/ggml.c.o ../ggml.o | |
[0mCMake Deprecation Warning at CMakeLists.txt:1 (cmake_minimum_required): | |
Compatibility with CMake < 3.5 will be removed from a future version of | |
CMake. | |
Update the VERSION argument <min> value or use a ...<max> suffix to tell | |
CMake that the project does not need compatibility with older versions. | |
[0m | |
-- The C compiler identification is GNU 10.2.1 | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Found Git: /usr/bin/git (found version "2.30.2") | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Looking for pthread_create in pthreads | |
-- Looking for pthread_create in pthreads - not found | |
-- Looking for pthread_create in pthread | |
-- Looking for pthread_create in pthread - found | |
-- Found Threads: TRUE | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
-- Linux detected | |
-- Found CUDAToolkit: /usr/local/cuda/include (found version "12.1.105") | |
-- cuBLAS found | |
-- The CUDA compiler identification is NVIDIA 12.1.105 | |
-- Detecting CUDA compiler ABI info | |
-- Detecting CUDA compiler ABI info - done | |
-- Check for working CUDA compiler: /usr/local/cuda/bin/nvcc - skipped | |
-- Detecting CUDA compile features | |
-- Detecting CUDA compile features - done | |
-- GGML CUDA sources found, configuring CUDA architecture | |
-- x86 detected | |
-- Linux detected | |
-- Configuring done (1.8s) | |
-- Generating done (0.1s) | |
-- Build files have been written to: /build/go-ggml-transformers/build | |
make[2]: Entering directory '/build/go-ggml-transformers/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-ggml-transformers/ggml.cpp -B/build/go-ggml-transformers/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
make -f CMakeFiles/Makefile2 ggml | |
make[3]: Entering directory '/build/go-ggml-transformers/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-ggml-transformers/ggml.cpp -B/build/go-ggml-transformers/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-ggml-transformers/build/CMakeFiles 3 | |
make -f CMakeFiles/Makefile2 src/CMakeFiles/ggml.dir/all | |
make[4]: Entering directory '/build/go-ggml-transformers/build' | |
make -f src/CMakeFiles/ggml.dir/build.make src/CMakeFiles/ggml.dir/depend | |
make[5]: Entering directory '/build/go-ggml-transformers/build' | |
cd /build/go-ggml-transformers/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggml-transformers/ggml.cpp /build/go-ggml-transformers/ggml.cpp/src /build/go-ggml-transformers/build /build/go-ggml-transformers/build/src /build/go-ggml-transformers/build/src/CMakeFiles/ggml.dir/DependInfo.cmake "--color=" | |
make[5]: Leaving directory '/build/go-ggml-transformers/build' | |
make -f src/CMakeFiles/ggml.dir/build.make src/CMakeFiles/ggml.dir/build | |
make[5]: Entering directory '/build/go-ggml-transformers/build' | |
[ 33%] [32mBuilding C object src/CMakeFiles/ggml.dir/ggml.c.o[0m | |
cd /build/go-ggml-transformers/build/src && /usr/bin/cc -DGGML_USE_CUBLAS -I/build/go-ggml-transformers/ggml.cpp/src/. -I/build/go-ggml-transformers/ggml.cpp/src/../include -I/build/go-ggml-transformers/ggml.cpp/src/../include/ggml -isystem /usr/local/cuda/include -Wall -Wextra -Wpedantic -Wshadow -Wcast-qual -Wstrict-prototypes -Wpointer-arith -Wdouble-promotion -Wno-unused-function -Werror=vla -mavx -mavx2 -mfma -mf16c -msse3 -O3 -DNDEBUG -std=gnu11 -MD -MT src/CMakeFiles/ggml.dir/ggml.c.o -MF CMakeFiles/ggml.dir/ggml.c.o.d -o CMakeFiles/ggml.dir/ggml.c.o -c /build/go-ggml-transformers/ggml.cpp/src/ggml.c | |
[ 66%] [32mBuilding CUDA object src/CMakeFiles/ggml.dir/ggml-cuda.cu.o[0m | |
cd /build/go-ggml-transformers/build/src && /usr/local/cuda/bin/nvcc -forward-unknown-to-host-compiler -DGGML_USE_CUBLAS --options-file CMakeFiles/ggml.dir/includes_CUDA.rsp -O3 -DNDEBUG -std=c++11 -MD -MT src/CMakeFiles/ggml.dir/ggml-cuda.cu.o -MF CMakeFiles/ggml.dir/ggml-cuda.cu.o.d -x cu -c /build/go-ggml-transformers/ggml.cpp/src/ggml-cuda.cu -o CMakeFiles/ggml.dir/ggml-cuda.cu.o | |
[100%] [32m[1mLinking CUDA static library libggml.a[0m | |
cd /build/go-ggml-transformers/build/src && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/ggml.dir/cmake_clean_target.cmake | |
cd /build/go-ggml-transformers/build/src && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/ggml.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libggml.a CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-cuda.cu.o" | |
/usr/bin/ranlib libggml.a | |
make[5]: Leaving directory '/build/go-ggml-transformers/build' | |
[100%] Built target ggml | |
make[4]: Leaving directory '/build/go-ggml-transformers/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-ggml-transformers/build/CMakeFiles 0 | |
make[3]: Leaving directory '/build/go-ggml-transformers/build' | |
make[2]: Leaving directory '/build/go-ggml-transformers/build' | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native gpt2.cpp ggml.o -o gpt2.o -c | |
In file included from [01m[Kgpt2.cpp:16[m[K: | |
[01m[Kggml.cpp/examples/gpt-2/main.cpp:[m[K In function '[01m[Kint main_gpt2(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/gpt-2/main.cpp:770:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
770 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-2/main.cpp:786:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
786 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-2/main.cpp:808:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
808 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-2/main.cpp:810:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
810 | if ([01;35m[Kembd.size() >= params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:[m[K In function '[01m[Kint gpt2_predict(void*, void*, char*)[m[K': | |
[01m[Kgpt2.cpp:68:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
68 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:84:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
84 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:106:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
106 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:108:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
108 | if ([01;35m[Kembd.size() >= params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:34:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
34 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kgpt2.cpp:43:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
43 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
[01m[Kg++:[m[K [01;35m[Kwarning: [m[Kggml.o: linker input file unused because linking not done | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native replit.cpp -o replit.o -c | |
In file included from [01m[Kreplit.cpp:21[m[K: | |
[01m[Kggml.cpp/examples/replit/main.cpp:[m[K In function '[01m[Kstd::pair<std::vector<long unsigned int>, float> encode_word(const string&, const piece_map_t&)[m[K': | |
[01m[Kggml.cpp/examples/replit/main.cpp:54:39:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::__cxx11::basic_string<char>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
54 | for (int start_idx = 0; [01;35m[Kstart_idx < word.length()[m[K; ++start_idx) { | |
| [01;35m[K~~~~~~~~~~^~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:56:51:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::__cxx11::basic_string<char>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
56 | for (int end_idx = start_idx + 1; [01;35m[Kend_idx <= word.length()[m[K; ++end_idx) { | |
| [01;35m[K~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:[m[K In function '[01m[Kbool replit_tokenizer_load(replit_tokenizer&, std::istream&, int)[m[K': | |
[01m[Kggml.cpp/examples/replit/main.cpp:94:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::size_t[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint[m[K' [[01;35m[K-Wsign-compare[m[K] | |
94 | for (std::size_t i = 0; [01;35m[Ki < max_vocab_size[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:[m[K In function '[01m[Kbool replit_model_load(const string&, replit_model&, replit_tokenizer&)[m[K': | |
[01m[Kggml.cpp/examples/replit/main.cpp:345:56:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 4 has type '[01m[Klong int[m[K' [[01;35m[K-Wformat=[m[K] | |
345 | printf("%s: memory_size = %8.2f MB, n_mem = [01;35m[K%lld[m[K\n", __func__, memory_size / 1024.0 / 1024.0, [32m[Kn_mem[m[K); | |
| [01;35m[K~~~^[m[K [32m[K~~~~~[m[K | |
| [01;35m[K|[m[K [32m[K|[m[K | |
| [01;35m[Klong long int[m[K [32m[Klong int[m[K | |
| [32m[K%ld[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:[m[K In function '[01m[Kint main_replit(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/replit/main.cpp:704:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
704 | for (int i = 0; [01;35m[Ki < embd_inp.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:718:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
718 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:734:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
734 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:757:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
757 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/replit/main.cpp:759:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
759 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:[m[K In function '[01m[Kint replit_predict(void*, void*, char*)[m[K': | |
[01m[Kreplit.cpp:64:21:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
64 | for (int i = 0; [01;35m[Ki < embd_inp.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:65:31:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%d[m[K' expects argument of type '[01m[Kint[m[K', but argument 4 has type '[01m[K__gnu_cxx::__alloc_traits<std::allocator<long unsigned int>, long unsigned int>::value_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wformat=[m[K] | |
65 | printf("%s: token[%d] = [01;35m[K%6d[m[K\n", __func__, i, embd_inp[i]); | |
| [01;35m[K~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Kint[m[K | |
| [32m[K%6ld[m[K | |
[01m[Kreplit.cpp:80:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
80 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:96:11:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
96 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:120:25:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<long unsigned int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
120 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:122:25:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
122 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:39:17:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
39 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kreplit.cpp:49:11:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
49 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native gptneox.cpp -o gptneox.o -c | |
In file included from [01m[Kgptneox.cpp:19[m[K: | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:[m[K In function '[01m[Kint main_gptneox(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:728:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
728 | for (int i = 0; [01;35m[Ki < embd_inp.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:739:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
739 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:755:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
755 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:777:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
777 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/gpt-neox/main.cpp:779:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
779 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:[m[K In function '[01m[Kint gpt_neox_predict(void*, void*, char*)[m[K': | |
[01m[Kgptneox.cpp:71:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
71 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:87:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
87 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:109:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
109 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:111:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
111 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:36:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
36 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kgptneox.cpp:48:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
48 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native dolly.cpp -o dolly.o -c | |
In file included from [01m[Kdolly.cpp:18[m[K: | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:[m[K In function '[01m[Kint main_dolly(int, char**)[m[K': | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:731:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
731 | for (int i = 0; [01;35m[Ki < embd_inp.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:744:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
744 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:760:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
760 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:783:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
783 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kggml.cpp/examples/dolly-v2/main.cpp:785:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
785 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kdolly.cpp:[m[K In function '[01m[Kint dolly_predict(void*, void*, char*)[m[K': | |
[01m[Kdolly.cpp:70:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
70 | for (int i = embd.size(); [01;35m[Ki < embd_inp.size() + params.n_predict[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kdolly.cpp:86:15:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
86 | if ([01;35m[Ki >= embd_inp.size()[m[K) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kdolly.cpp:109:31:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
109 | for (int k = i; [01;35m[Kk < embd_inp.size()[m[K; k++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kdolly.cpp:111:33:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kstd::vector<int>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} and '[01m[Kint32_t[m[K' {aka '[01m[Kint[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
111 | if ([01;35m[Kembd.size() > params.n_batch[m[K) { | |
| [01;35m[K~~~~~~~~~~~~^~~~~~~~~~~~~~~~[m[K | |
[01m[Kdolly.cpp:46:13:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_load_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
46 | int64_t [01;35m[Kt_load_us[m[K = 0; | |
| [01;35m[K^~~~~~~~~[m[K | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native -c ggml.cpp/examples/common-ggml.cpp -o common-ggml.o | |
g++ -I. -I./ggml.cpp/include -I./ggml.cpp/include/ggml/ -I./ggml.cpp/examples/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native -c ggml.cpp/examples/common.cpp -o common.o | |
ar src libtransformers.a replit.o gptj.o mpt.o gptneox.o starcoder.o gpt2.o dolly.o falcon.o ggml.o common-ggml.o common.o ggml-cuda.o | |
make[1]: Leaving directory '/build/go-ggml-transformers' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/falcon-ggml ./cmd/grpc/falcon-ggml/ | |
make -C go-bert libgobert.a | |
make[1]: Entering directory '/build/go-bert' | |
I go-gpt4all-j build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I. -I./bert.cpp/ggml/include/ggml/ -I./bert.cpp/ -I -O3 -DNDEBUG -std=c11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wno-unused-function -pthread -march=native -mtune=native | |
I CXXFLAGS: -I. -I./bert.cpp/ggml/include/ggml/ -I./bert.cpp/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native | |
I LDFLAGS: | |
I CMAKEFLAGS: | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
cd bert.cpp && mkdir build | |
sed "s/#include <regex>/#include <regex>\n#include <unordered_map>/" bert.cpp/bert.cpp > bert.cpp/bert.tmp && mv bert.cpp/bert.tmp bert.cpp/bert.cpp | |
cd bert.cpp/build && cmake .. -DBUILD_SHARED_LIBS=OFF -DCMAKE_BUILD_TYPE=Release && make | |
-- The C compiler identification is GNU 10.2.1 | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Check if compiler accepts -pthread | |
-- Check if compiler accepts -pthread - yes | |
-- Found Threads: TRUE | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
[0mCMake Deprecation Warning at ggml/CMakeLists.txt:1 (cmake_minimum_required): | |
Compatibility with CMake < 3.5 will be removed from a future version of | |
CMake. | |
Update the VERSION argument <min> value or use a ...<max> suffix to tell | |
CMake that the project does not need compatibility with older versions. | |
[0m | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
-- Linux detected | |
-- Configuring done (0.4s) | |
-- Generating done (0.0s) | |
-- Build files have been written to: /build/go-bert/bert.cpp/build | |
make[2]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[3]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 8%] [32mBuilding C object ggml/src/CMakeFiles/ggml.dir/ggml.c.o[0m | |
[ 16%] [32m[1mLinking C static library libggml.a[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[ 16%] Built target ggml | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 25%] [32mBuilding CXX object CMakeFiles/bert.dir/bert.cpp.o[0m | |
[ 33%] [32m[1mLinking CXX static library libbert.a[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[ 33%] Built target bert | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 41%] [32mBuilding CXX object examples/CMakeFiles/server.dir/server.cpp.o[0m | |
[ 50%] [32m[1mLinking CXX executable ../bin/server[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[ 50%] Built target server | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 58%] [32mBuilding CXX object examples/CMakeFiles/main.dir/main.cpp.o[0m | |
[ 66%] [32m[1mLinking CXX executable ../bin/main[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[ 66%] Built target main | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 75%] [32mBuilding CXX object examples/CMakeFiles/test_tokenizer.dir/test_tokenizer.cpp.o[0m | |
[ 83%] [32m[1mLinking CXX executable ../bin/test_tokenizer[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[ 83%] Built target test_tokenizer | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
[ 91%] [32mBuilding CXX object models/CMakeFiles/quantize.dir/quantize.cpp.o[0m | |
[100%] [32m[1mLinking CXX executable ../bin/quantize[0m | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[100%] Built target quantize | |
make[3]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[2]: Leaving directory '/build/go-bert/bert.cpp/build' | |
cp bert.cpp/build/CMakeFiles/bert.dir/bert.cpp.o bert.o | |
g++ -I. -I./bert.cpp/ggml/include/ggml/ -I./bert.cpp/ -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native gobert.cpp -o gobert.o -c | |
In file included from [01m[Kgobert.cpp:6[m[K: | |
[01m[K./bert.cpp/bert.cpp:[m[K In function '[01m[Kbert_ctx* bert_load_from_file(const char*)[m[K': | |
[01m[K./bert.cpp/bert.cpp:470:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
470 | [01;35m[K.[m[Kmem_size = model_mem_req, | |
| [01;35m[K^[m[K | |
[01m[K./bert.cpp/bert.cpp:471:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
471 | [01;35m[K.[m[Kmem_buffer = NULL, | |
| [01;35m[K^[m[K | |
[01m[K./bert.cpp/bert.cpp:472:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
472 | [01;35m[K.[m[Kno_alloc = false, | |
| [01;35m[K^[m[K | |
[01m[K./bert.cpp/bert.cpp:610:89:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 5 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
610 | fprintf(stderr, "%s: tensor '%s' has wrong shape in model file: got [[01;35m[K%lld[m[K, %lld], expected [%lld, %lld]\n", | |
| [01;35m[K~~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Klong long int[m[K | |
| [32m[K%ld[m[K | |
611 | __func__, name.data(), [32m[Ktensor->ne[0][m[K, tensor->ne[1], ne[0], ne[1]); | |
| [32m[K~~~~~~~~~~~~~[m[K | |
| [32m[K|[m[K | |
| [32m[Kint64_t {aka long int}[m[K | |
[01m[K./bert.cpp/bert.cpp:610:95:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 6 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
610 | fprintf(stderr, "%s: tensor '%s' has wrong shape in model file: got [%lld, [01;35m[K%lld[m[K], expected [%lld, %lld]\n", | |
| [01;35m[K~~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Klong long int[m[K | |
| [32m[K%ld[m[K | |
611 | __func__, name.data(), tensor->ne[0], [32m[Ktensor->ne[1][m[K, ne[0], ne[1]); | |
| [32m[K~~~~~~~~~~~~~[m[K | |
| [32m[K|[m[K | |
| [32m[Kint64_t {aka long int}[m[K | |
[01m[K./bert.cpp/bert.cpp:610:112:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 7 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
610 | fprintf(stderr, "%s: tensor '%s' has wrong shape in model file: got [%lld, %lld], expected [[01;35m[K%lld[m[K, %lld]\n", | |
| [01;35m[K~~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Klong long int[m[K | |
| [32m[K%ld[m[K | |
611 | __func__, name.data(), tensor->ne[0], tensor->ne[1], [32m[Kne[0][m[K, ne[1]); | |
| [32m[K~~~~~[m[K | |
| [32m[K|[m[K | |
| [32m[Kint64_t {aka long int}[m[K | |
[01m[K./bert.cpp/bert.cpp:610:118:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 8 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
610 | fprintf(stderr, "%s: tensor '%s' has wrong shape in model file: got [%lld, %lld], expected [%lld, [01;35m[K%lld[m[K]\n", | |
| [01;35m[K~~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Klong long int[m[K | |
| [32m[K%ld[m[K | |
611 | __func__, name.data(), tensor->ne[0], tensor->ne[1], ne[0], [32m[Kne[1][m[K); | |
| [32m[K~~~~~[m[K | |
| [32m[K|[m[K | |
| [32m[Kint64_t {aka long int}[m[K | |
[01m[K./bert.cpp/bert.cpp:624:37:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 3 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
624 | printf("%24s - [[01;35m[K%5lld[m[K, %5lld], type = %6s, %6.2f MB, %9zu bytes\n", name.data(), [32m[Kne[0][m[K, ne[1], ftype_str[ftype], ggml_nbytes(tensor) / 1024.0 / 1024.0, ggml_nbytes(tensor)); | |
| [01;35m[K~~~~^[m[K [32m[K~~~~~[m[K | |
| [01;35m[K|[m[K [32m[K|[m[K | |
| [01;35m[Klong long int[m[K [32m[Kint64_t {aka long int}[m[K | |
| [32m[K%5ld[m[K | |
[01m[K./bert.cpp/bert.cpp:624:44:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 4 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
624 | printf("%24s - [%5lld, [01;35m[K%5lld[m[K], type = %6s, %6.2f MB, %9zu bytes\n", name.data(), ne[0], [32m[Kne[1][m[K, ftype_str[ftype], ggml_nbytes(tensor) / 1024.0 / 1024.0, ggml_nbytes(tensor)); | |
| [01;35m[K~~~~^[m[K [32m[K~~~~~[m[K | |
| [01;35m[K|[m[K [32m[K|[m[K | |
| [01;35m[Klong long int[m[K [32m[Kint64_t {aka long int}[m[K | |
| [32m[K%5ld[m[K | |
[01m[K./bert.cpp/bert.cpp:655:101:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%llu[m[K' expects argument of type '[01m[Klong long unsigned int[m[K', but argument 6 has type '[01m[Klong unsigned int[m[K' [[01;35m[K-Wformat=[m[K] | |
655 | fprintf(stderr, "%s: tensor '%s' has wrong size in model file: got %zu, expected [01;35m[K%llu[m[K\n", | |
| [01;35m[K~~~^[m[K | |
| [01;35m[K|[m[K | |
| [01;35m[Klong long unsigned int[m[K | |
| [32m[K%lu[m[K | |
656 | __func__, name.data(), ggml_nbytes(tensor), [32m[Knelements * bpe[m[K); | |
| [32m[K~~~~~~~~~~~~~~~[m[K | |
| [32m[K|[m[K | |
| [32m[Klong unsigned int[m[K | |
[01m[K./bert.cpp/bert.cpp:692:56:[m[K [01;35m[Kwarning: [m[Kformat '[01m[K%lld[m[K' expects argument of type '[01m[Klong long int[m[K', but argument 4 has type '[01m[Kint64_t[m[K' {aka '[01m[Klong int[m[K'} [[01;35m[K-Wformat=[m[K] | |
692 | printf("%s: mem_per_token %zd KB, mem_per_input [01;35m[K%lld[m[K MB\n", __func__, new_bert->mem_per_token / (1 << 10), [32m[Knew_bert->mem_per_input / (1 << 20)[m[K); | |
| [01;35m[K~~~^[m[K [32m[K~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
| [01;35m[K|[m[K [32m[K|[m[K | |
| [01;35m[Klong long int[m[K [32m[Kint64_t {aka long int}[m[K | |
| [32m[K%ld[m[K | |
[01m[K./bert.cpp/bert.cpp:[m[K In function '[01m[Kvoid bert_eval_batch(bert_ctx*, int32_t, int32_t, bert_vocab_id**, int32_t*, float**)[m[K': | |
[01m[K./bert.cpp/bert.cpp:776:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
776 | [01;35m[K.[m[Kmem_size = buf_compute.size, | |
| [01;35m[K^[m[K | |
[01m[K./bert.cpp/bert.cpp:777:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
777 | [01;35m[K.[m[Kmem_buffer = buf_compute.data, | |
| [01;35m[K^[m[K | |
[01m[K./bert.cpp/bert.cpp:778:13:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
778 | [01;35m[K.[m[Kno_alloc = false, | |
| [01;35m[K^[m[K | |
[01m[Kgobert.cpp:[m[K In function '[01m[Kint bert_token_embeddings(void*, void*, int*, int, float*)[m[K': | |
[01m[Kgobert.cpp:32:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<float>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
32 | for (int i = 0; [01;35m[Ki < embeddings.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgobert.cpp:19:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
19 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[Kgobert.cpp:24:9:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[KN[m[K' [[01;35m[K-Wunused-variable[m[K] | |
24 | int [01;35m[KN[m[K = bert_n_max_tokens(bctx); | |
| [01;35m[K^[m[K | |
[01m[Kgobert.cpp:[m[K In function '[01m[Kint bert_embeddings(void*, void*, float*)[m[K': | |
[01m[Kgobert.cpp:53:23:[m[K [01;35m[Kwarning: [m[Kcomparison of integer expressions of different signedness: '[01m[Kint[m[K' and '[01m[Kstd::vector<float>::size_type[m[K' {aka '[01m[Klong unsigned int[m[K'} [[01;35m[K-Wsign-compare[m[K] | |
53 | for (int i = 0; [01;35m[Ki < embeddings.size()[m[K; i++) { | |
| [01;35m[K~~^~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kgobert.cpp:39:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kt_main_start_us[m[K' [[01;35m[K-Wunused-variable[m[K] | |
39 | const int64_t [01;35m[Kt_main_start_us[m[K = ggml_time_us(); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
cd bert.cpp/build && make VERBOSE=1 ggml && cp -rf ggml/src/CMakeFiles/ggml.dir/ggml.c.o ../../ggml.o | |
make[2]: Entering directory '/build/go-bert/bert.cpp/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-bert/bert.cpp -B/build/go-bert/bert.cpp/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
make -f CMakeFiles/Makefile2 ggml | |
make[3]: Entering directory '/build/go-bert/bert.cpp/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-bert/bert.cpp -B/build/go-bert/bert.cpp/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-bert/bert.cpp/build/CMakeFiles 2 | |
make -f CMakeFiles/Makefile2 ggml/src/CMakeFiles/ggml.dir/all | |
make[4]: Entering directory '/build/go-bert/bert.cpp/build' | |
make -f ggml/src/CMakeFiles/ggml.dir/build.make ggml/src/CMakeFiles/ggml.dir/depend | |
make[5]: Entering directory '/build/go-bert/bert.cpp/build' | |
cd /build/go-bert/bert.cpp/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-bert/bert.cpp /build/go-bert/bert.cpp/ggml/src /build/go-bert/bert.cpp/build /build/go-bert/bert.cpp/build/ggml/src /build/go-bert/bert.cpp/build/ggml/src/CMakeFiles/ggml.dir/DependInfo.cmake "--color=" | |
Dependencies file "ggml/src/CMakeFiles/ggml.dir/ggml.c.o.d" is newer than depends file "/build/go-bert/bert.cpp/build/ggml/src/CMakeFiles/ggml.dir/compiler_depend.internal". | |
[35m[1mConsolidate compiler generated dependencies of target ggml[0m | |
make[5]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make -f ggml/src/CMakeFiles/ggml.dir/build.make ggml/src/CMakeFiles/ggml.dir/build | |
make[5]: Entering directory '/build/go-bert/bert.cpp/build' | |
make[5]: Nothing to be done for 'ggml/src/CMakeFiles/ggml.dir/build'. | |
make[5]: Leaving directory '/build/go-bert/bert.cpp/build' | |
[100%] Built target ggml | |
make[4]: Leaving directory '/build/go-bert/bert.cpp/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-bert/bert.cpp/build/CMakeFiles 0 | |
make[3]: Leaving directory '/build/go-bert/bert.cpp/build' | |
make[2]: Leaving directory '/build/go-bert/bert.cpp/build' | |
ar src libgobert.a gobert.o ggml.o | |
make[1]: Leaving directory '/build/go-bert' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-bert LIBRARY_PATH=/build/go-bert \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/bert-embeddings ./cmd/grpc/bert-embeddings/ | |
make -C go-ggllm BUILD_TYPE=cublas libggllm.a | |
make[1]: Entering directory '/build/go-ggllm' | |
I ggllm.cpp build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I./ggllm.cpp -I. -O3 -DNDEBUG -std=c11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wno-unused-function -pthread -march=native -mtune=native | |
I CXXFLAGS: -I./ggllm.cpp -I. -I./ggllm.cpp/examples -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -pthread | |
I CGO_LDFLAGS: | |
I LDFLAGS: | |
I BUILD_TYPE: cublas | |
I CMAKE_ARGS: -DLLAMA_CUBLAS=ON | |
I EXTRA_TARGETS: ggllm.cpp/ggml-cuda.o | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
cd ggllm.cpp && patch -p1 < ../patches/1902-cuda.patch | |
patching file examples/falcon_common.cpp | |
patching file libfalcon.cpp | |
patching file libfalcon.h | |
touch prepare | |
mkdir -p build | |
cd build && cmake ../ggllm.cpp -DLLAMA_CUBLAS=ON && VERBOSE=1 cmake --build . --config Release && cp -rf CMakeFiles/ggml.dir/ggml.c.o ../ggllm.cpp/ggml.o | |
-- The C compiler identification is GNU 10.2.1 | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Found Git: /usr/bin/git (found version "2.30.2") | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Check if compiler accepts -pthread | |
-- Check if compiler accepts -pthread - yes | |
-- Found Threads: TRUE | |
-- Found CUDAToolkit: /usr/local/cuda/include (found version "12.1.105") | |
-- cuBLAS found | |
-- The CUDA compiler identification is NVIDIA 12.1.105 | |
-- Detecting CUDA compiler ABI info | |
-- Detecting CUDA compiler ABI info - done | |
-- Check for working CUDA compiler: /usr/local/cuda/bin/nvcc - skipped | |
-- Detecting CUDA compile features | |
-- Detecting CUDA compile features - done | |
-- CUDA found, version: 12.1.105 | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
-- GGML CUDA sources found, configuring CUDA architecture | |
-- Configuring done (1.7s) | |
[33mCMake Warning (dev) in CMakeLists.txt: | |
Policy CMP0104 is not set: CMAKE_CUDA_ARCHITECTURES now detected for NVCC, | |
empty CUDA_ARCHITECTURES not allowed. Run "cmake --help-policy CMP0104" | |
for policy details. Use the cmake_policy command to set the policy and | |
suppress this warning. | |
CUDA_ARCHITECTURES is empty for target "ggml_static". | |
This warning is for project developers. Use -Wno-dev to suppress it. | |
[0m | |
[33mCMake Warning (dev) in CMakeLists.txt: | |
Policy CMP0104 is not set: CMAKE_CUDA_ARCHITECTURES now detected for NVCC, | |
empty CUDA_ARCHITECTURES not allowed. Run "cmake --help-policy CMP0104" | |
for policy details. Use the cmake_policy command to set the policy and | |
suppress this warning. | |
CUDA_ARCHITECTURES is empty for target "ggml_static". | |
This warning is for project developers. Use -Wno-dev to suppress it. | |
[0m | |
-- Generating done (0.0s) | |
-- Build files have been written to: /build/go-ggllm/build | |
Change Dir: '/build/go-ggllm/build' | |
Run Build Command(s): /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E env VERBOSE=1 /usr/bin/gmake -f Makefile | |
gmake[2]: Entering directory '/build/go-ggllm/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-ggllm/ggllm.cpp -B/build/go-ggllm/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-ggllm/build/CMakeFiles /build/go-ggllm/build//CMakeFiles/progress.marks | |
/usr/bin/gmake -f CMakeFiles/Makefile2 all | |
gmake[3]: Entering directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/BUILD_INFO.dir/build.make CMakeFiles/BUILD_INFO.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/BUILD_INFO.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/BUILD_INFO.dir/build.make CMakeFiles/BUILD_INFO.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
gmake[4]: Nothing to be done for 'CMakeFiles/BUILD_INFO.dir/build'. | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 2%] Built target BUILD_INFO | |
/usr/bin/gmake -f CMakeFiles/ggml.dir/build.make CMakeFiles/ggml.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/ggml.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/ggml.dir/build.make CMakeFiles/ggml.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 4%] [32mBuilding C object CMakeFiles/ggml.dir/ggml.c.o[0m | |
/usr/bin/cc -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-ggllm/ggllm.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu11 -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/ggml.dir/ggml.c.o -MF CMakeFiles/ggml.dir/ggml.c.o.d -o CMakeFiles/ggml.dir/ggml.c.o -c /build/go-ggllm/ggllm.cpp/ggml.c | |
[01m[K/build/go-ggllm/ggllm.cpp/ggml.c:[m[K In function ‘[01m[Kggml_compute_forward_mul_mat_f32[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/ggml.c:10924:19:[m[K [01;35m[Kwarning: [m[Kunused variable ‘[01m[Kne10[m[K’ [[01;35m[K-Wunused-variable[m[K] | |
10924 | const int64_t [01;35m[Kne10[m[K = src1->ne[0]; | |
| [01;35m[K^~~~[m[K | |
[ 6%] [32mBuilding CUDA object CMakeFiles/ggml.dir/ggml-cuda.cu.o[0m | |
/usr/local/cuda/bin/nvcc -forward-unknown-to-host-compiler -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 --options-file CMakeFiles/ggml.dir/includes_CUDA.rsp -O3 -DNDEBUG -std=c++11 -mf16c -mfma -mavx -mavx2 -Xcompiler -pthread -MD -MT CMakeFiles/ggml.dir/ggml-cuda.cu.o -MF CMakeFiles/ggml.dir/ggml-cuda.cu.o.d -x cu -c /build/go-ggllm/ggllm.cpp/ggml-cuda.cu -o CMakeFiles/ggml.dir/ggml-cuda.cu.o | |
[ 8%] [32mBuilding C object CMakeFiles/ggml.dir/k_quants.c.o[0m | |
/usr/bin/cc -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-ggllm/ggllm.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu11 -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/ggml.dir/k_quants.c.o -MF CMakeFiles/ggml.dir/k_quants.c.o.d -o CMakeFiles/ggml.dir/k_quants.c.o -c /build/go-ggllm/ggllm.cpp/k_quants.c | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 8%] Built target ggml | |
/usr/bin/gmake -f CMakeFiles/ggml_static.dir/build.make CMakeFiles/ggml_static.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/ggml_static.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/ggml_static.dir/build.make CMakeFiles/ggml_static.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 10%] [32m[1mLinking CUDA static library libggml_static.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/ggml_static.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/ggml_static.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libggml_static.a CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-cuda.cu.o" CMakeFiles/ggml.dir/k_quants.c.o | |
/usr/bin/ranlib libggml_static.a | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 10%] Built target ggml_static | |
/usr/bin/gmake -f CMakeFiles/llama.dir/build.make CMakeFiles/llama.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/llama.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/llama.dir/build.make CMakeFiles/llama.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 12%] [32mBuilding CXX object CMakeFiles/llama.dir/llama.cpp.o[0m | |
/usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-ggllm/ggllm.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/llama.dir/llama.cpp.o -MF CMakeFiles/llama.dir/llama.cpp.o.d -o CMakeFiles/llama.dir/llama.cpp.o -c /build/go-ggllm/ggllm.cpp/llama.cpp | |
[ 14%] [32m[1mLinking CXX static library libllama.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama.a CMakeFiles/llama.dir/llama.cpp.o CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-cuda.cu.o" CMakeFiles/ggml.dir/k_quants.c.o | |
/usr/bin/ranlib libllama.a | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 14%] Built target llama | |
/usr/bin/gmake -f CMakeFiles/cmpnct_unicode.dir/build.make CMakeFiles/cmpnct_unicode.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/cmpnct_unicode.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/cmpnct_unicode.dir/build.make CMakeFiles/cmpnct_unicode.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 16%] [32mBuilding CXX object CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o[0m | |
/usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o -MF CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o.d -o CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o -c /build/go-ggllm/ggllm.cpp/cmpnct_unicode.cpp | |
[ 18%] [32m[1mLinking CXX static library libcmpnct_unicode.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/cmpnct_unicode.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/cmpnct_unicode.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libcmpnct_unicode.a CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o | |
/usr/bin/ranlib libcmpnct_unicode.a | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 18%] Built target cmpnct_unicode | |
/usr/bin/gmake -f CMakeFiles/falcon.dir/build.make CMakeFiles/falcon.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp /build/go-ggllm/build /build/go-ggllm/build /build/go-ggllm/build/CMakeFiles/falcon.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f CMakeFiles/falcon.dir/build.make CMakeFiles/falcon.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 20%] [32mBuilding CXX object CMakeFiles/falcon.dir/libfalcon.cpp.o[0m | |
/usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-ggllm/ggllm.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/falcon.dir/libfalcon.cpp.o -MF CMakeFiles/falcon.dir/libfalcon.cpp.o.d -o CMakeFiles/falcon.dir/libfalcon.cpp.o -c /build/go-ggllm/ggllm.cpp/libfalcon.cpp | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:[m[K In function ‘[01m[Kbool falcon_eval_internal(falcon_context&, const falcon_token*, int, int, int, const char*, int)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:2076:20:[m[K [01;35m[Kwarning: [m[Kvariable ‘[01m[Koffload_func_nr[m[K’ set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
2076 | offload_func_t [01;35m[Koffload_func_nr[m[K = llama_nop; // nr = non-repeating | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:2077:20:[m[K [01;35m[Kwarning: [m[Kvariable ‘[01m[Koffload_func_kqv[m[K’ set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
2077 | offload_func_t [01;35m[Koffload_func_kqv[m[K = llama_nop; | |
| [01;35m[K^~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:2316:20:[m[K [01;35m[Kwarning: [m[Kvariable ‘[01m[Koffload_func[m[K’ set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
2316 | offload_func_t [01;35m[Koffload_func[m[K = llama_nop; | |
| [01;35m[K^~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:[m[K In function ‘[01m[Ksize_t falcon_copy_state_data(falcon_context*, uint8_t*)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:4114:22:[m[K [01;35m[Kwarning: [m[Kunused variable ‘[01m[Kn_embd[m[K’ [[01;35m[K-Wunused-variable[m[K] | |
4114 | const int [01;35m[Kn_embd[m[K = hparams.n_embd; | |
| [01;35m[K^~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:[m[K In function ‘[01m[Ksize_t falcon_set_state_data(falcon_context*, uint8_t*)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/libfalcon.cpp:4230:22:[m[K [01;35m[Kwarning: [m[Kunused variable ‘[01m[Kn_embd[m[K’ [[01;35m[K-Wunused-variable[m[K] | |
4230 | const int [01;35m[Kn_embd[m[K = hparams.n_embd; | |
| [01;35m[K^~~~~~[m[K | |
[ 22%] [32m[1mLinking CXX static library libfalcon.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/falcon.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libfalcon.a CMakeFiles/falcon.dir/libfalcon.cpp.o CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-cuda.cu.o" CMakeFiles/ggml.dir/k_quants.c.o | |
/usr/bin/ranlib libfalcon.a | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 22%] Built target falcon | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-fns.dir/build.make tests/CMakeFiles/test-quantize-fns.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/tests /build/go-ggllm/build /build/go-ggllm/build/tests /build/go-ggllm/build/tests/CMakeFiles/test-quantize-fns.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-fns.dir/build.make tests/CMakeFiles/test-quantize-fns.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 24%] [32mBuilding CXX object tests/CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o[0m | |
cd /build/go-ggllm/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o -MF CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o.d -o CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o -c /build/go-ggllm/ggllm.cpp/tests/test-quantize-fns.cpp | |
[ 26%] [32m[1mLinking CXX executable ../bin/test-quantize-fns[0m | |
cd /build/go-ggllm/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-quantize-fns.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o" -o ../bin/test-quantize-fns -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 26%] Built target test-quantize-fns | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-perf.dir/build.make tests/CMakeFiles/test-quantize-perf.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/tests /build/go-ggllm/build /build/go-ggllm/build/tests /build/go-ggllm/build/tests/CMakeFiles/test-quantize-perf.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-perf.dir/build.make tests/CMakeFiles/test-quantize-perf.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 28%] [32mBuilding CXX object tests/CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o[0m | |
cd /build/go-ggllm/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o -MF CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o.d -o CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o -c /build/go-ggllm/ggllm.cpp/tests/test-quantize-perf.cpp | |
[ 30%] [32m[1mLinking CXX executable ../bin/test-quantize-perf[0m | |
cd /build/go-ggllm/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-quantize-perf.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o" -o ../bin/test-quantize-perf -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 30%] Built target test-quantize-perf | |
/usr/bin/gmake -f tests/CMakeFiles/test-sampling.dir/build.make tests/CMakeFiles/test-sampling.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/tests /build/go-ggllm/build /build/go-ggllm/build/tests /build/go-ggllm/build/tests/CMakeFiles/test-sampling.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-sampling.dir/build.make tests/CMakeFiles/test-sampling.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 32%] [32mBuilding CXX object tests/CMakeFiles/test-sampling.dir/test-sampling.cpp.o[0m | |
cd /build/go-ggllm/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-sampling.dir/test-sampling.cpp.o -MF CMakeFiles/test-sampling.dir/test-sampling.cpp.o.d -o CMakeFiles/test-sampling.dir/test-sampling.cpp.o -c /build/go-ggllm/ggllm.cpp/tests/test-sampling.cpp | |
In file included from [01m[K/build/go-ggllm/ggllm.cpp/tests/test-sampling.cpp:2[m[K: | |
[01m[K/build/go-ggllm/ggllm.cpp/./libfalcon.h:252:24:[m[K [01;35m[Kwarning: [m[K‘[01m[KFINETUNE_NAME[m[K’ defined but not used [[01;35m[K-Wunused-variable[m[K] | |
252 | static const char *[01;35m[KFINETUNE_NAME[m[K[6] = { "UNSPECIFIED", "NONE", "ALPACA", "OPENASSISTANT", "WIZARD", "FALCONINSTRUCT" }; | |
| [01;35m[K^~~~~~~~~~~~~[m[K | |
[ 34%] [32m[1mLinking CXX executable ../bin/test-sampling[0m | |
cd /build/go-ggllm/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-sampling.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-sampling.dir/test-sampling.cpp.o" -o ../bin/test-sampling -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 34%] Built target test-sampling | |
/usr/bin/gmake -f examples/CMakeFiles/common.dir/build.make examples/CMakeFiles/common.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples /build/go-ggllm/build /build/go-ggllm/build/examples /build/go-ggllm/build/examples/CMakeFiles/common.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/CMakeFiles/common.dir/build.make examples/CMakeFiles/common.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 36%] [32mBuilding CXX object examples/CMakeFiles/common.dir/common.cpp.o[0m | |
cd /build/go-ggllm/build/examples && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/CMakeFiles/common.dir/common.cpp.o -MF CMakeFiles/common.dir/common.cpp.o.d -o CMakeFiles/common.dir/common.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/common.cpp | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 36%] Built target common | |
/usr/bin/gmake -f examples/CMakeFiles/falcon_common.dir/build.make examples/CMakeFiles/falcon_common.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples /build/go-ggllm/build /build/go-ggllm/build/examples /build/go-ggllm/build/examples/CMakeFiles/falcon_common.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/CMakeFiles/falcon_common.dir/build.make examples/CMakeFiles/falcon_common.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 38%] [32mBuilding CXX object examples/CMakeFiles/falcon_common.dir/falcon_common.cpp.o[0m | |
cd /build/go-ggllm/build/examples && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/CMakeFiles/falcon_common.dir/falcon_common.cpp.o -MF CMakeFiles/falcon_common.dir/falcon_common.cpp.o.d -o CMakeFiles/falcon_common.dir/falcon_common.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/falcon_common.cpp | |
In file included from [01m[K/build/go-ggllm/ggllm.cpp/examples/falcon_common.h:6[m[K, | |
from [01m[K/build/go-ggllm/ggllm.cpp/examples/falcon_common.cpp:1[m[K: | |
[01m[K/build/go-ggllm/ggllm.cpp/./libfalcon.h:252:24:[m[K [01;35m[Kwarning: [m[K‘[01m[KFINETUNE_NAME[m[K’ defined but not used [[01;35m[K-Wunused-variable[m[K] | |
252 | static const char *[01;35m[KFINETUNE_NAME[m[K[6] = { "UNSPECIFIED", "NONE", "ALPACA", "OPENASSISTANT", "WIZARD", "FALCONINSTRUCT" }; | |
| [01;35m[K^~~~~~~~~~~~~[m[K | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 38%] Built target falcon_common | |
/usr/bin/gmake -f examples/main/CMakeFiles/main.dir/build.make examples/main/CMakeFiles/main.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/main /build/go-ggllm/build /build/go-ggllm/build/examples/main /build/go-ggllm/build/examples/main/CMakeFiles/main.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/main/CMakeFiles/main.dir/build.make examples/main/CMakeFiles/main.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 40%] [32mBuilding CXX object examples/main/CMakeFiles/main.dir/main.cpp.o[0m | |
cd /build/go-ggllm/build/examples/main && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/main/CMakeFiles/main.dir/main.cpp.o -MF CMakeFiles/main.dir/main.cpp.o.d -o CMakeFiles/main.dir/main.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/main/main.cpp | |
[ 42%] [32m[1mLinking CXX executable ../../bin/main[0m | |
cd /build/go-ggllm/build/examples/main && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/main.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/main.dir/main.cpp.o ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/main -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 42%] Built target main | |
/usr/bin/gmake -f examples/falcon/CMakeFiles/falcon_main.dir/build.make examples/falcon/CMakeFiles/falcon_main.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/falcon /build/go-ggllm/build /build/go-ggllm/build/examples/falcon /build/go-ggllm/build/examples/falcon/CMakeFiles/falcon_main.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/falcon/CMakeFiles/falcon_main.dir/build.make examples/falcon/CMakeFiles/falcon_main.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 44%] [32mBuilding CXX object examples/falcon/CMakeFiles/falcon_main.dir/falcon_main.cpp.o[0m | |
cd /build/go-ggllm/build/examples/falcon && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/falcon/CMakeFiles/falcon_main.dir/falcon_main.cpp.o -MF CMakeFiles/falcon_main.dir/falcon_main.cpp.o.d -o CMakeFiles/falcon_main.dir/falcon_main.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/falcon/falcon_main.cpp | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/falcon/falcon_main.cpp:[m[K In function ‘[01m[Kint main(int, char**)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/falcon/falcon_main.cpp:963:27:[m[K [01;35m[Kwarning: [m[Ksuggest parentheses around ‘[01m[K&&[m[K’ within ‘[01m[K||[m[K’ [[01;35m[K-Wparentheses[m[K] | |
963 | if ([01;35m[K!embd.empty() && embd.back() == falcon_token_eos()[m[K || stopword_fulfilled) | |
| [01;35m[K~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/falcon/falcon_main.cpp:433:10:[m[K [01;35m[Kwarning: [m[Kunused variable ‘[01m[Kfalcon_token_newline[m[K’ [[01;35m[K-Wunused-variable[m[K] | |
433 | auto [01;35m[Kfalcon_token_newline[m[K = falcon_token_nl(); | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[ 46%] [32m[1mLinking CXX executable ../../bin/falcon_main[0m | |
cd /build/go-ggllm/build/examples/falcon && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon_main.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/falcon_main.dir/falcon_main.cpp.o ../CMakeFiles/falcon_common.dir/falcon_common.cpp.o -o ../../bin/falcon_main -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libfalcon.a -pthread ../../libcmpnct_unicode.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 46%] Built target falcon_main | |
/usr/bin/gmake -f examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/build.make examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/falcon_quantize /build/go-ggllm/build /build/go-ggllm/build/examples/falcon_quantize /build/go-ggllm/build/examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/build.make examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 48%] [32mBuilding CXX object examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/quantize.cpp.o[0m | |
cd /build/go-ggllm/build/examples/falcon_quantize && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/falcon_quantize/CMakeFiles/falcon_quantize.dir/quantize.cpp.o -MF CMakeFiles/falcon_quantize.dir/quantize.cpp.o.d -o CMakeFiles/falcon_quantize.dir/quantize.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/falcon_quantize/quantize.cpp | |
In file included from [01m[K/build/go-ggllm/ggllm.cpp/examples/falcon_quantize/quantize.cpp:3[m[K: | |
[01m[K/build/go-ggllm/ggllm.cpp/./libfalcon.h:252:24:[m[K [01;35m[Kwarning: [m[K‘[01m[KFINETUNE_NAME[m[K’ defined but not used [[01;35m[K-Wunused-variable[m[K] | |
252 | static const char *[01;35m[KFINETUNE_NAME[m[K[6] = { "UNSPECIFIED", "NONE", "ALPACA", "OPENASSISTANT", "WIZARD", "FALCONINSTRUCT" }; | |
| [01;35m[K^~~~~~~~~~~~~[m[K | |
[ 51%] [32m[1mLinking CXX executable ../../bin/falcon_quantize[0m | |
cd /build/go-ggllm/build/examples/falcon_quantize && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon_quantize.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/falcon_quantize.dir/quantize.cpp.o -o ../../bin/falcon_quantize -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libfalcon.a -pthread ../../libcmpnct_unicode.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 51%] Built target falcon_quantize | |
/usr/bin/gmake -f examples/quantize/CMakeFiles/quantize.dir/build.make examples/quantize/CMakeFiles/quantize.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/quantize /build/go-ggllm/build /build/go-ggllm/build/examples/quantize /build/go-ggllm/build/examples/quantize/CMakeFiles/quantize.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/quantize/CMakeFiles/quantize.dir/build.make examples/quantize/CMakeFiles/quantize.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 53%] [32mBuilding CXX object examples/quantize/CMakeFiles/quantize.dir/quantize.cpp.o[0m | |
cd /build/go-ggllm/build/examples/quantize && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/quantize/CMakeFiles/quantize.dir/quantize.cpp.o -MF CMakeFiles/quantize.dir/quantize.cpp.o.d -o CMakeFiles/quantize.dir/quantize.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/quantize/quantize.cpp | |
[ 55%] [32m[1mLinking CXX executable ../../bin/quantize[0m | |
cd /build/go-ggllm/build/examples/quantize && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/quantize.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/quantize.dir/quantize.cpp.o -o ../../bin/quantize -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 55%] Built target quantize | |
/usr/bin/gmake -f examples/quantize-stats/CMakeFiles/quantize-stats.dir/build.make examples/quantize-stats/CMakeFiles/quantize-stats.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/quantize-stats /build/go-ggllm/build /build/go-ggllm/build/examples/quantize-stats /build/go-ggllm/build/examples/quantize-stats/CMakeFiles/quantize-stats.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/quantize-stats/CMakeFiles/quantize-stats.dir/build.make examples/quantize-stats/CMakeFiles/quantize-stats.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 57%] [32mBuilding CXX object examples/quantize-stats/CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o[0m | |
cd /build/go-ggllm/build/examples/quantize-stats && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/quantize-stats/CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o -MF CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o.d -o CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/quantize-stats/quantize-stats.cpp | |
[ 59%] [32m[1mLinking CXX executable ../../bin/quantize-stats[0m | |
cd /build/go-ggllm/build/examples/quantize-stats && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/quantize-stats.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o" -o ../../bin/quantize-stats -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 59%] Built target quantize-stats | |
/usr/bin/gmake -f examples/perplexity/CMakeFiles/perplexity.dir/build.make examples/perplexity/CMakeFiles/perplexity.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/perplexity /build/go-ggllm/build /build/go-ggllm/build/examples/perplexity /build/go-ggllm/build/examples/perplexity/CMakeFiles/perplexity.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/perplexity/CMakeFiles/perplexity.dir/build.make examples/perplexity/CMakeFiles/perplexity.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 61%] [32mBuilding CXX object examples/perplexity/CMakeFiles/perplexity.dir/perplexity.cpp.o[0m | |
cd /build/go-ggllm/build/examples/perplexity && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/perplexity/CMakeFiles/perplexity.dir/perplexity.cpp.o -MF CMakeFiles/perplexity.dir/perplexity.cpp.o.d -o CMakeFiles/perplexity.dir/perplexity.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/perplexity/perplexity.cpp | |
[ 63%] [32m[1mLinking CXX executable ../../bin/perplexity[0m | |
cd /build/go-ggllm/build/examples/perplexity && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/perplexity.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/perplexity.dir/perplexity.cpp.o ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/perplexity -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 63%] Built target perplexity | |
/usr/bin/gmake -f examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/build.make examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/falcon_perplexity /build/go-ggllm/build /build/go-ggllm/build/examples/falcon_perplexity /build/go-ggllm/build/examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/build.make examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 65%] [32mBuilding CXX object examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/falcon_perplexity.cpp.o[0m | |
cd /build/go-ggllm/build/examples/falcon_perplexity && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/falcon_perplexity/CMakeFiles/falcon_perplexity.dir/falcon_perplexity.cpp.o -MF CMakeFiles/falcon_perplexity.dir/falcon_perplexity.cpp.o.d -o CMakeFiles/falcon_perplexity.dir/falcon_perplexity.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/falcon_perplexity/falcon_perplexity.cpp | |
In file included from [01m[K/build/go-ggllm/ggllm.cpp/examples/falcon_common.h:6[m[K, | |
from [01m[K/build/go-ggllm/ggllm.cpp/examples/falcon_perplexity/falcon_perplexity.cpp:1[m[K: | |
[01m[K/build/go-ggllm/ggllm.cpp/./libfalcon.h:252:24:[m[K [01;35m[Kwarning: [m[K‘[01m[KFINETUNE_NAME[m[K’ defined but not used [[01;35m[K-Wunused-variable[m[K] | |
252 | static const char *[01;35m[KFINETUNE_NAME[m[K[6] = { "UNSPECIFIED", "NONE", "ALPACA", "OPENASSISTANT", "WIZARD", "FALCONINSTRUCT" }; | |
| [01;35m[K^~~~~~~~~~~~~[m[K | |
[ 67%] [32m[1mLinking CXX executable ../../bin/falcon_perplexity[0m | |
cd /build/go-ggllm/build/examples/falcon_perplexity && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon_perplexity.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/falcon_perplexity.dir/falcon_perplexity.cpp.o ../CMakeFiles/falcon_common.dir/falcon_common.cpp.o -o ../../bin/falcon_perplexity -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libfalcon.a -pthread ../../libcmpnct_unicode.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 67%] Built target falcon_perplexity | |
/usr/bin/gmake -f examples/embedding/CMakeFiles/embedding.dir/build.make examples/embedding/CMakeFiles/embedding.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/embedding /build/go-ggllm/build /build/go-ggllm/build/examples/embedding /build/go-ggllm/build/examples/embedding/CMakeFiles/embedding.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/embedding/CMakeFiles/embedding.dir/build.make examples/embedding/CMakeFiles/embedding.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 69%] [32mBuilding CXX object examples/embedding/CMakeFiles/embedding.dir/embedding.cpp.o[0m | |
cd /build/go-ggllm/build/examples/embedding && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/embedding/CMakeFiles/embedding.dir/embedding.cpp.o -MF CMakeFiles/embedding.dir/embedding.cpp.o.d -o CMakeFiles/embedding.dir/embedding.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/embedding/embedding.cpp | |
[ 71%] [32m[1mLinking CXX executable ../../bin/embedding[0m | |
cd /build/go-ggllm/build/examples/embedding && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/embedding.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/embedding.dir/embedding.cpp.o ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/embedding -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 71%] Built target embedding | |
/usr/bin/gmake -f examples/save-load-state/CMakeFiles/save-load-state.dir/build.make examples/save-load-state/CMakeFiles/save-load-state.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/save-load-state /build/go-ggllm/build /build/go-ggllm/build/examples/save-load-state /build/go-ggllm/build/examples/save-load-state/CMakeFiles/save-load-state.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/save-load-state/CMakeFiles/save-load-state.dir/build.make examples/save-load-state/CMakeFiles/save-load-state.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 73%] [32mBuilding CXX object examples/save-load-state/CMakeFiles/save-load-state.dir/save-load-state.cpp.o[0m | |
cd /build/go-ggllm/build/examples/save-load-state && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/save-load-state/CMakeFiles/save-load-state.dir/save-load-state.cpp.o -MF CMakeFiles/save-load-state.dir/save-load-state.cpp.o.d -o CMakeFiles/save-load-state.dir/save-load-state.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/save-load-state/save-load-state.cpp | |
[ 75%] [32m[1mLinking CXX executable ../../bin/save-load-state[0m | |
cd /build/go-ggllm/build/examples/save-load-state && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/save-load-state.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/save-load-state.dir/save-load-state.cpp.o" ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/save-load-state -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 75%] Built target save-load-state | |
/usr/bin/gmake -f examples/benchmark/CMakeFiles/benchmark.dir/build.make examples/benchmark/CMakeFiles/benchmark.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/benchmark /build/go-ggllm/build /build/go-ggllm/build/examples/benchmark /build/go-ggllm/build/examples/benchmark/CMakeFiles/benchmark.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/benchmark/CMakeFiles/benchmark.dir/build.make examples/benchmark/CMakeFiles/benchmark.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 77%] [32mBuilding CXX object examples/benchmark/CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o[0m | |
cd /build/go-ggllm/build/examples/benchmark && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/benchmark/CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o -MF CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o.d -o CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/benchmark/benchmark-matmult.cpp | |
[ 79%] [32m[1mLinking CXX executable ../../bin/benchmark[0m | |
cd /build/go-ggllm/build/examples/benchmark && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/benchmark.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o" ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/benchmark -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 79%] Built target benchmark | |
/usr/bin/gmake -f examples/baby-llama/CMakeFiles/baby-llama.dir/build.make examples/baby-llama/CMakeFiles/baby-llama.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/baby-llama /build/go-ggllm/build /build/go-ggllm/build/examples/baby-llama /build/go-ggllm/build/examples/baby-llama/CMakeFiles/baby-llama.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/baby-llama/CMakeFiles/baby-llama.dir/build.make examples/baby-llama/CMakeFiles/baby-llama.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 81%] [32mBuilding CXX object examples/baby-llama/CMakeFiles/baby-llama.dir/baby-llama.cpp.o[0m | |
cd /build/go-ggllm/build/examples/baby-llama && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/baby-llama/CMakeFiles/baby-llama.dir/baby-llama.cpp.o -MF CMakeFiles/baby-llama.dir/baby-llama.cpp.o.d -o CMakeFiles/baby-llama.dir/baby-llama.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/baby-llama/baby-llama.cpp | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/baby-llama/baby-llama.cpp:[m[K In function ‘[01m[Kint main(int, char**)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/baby-llama/baby-llama.cpp:1602:32:[m[K [01;35m[Kwarning: [m[Kvariable ‘[01m[Kopt_params_adam[m[K’ set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
1602 | struct ggml_opt_params [01;35m[Kopt_params_adam[m[K = ggml_opt_default_params(GGML_OPT_ADAM); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[ 83%] [32m[1mLinking CXX executable ../../bin/baby-llama[0m | |
cd /build/go-ggllm/build/examples/baby-llama && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/baby-llama.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/baby-llama.dir/baby-llama.cpp.o" ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/baby-llama -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 83%] Built target baby-llama | |
/usr/bin/gmake -f examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build.make examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch /build/go-ggllm/build /build/go-ggllm/build/examples/train-text-from-scratch /build/go-ggllm/build/examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build.make examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 85%] [32mBuilding CXX object examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o[0m | |
cd /build/go-ggllm/build/examples/train-text-from-scratch && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o -MF CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o.d -o CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:[m[K In function ‘[01m[Kvoid write_tensor(llama_file*, ggml_tensor*)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:2371:21:[m[K [01;35m[Kwarning: [m[Ksuggest parentheses around ‘[01m[K-[m[K’ in operand of ‘[01m[K&[m[K’ [[01;35m[K-Wparentheses[m[K] | |
2371 | file->seek([01;35m[K0-file->tell()[m[K & 31, SEEK_CUR); | |
| [01;35m[K~^~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:2386:17:[m[K [01;35m[Kwarning: [m[Ksuggest parentheses around ‘[01m[K-[m[K’ in operand of ‘[01m[K&[m[K’ [[01;35m[K-Wparentheses[m[K] | |
2386 | file->seek([01;35m[K0-file->tell()[m[K & 31, SEEK_CUR); | |
| [01;35m[K~^~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:[m[K In function ‘[01m[Kvoid read_tensor(llama_file*, ggml_tensor*)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:2407:17:[m[K [01;35m[Kwarning: [m[Ksuggest parentheses around ‘[01m[K-[m[K’ in operand of ‘[01m[K&[m[K’ [[01;35m[K-Wparentheses[m[K] | |
2407 | file->seek([01;35m[K0-file->tell()[m[K & 31, SEEK_CUR); | |
| [01;35m[K~^~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:[m[K In function ‘[01m[Kvoid init_model(my_llama_model*)[m[K’: | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:305:16:[m[K [01;35m[Kwarning: [m[K‘[01m[Kchar* strncpy(char*, const char*, size_t)[m[K’ specified bound 64 equals destination size [[01;35m[K-Wstringop-truncation[m[K] | |
305 | [01;35m[Kstrncpy(layer.w1->name, (layers_i + ".feed_forward.w1.weight").c_str(), sizeof(layer.w1->name))[m[K; | |
| [01;35m[K~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:306:16:[m[K [01;35m[Kwarning: [m[K‘[01m[Kchar* strncpy(char*, const char*, size_t)[m[K’ specified bound 64 equals destination size [[01;35m[K-Wstringop-truncation[m[K] | |
306 | [01;35m[Kstrncpy(layer.w2->name, (layers_i + ".feed_forward.w2.weight").c_str(), sizeof(layer.w2->name))[m[K; | |
| [01;35m[K~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-ggllm/ggllm.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp:307:16:[m[K [01;35m[Kwarning: [m[K‘[01m[Kchar* strncpy(char*, const char*, size_t)[m[K’ specified bound 64 equals destination size [[01;35m[K-Wstringop-truncation[m[K] | |
307 | [01;35m[Kstrncpy(layer.w3->name, (layers_i + ".feed_forward.w3.weight").c_str(), sizeof(layer.w3->name))[m[K; | |
| [01;35m[K~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[ 87%] [32m[1mLinking CXX executable ../../bin/train-text-from-scratch[0m | |
cd /build/go-ggllm/build/examples/train-text-from-scratch && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/train-text-from-scratch.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o" ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/train-text-from-scratch -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 87%] Built target train-text-from-scratch | |
/usr/bin/gmake -f examples/simple/CMakeFiles/simple.dir/build.make examples/simple/CMakeFiles/simple.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/examples/simple /build/go-ggllm/build /build/go-ggllm/build/examples/simple /build/go-ggllm/build/examples/simple/CMakeFiles/simple.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f examples/simple/CMakeFiles/simple.dir/build.make examples/simple/CMakeFiles/simple.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 89%] [32mBuilding CXX object examples/simple/CMakeFiles/simple.dir/simple.cpp.o[0m | |
cd /build/go-ggllm/build/examples/simple && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/examples -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/simple/CMakeFiles/simple.dir/simple.cpp.o -MF CMakeFiles/simple.dir/simple.cpp.o.d -o CMakeFiles/simple.dir/simple.cpp.o -c /build/go-ggllm/ggllm.cpp/examples/simple/simple.cpp | |
[ 91%] [32m[1mLinking CXX executable ../../bin/simple[0m | |
cd /build/go-ggllm/build/examples/simple && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/simple.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/simple.dir/simple.cpp.o ../CMakeFiles/common.dir/common.cpp.o -o ../../bin/simple -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 91%] Built target simple | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/vdot.dir/build.make pocs/vdot/CMakeFiles/vdot.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/pocs/vdot /build/go-ggllm/build /build/go-ggllm/build/pocs/vdot /build/go-ggllm/build/pocs/vdot/CMakeFiles/vdot.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/vdot.dir/build.make pocs/vdot/CMakeFiles/vdot.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 93%] [32mBuilding CXX object pocs/vdot/CMakeFiles/vdot.dir/vdot.cpp.o[0m | |
cd /build/go-ggllm/build/pocs/vdot && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/pocs -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT pocs/vdot/CMakeFiles/vdot.dir/vdot.cpp.o -MF CMakeFiles/vdot.dir/vdot.cpp.o.d -o CMakeFiles/vdot.dir/vdot.cpp.o -c /build/go-ggllm/ggllm.cpp/pocs/vdot/vdot.cpp | |
[ 95%] [32m[1mLinking CXX executable ../../bin/vdot[0m | |
cd /build/go-ggllm/build/pocs/vdot && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/vdot.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/vdot.dir/vdot.cpp.o ../../examples/CMakeFiles/common.dir/common.cpp.o -o ../../bin/vdot -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[ 95%] Built target vdot | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/q8dot.dir/build.make pocs/vdot/CMakeFiles/q8dot.dir/depend | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
cd /build/go-ggllm/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-ggllm/ggllm.cpp /build/go-ggllm/ggllm.cpp/pocs/vdot /build/go-ggllm/build /build/go-ggllm/build/pocs/vdot /build/go-ggllm/build/pocs/vdot/CMakeFiles/q8dot.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/q8dot.dir/build.make pocs/vdot/CMakeFiles/q8dot.dir/build | |
gmake[4]: Entering directory '/build/go-ggllm/build' | |
[ 97%] [32mBuilding CXX object pocs/vdot/CMakeFiles/q8dot.dir/q8dot.cpp.o[0m | |
cd /build/go-ggllm/build/pocs/vdot && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_DMMV_Y=1 -DGGML_PERF=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/usr/local/cuda/include -I/build/go-ggllm/ggllm.cpp/pocs -I/build/go-ggllm/ggllm.cpp/examples/. -I/build/go-ggllm/ggllm.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT pocs/vdot/CMakeFiles/q8dot.dir/q8dot.cpp.o -MF CMakeFiles/q8dot.dir/q8dot.cpp.o.d -o CMakeFiles/q8dot.dir/q8dot.cpp.o -c /build/go-ggllm/ggllm.cpp/pocs/vdot/q8dot.cpp | |
[100%] [32m[1mLinking CXX executable ../../bin/q8dot[0m | |
cd /build/go-ggllm/build/pocs/vdot && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/q8dot.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/q8dot.dir/q8dot.cpp.o ../../examples/CMakeFiles/common.dir/common.cpp.o -o ../../bin/q8dot -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-ggllm/build' | |
[100%] Built target q8dot | |
gmake[3]: Leaving directory '/build/go-ggllm/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-ggllm/build/CMakeFiles 0 | |
gmake[2]: Leaving directory '/build/go-ggllm/build' | |
cd build && cp -rf CMakeFiles/cmpnct_unicode.dir/cmpnct_unicode.cpp.o ../ggllm.cpp/cmpnct_unicode.o | |
cd build && cp -rf CMakeFiles/llama.dir/llama.cpp.o ../ggllm.cpp/llama.o | |
cd build && cp -rf CMakeFiles/falcon.dir/libfalcon.cpp.o ../ggllm.cpp/libfalcon.o | |
cd build && cp -rf examples/CMakeFiles/falcon_common.dir/falcon_common.cpp.o ../ggllm.cpp/falcon_common.o | |
g++ -I./ggllm.cpp -I. -I./ggllm.cpp/examples -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -pthread -I./ggllm.cpp -I./ggllm.cpp/examples falcon_binding.cpp -o falcon_binding.o -c | |
[01m[Kfalcon_binding.cpp:[m[K In function '[01m[Kint falcon_predict(void*, void*, char*, bool)[m[K': | |
[01m[Kfalcon_binding.cpp:468:48:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst char*[m[K' to type '[01m[Kchar*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
468 | if (!returntokenCallback(state_pr, [01;35m[K(char*)token_str[m[K)) { | |
| [01;35m[K^~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon_binding.cpp:517:27:[m[K [01;35m[Kwarning: [m[Ksuggest parentheses around '[01m[K&&[m[K' within '[01m[K||[m[K' [[01;35m[K-Wparentheses[m[K] | |
517 | if ([01;35m[K!embd.empty() && embd.back() == falcon_token_eos()[m[K || stopword_fulfilled) | |
| [01;35m[K~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon_binding.cpp:186:10:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kfalcon_token_newline[m[K' [[01;35m[K-Wunused-variable[m[K] | |
186 | auto [01;35m[Kfalcon_token_newline[m[K = falcon_token_nl(); | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kfalcon_binding.cpp:231:10:[m[K [01;35m[Kwarning: [m[Kvariable '[01m[Kinput_echo[m[K' set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
231 | bool [01;35m[Kinput_echo[m[K = true; | |
| [01;35m[K^~~~~~~~~~[m[K | |
cd build && cp -rf CMakeFiles/ggml.dir/k_quants.c.o ../ggllm.cpp/k_quants.o | |
cd build && cp -rf CMakeFiles/ggml.dir/ggml-cuda.cu.o ../ggllm.cpp/ggml-cuda.o | |
ar src libggllm.a ggllm.cpp/libfalcon.o ggllm.cpp/cmpnct_unicode.o ggllm.cpp/ggml.o ggllm.cpp/k_quants.o ggllm.cpp/ggml-cuda.o ggllm.cpp/falcon_common.o falcon_binding.o | |
make[1]: Leaving directory '/build/go-ggllm' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggllm LIBRARY_PATH=/build/go-ggllm \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/falcon ./cmd/grpc/falcon/ | |
cd bloomz && make libbloomz.a | |
make[1]: Entering directory '/build/bloomz' | |
I llama.cpp build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I. -O3 -DNDEBUG -std=c11 -fPIC -pthread -mavx -mavx2 -mfma -mf16c -msse3 | |
I CXXFLAGS: -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread | |
I LDFLAGS: | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
cc -I. -O3 -DNDEBUG -std=c11 -fPIC -pthread -mavx -mavx2 -mfma -mf16c -msse3 -c ggml.c -o ggml.o | |
g++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread -c utils.cpp -o utils.o | |
g++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread bloomz.cpp ggml.o utils.o -o bloomz.o -c | |
[01m[Kg++:[m[K [01;35m[Kwarning: [m[Kggml.o: linker input file unused because linking not done | |
[01m[Kg++:[m[K [01;35m[Kwarning: [m[Kutils.o: linker input file unused because linking not done | |
ar src libbloomz.a bloomz.o ggml.o utils.o | |
make[1]: Leaving directory '/build/bloomz' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/bloomz LIBRARY_PATH=/build/bloomz \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/bloomz ./cmd/grpc/bloomz/ | |
make -C go-llama BUILD_TYPE=cublas libbinding.a | |
make[1]: Entering directory '/build/go-llama' | |
I llama.cpp build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I./llama.cpp -I. -O3 -DNDEBUG -std=c11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wno-unused-function -pthread -march=native -mtune=native | |
I CXXFLAGS: -I./llama.cpp -I. -I./llama.cpp/examples -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -pthread | |
I CGO_LDFLAGS: | |
I LDFLAGS: | |
I BUILD_TYPE: cublas | |
I CMAKE_ARGS: -DLLAMA_CUBLAS=ON | |
I EXTRA_TARGETS: llama.cpp/ggml-cuda.o | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
cd llama.cpp && patch -p1 < ../patches/1902-cuda.patch | |
patching file examples/common.cpp | |
patching file examples/common.h | |
touch prepare | |
mkdir -p build | |
cd build && cmake ../llama.cpp -DLLAMA_CUBLAS=ON && VERBOSE=1 cmake --build . --config Release && cp -rf CMakeFiles/ggml.dir/ggml.c.o ../llama.cpp/ggml.o | |
-- The C compiler identification is GNU 10.2.1 | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Found Git: /usr/bin/git (found version "2.30.2") | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Check if compiler accepts -pthread | |
-- Check if compiler accepts -pthread - yes | |
-- Found Threads: TRUE | |
-- Found CUDAToolkit: /usr/local/cuda/include (found version "12.1.105") | |
-- cuBLAS found | |
-- The CUDA compiler identification is NVIDIA 12.1.105 | |
-- Detecting CUDA compiler ABI info | |
-- Detecting CUDA compiler ABI info - done | |
-- Check for working CUDA compiler: /usr/local/cuda/bin/nvcc - skipped | |
-- Detecting CUDA compile features | |
-- Detecting CUDA compile features - done | |
-- Using CUDA architectures: 52;61;70 | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
-- Configuring done (1.7s) | |
-- Generating done (0.1s) | |
-- Build files have been written to: /build/go-llama/build | |
Change Dir: '/build/go-llama/build' | |
Run Build Command(s): /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E env VERBOSE=1 /usr/bin/gmake -f Makefile | |
gmake[2]: Entering directory '/build/go-llama/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/go-llama/llama.cpp -B/build/go-llama/build --check-build-system CMakeFiles/Makefile.cmake 0 | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-llama/build/CMakeFiles /build/go-llama/build//CMakeFiles/progress.marks | |
/usr/bin/gmake -f CMakeFiles/Makefile2 all | |
gmake[3]: Entering directory '/build/go-llama/build' | |
/usr/bin/gmake -f CMakeFiles/BUILD_INFO.dir/build.make CMakeFiles/BUILD_INFO.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp /build/go-llama/build /build/go-llama/build /build/go-llama/build/CMakeFiles/BUILD_INFO.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f CMakeFiles/BUILD_INFO.dir/build.make CMakeFiles/BUILD_INFO.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
gmake[4]: Nothing to be done for 'CMakeFiles/BUILD_INFO.dir/build'. | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 2%] Built target BUILD_INFO | |
/usr/bin/gmake -f CMakeFiles/ggml.dir/build.make CMakeFiles/ggml.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp /build/go-llama/build /build/go-llama/build /build/go-llama/build/CMakeFiles/ggml.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f CMakeFiles/ggml.dir/build.make CMakeFiles/ggml.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 4%] [32mBuilding C object CMakeFiles/ggml.dir/ggml.c.o[0m | |
/usr/bin/cc -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu11 -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/ggml.dir/ggml.c.o -MF CMakeFiles/ggml.dir/ggml.c.o.d -o CMakeFiles/ggml.dir/ggml.c.o -c /build/go-llama/llama.cpp/ggml.c | |
[ 6%] [32mBuilding C object CMakeFiles/ggml.dir/ggml-alloc.c.o[0m | |
/usr/bin/cc -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu11 -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/ggml.dir/ggml-alloc.c.o -MF CMakeFiles/ggml.dir/ggml-alloc.c.o.d -o CMakeFiles/ggml.dir/ggml-alloc.c.o -c /build/go-llama/llama.cpp/ggml-alloc.c | |
[ 8%] [32mBuilding CUDA object CMakeFiles/ggml.dir/ggml-cuda.cu.o[0m | |
/usr/local/cuda/bin/nvcc -forward-unknown-to-host-compiler -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 --options-file CMakeFiles/ggml.dir/includes_CUDA.rsp -O3 -DNDEBUG -std=c++11 "--generate-code=arch=compute_52,code=[compute_52,sm_52]" "--generate-code=arch=compute_61,code=[compute_61,sm_61]" "--generate-code=arch=compute_70,code=[compute_70,sm_70]" -mf16c -mfma -mavx -mavx2 -Xcompiler -pthread -MD -MT CMakeFiles/ggml.dir/ggml-cuda.cu.o -MF CMakeFiles/ggml.dir/ggml-cuda.cu.o.d -x cu -c /build/go-llama/llama.cpp/ggml-cuda.cu -o CMakeFiles/ggml.dir/ggml-cuda.cu.o | |
[ 10%] [32mBuilding C object CMakeFiles/ggml.dir/k_quants.c.o[0m | |
/usr/bin/cc -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu11 -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/ggml.dir/k_quants.c.o -MF CMakeFiles/ggml.dir/k_quants.c.o.d -o CMakeFiles/ggml.dir/k_quants.c.o -c /build/go-llama/llama.cpp/k_quants.c | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 10%] Built target ggml | |
/usr/bin/gmake -f CMakeFiles/ggml_static.dir/build.make CMakeFiles/ggml_static.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp /build/go-llama/build /build/go-llama/build /build/go-llama/build/CMakeFiles/ggml_static.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f CMakeFiles/ggml_static.dir/build.make CMakeFiles/ggml_static.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 12%] [32m[1mLinking CUDA static library libggml_static.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/ggml_static.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/ggml_static.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libggml_static.a CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-alloc.c.o" "CMakeFiles/ggml.dir/ggml-cuda.cu.o" CMakeFiles/ggml.dir/k_quants.c.o | |
/usr/bin/ranlib libggml_static.a | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 12%] Built target ggml_static | |
/usr/bin/gmake -f CMakeFiles/llama.dir/build.make CMakeFiles/llama.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp /build/go-llama/build /build/go-llama/build /build/go-llama/build/CMakeFiles/llama.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f CMakeFiles/llama.dir/build.make CMakeFiles/llama.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 14%] [32mBuilding CXX object CMakeFiles/llama.dir/llama.cpp.o[0m | |
/usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -isystem /usr/local/cuda/include -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -pthread -MD -MT CMakeFiles/llama.dir/llama.cpp.o -MF CMakeFiles/llama.dir/llama.cpp.o.d -o CMakeFiles/llama.dir/llama.cpp.o -c /build/go-llama/llama.cpp/llama.cpp | |
[ 16%] [32m[1mLinking CXX static library libllama.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama.a CMakeFiles/llama.dir/llama.cpp.o CMakeFiles/ggml.dir/ggml.c.o "CMakeFiles/ggml.dir/ggml-alloc.c.o" "CMakeFiles/ggml.dir/ggml-cuda.cu.o" CMakeFiles/ggml.dir/k_quants.c.o | |
/usr/bin/ranlib libllama.a | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 16%] Built target llama | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-fns.dir/build.make tests/CMakeFiles/test-quantize-fns.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/tests /build/go-llama/build /build/go-llama/build/tests /build/go-llama/build/tests/CMakeFiles/test-quantize-fns.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-fns.dir/build.make tests/CMakeFiles/test-quantize-fns.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 18%] [32mBuilding CXX object tests/CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o[0m | |
cd /build/go-llama/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o -MF CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o.d -o CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o -c /build/go-llama/llama.cpp/tests/test-quantize-fns.cpp | |
[ 20%] [32m[1mLinking CXX executable ../bin/test-quantize-fns[0m | |
cd /build/go-llama/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-quantize-fns.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-quantize-fns.dir/test-quantize-fns.cpp.o" -o ../bin/test-quantize-fns -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 20%] Built target test-quantize-fns | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-perf.dir/build.make tests/CMakeFiles/test-quantize-perf.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/tests /build/go-llama/build /build/go-llama/build/tests /build/go-llama/build/tests/CMakeFiles/test-quantize-perf.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-quantize-perf.dir/build.make tests/CMakeFiles/test-quantize-perf.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 22%] [32mBuilding CXX object tests/CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o[0m | |
cd /build/go-llama/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o -MF CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o.d -o CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o -c /build/go-llama/llama.cpp/tests/test-quantize-perf.cpp | |
[ 24%] [32m[1mLinking CXX executable ../bin/test-quantize-perf[0m | |
cd /build/go-llama/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-quantize-perf.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-quantize-perf.dir/test-quantize-perf.cpp.o" -o ../bin/test-quantize-perf -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 24%] Built target test-quantize-perf | |
/usr/bin/gmake -f tests/CMakeFiles/test-sampling.dir/build.make tests/CMakeFiles/test-sampling.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/tests /build/go-llama/build /build/go-llama/build/tests /build/go-llama/build/tests/CMakeFiles/test-sampling.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-sampling.dir/build.make tests/CMakeFiles/test-sampling.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 26%] [32mBuilding CXX object tests/CMakeFiles/test-sampling.dir/test-sampling.cpp.o[0m | |
cd /build/go-llama/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-sampling.dir/test-sampling.cpp.o -MF CMakeFiles/test-sampling.dir/test-sampling.cpp.o.d -o CMakeFiles/test-sampling.dir/test-sampling.cpp.o -c /build/go-llama/llama.cpp/tests/test-sampling.cpp | |
[ 28%] [32m[1mLinking CXX executable ../bin/test-sampling[0m | |
cd /build/go-llama/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-sampling.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-sampling.dir/test-sampling.cpp.o" -o ../bin/test-sampling -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 28%] Built target test-sampling | |
/usr/bin/gmake -f tests/CMakeFiles/test-tokenizer-0.dir/build.make tests/CMakeFiles/test-tokenizer-0.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/tests /build/go-llama/build /build/go-llama/build/tests /build/go-llama/build/tests/CMakeFiles/test-tokenizer-0.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-tokenizer-0.dir/build.make tests/CMakeFiles/test-tokenizer-0.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 30%] [32mBuilding CXX object tests/CMakeFiles/test-tokenizer-0.dir/test-tokenizer-0.cpp.o[0m | |
cd /build/go-llama/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-tokenizer-0.dir/test-tokenizer-0.cpp.o -MF CMakeFiles/test-tokenizer-0.dir/test-tokenizer-0.cpp.o.d -o CMakeFiles/test-tokenizer-0.dir/test-tokenizer-0.cpp.o -c /build/go-llama/llama.cpp/tests/test-tokenizer-0.cpp | |
[01m[K/build/go-llama/llama.cpp/tests/test-tokenizer-0.cpp:19:2:[m[K [01;35m[Kwarning: [m[Kextra ‘[01m[K;[m[K’ [[01;35m[K-Wpedantic[m[K] | |
19 | }[01;35m[K;[m[K | |
| [01;35m[K^[m[K | |
[ 32%] [32m[1mLinking CXX executable ../bin/test-tokenizer-0[0m | |
cd /build/go-llama/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-tokenizer-0.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-tokenizer-0.dir/test-tokenizer-0.cpp.o" -o ../bin/test-tokenizer-0 -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 32%] Built target test-tokenizer-0 | |
/usr/bin/gmake -f tests/CMakeFiles/test-grad0.dir/build.make tests/CMakeFiles/test-grad0.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/tests /build/go-llama/build /build/go-llama/build/tests /build/go-llama/build/tests/CMakeFiles/test-grad0.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f tests/CMakeFiles/test-grad0.dir/build.make tests/CMakeFiles/test-grad0.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 34%] [32mBuilding CXX object tests/CMakeFiles/test-grad0.dir/test-grad0.cpp.o[0m | |
cd /build/go-llama/build/tests && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT tests/CMakeFiles/test-grad0.dir/test-grad0.cpp.o -MF CMakeFiles/test-grad0.dir/test-grad0.cpp.o.d -o CMakeFiles/test-grad0.dir/test-grad0.cpp.o -c /build/go-llama/llama.cpp/tests/test-grad0.cpp | |
[ 36%] [32m[1mLinking CXX executable ../bin/test-grad0[0m | |
cd /build/go-llama/build/tests && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/test-grad0.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/test-grad0.dir/test-grad0.cpp.o" -o ../bin/test-grad0 -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../libllama.a -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 36%] Built target test-grad0 | |
/usr/bin/gmake -f examples/CMakeFiles/common.dir/build.make examples/CMakeFiles/common.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples /build/go-llama/build /build/go-llama/build/examples /build/go-llama/build/examples/CMakeFiles/common.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/CMakeFiles/common.dir/build.make examples/CMakeFiles/common.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 38%] [32mBuilding CXX object examples/CMakeFiles/common.dir/common.cpp.o[0m | |
cd /build/go-llama/build/examples && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/CMakeFiles/common.dir/common.cpp.o -MF CMakeFiles/common.dir/common.cpp.o.d -o CMakeFiles/common.dir/common.cpp.o -c /build/go-llama/llama.cpp/examples/common.cpp | |
[ 40%] [32mBuilding CXX object examples/CMakeFiles/common.dir/grammar-parser.cpp.o[0m | |
cd /build/go-llama/build/examples && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/CMakeFiles/common.dir/grammar-parser.cpp.o -MF CMakeFiles/common.dir/grammar-parser.cpp.o.d -o CMakeFiles/common.dir/grammar-parser.cpp.o -c /build/go-llama/llama.cpp/examples/grammar-parser.cpp | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 40%] Built target common | |
/usr/bin/gmake -f examples/main/CMakeFiles/main.dir/build.make examples/main/CMakeFiles/main.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/main /build/go-llama/build /build/go-llama/build/examples/main /build/go-llama/build/examples/main/CMakeFiles/main.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/main/CMakeFiles/main.dir/build.make examples/main/CMakeFiles/main.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 42%] [32mBuilding CXX object examples/main/CMakeFiles/main.dir/main.cpp.o[0m | |
cd /build/go-llama/build/examples/main && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/main/CMakeFiles/main.dir/main.cpp.o -MF CMakeFiles/main.dir/main.cpp.o.d -o CMakeFiles/main.dir/main.cpp.o -c /build/go-llama/llama.cpp/examples/main/main.cpp | |
[ 44%] [32m[1mLinking CXX executable ../../bin/main[0m | |
cd /build/go-llama/build/examples/main && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/main.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/main.dir/main.cpp.o ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/main -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 44%] Built target main | |
/usr/bin/gmake -f examples/quantize/CMakeFiles/quantize.dir/build.make examples/quantize/CMakeFiles/quantize.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/quantize /build/go-llama/build /build/go-llama/build/examples/quantize /build/go-llama/build/examples/quantize/CMakeFiles/quantize.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/quantize/CMakeFiles/quantize.dir/build.make examples/quantize/CMakeFiles/quantize.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 46%] [32mBuilding CXX object examples/quantize/CMakeFiles/quantize.dir/quantize.cpp.o[0m | |
cd /build/go-llama/build/examples/quantize && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/quantize/CMakeFiles/quantize.dir/quantize.cpp.o -MF CMakeFiles/quantize.dir/quantize.cpp.o.d -o CMakeFiles/quantize.dir/quantize.cpp.o -c /build/go-llama/llama.cpp/examples/quantize/quantize.cpp | |
[ 48%] [32m[1mLinking CXX executable ../../bin/quantize[0m | |
cd /build/go-llama/build/examples/quantize && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/quantize.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/quantize.dir/quantize.cpp.o -o ../../bin/quantize -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 48%] Built target quantize | |
/usr/bin/gmake -f examples/quantize-stats/CMakeFiles/quantize-stats.dir/build.make examples/quantize-stats/CMakeFiles/quantize-stats.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/quantize-stats /build/go-llama/build /build/go-llama/build/examples/quantize-stats /build/go-llama/build/examples/quantize-stats/CMakeFiles/quantize-stats.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/quantize-stats/CMakeFiles/quantize-stats.dir/build.make examples/quantize-stats/CMakeFiles/quantize-stats.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 50%] [32mBuilding CXX object examples/quantize-stats/CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o[0m | |
cd /build/go-llama/build/examples/quantize-stats && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/quantize-stats/CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o -MF CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o.d -o CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o -c /build/go-llama/llama.cpp/examples/quantize-stats/quantize-stats.cpp | |
[ 52%] [32m[1mLinking CXX executable ../../bin/quantize-stats[0m | |
cd /build/go-llama/build/examples/quantize-stats && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/quantize-stats.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/quantize-stats.dir/quantize-stats.cpp.o" -o ../../bin/quantize-stats -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 52%] Built target quantize-stats | |
/usr/bin/gmake -f examples/perplexity/CMakeFiles/perplexity.dir/build.make examples/perplexity/CMakeFiles/perplexity.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/perplexity /build/go-llama/build /build/go-llama/build/examples/perplexity /build/go-llama/build/examples/perplexity/CMakeFiles/perplexity.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/perplexity/CMakeFiles/perplexity.dir/build.make examples/perplexity/CMakeFiles/perplexity.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 54%] [32mBuilding CXX object examples/perplexity/CMakeFiles/perplexity.dir/perplexity.cpp.o[0m | |
cd /build/go-llama/build/examples/perplexity && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/perplexity/CMakeFiles/perplexity.dir/perplexity.cpp.o -MF CMakeFiles/perplexity.dir/perplexity.cpp.o.d -o CMakeFiles/perplexity.dir/perplexity.cpp.o -c /build/go-llama/llama.cpp/examples/perplexity/perplexity.cpp | |
[ 56%] [32m[1mLinking CXX executable ../../bin/perplexity[0m | |
cd /build/go-llama/build/examples/perplexity && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/perplexity.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/perplexity.dir/perplexity.cpp.o ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/perplexity -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 56%] Built target perplexity | |
/usr/bin/gmake -f examples/embedding/CMakeFiles/embedding.dir/build.make examples/embedding/CMakeFiles/embedding.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/embedding /build/go-llama/build /build/go-llama/build/examples/embedding /build/go-llama/build/examples/embedding/CMakeFiles/embedding.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/embedding/CMakeFiles/embedding.dir/build.make examples/embedding/CMakeFiles/embedding.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 58%] [32mBuilding CXX object examples/embedding/CMakeFiles/embedding.dir/embedding.cpp.o[0m | |
cd /build/go-llama/build/examples/embedding && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/embedding/CMakeFiles/embedding.dir/embedding.cpp.o -MF CMakeFiles/embedding.dir/embedding.cpp.o.d -o CMakeFiles/embedding.dir/embedding.cpp.o -c /build/go-llama/llama.cpp/examples/embedding/embedding.cpp | |
[ 60%] [32m[1mLinking CXX executable ../../bin/embedding[0m | |
cd /build/go-llama/build/examples/embedding && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/embedding.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/embedding.dir/embedding.cpp.o ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/embedding -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 60%] Built target embedding | |
/usr/bin/gmake -f examples/save-load-state/CMakeFiles/save-load-state.dir/build.make examples/save-load-state/CMakeFiles/save-load-state.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/save-load-state /build/go-llama/build /build/go-llama/build/examples/save-load-state /build/go-llama/build/examples/save-load-state/CMakeFiles/save-load-state.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/save-load-state/CMakeFiles/save-load-state.dir/build.make examples/save-load-state/CMakeFiles/save-load-state.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 62%] [32mBuilding CXX object examples/save-load-state/CMakeFiles/save-load-state.dir/save-load-state.cpp.o[0m | |
cd /build/go-llama/build/examples/save-load-state && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/save-load-state/CMakeFiles/save-load-state.dir/save-load-state.cpp.o -MF CMakeFiles/save-load-state.dir/save-load-state.cpp.o.d -o CMakeFiles/save-load-state.dir/save-load-state.cpp.o -c /build/go-llama/llama.cpp/examples/save-load-state/save-load-state.cpp | |
[ 64%] [32m[1mLinking CXX executable ../../bin/save-load-state[0m | |
cd /build/go-llama/build/examples/save-load-state && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/save-load-state.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/save-load-state.dir/save-load-state.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/save-load-state -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 64%] Built target save-load-state | |
/usr/bin/gmake -f examples/benchmark/CMakeFiles/benchmark.dir/build.make examples/benchmark/CMakeFiles/benchmark.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/benchmark /build/go-llama/build /build/go-llama/build/examples/benchmark /build/go-llama/build/examples/benchmark/CMakeFiles/benchmark.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/benchmark/CMakeFiles/benchmark.dir/build.make examples/benchmark/CMakeFiles/benchmark.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 66%] [32mBuilding CXX object examples/benchmark/CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o[0m | |
cd /build/go-llama/build/examples/benchmark && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/benchmark/CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o -MF CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o.d -o CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o -c /build/go-llama/llama.cpp/examples/benchmark/benchmark-matmult.cpp | |
[ 68%] [32m[1mLinking CXX executable ../../bin/benchmark[0m | |
cd /build/go-llama/build/examples/benchmark && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/benchmark.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/benchmark.dir/benchmark-matmult.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/benchmark -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 68%] Built target benchmark | |
/usr/bin/gmake -f examples/baby-llama/CMakeFiles/baby-llama.dir/build.make examples/baby-llama/CMakeFiles/baby-llama.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/baby-llama /build/go-llama/build /build/go-llama/build/examples/baby-llama /build/go-llama/build/examples/baby-llama/CMakeFiles/baby-llama.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/baby-llama/CMakeFiles/baby-llama.dir/build.make examples/baby-llama/CMakeFiles/baby-llama.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 70%] [32mBuilding CXX object examples/baby-llama/CMakeFiles/baby-llama.dir/baby-llama.cpp.o[0m | |
cd /build/go-llama/build/examples/baby-llama && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/baby-llama/CMakeFiles/baby-llama.dir/baby-llama.cpp.o -MF CMakeFiles/baby-llama.dir/baby-llama.cpp.o.d -o CMakeFiles/baby-llama.dir/baby-llama.cpp.o -c /build/go-llama/llama.cpp/examples/baby-llama/baby-llama.cpp | |
[01m[K/build/go-llama/llama.cpp/examples/baby-llama/baby-llama.cpp:[m[K In function ‘[01m[Kint main(int, char**)[m[K’: | |
[01m[K/build/go-llama/llama.cpp/examples/baby-llama/baby-llama.cpp:1620:32:[m[K [01;35m[Kwarning: [m[Kvariable ‘[01m[Kopt_params_adam[m[K’ set but not used [[01;35m[K-Wunused-but-set-variable[m[K] | |
1620 | struct ggml_opt_params [01;35m[Kopt_params_adam[m[K = ggml_opt_default_params(GGML_OPT_ADAM); | |
| [01;35m[K^~~~~~~~~~~~~~~[m[K | |
[ 72%] [32m[1mLinking CXX executable ../../bin/baby-llama[0m | |
cd /build/go-llama/build/examples/baby-llama && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/baby-llama.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/baby-llama.dir/baby-llama.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/baby-llama -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 72%] Built target baby-llama | |
/usr/bin/gmake -f examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build.make examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/train-text-from-scratch /build/go-llama/build /build/go-llama/build/examples/train-text-from-scratch /build/go-llama/build/examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build.make examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 74%] [32mBuilding CXX object examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o[0m | |
cd /build/go-llama/build/examples/train-text-from-scratch && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/train-text-from-scratch/CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o -MF CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o.d -o CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o -c /build/go-llama/llama.cpp/examples/train-text-from-scratch/train-text-from-scratch.cpp | |
[ 76%] [32m[1mLinking CXX executable ../../bin/train-text-from-scratch[0m | |
cd /build/go-llama/build/examples/train-text-from-scratch && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/train-text-from-scratch.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/train-text-from-scratch.dir/train-text-from-scratch.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/train-text-from-scratch -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 76%] Built target train-text-from-scratch | |
/usr/bin/gmake -f examples/simple/CMakeFiles/simple.dir/build.make examples/simple/CMakeFiles/simple.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/simple /build/go-llama/build /build/go-llama/build/examples/simple /build/go-llama/build/examples/simple/CMakeFiles/simple.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/simple/CMakeFiles/simple.dir/build.make examples/simple/CMakeFiles/simple.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 78%] [32mBuilding CXX object examples/simple/CMakeFiles/simple.dir/simple.cpp.o[0m | |
cd /build/go-llama/build/examples/simple && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/simple/CMakeFiles/simple.dir/simple.cpp.o -MF CMakeFiles/simple.dir/simple.cpp.o.d -o CMakeFiles/simple.dir/simple.cpp.o -c /build/go-llama/llama.cpp/examples/simple/simple.cpp | |
[ 80%] [32m[1mLinking CXX executable ../../bin/simple[0m | |
cd /build/go-llama/build/examples/simple && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/simple.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/simple.dir/simple.cpp.o ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/simple -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 80%] Built target simple | |
/usr/bin/gmake -f examples/embd-input/CMakeFiles/embdinput.dir/build.make examples/embd-input/CMakeFiles/embdinput.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/embd-input /build/go-llama/build /build/go-llama/build/examples/embd-input /build/go-llama/build/examples/embd-input/CMakeFiles/embdinput.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/embd-input/CMakeFiles/embdinput.dir/build.make examples/embd-input/CMakeFiles/embdinput.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 82%] [32mBuilding CXX object examples/embd-input/CMakeFiles/embdinput.dir/embd-input-lib.cpp.o[0m | |
cd /build/go-llama/build/examples/embd-input && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/embd-input/CMakeFiles/embdinput.dir/embd-input-lib.cpp.o -MF CMakeFiles/embdinput.dir/embd-input-lib.cpp.o.d -o CMakeFiles/embdinput.dir/embd-input-lib.cpp.o -c /build/go-llama/llama.cpp/examples/embd-input/embd-input-lib.cpp | |
[ 84%] [32m[1mLinking CXX static library libembdinput.a[0m | |
cd /build/go-llama/build/examples/embd-input && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/embdinput.dir/cmake_clean_target.cmake | |
cd /build/go-llama/build/examples/embd-input && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/embdinput.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libembdinput.a "CMakeFiles/embdinput.dir/embd-input-lib.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" | |
/usr/bin/ranlib libembdinput.a | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 84%] Built target embdinput | |
/usr/bin/gmake -f examples/embd-input/CMakeFiles/embd-input-test.dir/build.make examples/embd-input/CMakeFiles/embd-input-test.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/embd-input /build/go-llama/build /build/go-llama/build/examples/embd-input /build/go-llama/build/examples/embd-input/CMakeFiles/embd-input-test.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/embd-input/CMakeFiles/embd-input-test.dir/build.make examples/embd-input/CMakeFiles/embd-input-test.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 86%] [32mBuilding CXX object examples/embd-input/CMakeFiles/embd-input-test.dir/embd-input-test.cpp.o[0m | |
cd /build/go-llama/build/examples/embd-input && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/embd-input/CMakeFiles/embd-input-test.dir/embd-input-test.cpp.o -MF CMakeFiles/embd-input-test.dir/embd-input-test.cpp.o.d -o CMakeFiles/embd-input-test.dir/embd-input-test.cpp.o -c /build/go-llama/llama.cpp/examples/embd-input/embd-input-test.cpp | |
[ 88%] [32m[1mLinking CXX executable ../../bin/embd-input-test[0m | |
cd /build/go-llama/build/examples/embd-input && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/embd-input-test.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG "CMakeFiles/embd-input-test.dir/embd-input-test.cpp.o" ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/embd-input-test -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a libembdinput.a -pthread ../../libllama.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -pthread -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 88%] Built target embd-input-test | |
/usr/bin/gmake -f examples/server/CMakeFiles/server.dir/build.make examples/server/CMakeFiles/server.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/examples/server /build/go-llama/build /build/go-llama/build/examples/server /build/go-llama/build/examples/server/CMakeFiles/server.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f examples/server/CMakeFiles/server.dir/build.make examples/server/CMakeFiles/server.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 90%] [32mBuilding CXX object examples/server/CMakeFiles/server.dir/server.cpp.o[0m | |
cd /build/go-llama/build/examples/server && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -DSERVER_VERBOSE=1 -I/build/go-llama/llama.cpp/examples -I/build/go-llama/llama.cpp/examples/server -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT examples/server/CMakeFiles/server.dir/server.cpp.o -MF CMakeFiles/server.dir/server.cpp.o.d -o CMakeFiles/server.dir/server.cpp.o -c /build/go-llama/llama.cpp/examples/server/server.cpp | |
[ 92%] [32m[1mLinking CXX executable ../../bin/server[0m | |
cd /build/go-llama/build/examples/server && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/server.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/server.dir/server.cpp.o ../CMakeFiles/common.dir/common.cpp.o "../CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/server -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib: ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 92%] Built target server | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/vdot.dir/build.make pocs/vdot/CMakeFiles/vdot.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/pocs/vdot /build/go-llama/build /build/go-llama/build/pocs/vdot /build/go-llama/build/pocs/vdot/CMakeFiles/vdot.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/vdot.dir/build.make pocs/vdot/CMakeFiles/vdot.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 94%] [32mBuilding CXX object pocs/vdot/CMakeFiles/vdot.dir/vdot.cpp.o[0m | |
cd /build/go-llama/build/pocs/vdot && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/pocs -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT pocs/vdot/CMakeFiles/vdot.dir/vdot.cpp.o -MF CMakeFiles/vdot.dir/vdot.cpp.o.d -o CMakeFiles/vdot.dir/vdot.cpp.o -c /build/go-llama/llama.cpp/pocs/vdot/vdot.cpp | |
[ 96%] [32m[1mLinking CXX executable ../../bin/vdot[0m | |
cd /build/go-llama/build/pocs/vdot && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/vdot.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/vdot.dir/vdot.cpp.o ../../examples/CMakeFiles/common.dir/common.cpp.o "../../examples/CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/vdot -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[ 96%] Built target vdot | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/q8dot.dir/build.make pocs/vdot/CMakeFiles/q8dot.dir/depend | |
gmake[4]: Entering directory '/build/go-llama/build' | |
cd /build/go-llama/build && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/go-llama/llama.cpp /build/go-llama/llama.cpp/pocs/vdot /build/go-llama/build /build/go-llama/build/pocs/vdot /build/go-llama/build/pocs/vdot/CMakeFiles/q8dot.dir/DependInfo.cmake "--color=" | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
/usr/bin/gmake -f pocs/vdot/CMakeFiles/q8dot.dir/build.make pocs/vdot/CMakeFiles/q8dot.dir/build | |
gmake[4]: Entering directory '/build/go-llama/build' | |
[ 98%] [32mBuilding CXX object pocs/vdot/CMakeFiles/q8dot.dir/q8dot.cpp.o[0m | |
cd /build/go-llama/build/pocs/vdot && /usr/bin/c++ -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMQ_Y=64 -DGGML_CUDA_MMV_Y=1 -DGGML_USE_CUBLAS -DGGML_USE_K_QUANTS -DK_QUANTS_PER_ITERATION=2 -I/build/go-llama/llama.cpp/pocs -I/build/go-llama/llama.cpp/examples/. -I/build/go-llama/llama.cpp/. -O3 -DNDEBUG -std=gnu++11 -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -mf16c -mfma -mavx -mavx2 -MD -MT pocs/vdot/CMakeFiles/q8dot.dir/q8dot.cpp.o -MF CMakeFiles/q8dot.dir/q8dot.cpp.o.d -o CMakeFiles/q8dot.dir/q8dot.cpp.o -c /build/go-llama/llama.cpp/pocs/vdot/q8dot.cpp | |
[100%] [32m[1mLinking CXX executable ../../bin/q8dot[0m | |
cd /build/go-llama/build/pocs/vdot && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/q8dot.dir/link.txt --verbose=1 | |
/usr/bin/c++ -O3 -DNDEBUG CMakeFiles/q8dot.dir/q8dot.cpp.o ../../examples/CMakeFiles/common.dir/common.cpp.o "../../examples/CMakeFiles/common.dir/grammar-parser.cpp.o" -o ../../bin/q8dot -L/usr/local/cuda/targets/x86_64-linux/lib/stubs -L/usr/local/cuda/targets/x86_64-linux/lib -Wl,-rpath,/usr/local/cuda-12.1/targets/x86_64-linux/lib ../../libllama.a -pthread -pthread /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcudart.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublas.so /usr/local/cuda-12.1/targets/x86_64-linux/lib/libculibos.a /usr/local/cuda-12.1/targets/x86_64-linux/lib/libcublasLt.so -lcudadevrt -lcudart_static -lrt -lpthread -ldl | |
gmake[4]: Leaving directory '/build/go-llama/build' | |
[100%] Built target q8dot | |
gmake[3]: Leaving directory '/build/go-llama/build' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/go-llama/build/CMakeFiles 0 | |
gmake[2]: Leaving directory '/build/go-llama/build' | |
cd build && cp -rf CMakeFiles/llama.dir/llama.cpp.o ../llama.cpp/llama.o | |
cd build && cp -rf examples/CMakeFiles/common.dir/common.cpp.o ../llama.cpp/common.o | |
cd build && cp -rf examples/CMakeFiles/common.dir/grammar-parser.cpp.o ../llama.cpp/grammar-parser.o | |
cd build && cp -rf CMakeFiles/ggml.dir/ggml-alloc.c.o ../llama.cpp/ggml-alloc.o | |
g++ -I./llama.cpp -I. -I./llama.cpp/examples -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -pthread -I./llama.cpp -I./llama.cpp/examples binding.cpp -o binding.o -c | |
[01m[Kbinding.cpp:[m[K In function '[01m[Kint llama_predict(void*, void*, char*, bool)[m[K': | |
[01m[Kbinding.cpp:533:42:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst char*[m[K' to type '[01m[Kchar*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
533 | if (!tokenCallback(state_pr, [01;35m[K(char*)token_str[m[K)) { | |
| [01;35m[K^~~~~~~~~~~~~~~~[m[K | |
[01m[Kbinding.cpp:591:1:[m[K [01;35m[Kwarning: [m[Klabel '[01m[Kend[m[K' defined but not used [[01;35m[K-Wunused-label[m[K] | |
591 | [01;35m[Kend[m[K: | |
| [01;35m[K^~~[m[K | |
[01m[Kbinding.cpp:[m[K In function '[01m[Kvoid llama_binding_free_model(void*)[m[K': | |
[01m[Kbinding.cpp:613:5:[m[K [01;35m[Kwarning: [m[Kpossible problem detected in invocation of '[01m[Koperator delete[m[K' [[01;35m[K-Wdelete-incomplete[m[K] | |
613 | [01;35m[Kdelete ctx->model[m[K; | |
| [01;35m[K^~~~~~~~~~~~~~~~~[m[K | |
[01m[Kbinding.cpp:613:17:[m[K [01;35m[Kwarning: [m[Kinvalid use of incomplete type '[01m[Kstruct llama_model[m[K' | |
613 | delete [01;35m[Kctx->model[m[K; | |
| [01;35m[K~~~~~^~~~~[m[K | |
In file included from [01m[K./llama.cpp/examples/common.h:5[m[K, | |
from [01m[Kbinding.cpp:1[m[K: | |
[01m[K./llama.cpp/llama.h:70:12:[m[K [01;36m[Knote: [m[Kforward declaration of '[01m[Kstruct llama_model[m[K' | |
70 | struct [01;36m[Kllama_model[m[K; | |
| [01;36m[K^~~~~~~~~~~[m[K | |
[01m[Kbinding.cpp:613:5:[m[K [01;36m[Knote: [m[Kneither the destructor nor the class-specific '[01m[Koperator delete[m[K' will be called, even if they are declared when the class is defined | |
613 | [01;36m[Kdelete ctx->model[m[K; | |
| [01;36m[K^~~~~~~~~~~~~~~~~[m[K | |
cd build && cp -rf CMakeFiles/ggml.dir/k_quants.c.o ../llama.cpp/k_quants.o | |
cd build && cp -rf CMakeFiles/ggml.dir/ggml-cuda.cu.o ../llama.cpp/ggml-cuda.o | |
ar src libbinding.a llama.cpp/ggml.o llama.cpp/k_quants.o llama.cpp/ggml-cuda.o llama.cpp/ggml-alloc.o llama.cpp/common.o llama.cpp/grammar-parser.o llama.cpp/llama.o binding.o | |
make[1]: Leaving directory '/build/go-llama' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-llama LIBRARY_PATH=/build/go-llama \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/llama ./cmd/grpc/llama/ | |
# github.com/go-skynet/go-llama.cpp | |
binding.cpp: In function 'void llama_binding_free_model(void*)': | |
binding.cpp:613:5: warning: possible problem detected in invocation of 'operator delete' [-Wdelete-incomplete] | |
613 | delete ctx->model; | |
| ^~~~~~~~~~~~~~~~~ | |
binding.cpp:613:17: warning: invalid use of incomplete type 'struct llama_model' | |
613 | delete ctx->model; | |
| ~~~~~^~~~~ | |
In file included from go-llama/llama.cpp/examples/common.h:5, | |
from binding.cpp:1: | |
go-llama/llama.cpp/llama.h:70:12: note: forward declaration of 'struct llama_model' | |
70 | struct llama_model; | |
| ^~~~~~~~~~~ | |
binding.cpp:613:5: note: neither the destructor nor the class-specific 'operator delete' will be called, even if they are declared when the class is defined | |
613 | delete ctx->model; | |
| ^~~~~~~~~~~~~~~~~ | |
make -C gpt4all/gpt4all-bindings/golang/ libgpt4all.a | |
make[1]: Entering directory '/build/gpt4all/gpt4all-bindings/golang' | |
I go-gpt4all build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I. -I../../gpt4all-backend/llama.cpp -I../../gpt4all-backend -I -O3 -DNDEBUG -std=c11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -Wno-unused-function -pthread -march=native -mtune=native | |
I CXXFLAGS: -I. -I../../gpt4all-backend/llama.cpp -I../../gpt4all-backend -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native | |
I LDFLAGS: | |
I CMAKEFLAGS: | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
g++ -I. -I../../gpt4all-backend/llama.cpp -I../../gpt4all-backend -O3 -DNDEBUG -std=c++17 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native binding.cpp -o binding.o -c | |
[01m[Kbinding.cpp:[m[K In lambda function: | |
[01m[Kbinding.cpp:46:33:[m[K [01;35m[Kwarning: [m[Kunused parameter '[01m[Ktoken_id[m[K' [[01;35m[K-Wunused-parameter[m[K] | |
46 | auto lambda_prompt = []([01;35m[Kint token_id[m[K) { | |
| [01;35m[K~~~~^~~~~~~~[m[K | |
[01m[Kbinding.cpp:[m[K In lambda function: | |
[01m[Kbinding.cpp:54:20:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst char*[m[K' to type '[01m[Kchar*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
54 | res.append([01;35m[K(char*)responsechars[m[K); | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kbinding.cpp:55:39:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst char*[m[K' to type '[01m[Kchar*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
55 | return !!getTokenCallback(mm, [01;35m[K(char*)responsechars[m[K); | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[Kbinding.cpp:53:35:[m[K [01;35m[Kwarning: [m[Kunused parameter '[01m[Ktoken_id[m[K' [[01;35m[K-Wunused-parameter[m[K] | |
53 | auto lambda_response = []([01;35m[Kint token_id[m[K, const char *responsechars) { | |
| [01;35m[K~~~~^~~~~~~~[m[K | |
[01m[Kbinding.cpp:[m[K In function '[01m[Kvoid model_prompt(const char*, void*, char*, int, float, int, int, int, float, float, int, float)[m[K': | |
[01m[Kbinding.cpp:64:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
64 | [01;35m[K.[m[Klogits = NULL, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:65:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
65 | [01;35m[K.[m[Klogits_size = 0, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:66:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
66 | [01;35m[K.[m[Ktokens = NULL, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:67:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
67 | [01;35m[K.[m[Ktokens_size = 0, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:68:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
68 | [01;35m[K.[m[Kn_past = 0, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:69:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
69 | [01;35m[K.[m[Kn_ctx = 1024, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:70:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
70 | [01;35m[K.[m[Kn_predict = 50, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:71:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
71 | [01;35m[K.[m[Ktop_k = 10, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:72:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
72 | [01;35m[K.[m[Ktop_p = 0.9, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:73:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
73 | [01;35m[K.[m[Ktemp = 1.0, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:74:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
74 | [01;35m[K.[m[Kn_batch = 1, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:75:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
75 | [01;35m[K.[m[Krepeat_penalty = 1.2, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:76:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
76 | [01;35m[K.[m[Krepeat_last_n = 10, | |
| [01;35m[K^[m[K | |
[01m[Kbinding.cpp:77:9:[m[K [01;35m[Kwarning: [m[KC++ designated initializers only available with '[01m[K-std=c++2a[m[K' or '[01m[K-std=gnu++2a[m[K' [[01;35m[K-Wpedantic[m[K] | |
77 | [01;35m[K.[m[Kcontext_erase = 0.5 | |
| [01;35m[K^[m[K | |
mkdir buildllm | |
cd buildllm && cmake ../../../gpt4all-backend/ && make | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- The C compiler identification is GNU 10.2.1 | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Interprocedural optimization support detected | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Check if compiler accepts -pthread | |
-- Check if compiler accepts -pthread - yes | |
-- Found Threads: TRUE | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- Configuring ggml implementation target llama-mainline-default in /build/gpt4all/gpt4all-backend/llama.cpp-mainline | |
-- x86 detected | |
-- Configuring ggml implementation target llama-230511-default in /build/gpt4all/gpt4all-backend/llama.cpp-230511 | |
-- x86 detected | |
-- Configuring ggml implementation target llama-230519-default in /build/gpt4all/gpt4all-backend/llama.cpp-230519 | |
-- x86 detected | |
-- Configuring model implementation target llamamodel-mainline-default | |
-- Configuring model implementation target replit-mainline-default | |
-- Configuring model implementation target llamamodel-230519-default | |
-- Configuring model implementation target llamamodel-230511-default | |
-- Configuring model implementation target gptj-default | |
-- Configuring model implementation target falcon-default | |
-- Configuring model implementation target mpt-default | |
-- Configuring model implementation target bert-default | |
-- Configuring model implementation target starcoder-default | |
-- Configuring ggml implementation target llama-mainline-avxonly in /build/gpt4all/gpt4all-backend/llama.cpp-mainline | |
-- x86 detected | |
-- Configuring ggml implementation target llama-230511-avxonly in /build/gpt4all/gpt4all-backend/llama.cpp-230511 | |
-- x86 detected | |
-- Configuring ggml implementation target llama-230519-avxonly in /build/gpt4all/gpt4all-backend/llama.cpp-230519 | |
-- x86 detected | |
-- Configuring model implementation target llamamodel-mainline-avxonly | |
-- Configuring model implementation target replit-mainline-avxonly | |
-- Configuring model implementation target llamamodel-230519-avxonly | |
-- Configuring model implementation target llamamodel-230511-avxonly | |
-- Configuring model implementation target gptj-avxonly | |
-- Configuring model implementation target falcon-avxonly | |
-- Configuring model implementation target mpt-avxonly | |
-- Configuring model implementation target bert-avxonly | |
-- Configuring model implementation target starcoder-avxonly | |
-- Configuring done (0.7s) | |
-- Generating done (0.0s) | |
-- Build files have been written to: /build/gpt4all/gpt4all-bindings/golang/buildllm | |
make[2]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -S/build/gpt4all/gpt4all-backend -B/build/gpt4all/gpt4all-bindings/golang/buildllm --check-build-system CMakeFiles/Makefile.cmake 0 | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles /build/gpt4all/gpt4all-bindings/golang/buildllm//CMakeFiles/progress.marks | |
make -f CMakeFiles/Makefile2 all | |
make[3]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-mainline-default.dir/build.make CMakeFiles/ggml-mainline-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-mainline-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-mainline-default.dir/build.make CMakeFiles/ggml-mainline-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 1%] [32mBuilding C object CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/ggml.c.o[0m | |
/usr/bin/cc -DGGML_USE_K_QUANTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx2 -pthread -MD -MT CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/ggml.c.o -MF CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/ggml.c.o.d -o CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c | |
[ 2%] [32mBuilding C object CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/k_quants.c.o[0m | |
/usr/bin/cc -DGGML_USE_K_QUANTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx2 -pthread -MD -MT CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/k_quants.c.o -MF CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/k_quants.c.o.d -o CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/k_quants.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/k_quants.c | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 2%] Built target ggml-mainline-default | |
make -f CMakeFiles/llama-mainline-default.dir/build.make CMakeFiles/llama-mainline-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-mainline-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-mainline-default.dir/build.make CMakeFiles/llama-mainline-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 3%] [32mBuilding CXX object CMakeFiles/llama-mainline-default.dir/llama.cpp-mainline/llama.cpp.o[0m | |
/usr/bin/c++ -DGGML_USE_K_QUANTS -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-mainline-default.dir/llama.cpp-mainline/llama.cpp.o -MF CMakeFiles/llama-mainline-default.dir/llama.cpp-mainline/llama.cpp.o.d -o CMakeFiles/llama-mainline-default.dir/llama.cpp-mainline/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.cpp | |
[ 4%] [32m[1mLinking CXX static library libllama-mainline-default.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-mainline-default.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-mainline-default.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-mainline-default.a "CMakeFiles/llama-mainline-default.dir/llama.cpp-mainline/llama.cpp.o" "CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/ggml.c.o" "CMakeFiles/ggml-mainline-default.dir/llama.cpp-mainline/k_quants.c.o" | |
/usr/bin/ranlib libllama-mainline-default.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 4%] Built target llama-mainline-default | |
make -f CMakeFiles/ggml-230511-default.dir/build.make CMakeFiles/ggml-230511-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-230511-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-230511-default.dir/build.make CMakeFiles/ggml-230511-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 5%] [32mBuilding C object CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o[0m | |
/usr/bin/cc -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx2 -pthread -MD -MT CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o -MF CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o.d -o CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi_f32[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9357:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kne2_ne3[m[K' [[01;35m[K-Wunused-variable[m[K] | |
9357 | const int [01;35m[Kne2_ne3[m[K = n/ne1; // ne2*ne3 | |
| [01;35m[K^~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi_f16[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9419:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kne2[m[K' [[01;35m[K-Wunused-variable[m[K] | |
9419 | const int [01;35m[Kne2[m[K = src0->ne[2]; // n_head -> this is k | |
| [01;35m[K^~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9468:5:[m[K [01;35m[Kwarning: [m[Kenumeration value '[01m[KGGML_TYPE_Q4_3[m[K' not handled in switch [[01;35m[K-Wswitch[m[K] | |
9468 | [01;35m[Kswitch[m[K (src0->type) { | |
| [01;35m[K^~~~~~[m[K | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 5%] Built target ggml-230511-default | |
make -f CMakeFiles/llama-230511-default.dir/build.make CMakeFiles/llama-230511-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-230511-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-230511-default.dir/build.make CMakeFiles/llama-230511-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 6%] [32mBuilding CXX object CMakeFiles/llama-230511-default.dir/llama.cpp-230511/llama.cpp.o[0m | |
/usr/bin/c++ -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-230511-default.dir/llama.cpp-230511/llama.cpp.o -MF CMakeFiles/llama-230511-default.dir/llama.cpp-230511/llama.cpp.o.d -o CMakeFiles/llama-230511-default.dir/llama.cpp-230511/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230511/llama.cpp | |
[ 7%] [32m[1mLinking CXX static library libllama-230511-default.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-230511-default.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-230511-default.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-230511-default.a "CMakeFiles/llama-230511-default.dir/llama.cpp-230511/llama.cpp.o" "CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o" | |
/usr/bin/ranlib libllama-230511-default.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 7%] Built target llama-230511-default | |
make -f CMakeFiles/ggml-230519-default.dir/build.make CMakeFiles/ggml-230519-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-230519-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-230519-default.dir/build.make CMakeFiles/ggml-230519-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 8%] [32mBuilding C object CMakeFiles/ggml-230519-default.dir/llama.cpp-230519/ggml.c.o[0m | |
/usr/bin/cc -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -mf16c -mfma -mavx2 -pthread -MD -MT CMakeFiles/ggml-230519-default.dir/llama.cpp-230519/ggml.c.o -MF CMakeFiles/ggml-230519-default.dir/llama.cpp-230519/ggml.c.o.d -o CMakeFiles/ggml-230519-default.dir/llama.cpp-230519/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 8%] Built target ggml-230519-default | |
make -f CMakeFiles/llama-230519-default.dir/build.make CMakeFiles/llama-230519-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-230519-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-230519-default.dir/build.make CMakeFiles/llama-230519-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 10%] [32mBuilding CXX object CMakeFiles/llama-230519-default.dir/llama.cpp-230519/llama.cpp.o[0m | |
/usr/bin/c++ -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-230519-default.dir/llama.cpp-230519/llama.cpp.o -MF CMakeFiles/llama-230519-default.dir/llama.cpp-230519/llama.cpp.o.d -o CMakeFiles/llama-230519-default.dir/llama.cpp-230519/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:[m[K In function '[01m[Ksize_t llama_set_state_data(llama_context*, const uint8_t*)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:2685:27:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst uint8_t*[m[K' {aka '[01m[Kconst unsigned char*[m[K'} to type '[01m[Kvoid*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
2685 | kin3d->data = [01;35m[K(void *) inp[m[K; | |
| [01;35m[K^~~~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:2689:27:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst uint8_t*[m[K' {aka '[01m[Kconst unsigned char*[m[K'} to type '[01m[Kvoid*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
2689 | vin3d->data = [01;35m[K(void *) inp[m[K; | |
| [01;35m[K^~~~~~~~~~~~[m[K | |
[ 11%] [32m[1mLinking CXX static library libllama-230519-default.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-230519-default.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-230519-default.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-230519-default.a "CMakeFiles/llama-230519-default.dir/llama.cpp-230519/llama.cpp.o" "CMakeFiles/ggml-230519-default.dir/llama.cpp-230519/ggml.c.o" | |
/usr/bin/ranlib libllama-230519-default.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 11%] Built target llama-230519-default | |
make -f CMakeFiles/llamamodel-mainline-default.dir/build.make CMakeFiles/llamamodel-mainline-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-mainline-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-mainline-default.dir/build.make CMakeFiles/llamamodel-mainline-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 12%] [32mBuilding CXX object CMakeFiles/llamamodel-mainline-default.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=999999 -DLLAMA_VERSIONS=">=3" -Dllamamodel_mainline_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-mainline-default.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-mainline-default.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-mainline-default.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:[m[K In member function '[01m[Kvirtual bool LLamaModel::loadModel(const string&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:159:71:[m[K [01;35m[Kwarning: [m[K'[01m[Kllama_context* llama_init_from_file(const char*, llama_context_params)[m[K' is deprecated: please use llama_load_model_from_file combined with llama_new_context_with_model instead [[01;35m[K-Wdeprecated-declarations[m[K] | |
159 | d_ptr->ctx = llama_init_from_file(modelPath.c_str(), d_ptr->params[01;35m[K)[m[K; | |
| [01;35m[K^[m[K | |
In file included from [01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:28[m[K: | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.h:161:49:[m[K [01;36m[Knote: [m[Kdeclared here | |
161 | LLAMA_API DEPRECATED(struct llama_context * [01;36m[Kllama_init_from_file[m[K( | |
| [01;36m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.h:30:36:[m[K [01;36m[Knote: [m[Kin definition of macro '[01m[KDEPRECATED[m[K' | |
30 | # define DEPRECATED(func, hint) [01;36m[Kfunc[m[K __attribute__((deprecated(hint))) | |
| [01;36m[K^~~~[m[K | |
[ 13%] [32mBuilding CXX object CMakeFiles/llamamodel-mainline-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=999999 -DLLAMA_VERSIONS=">=3" -Dllamamodel_mainline_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-mainline-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-mainline-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-mainline-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 14%] [32m[1mLinking CXX shared library libllamamodel-mainline-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-mainline-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-mainline-default.so -o libllamamodel-mainline-default.so "CMakeFiles/llamamodel-mainline-default.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-mainline-default.dir/llmodel_shared.cpp.o" libllama-mainline-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 14%] Built target llamamodel-mainline-default | |
make -f CMakeFiles/replit-mainline-default.dir/build.make CMakeFiles/replit-mainline-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/replit-mainline-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/replit-mainline-default.dir/build.make CMakeFiles/replit-mainline-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 15%] [32mBuilding CXX object CMakeFiles/replit-mainline-default.dir/replit.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dreplit_mainline_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-default.dir/replit.cpp.o -MF CMakeFiles/replit-mainline-default.dir/replit.cpp.o.d -o CMakeFiles/replit-mainline-default.dir/replit.cpp.o -c /build/gpt4all/gpt4all-backend/replit.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:[m[K In function '[01m[Kbool replit_eval(const replit_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::n_nodes[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
555 | struct ggml_cgraph gf = {.n_threads = n_threads[01;35m[K}[m[K; | |
| [01;35m[K^[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::n_leafs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::work_size[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::work[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::nodes[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::grads[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::leafs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_runs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_cycles[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_time_us[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[ 16%] [32mBuilding CXX object CMakeFiles/replit-mainline-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dreplit_mainline_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-default.dir/utils.cpp.o -MF CMakeFiles/replit-mainline-default.dir/utils.cpp.o.d -o CMakeFiles/replit-mainline-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 17%] [32mBuilding CXX object CMakeFiles/replit-mainline-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dreplit_mainline_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/replit-mainline-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/replit-mainline-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 18%] [32m[1mLinking CXX shared library libreplit-mainline-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/replit-mainline-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libreplit-mainline-default.so -o libreplit-mainline-default.so "CMakeFiles/replit-mainline-default.dir/replit.cpp.o" "CMakeFiles/replit-mainline-default.dir/utils.cpp.o" "CMakeFiles/replit-mainline-default.dir/llmodel_shared.cpp.o" libllama-mainline-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 18%] Built target replit-mainline-default | |
make -f CMakeFiles/llamamodel-230519-default.dir/build.make CMakeFiles/llamamodel-230519-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-230519-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-230519-default.dir/build.make CMakeFiles/llamamodel-230519-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 20%] [32mBuilding CXX object CMakeFiles/llamamodel-230519-default.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=230519 -DLLAMA_VERSIONS===2 -Dllamamodel_230519_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230519-default.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-230519-default.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-230519-default.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[ 21%] [32mBuilding CXX object CMakeFiles/llamamodel-230519-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=230519 -DLLAMA_VERSIONS===2 -Dllamamodel_230519_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230519-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-230519-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-230519-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 22%] [32m[1mLinking CXX shared library libllamamodel-230519-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-230519-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-230519-default.so -o libllamamodel-230519-default.so "CMakeFiles/llamamodel-230519-default.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-230519-default.dir/llmodel_shared.cpp.o" libllama-230519-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 22%] Built target llamamodel-230519-default | |
make -f CMakeFiles/llamamodel-230511-default.dir/build.make CMakeFiles/llamamodel-230511-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-230511-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-230511-default.dir/build.make CMakeFiles/llamamodel-230511-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 23%] [32mBuilding CXX object CMakeFiles/llamamodel-230511-default.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=230511 -DLLAMA_VERSIONS="<=1" -Dllamamodel_230511_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230511-default.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-230511-default.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-230511-default.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[ 24%] [32mBuilding CXX object CMakeFiles/llamamodel-230511-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -DLLAMA_DATE=230511 -DLLAMA_VERSIONS="<=1" -Dllamamodel_230511_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230511-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-230511-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-230511-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 25%] [32m[1mLinking CXX shared library libllamamodel-230511-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-230511-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-230511-default.so -o libllamamodel-230511-default.so "CMakeFiles/llamamodel-230511-default.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-230511-default.dir/llmodel_shared.cpp.o" libllama-230511-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 25%] Built target llamamodel-230511-default | |
make -f CMakeFiles/gptj-default.dir/build.make CMakeFiles/gptj-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/gptj-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/gptj-default.dir/build.make CMakeFiles/gptj-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 26%] [32mBuilding CXX object CMakeFiles/gptj-default.dir/gptj.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dgptj_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-default.dir/gptj.cpp.o -MF CMakeFiles/gptj-default.dir/gptj.cpp.o.d -o CMakeFiles/gptj-default.dir/gptj.cpp.o -c /build/gpt4all/gpt4all-backend/gptj.cpp | |
[ 27%] [32mBuilding CXX object CMakeFiles/gptj-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dgptj_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-default.dir/utils.cpp.o -MF CMakeFiles/gptj-default.dir/utils.cpp.o.d -o CMakeFiles/gptj-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 28%] [32mBuilding CXX object CMakeFiles/gptj-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dgptj_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/gptj-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/gptj-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 30%] [32m[1mLinking CXX shared library libgptj-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/gptj-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libgptj-default.so -o libgptj-default.so "CMakeFiles/gptj-default.dir/gptj.cpp.o" "CMakeFiles/gptj-default.dir/utils.cpp.o" "CMakeFiles/gptj-default.dir/llmodel_shared.cpp.o" "CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o" -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 30%] Built target gptj-default | |
make -f CMakeFiles/falcon-default.dir/build.make CMakeFiles/falcon-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/falcon-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/falcon-default.dir/build.make CMakeFiles/falcon-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 31%] [32mBuilding CXX object CMakeFiles/falcon-default.dir/falcon.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dfalcon_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-default.dir/falcon.cpp.o -MF CMakeFiles/falcon-default.dir/falcon.cpp.o.d -o CMakeFiles/falcon-default.dir/falcon.cpp.o -c /build/gpt4all/gpt4all-backend/falcon.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:[m[K In function '[01m[Kbool falcon_model_load(const string&, falcon_model&, gpt_vocab&, size_t*)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:199:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_ctx[m[K' [[01;35m[K-Wunused-variable[m[K] | |
199 | const int [01;35m[Kn_ctx[m[K = hparams.n_ctx; | |
| [01;35m[K^~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:340:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_head_kv[m[K' [[01;35m[K-Wunused-variable[m[K] | |
340 | const int [01;35m[Kn_head_kv[m[K = hparams.n_head_kv; | |
| [01;35m[K^~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:344:23:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_elements[m[K' [[01;35m[K-Wunused-variable[m[K] | |
344 | const int64_t [01;35m[Kn_elements[m[K = head_dim*n_mem; | |
| [01;35m[K^~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:[m[K In function '[01m[Kbool falcon_eval(const falcon_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:465:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kversion[m[K' [[01;35m[K-Wunused-variable[m[K] | |
465 | const int [01;35m[Kversion[m[K = hparams.falcon_version; | |
| [01;35m[K^~~~~~~[m[K | |
[ 32%] [32mBuilding CXX object CMakeFiles/falcon-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dfalcon_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-default.dir/utils.cpp.o -MF CMakeFiles/falcon-default.dir/utils.cpp.o.d -o CMakeFiles/falcon-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 33%] [32mBuilding CXX object CMakeFiles/falcon-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dfalcon_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/falcon-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/falcon-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 34%] [32m[1mLinking CXX shared library libfalcon-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libfalcon-default.so -o libfalcon-default.so "CMakeFiles/falcon-default.dir/falcon.cpp.o" "CMakeFiles/falcon-default.dir/utils.cpp.o" "CMakeFiles/falcon-default.dir/llmodel_shared.cpp.o" libllama-mainline-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 34%] Built target falcon-default | |
make -f CMakeFiles/mpt-default.dir/build.make CMakeFiles/mpt-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/mpt-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/mpt-default.dir/build.make CMakeFiles/mpt-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 35%] [32mBuilding CXX object CMakeFiles/mpt-default.dir/mpt.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dmpt_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-default.dir/mpt.cpp.o -MF CMakeFiles/mpt-default.dir/mpt.cpp.o.d -o CMakeFiles/mpt-default.dir/mpt.cpp.o -c /build/gpt4all/gpt4all-backend/mpt.cpp | |
[ 36%] [32mBuilding CXX object CMakeFiles/mpt-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dmpt_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-default.dir/utils.cpp.o -MF CMakeFiles/mpt-default.dir/utils.cpp.o.d -o CMakeFiles/mpt-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 37%] [32mBuilding CXX object CMakeFiles/mpt-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dmpt_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/mpt-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/mpt-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 38%] [32m[1mLinking CXX shared library libmpt-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/mpt-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libmpt-default.so -o libmpt-default.so "CMakeFiles/mpt-default.dir/mpt.cpp.o" "CMakeFiles/mpt-default.dir/utils.cpp.o" "CMakeFiles/mpt-default.dir/llmodel_shared.cpp.o" "CMakeFiles/ggml-230511-default.dir/llama.cpp-230511/ggml.c.o" -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 38%] Built target mpt-default | |
make -f CMakeFiles/bert-default.dir/build.make CMakeFiles/bert-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/bert-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/bert-default.dir/build.make CMakeFiles/bert-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 40%] [32mBuilding CXX object CMakeFiles/bert-default.dir/bert.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dbert_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-default.dir/bert.cpp.o -MF CMakeFiles/bert-default.dir/bert.cpp.o.d -o CMakeFiles/bert-default.dir/bert.cpp.o -c /build/gpt4all/gpt4all-backend/bert.cpp | |
[ 41%] [32mBuilding CXX object CMakeFiles/bert-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dbert_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-default.dir/utils.cpp.o -MF CMakeFiles/bert-default.dir/utils.cpp.o.d -o CMakeFiles/bert-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 42%] [32mBuilding CXX object CMakeFiles/bert-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dbert_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/bert-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/bert-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 43%] [32m[1mLinking CXX shared library libbert-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/bert-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libbert-default.so -o libbert-default.so "CMakeFiles/bert-default.dir/bert.cpp.o" "CMakeFiles/bert-default.dir/utils.cpp.o" "CMakeFiles/bert-default.dir/llmodel_shared.cpp.o" libllama-mainline-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 43%] Built target bert-default | |
make -f CMakeFiles/starcoder-default.dir/build.make CMakeFiles/starcoder-default.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/starcoder-default.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/starcoder-default.dir/build.make CMakeFiles/starcoder-default.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 44%] [32mBuilding CXX object CMakeFiles/starcoder-default.dir/starcoder.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dstarcoder_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-default.dir/starcoder.cpp.o -MF CMakeFiles/starcoder-default.dir/starcoder.cpp.o.d -o CMakeFiles/starcoder-default.dir/starcoder.cpp.o -c /build/gpt4all/gpt4all-backend/starcoder.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/starcoder.cpp:[m[K In function '[01m[Kbool starcoder_eval(const starcoder_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/starcoder.cpp:470:18:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Khead_dim[m[K' [[01;35m[K-Wunused-variable[m[K] | |
470 | const size_t [01;35m[Khead_dim[m[K = n_embd / n_head; | |
| [01;35m[K^~~~~~~~[m[K | |
[ 45%] [32mBuilding CXX object CMakeFiles/starcoder-default.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dstarcoder_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-default.dir/utils.cpp.o -MF CMakeFiles/starcoder-default.dir/utils.cpp.o.d -o CMakeFiles/starcoder-default.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 46%] [32mBuilding CXX object CMakeFiles/starcoder-default.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"default\" -Dstarcoder_default_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-default.dir/llmodel_shared.cpp.o -MF CMakeFiles/starcoder-default.dir/llmodel_shared.cpp.o.d -o CMakeFiles/starcoder-default.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 47%] [32m[1mLinking CXX shared library libstarcoder-default.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/starcoder-default.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libstarcoder-default.so -o libstarcoder-default.so "CMakeFiles/starcoder-default.dir/starcoder.cpp.o" "CMakeFiles/starcoder-default.dir/utils.cpp.o" "CMakeFiles/starcoder-default.dir/llmodel_shared.cpp.o" libllama-mainline-default.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 47%] Built target starcoder-default | |
make -f CMakeFiles/ggml-mainline-avxonly.dir/build.make CMakeFiles/ggml-mainline-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-mainline-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-mainline-avxonly.dir/build.make CMakeFiles/ggml-mainline-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 48%] [32mBuilding C object CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/ggml.c.o[0m | |
/usr/bin/cc -DGGML_USE_K_QUANTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -MD -MT CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/ggml.c.o -MF CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/ggml.c.o.d -o CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c:[m[K In function '[01m[Kquantize_row_q8_0[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c:1096:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1096 | const int [01;35m[Knb[m[K = k / QK8_0; | |
| [01;35m[K^~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c:[m[K In function '[01m[Kquantize_row_q8_1[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml.c:1291:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1291 | const int [01;35m[Knb[m[K = k / QK8_1; | |
| [01;35m[K^~[m[K | |
[ 50%] [32mBuilding C object CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/k_quants.c.o[0m | |
/usr/bin/cc -DGGML_USE_K_QUANTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -MD -MT CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/k_quants.c.o -MF CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/k_quants.c.o.d -o CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/k_quants.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/k_quants.c | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 50%] Built target ggml-mainline-avxonly | |
make -f CMakeFiles/llama-mainline-avxonly.dir/build.make CMakeFiles/llama-mainline-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-mainline-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-mainline-avxonly.dir/build.make CMakeFiles/llama-mainline-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 51%] [32mBuilding CXX object CMakeFiles/llama-mainline-avxonly.dir/llama.cpp-mainline/llama.cpp.o[0m | |
/usr/bin/c++ -DGGML_USE_K_QUANTS -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-mainline-avxonly.dir/llama.cpp-mainline/llama.cpp.o -MF CMakeFiles/llama-mainline-avxonly.dir/llama.cpp-mainline/llama.cpp.o.d -o CMakeFiles/llama-mainline-avxonly.dir/llama.cpp-mainline/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.cpp | |
[ 52%] [32m[1mLinking CXX static library libllama-mainline-avxonly.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-mainline-avxonly.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-mainline-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-mainline-avxonly.a "CMakeFiles/llama-mainline-avxonly.dir/llama.cpp-mainline/llama.cpp.o" "CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/ggml.c.o" "CMakeFiles/ggml-mainline-avxonly.dir/llama.cpp-mainline/k_quants.c.o" | |
/usr/bin/ranlib libllama-mainline-avxonly.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 52%] Built target llama-mainline-avxonly | |
make -f CMakeFiles/ggml-230511-avxonly.dir/build.make CMakeFiles/ggml-230511-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-230511-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-230511-avxonly.dir/build.make CMakeFiles/ggml-230511-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 53%] [32mBuilding C object CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o[0m | |
/usr/bin/cc -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -MD -MT CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o -MF CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o.d -o CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kquantize_row_q4_0[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:781:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
781 | const int [01;35m[Knb[m[K = k / QK4_0; | |
| [01;35m[K^~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kquantize_row_q4_1[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:1129:27:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Ky[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1129 | block_q4_1 * restrict [01;35m[Ky[m[K = vy; | |
| [01;35m[K^[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:1127:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1127 | const int [01;35m[Knb[m[K = k / QK4_1; | |
| [01;35m[K^~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kquantize_row_q8_1[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:1507:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1507 | const int [01;35m[Knb[m[K = k / QK8_1; | |
| [01;35m[K^~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi_f32[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9357:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kne2_ne3[m[K' [[01;35m[K-Wunused-variable[m[K] | |
9357 | const int [01;35m[Kne2_ne3[m[K = n/ne1; // ne2*ne3 | |
| [01;35m[K^~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi_f16[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9419:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kne2[m[K' [[01;35m[K-Wunused-variable[m[K] | |
9419 | const int [01;35m[Kne2[m[K = src0->ne[2]; // n_head -> this is k | |
| [01;35m[K^~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:[m[K In function '[01m[Kggml_compute_forward_alibi[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230511/ggml.c:9468:5:[m[K [01;35m[Kwarning: [m[Kenumeration value '[01m[KGGML_TYPE_Q4_3[m[K' not handled in switch [[01;35m[K-Wswitch[m[K] | |
9468 | [01;35m[Kswitch[m[K (src0->type) { | |
| [01;35m[K^~~~~~[m[K | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 53%] Built target ggml-230511-avxonly | |
make -f CMakeFiles/llama-230511-avxonly.dir/build.make CMakeFiles/llama-230511-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-230511-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-230511-avxonly.dir/build.make CMakeFiles/llama-230511-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 54%] [32mBuilding CXX object CMakeFiles/llama-230511-avxonly.dir/llama.cpp-230511/llama.cpp.o[0m | |
/usr/bin/c++ -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-230511-avxonly.dir/llama.cpp-230511/llama.cpp.o -MF CMakeFiles/llama-230511-avxonly.dir/llama.cpp-230511/llama.cpp.o.d -o CMakeFiles/llama-230511-avxonly.dir/llama.cpp-230511/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230511/llama.cpp | |
[ 55%] [32m[1mLinking CXX static library libllama-230511-avxonly.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-230511-avxonly.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-230511-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-230511-avxonly.a "CMakeFiles/llama-230511-avxonly.dir/llama.cpp-230511/llama.cpp.o" "CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o" | |
/usr/bin/ranlib libllama-230511-avxonly.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 55%] Built target llama-230511-avxonly | |
make -f CMakeFiles/ggml-230519-avxonly.dir/build.make CMakeFiles/ggml-230519-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/ggml-230519-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/ggml-230519-avxonly.dir/build.make CMakeFiles/ggml-230519-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 56%] [32mBuilding C object CMakeFiles/ggml-230519-avxonly.dir/llama.cpp-230519/ggml.c.o[0m | |
/usr/bin/cc -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu11 -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -MD -MT CMakeFiles/ggml-230519-avxonly.dir/llama.cpp-230519/ggml.c.o -MF CMakeFiles/ggml-230519-avxonly.dir/llama.cpp-230519/ggml.c.o.d -o CMakeFiles/ggml-230519-avxonly.dir/llama.cpp-230519/ggml.c.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c:[m[K In function '[01m[Kquantize_row_q8_0[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c:1025:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1025 | const int [01;35m[Knb[m[K = k / QK8_0; | |
| [01;35m[K^~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c:[m[K In function '[01m[Kquantize_row_q8_1[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/ggml.c:1187:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Knb[m[K' [[01;35m[K-Wunused-variable[m[K] | |
1187 | const int [01;35m[Knb[m[K = k / QK8_1; | |
| [01;35m[K^~[m[K | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 56%] Built target ggml-230519-avxonly | |
make -f CMakeFiles/llama-230519-avxonly.dir/build.make CMakeFiles/llama-230519-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llama-230519-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llama-230519-avxonly.dir/build.make CMakeFiles/llama-230519-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 57%] [32mBuilding CXX object CMakeFiles/llama-230519-avxonly.dir/llama.cpp-230519/llama.cpp.o[0m | |
/usr/bin/c++ -DLLAMA_BUILD -DLLAMA_SHARED -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/llama-230519-avxonly.dir/llama.cpp-230519/llama.cpp.o -MF CMakeFiles/llama-230519-avxonly.dir/llama.cpp-230519/llama.cpp.o.d -o CMakeFiles/llama-230519-avxonly.dir/llama.cpp-230519/llama.cpp.o -c /build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:[m[K In function '[01m[Ksize_t llama_set_state_data(llama_context*, const uint8_t*)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:2685:27:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst uint8_t*[m[K' {aka '[01m[Kconst unsigned char*[m[K'} to type '[01m[Kvoid*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
2685 | kin3d->data = [01;35m[K(void *) inp[m[K; | |
| [01;35m[K^~~~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-230519/llama.cpp:2689:27:[m[K [01;35m[Kwarning: [m[Kcast from type '[01m[Kconst uint8_t*[m[K' {aka '[01m[Kconst unsigned char*[m[K'} to type '[01m[Kvoid*[m[K' casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
2689 | vin3d->data = [01;35m[K(void *) inp[m[K; | |
| [01;35m[K^~~~~~~~~~~~[m[K | |
[ 58%] [32m[1mLinking CXX static library libllama-230519-avxonly.a[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -P CMakeFiles/llama-230519-avxonly.dir/cmake_clean_target.cmake | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llama-230519-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/ar qc libllama-230519-avxonly.a "CMakeFiles/llama-230519-avxonly.dir/llama.cpp-230519/llama.cpp.o" "CMakeFiles/ggml-230519-avxonly.dir/llama.cpp-230519/ggml.c.o" | |
/usr/bin/ranlib libllama-230519-avxonly.a | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 58%] Built target llama-230519-avxonly | |
make -f CMakeFiles/llamamodel-mainline-avxonly.dir/build.make CMakeFiles/llamamodel-mainline-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-mainline-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-mainline-avxonly.dir/build.make CMakeFiles/llamamodel-mainline-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 60%] [32mBuilding CXX object CMakeFiles/llamamodel-mainline-avxonly.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=999999 -DLLAMA_VERSIONS=">=3" -Dllamamodel_mainline_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-mainline-avxonly.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-mainline-avxonly.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-mainline-avxonly.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:[m[K In member function '[01m[Kvirtual bool LLamaModel::loadModel(const string&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:159:71:[m[K [01;35m[Kwarning: [m[K'[01m[Kllama_context* llama_init_from_file(const char*, llama_context_params)[m[K' is deprecated: please use llama_load_model_from_file combined with llama_new_context_with_model instead [[01;35m[K-Wdeprecated-declarations[m[K] | |
159 | d_ptr->ctx = llama_init_from_file(modelPath.c_str(), d_ptr->params[01;35m[K)[m[K; | |
| [01;35m[K^[m[K | |
In file included from [01m[K/build/gpt4all/gpt4all-backend/llamamodel.cpp:28[m[K: | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.h:161:49:[m[K [01;36m[Knote: [m[Kdeclared here | |
161 | LLAMA_API DEPRECATED(struct llama_context * [01;36m[Kllama_init_from_file[m[K( | |
| [01;36m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/llama.cpp-mainline/llama.h:30:36:[m[K [01;36m[Knote: [m[Kin definition of macro '[01m[KDEPRECATED[m[K' | |
30 | # define DEPRECATED(func, hint) [01;36m[Kfunc[m[K __attribute__((deprecated(hint))) | |
| [01;36m[K^~~~[m[K | |
[ 61%] [32mBuilding CXX object CMakeFiles/llamamodel-mainline-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=999999 -DLLAMA_VERSIONS=">=3" -Dllamamodel_mainline_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-mainline-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-mainline-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-mainline-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 62%] [32m[1mLinking CXX shared library libllamamodel-mainline-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-mainline-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-mainline-avxonly.so -o libllamamodel-mainline-avxonly.so "CMakeFiles/llamamodel-mainline-avxonly.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-mainline-avxonly.dir/llmodel_shared.cpp.o" libllama-mainline-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 62%] Built target llamamodel-mainline-avxonly | |
make -f CMakeFiles/replit-mainline-avxonly.dir/build.make CMakeFiles/replit-mainline-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/replit-mainline-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/replit-mainline-avxonly.dir/build.make CMakeFiles/replit-mainline-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 63%] [32mBuilding CXX object CMakeFiles/replit-mainline-avxonly.dir/replit.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dreplit_mainline_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-avxonly.dir/replit.cpp.o -MF CMakeFiles/replit-mainline-avxonly.dir/replit.cpp.o.d -o CMakeFiles/replit-mainline-avxonly.dir/replit.cpp.o -c /build/gpt4all/gpt4all-backend/replit.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:[m[K In function '[01m[Kbool replit_eval(const replit_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::n_nodes[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
555 | struct ggml_cgraph gf = {.n_threads = n_threads[01;35m[K}[m[K; | |
| [01;35m[K^[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::n_leafs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::work_size[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::work[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::nodes[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::grads[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::leafs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_runs[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_cycles[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[01m[K/build/gpt4all/gpt4all-backend/replit.cpp:555:52:[m[K [01;35m[Kwarning: [m[Kmissing initializer for member '[01m[Kggml_cgraph::perf_time_us[m[K' [[01;35m[K-Wmissing-field-initializers[m[K] | |
[ 64%] [32mBuilding CXX object CMakeFiles/replit-mainline-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dreplit_mainline_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-avxonly.dir/utils.cpp.o -MF CMakeFiles/replit-mainline-avxonly.dir/utils.cpp.o.d -o CMakeFiles/replit-mainline-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 65%] [32mBuilding CXX object CMakeFiles/replit-mainline-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dreplit_mainline_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/replit-mainline-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/replit-mainline-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/replit-mainline-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 66%] [32m[1mLinking CXX shared library libreplit-mainline-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/replit-mainline-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libreplit-mainline-avxonly.so -o libreplit-mainline-avxonly.so "CMakeFiles/replit-mainline-avxonly.dir/replit.cpp.o" "CMakeFiles/replit-mainline-avxonly.dir/utils.cpp.o" "CMakeFiles/replit-mainline-avxonly.dir/llmodel_shared.cpp.o" libllama-mainline-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 66%] Built target replit-mainline-avxonly | |
make -f CMakeFiles/llamamodel-230519-avxonly.dir/build.make CMakeFiles/llamamodel-230519-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-230519-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-230519-avxonly.dir/build.make CMakeFiles/llamamodel-230519-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 67%] [32mBuilding CXX object CMakeFiles/llamamodel-230519-avxonly.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=230519 -DLLAMA_VERSIONS===2 -Dllamamodel_230519_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230519-avxonly.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-230519-avxonly.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-230519-avxonly.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[ 68%] [32mBuilding CXX object CMakeFiles/llamamodel-230519-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=230519 -DLLAMA_VERSIONS===2 -Dllamamodel_230519_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230519 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230519-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-230519-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-230519-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 70%] [32m[1mLinking CXX shared library libllamamodel-230519-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-230519-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-230519-avxonly.so -o libllamamodel-230519-avxonly.so "CMakeFiles/llamamodel-230519-avxonly.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-230519-avxonly.dir/llmodel_shared.cpp.o" libllama-230519-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 70%] Built target llamamodel-230519-avxonly | |
make -f CMakeFiles/llamamodel-230511-avxonly.dir/build.make CMakeFiles/llamamodel-230511-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llamamodel-230511-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llamamodel-230511-avxonly.dir/build.make CMakeFiles/llamamodel-230511-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 71%] [32mBuilding CXX object CMakeFiles/llamamodel-230511-avxonly.dir/llamamodel.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=230511 -DLLAMA_VERSIONS="<=1" -Dllamamodel_230511_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230511-avxonly.dir/llamamodel.cpp.o -MF CMakeFiles/llamamodel-230511-avxonly.dir/llamamodel.cpp.o.d -o CMakeFiles/llamamodel-230511-avxonly.dir/llamamodel.cpp.o -c /build/gpt4all/gpt4all-backend/llamamodel.cpp | |
[ 72%] [32mBuilding CXX object CMakeFiles/llamamodel-230511-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -DLLAMA_DATE=230511 -DLLAMA_VERSIONS="<=1" -Dllamamodel_230511_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llamamodel-230511-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/llamamodel-230511-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llamamodel-230511-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 73%] [32m[1mLinking CXX shared library libllamamodel-230511-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llamamodel-230511-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllamamodel-230511-avxonly.so -o libllamamodel-230511-avxonly.so "CMakeFiles/llamamodel-230511-avxonly.dir/llamamodel.cpp.o" "CMakeFiles/llamamodel-230511-avxonly.dir/llmodel_shared.cpp.o" libllama-230511-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 73%] Built target llamamodel-230511-avxonly | |
make -f CMakeFiles/gptj-avxonly.dir/build.make CMakeFiles/gptj-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/gptj-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/gptj-avxonly.dir/build.make CMakeFiles/gptj-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 74%] [32mBuilding CXX object CMakeFiles/gptj-avxonly.dir/gptj.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dgptj_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-avxonly.dir/gptj.cpp.o -MF CMakeFiles/gptj-avxonly.dir/gptj.cpp.o.d -o CMakeFiles/gptj-avxonly.dir/gptj.cpp.o -c /build/gpt4all/gpt4all-backend/gptj.cpp | |
[ 75%] [32mBuilding CXX object CMakeFiles/gptj-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dgptj_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-avxonly.dir/utils.cpp.o -MF CMakeFiles/gptj-avxonly.dir/utils.cpp.o.d -o CMakeFiles/gptj-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 76%] [32mBuilding CXX object CMakeFiles/gptj-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dgptj_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/gptj-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/gptj-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/gptj-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 77%] [32m[1mLinking CXX shared library libgptj-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/gptj-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libgptj-avxonly.so -o libgptj-avxonly.so "CMakeFiles/gptj-avxonly.dir/gptj.cpp.o" "CMakeFiles/gptj-avxonly.dir/utils.cpp.o" "CMakeFiles/gptj-avxonly.dir/llmodel_shared.cpp.o" "CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o" -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 77%] Built target gptj-avxonly | |
make -f CMakeFiles/falcon-avxonly.dir/build.make CMakeFiles/falcon-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/falcon-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/falcon-avxonly.dir/build.make CMakeFiles/falcon-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 78%] [32mBuilding CXX object CMakeFiles/falcon-avxonly.dir/falcon.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dfalcon_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-avxonly.dir/falcon.cpp.o -MF CMakeFiles/falcon-avxonly.dir/falcon.cpp.o.d -o CMakeFiles/falcon-avxonly.dir/falcon.cpp.o -c /build/gpt4all/gpt4all-backend/falcon.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:[m[K In function '[01m[Kbool falcon_model_load(const string&, falcon_model&, gpt_vocab&, size_t*)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:199:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_ctx[m[K' [[01;35m[K-Wunused-variable[m[K] | |
199 | const int [01;35m[Kn_ctx[m[K = hparams.n_ctx; | |
| [01;35m[K^~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:340:19:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_head_kv[m[K' [[01;35m[K-Wunused-variable[m[K] | |
340 | const int [01;35m[Kn_head_kv[m[K = hparams.n_head_kv; | |
| [01;35m[K^~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:344:23:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kn_elements[m[K' [[01;35m[K-Wunused-variable[m[K] | |
344 | const int64_t [01;35m[Kn_elements[m[K = head_dim*n_mem; | |
| [01;35m[K^~~~~~~~~~[m[K | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:[m[K In function '[01m[Kbool falcon_eval(const falcon_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/falcon.cpp:465:15:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Kversion[m[K' [[01;35m[K-Wunused-variable[m[K] | |
465 | const int [01;35m[Kversion[m[K = hparams.falcon_version; | |
| [01;35m[K^~~~~~~[m[K | |
[ 80%] [32mBuilding CXX object CMakeFiles/falcon-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dfalcon_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-avxonly.dir/utils.cpp.o -MF CMakeFiles/falcon-avxonly.dir/utils.cpp.o.d -o CMakeFiles/falcon-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 81%] [32mBuilding CXX object CMakeFiles/falcon-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dfalcon_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/falcon-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/falcon-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/falcon-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 82%] [32m[1mLinking CXX shared library libfalcon-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/falcon-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libfalcon-avxonly.so -o libfalcon-avxonly.so "CMakeFiles/falcon-avxonly.dir/falcon.cpp.o" "CMakeFiles/falcon-avxonly.dir/utils.cpp.o" "CMakeFiles/falcon-avxonly.dir/llmodel_shared.cpp.o" libllama-mainline-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 82%] Built target falcon-avxonly | |
make -f CMakeFiles/mpt-avxonly.dir/build.make CMakeFiles/mpt-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/mpt-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/mpt-avxonly.dir/build.make CMakeFiles/mpt-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 83%] [32mBuilding CXX object CMakeFiles/mpt-avxonly.dir/mpt.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dmpt_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-avxonly.dir/mpt.cpp.o -MF CMakeFiles/mpt-avxonly.dir/mpt.cpp.o.d -o CMakeFiles/mpt-avxonly.dir/mpt.cpp.o -c /build/gpt4all/gpt4all-backend/mpt.cpp | |
[ 84%] [32mBuilding CXX object CMakeFiles/mpt-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dmpt_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-avxonly.dir/utils.cpp.o -MF CMakeFiles/mpt-avxonly.dir/utils.cpp.o.d -o CMakeFiles/mpt-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 85%] [32mBuilding CXX object CMakeFiles/mpt-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dmpt_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-230511 -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -MD -MT CMakeFiles/mpt-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/mpt-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/mpt-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 86%] [32m[1mLinking CXX shared library libmpt-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/mpt-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libmpt-avxonly.so -o libmpt-avxonly.so "CMakeFiles/mpt-avxonly.dir/mpt.cpp.o" "CMakeFiles/mpt-avxonly.dir/utils.cpp.o" "CMakeFiles/mpt-avxonly.dir/llmodel_shared.cpp.o" "CMakeFiles/ggml-230511-avxonly.dir/llama.cpp-230511/ggml.c.o" -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 86%] Built target mpt-avxonly | |
make -f CMakeFiles/bert-avxonly.dir/build.make CMakeFiles/bert-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/bert-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/bert-avxonly.dir/build.make CMakeFiles/bert-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 87%] [32mBuilding CXX object CMakeFiles/bert-avxonly.dir/bert.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dbert_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-avxonly.dir/bert.cpp.o -MF CMakeFiles/bert-avxonly.dir/bert.cpp.o.d -o CMakeFiles/bert-avxonly.dir/bert.cpp.o -c /build/gpt4all/gpt4all-backend/bert.cpp | |
[ 88%] [32mBuilding CXX object CMakeFiles/bert-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dbert_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-avxonly.dir/utils.cpp.o -MF CMakeFiles/bert-avxonly.dir/utils.cpp.o.d -o CMakeFiles/bert-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 90%] [32mBuilding CXX object CMakeFiles/bert-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dbert_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/bert-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/bert-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/bert-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 91%] [32m[1mLinking CXX shared library libbert-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/bert-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libbert-avxonly.so -o libbert-avxonly.so "CMakeFiles/bert-avxonly.dir/bert.cpp.o" "CMakeFiles/bert-avxonly.dir/utils.cpp.o" "CMakeFiles/bert-avxonly.dir/llmodel_shared.cpp.o" libllama-mainline-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 91%] Built target bert-avxonly | |
make -f CMakeFiles/starcoder-avxonly.dir/build.make CMakeFiles/starcoder-avxonly.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/starcoder-avxonly.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/starcoder-avxonly.dir/build.make CMakeFiles/starcoder-avxonly.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 92%] [32mBuilding CXX object CMakeFiles/starcoder-avxonly.dir/starcoder.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dstarcoder_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-avxonly.dir/starcoder.cpp.o -MF CMakeFiles/starcoder-avxonly.dir/starcoder.cpp.o.d -o CMakeFiles/starcoder-avxonly.dir/starcoder.cpp.o -c /build/gpt4all/gpt4all-backend/starcoder.cpp | |
[01m[K/build/gpt4all/gpt4all-backend/starcoder.cpp:[m[K In function '[01m[Kbool starcoder_eval(const starcoder_model&, int, int, const std::vector<int>&, std::vector<float>&, size_t&)[m[K': | |
[01m[K/build/gpt4all/gpt4all-backend/starcoder.cpp:470:18:[m[K [01;35m[Kwarning: [m[Kunused variable '[01m[Khead_dim[m[K' [[01;35m[K-Wunused-variable[m[K] | |
470 | const size_t [01;35m[Khead_dim[m[K = n_embd / n_head; | |
| [01;35m[K^~~~~~~~[m[K | |
[ 93%] [32mBuilding CXX object CMakeFiles/starcoder-avxonly.dir/utils.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dstarcoder_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-avxonly.dir/utils.cpp.o -MF CMakeFiles/starcoder-avxonly.dir/utils.cpp.o.d -o CMakeFiles/starcoder-avxonly.dir/utils.cpp.o -c /build/gpt4all/gpt4all-backend/utils.cpp | |
[ 94%] [32mBuilding CXX object CMakeFiles/starcoder-avxonly.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DGGML_BUILD_VARIANT=\"avxonly\" -Dstarcoder_avxonly_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -I/build/gpt4all/gpt4all-backend/llama.cpp-mainline -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/starcoder-avxonly.dir/llmodel_shared.cpp.o -MF CMakeFiles/starcoder-avxonly.dir/llmodel_shared.cpp.o.d -o CMakeFiles/starcoder-avxonly.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 95%] [32m[1mLinking CXX shared library libstarcoder-avxonly.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/starcoder-avxonly.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libstarcoder-avxonly.so -o libstarcoder-avxonly.so "CMakeFiles/starcoder-avxonly.dir/starcoder.cpp.o" "CMakeFiles/starcoder-avxonly.dir/utils.cpp.o" "CMakeFiles/starcoder-avxonly.dir/llmodel_shared.cpp.o" libllama-mainline-avxonly.a -pthread | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 95%] Built target starcoder-avxonly | |
make -f CMakeFiles/llmodel.dir/build.make CMakeFiles/llmodel.dir/depend | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd /build/gpt4all/gpt4all-bindings/golang/buildllm && /usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_depends "Unix Makefiles" /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-backend /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles/llmodel.dir/DependInfo.cmake "--color=" | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
make -f CMakeFiles/llmodel.dir/build.make CMakeFiles/llmodel.dir/build | |
make[4]: Entering directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[ 96%] [32mBuilding CXX object CMakeFiles/llmodel.dir/llmodel.cpp.o[0m | |
/usr/bin/c++ -DLIB_FILE_EXT=\".so\" -Dllmodel_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llmodel.dir/llmodel.cpp.o -MF CMakeFiles/llmodel.dir/llmodel.cpp.o.d -o CMakeFiles/llmodel.dir/llmodel.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel.cpp | |
[ 97%] [32mBuilding CXX object CMakeFiles/llmodel.dir/llmodel_shared.cpp.o[0m | |
/usr/bin/c++ -DLIB_FILE_EXT=\".so\" -Dllmodel_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llmodel.dir/llmodel_shared.cpp.o -MF CMakeFiles/llmodel.dir/llmodel_shared.cpp.o.d -o CMakeFiles/llmodel.dir/llmodel_shared.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_shared.cpp | |
[ 98%] [32mBuilding CXX object CMakeFiles/llmodel.dir/llmodel_c.cpp.o[0m | |
/usr/bin/c++ -DLIB_FILE_EXT=\".so\" -Dllmodel_EXPORTS -I/build/gpt4all/gpt4all-bindings/golang/buildllm -O3 -DNDEBUG -std=gnu++2a -fPIC -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -MD -MT CMakeFiles/llmodel.dir/llmodel_c.cpp.o -MF CMakeFiles/llmodel.dir/llmodel_c.cpp.o.d -o CMakeFiles/llmodel.dir/llmodel_c.cpp.o -c /build/gpt4all/gpt4all-backend/llmodel_c.cpp | |
[100%] [32m[1mLinking CXX shared library libllmodel.so[0m | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_link_script CMakeFiles/llmodel.dir/link.txt --verbose=1 | |
/usr/bin/c++ -fPIC -O3 -DNDEBUG -shared -Wl,-soname,libllmodel.so.0 -o libllmodel.so.0.3.0 CMakeFiles/llmodel.dir/llmodel.cpp.o CMakeFiles/llmodel.dir/llmodel_shared.cpp.o CMakeFiles/llmodel.dir/llmodel_c.cpp.o | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_symlink_library libllmodel.so.0.3.0 libllmodel.so.0 libllmodel.so | |
make[4]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
[100%] Built target llmodel | |
make[3]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
/usr/local/lib/python3.9/dist-packages/cmake/data/bin/cmake -E cmake_progress_start /build/gpt4all/gpt4all-bindings/golang/buildllm/CMakeFiles 0 | |
make[2]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang/buildllm' | |
cd buildllm && cp -rf CMakeFiles/llmodel.dir/llmodel_c.cpp.o ../llmodel_c.o | |
cd buildllm && cp -rf CMakeFiles/llmodel.dir/llmodel.cpp.o ../llmodel.o | |
ar src libgpt4all.a llmodel.o binding.o | |
make[1]: Leaving directory '/build/gpt4all/gpt4all-bindings/golang' | |
mkdir -p backend-assets/gpt4all | |
cp: cannot stat 'gpt4all/gpt4all-bindings/golang/buildllm/*.dylib': No such file or directory | |
cp: cannot stat 'gpt4all/gpt4all-bindings/golang/buildllm/*.dll': No such file or directory | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/gpt4all/gpt4all-bindings/golang/ LIBRARY_PATH=/build/gpt4all/gpt4all-bindings/golang/ \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/gpt4all ./cmd/grpc/gpt4all/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/dolly ./cmd/grpc/dolly/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/gpt2 ./cmd/grpc/gpt2/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/gptj ./cmd/grpc/gptj/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/gptneox ./cmd/grpc/gptneox/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/mpt ./cmd/grpc/mpt/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/replit ./cmd/grpc/replit/ | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-ggml-transformers LIBRARY_PATH=/build/go-ggml-transformers \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/starcoder ./cmd/grpc/starcoder/ | |
cd go-rwkv && cd rwkv.cpp && cmake . -DRWKV_BUILD_SHARED_LIBRARY=OFF && cmake --build . && cp librwkv.a .. | |
-- The C compiler identification is GNU 10.2.1 | |
-- The CXX compiler identification is GNU 10.2.1 | |
-- Detecting C compiler ABI info | |
-- Detecting C compiler ABI info - done | |
-- Check for working C compiler: /usr/bin/cc - skipped | |
-- Detecting C compile features | |
-- Detecting C compile features - done | |
-- Detecting CXX compiler ABI info | |
-- Detecting CXX compiler ABI info - done | |
-- Check for working CXX compiler: /usr/bin/c++ - skipped | |
-- Detecting CXX compile features | |
-- Detecting CXX compile features - done | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD | |
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed | |
-- Check if compiler accepts -pthread | |
-- Check if compiler accepts -pthread - yes | |
-- Found Threads: TRUE | |
-- CMAKE_SYSTEM_PROCESSOR: x86_64 | |
-- x86 detected | |
-- Configuring done (0.4s) | |
-- Generating done (0.0s) | |
-- Build files have been written to: /build/go-rwkv/rwkv.cpp | |
gmake[1]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[2]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 7%] [32mBuilding C object CMakeFiles/ggml.dir/ggml/src/ggml.c.o[0m | |
[01m[K/build/go-rwkv/rwkv.cpp/ggml/src/ggml.c:[m[K In function ‘[01m[Kggml_compute_forward_win_part_f32[m[K’: | |
[01m[K/build/go-rwkv/rwkv.cpp/ggml/src/ggml.c:13064:19:[m[K [01;35m[Kwarning: [m[Kunused variable ‘[01m[Kne3[m[K’ [[01;35m[K-Wunused-variable[m[K] | |
13064 | const int64_t [01;35m[Kne3[m[K = dst->ne[3]; | |
| [01;35m[K^~~[m[K | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 7%] Built target ggml | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 15%] [32mBuilding CXX object CMakeFiles/rwkv.dir/rwkv.cpp.o[0m | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:[m[K In function ‘[01m[Kbool rwkv_fread_string(FILE*, size_t, std::string&)[m[K’: | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:149:18:[m[K [01;35m[Kwarning: [m[Kcast from type ‘[01m[Kconst char*[m[K’ to type ‘[01m[Kvoid*[m[K’ casts away qualifiers [[01;35m[K-Wcast-qual[m[K] | |
149 | return fread([01;35m[K(void *) dest.data()[m[K, length, 1, file) == 1; | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:[m[K At global scope: | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:223:21:[m[K [01;35m[Kwarning: [m[K‘[01m[Krwkv_type_to_string[m[K’ initialized and declared ‘[01m[Kextern[m[K’ | |
223 | extern const char * [01;35m[Krwkv_type_to_string[m[K[TYPE_COUNT + 1] = {"float32", "float16", "Q4_0", "Q4_1", "Q4_1_O", "Q4_2", "Q4_3", "Q5_0", "Q5_1", "Q8_0", "unknown"}; | |
| [01;35m[K^~~~~~~~~~~~~~~~~~~[m[K | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:[m[K In function ‘[01m[Kbool rwkv_gpu_offload_layers(const rwkv_context*, uint32_t)[m[K’: | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:1280:58:[m[K [01;35m[Kwarning: [m[Kunused parameter ‘[01m[Kctx[m[K’ [[01;35m[K-Wunused-parameter[m[K] | |
1280 | bool rwkv_gpu_offload_layers([01;35m[Kconst struct rwkv_context * ctx[m[K, const uint32_t n_gpu_layers) { | |
| [01;35m[K~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~[m[K | |
[01m[K/build/go-rwkv/rwkv.cpp/rwkv.cpp:1280:78:[m[K [01;35m[Kwarning: [m[Kunused parameter ‘[01m[Kn_gpu_layers[m[K’ [[01;35m[K-Wunused-parameter[m[K] | |
1280 | bool rwkv_gpu_offload_layers(const struct rwkv_context * ctx, [01;35m[Kconst uint32_t n_gpu_layers[m[K) { | |
| [01;35m[K~~~~~~~~~~~~~~~^~~~~~~~~~~~[m[K | |
[ 23%] [32m[1mLinking CXX static library librwkv.a[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 23%] Built target rwkv | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 30%] [32mBuilding C object tests/CMakeFiles/test_ggml_basics.dir/test_ggml_basics.c.o[0m | |
[ 38%] [32m[1mLinking CXX executable ../bin/test_ggml_basics[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 38%] Built target test_ggml_basics | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 46%] [32mBuilding C object tests/CMakeFiles/test_tiny_rwkv.dir/test_tiny_rwkv.c.o[0m | |
[01m[K/build/go-rwkv/rwkv.cpp/tests/test_tiny_rwkv.c:[m[K In function ‘[01m[Ktest_model[m[K’: | |
[01m[K/build/go-rwkv/rwkv.cpp/tests/test_tiny_rwkv.c:60:45:[m[K [01;35m[Kwarning: [m[Kimplicit conversion from ‘[01m[Kfloat[m[K’ to ‘[01m[Kdouble[m[K’ when passing argument to function [[01;35m[K-Wdouble-promotion[m[K] | |
60 | fprintf(stderr, "Difference sum: %f\n", [01;35m[Kdiff_sum[m[K); | |
| [01;35m[K^~~~~~~~[m[K | |
[01m[K/build/go-rwkv/rwkv.cpp/tests/test_tiny_rwkv.c:74:54:[m[K [01;35m[Kwarning: [m[Kimplicit conversion from ‘[01m[Kfloat[m[K’ to ‘[01m[Kdouble[m[K’ when passing argument to function [[01;35m[K-Wdouble-promotion[m[K] | |
74 | fprintf(stderr, "Sequence difference sum: %f\n", [01;35m[Kdiff_sum[m[K); | |
| [01;35m[K^~~~~~~~[m[K | |
[ 53%] [32m[1mLinking CXX executable ../bin/test_tiny_rwkv[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 53%] Built target test_tiny_rwkv | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 61%] [32mBuilding C object tests/CMakeFiles/test_context_cloning.dir/test_context_cloning.c.o[0m | |
[01m[K/build/go-rwkv/rwkv.cpp/tests/test_context_cloning.c:7:5:[m[K [01;35m[Kwarning: [m[Kfunction declaration isn’t a prototype [[01;35m[K-Wstrict-prototypes[m[K] | |
7 | int [01;35m[Kmain[m[K() { | |
| [01;35m[K^~~~[m[K | |
[ 69%] [32m[1mLinking CXX executable ../bin/test_context_cloning[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 69%] Built target test_context_cloning | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 76%] [32mBuilding C object extras/CMakeFiles/rwkv_cpu_info.dir/cpu_info.c.o[0m | |
[01m[K/build/go-rwkv/rwkv.cpp/extras/cpu_info.c:5:5:[m[K [01;35m[Kwarning: [m[Kfunction declaration isn’t a prototype [[01;35m[K-Wstrict-prototypes[m[K] | |
5 | int [01;35m[Kmain[m[K() { | |
| [01;35m[K^~~~[m[K | |
[ 84%] [32m[1mLinking CXX executable ../bin/rwkv_cpu_info[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[ 84%] Built target rwkv_cpu_info | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[3]: Entering directory '/build/go-rwkv/rwkv.cpp' | |
[ 92%] [32mBuilding C object extras/CMakeFiles/rwkv_quantize.dir/quantize.c.o[0m | |
[100%] [32m[1mLinking CXX executable ../bin/rwkv_quantize[0m | |
gmake[3]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
[100%] Built target rwkv_quantize | |
gmake[2]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
gmake[1]: Leaving directory '/build/go-rwkv/rwkv.cpp' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/go-rwkv LIBRARY_PATH=/build/go-rwkv \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/rwkv ./cmd/grpc/rwkv/ | |
cd whisper.cpp && make libwhisper.a | |
make[1]: Entering directory '/build/whisper.cpp' | |
I whisper.cpp build info: | |
I UNAME_S: Linux | |
I UNAME_P: unknown | |
I UNAME_M: x86_64 | |
I CFLAGS: -I. -O3 -DNDEBUG -std=c11 -fPIC -D_XOPEN_SOURCE=600 -pthread -mavx2 -mfma -mf16c -mavx -msse3 | |
I CXXFLAGS: -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -D_XOPEN_SOURCE=600 -pthread | |
I LDFLAGS: | |
I CC: cc (Debian 10.2.1-6) 10.2.1 20210110 | |
I CXX: g++ (Debian 10.2.1-6) 10.2.1 20210110 | |
cc -I. -O3 -DNDEBUG -std=c11 -fPIC -D_XOPEN_SOURCE=600 -pthread -mavx2 -mfma -mf16c -mavx -msse3 -c ggml.c -o ggml.o | |
g++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -D_XOPEN_SOURCE=600 -pthread -c whisper.cpp -o whisper.o | |
ar rcs libwhisper.a ggml.o whisper.o | |
make[1]: Leaving directory '/build/whisper.cpp' | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" C_INCLUDE_PATH=/build/whisper.cpp LIBRARY_PATH=/build/whisper.cpp \ | |
go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o backend-assets/grpc/whisper ./cmd/grpc/whisper/ | |
[32mI local-ai build info:(B[m | |
[32mI BUILD_TYPE: [33mcublas(B[m | |
[32mI GO_TAGS: [33m(B[m | |
[32mI LD_FLAGS: [33m-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"(B[m | |
CGO_LDFLAGS="-lcublas -lcudart -L/usr/local/cuda/lib64/" go build -ldflags "-X "github.com/go-skynet/LocalAI/internal.Version=v1.23.2" -X "github.com/go-skynet/LocalAI/internal.Commit=acd829a7a0e1623c0871c8b34c36c76afd4feac8"" -tags "" -o local-ai ./ | |
[90m5:19PM[0m [33mDBG[0m no galleries to load | |
[90m5:19PM[0m [32mINF[0m Starting LocalAI using 1 threads, with models path: /models | |
[90m5:19PM[0m [32mINF[0m LocalAI version: v1.23.2 (acd829a7a0e1623c0871c8b34c36c76afd4feac8) | |
[90m5:19PM[0m [33mDBG[0m Model: gpt-3.5-turbo (config: {PredictionOptions:{Model:open-llama-7b-q4_0.bin Language: N:0 TopP:0.7 TopK:80 Temperature:0.2 Maxtokens:0 Echo:false Batch:0 F16:false IgnoreEOS:false RepeatPenalty:0 Keep:0 MirostatETA:0 MirostatTAU:0 Mirostat:0 FrequencyPenalty:0 TFZ:0 TypicalP:0 Seed:0 NegativePrompt: RopeFreqBase:0 RopeFreqScale:0 NegativePromptScale:0} Name:gpt-3.5-turbo StopWords:[] Cutstrings:[] TrimSpace:[] ContextSize:1024 F16:true NUMA:false Threads:0 Debug:false Roles:map[] Embeddings:false Backend:llama TemplateConfig:{Chat:openllama-chat ChatMessage: Completion:openllama-completion Edit: Functions:} MirostatETA:0 MirostatTAU:0 Mirostat:0 NGPULayers:35 MMap:true MMlock:false LowVRAM:false TensorSplit: MainGPU: ImageGenerationAssets: PromptCachePath: PromptCacheAll:false PromptCacheRO:false Grammar: PromptStrings:[] InputStrings:[] InputToken:[] functionCallString: functionCallNameString: FunctionsConfig:{DisableNoAction:false NoActionFunctionName: NoActionDescriptionName:} SystemPrompt: RMSNormEps:0 NGQA:0}) | |
[90m5:19PM[0m [33mDBG[0m Extracting backend assets files to /tmp/localai/backend_data | |
[90m5:19PM[0m [33mDBG[0m Config overrides map[batch:512 f16:true gpu_layers:35 mmap:true] | |
[90m5:19PM[0m [33mDBG[0m Checking "open-llama-7b-q4_0.bin" exists and matches SHA | |
[90m5:20PM[0m [33mDBG[0m File "open-llama-7b-q4_0.bin" already exists and matches the SHA. Skipping download | |
[90m5:20PM[0m [33mDBG[0m Prompt template "openllama-completion" written | |
[90m5:20PM[0m [33mDBG[0m Prompt template "openllama-chat" written | |
[90m5:20PM[0m [33mDBG[0m Written config file /models/gpt-3.5-turbo.yaml | |
[90m[90m ┌───────────────────────────────────────────────────┐ | |
[90m │ [96mFiber v2.48.0[90m │ | |
[90m │ http://127.0.0.1:8080 │ | |
[90m │ (bound on host 0.0.0.0 and port 8080) │ | |
[90m │ │ | |
[90m │ Handlers ............ [96m32[90m Processes ........... [96m1[90m │ | |
[90m │ Prefork ....... Disabled PID ............. [96m11147[90m │ | |
[90m └───────────────────────────────────────────────────┘[0m | |
[127.0.0.1]:53244 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:51742 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:51422 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:52380 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:50440 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:45034 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:58848 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:39190 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:51318 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:40980 [92m 200 [0m - [96m GET [0m /readyz | |
[127.0.0.1]:52126 [92m 200 [0m - [96m GET [0m /readyz | |
[172.17.0.1]:58314 [92m 200 [0m - [96m GET [0m /v1/models | |
[90m5:31PM[0m [33mDBG[0m Request received: | |
[90m5:31PM[0m [33mDBG[0m Configuration read: &{PredictionOptions:{Model:luna-ai-llama2 Language: N:0 TopP:0.7 TopK:80 Temperature:0.9 Maxtokens:512 Echo:false Batch:0 F16:false IgnoreEOS:false RepeatPenalty:0 Keep:0 MirostatETA:0 MirostatTAU:0 Mirostat:0 FrequencyPenalty:0 TFZ:0 TypicalP:0 Seed:0 NegativePrompt: RopeFreqBase:0 RopeFreqScale:0 NegativePromptScale:0} Name: StopWords:[] Cutstrings:[] TrimSpace:[] ContextSize:512 F16:false NUMA:false Threads:1 Debug:true Roles:map[] Embeddings:false Backend: TemplateConfig:{Chat: ChatMessage: Completion: Edit: Functions:} MirostatETA:0 MirostatTAU:0 Mirostat:0 NGPULayers:0 MMap:false MMlock:false LowVRAM:false TensorSplit: MainGPU: ImageGenerationAssets: PromptCachePath: PromptCacheAll:false PromptCacheRO:false Grammar: PromptStrings:[] InputStrings:[] InputToken:[] functionCallString: functionCallNameString: FunctionsConfig:{DisableNoAction:false NoActionFunctionName: NoActionDescriptionName:} SystemPrompt: RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m Parameters: &{PredictionOptions:{Model:luna-ai-llama2 Language: N:0 TopP:0.7 TopK:80 Temperature:0.9 Maxtokens:512 Echo:false Batch:0 F16:false IgnoreEOS:false RepeatPenalty:0 Keep:0 MirostatETA:0 MirostatTAU:0 Mirostat:0 FrequencyPenalty:0 TFZ:0 TypicalP:0 Seed:0 NegativePrompt: RopeFreqBase:0 RopeFreqScale:0 NegativePromptScale:0} Name: StopWords:[] Cutstrings:[] TrimSpace:[] ContextSize:512 F16:false NUMA:false Threads:1 Debug:true Roles:map[] Embeddings:false Backend: TemplateConfig:{Chat: ChatMessage: Completion: Edit: Functions:} MirostatETA:0 MirostatTAU:0 Mirostat:0 NGPULayers:0 MMap:false MMlock:false LowVRAM:false TensorSplit: MainGPU: ImageGenerationAssets: PromptCachePath: PromptCacheAll:false PromptCacheRO:false Grammar: PromptStrings:[] InputStrings:[] InputToken:[] functionCallString: functionCallNameString: FunctionsConfig:{DisableNoAction:false NoActionFunctionName: NoActionDescriptionName:} SystemPrompt: RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m Prompt (before templating): How are you? | |
[90m5:31PM[0m [33mDBG[0m Template failed loading: failed loading a template for luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Prompt (after templating): How are you? | |
[90m5:31PM[0m [33mDBG[0m Loading model 'luna-ai-llama2' greedly from all the available backends: llama, gpt4all, falcon, gptneox, bert-embeddings, falcon-ggml, gptj, gpt2, dolly, mpt, replit, starcoder, bloomz, rwkv, whisper, stablediffusion, piper, /build/extra/grpc/huggingface/huggingface.py | |
[90m5:31PM[0m [33mDBG[0m [llama] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model llama from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model llama: {backendString:llama modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/llama | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:33457' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager3620426260 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:33457: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr 2023/11/09 17:31:16 gRPC Server listening at 127.0.0.1:33457 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath: RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr create_gpt_params: loading model /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr ggml_init_cublas: found 1 CUDA devices: | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr Device 0: NVIDIA GeForce RTX 3070 Ti Laptop GPU, compute capability 8.6 | |
[90m5:31PM[0m [33mDBG[0m [llama] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [gpt4all] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model gpt4all from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model gpt4all: {backendString:gpt4all modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/gpt4all | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:36555' | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr llama.cpp: loading model from /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr error loading model: unknown (magic, version) combination: 46554747, 00000002; is this really a GGML file? | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr llama_load_model_from_file: failed to load model | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr llama_init_from_gpt_params: error: failed to load model '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33457): stderr load_binding_model: error: unable to load model | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager1874264123 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:36555: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:36555): stderr 2023/11/09 17:31:17 gRPC Server listening at 127.0.0.1:36555 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:36555): stderr load_model: error 'Model format not supported (no matching implementation found)' | |
[90m5:31PM[0m [33mDBG[0m [gpt4all] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [falcon] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model falcon from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model falcon: {backendString:falcon modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/falcon | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:34157' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager2617030949 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:34157: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:34157): stderr 2023/11/09 17:31:18 gRPC Server listening at 127.0.0.1:34157 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:34157): stderr falcon.cpp: loading model from /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:34157): stderr error loading model: unknown (magic, version) combination: 46554747, 00000002; is this really a GGML file? | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:34157): stderr falcon_init_from_file: failed to load model | |
[90m5:31PM[0m [33mDBG[0m [falcon] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [gptneox] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model gptneox from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model gptneox: {backendString:gptneox modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/gptneox | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:39909' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager3713018119 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:39909: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:39909): stderr 2023/11/09 17:31:20 gRPC Server listening at 127.0.0.1:39909 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m [gptneox] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:39909): stderr gpt_neox_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m [bert-embeddings] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model bert-embeddings from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:39909): stderr gpt_neox_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model bert-embeddings: {backendString:bert-embeddings modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/bert-embeddings | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:46105' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager3859495679 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:46105: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46105): stderr 2023/11/09 17:31:21 gRPC Server listening at 127.0.0.1:46105 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46105): stderr bert_load_from_file: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m [bert-embeddings] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [falcon-ggml] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model falcon-ggml from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model falcon-ggml: {backendString:falcon-ggml modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46105): stderr bert_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/falcon-ggml | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:40631' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager2904493726 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:40631: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:40631): stderr 2023/11/09 17:31:22 gRPC Server listening at 127.0.0.1:40631 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:40631): stderr falcon_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:40631): stderr falcon_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m [falcon-ggml] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [gptj] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model gptj from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model gptj: {backendString:gptj modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/gptj | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:37217' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager2121419443 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:37217: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:37217): stderr 2023/11/09 17:31:23 gRPC Server listening at 127.0.0.1:37217 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m [gptj] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [gpt2] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:37217): stderr gptj_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m Loading model gpt2 from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model gpt2: {backendString:gpt2 modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:37217): stderr gptj_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/gpt2 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:46395' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager927256336 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:46395: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46395): stderr 2023/11/09 17:31:24 gRPC Server listening at 127.0.0.1:46395 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m [gpt2] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [dolly] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model dolly from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model dolly: {backendString:dolly modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46395): stderr gpt2_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46395): stderr gpt2_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/dolly | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:46733' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager3702163237 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:46733: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46733): stderr 2023/11/09 17:31:25 gRPC Server listening at 127.0.0.1:46733 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46733): stderr dollyv2_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:46733): stderr dolly_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m [dolly] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [mpt] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model mpt from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model mpt: {backendString:mpt modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/mpt | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:38383' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager371655322 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:38383: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38383): stderr 2023/11/09 17:31:26 gRPC Server listening at 127.0.0.1:38383 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m [mpt] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [replit] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model replit from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model replit: {backendString:replit modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38383): stderr mpt_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38383): stderr mpt_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/replit | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:33943' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager847316612 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:33943: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33943): stderr 2023/11/09 17:31:27 gRPC Server listening at 127.0.0.1:33943 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33943): stderr replit_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:33943): stderr replit_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m [replit] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [starcoder] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model starcoder from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model starcoder: {backendString:starcoder modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/starcoder | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:38617' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager1188628188 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:38617: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38617): stderr 2023/11/09 17:31:28 gRPC Server listening at 127.0.0.1:38617 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38617): stderr starcoder_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:38617): stderr starcoder_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m [starcoder] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [bloomz] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model bloomz from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model bloomz: {backendString:bloomz modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/bloomz | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:41523' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager1643108513 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:41523: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:41523): stderr 2023/11/09 17:31:29 gRPC Server listening at 127.0.0.1:41523 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:41523): stderr bloom_model_load: invalid model file '/models/luna-ai-llama2' (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:41523): stderr bloomz_bootstrap: failed to load model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m [bloomz] Fails: could not load model: rpc error: code = Unknown desc = failed loading model | |
[90m5:31PM[0m [33mDBG[0m [rwkv] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model rwkv from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model rwkv: {backendString:rwkv modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/rwkv | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:42799' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager880251401 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:42799: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr 2023/11/09 17:31:30 gRPC Server listening at 127.0.0.1:42799 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/rwkv.cpp/rwkv.cpp:250: header.magic == 0x67676d66 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr Invalid file header | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/rwkv.cpp/rwkv.cpp:1132: rwkv_fread_file_header(file.file, model.header) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/rwkv.cpp/rwkv.cpp:1266: rwkv_instance_from_file(file_path, *instance.get()) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr panic: runtime error: invalid memory address or nil pointer dereference | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr [signal SIGSEGV: segmentation violation code=0x1 addr=0x0 pc=0x524c54] | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr goroutine 50 [running]: | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/donomii/go-rwkv%2ecpp.(*Context).GetStateBufferElementCount.func1(0xc0002041d0?) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/wrapper.go:63 +0x14 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/donomii/go-rwkv%2ecpp.(*Context).GetStateBufferElementCount(0xc000214030?) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/wrapper.go:63 +0x19 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/donomii/go-rwkv%2ecpp.LoadFiles({0xc000214030?, 0xc000214038?}, {0xc000266030, 0x25}, 0x53?) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/go-rwkv/wrapper.go:131 +0x5d | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/go-skynet/LocalAI/pkg/grpc/llm/rwkv.(*LLM).Load(0xc0000d8290, 0xc00022c140) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/pkg/grpc/llm/rwkv/rwkv.go:25 +0xcf | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/go-skynet/LocalAI/pkg/grpc.(*server).LoadModel(0x910940?, {0xc00022c140?, 0x5cb826?}, 0x0?) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/pkg/grpc/server.go:42 +0x28 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr github.com/go-skynet/LocalAI/pkg/grpc/proto._Backend_LoadModel_Handler({0x8ee3c0?, 0xc0000b9d10}, {0x9d1c10, 0xc0002003c0}, 0xc000208150, 0x0) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /build/pkg/grpc/proto/backend_grpc.pb.go:236 +0x170 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr google.golang.org/grpc.(*Server).processUnaryRPC(0xc0001741e0, {0x9d4898, 0xc0002b8340}, 0xc000232000, 0xc00017c9f0, 0xc86530, 0x0) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /go/pkg/mod/google.golang.org/grpc@v1.57.0/server.go:1360 +0xe23 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr google.golang.org/grpc.(*Server).handleStream(0xc0001741e0, {0x9d4898, 0xc0002b8340}, 0xc000232000, 0x0) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /go/pkg/mod/google.golang.org/grpc@v1.57.0/server.go:1737 +0xa36 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr google.golang.org/grpc.(*Server).serveStreams.func1.1() | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /go/pkg/mod/google.golang.org/grpc@v1.57.0/server.go:982 +0x98 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr created by google.golang.org/grpc.(*Server).serveStreams.func1 | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:42799): stderr /go/pkg/mod/google.golang.org/grpc@v1.57.0/server.go:980 +0x18c | |
[90m5:31PM[0m [33mDBG[0m [rwkv] Fails: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF | |
[90m5:31PM[0m [33mDBG[0m [whisper] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model whisper from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model whisper: {backendString:whisper modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Process: /tmp/localai/backend_data/backend-assets/grpc/whisper | |
[90m5:31PM[0m [33mDBG[0m GRPC Service for luna-ai-llama2 will be running at: '127.0.0.1:35467' | |
[90m5:31PM[0m [33mDBG[0m GRPC Service state dir: /tmp/go-processmanager3358758835 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Started | |
rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 127.0.0.1:35467: connect: connection refused" | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:35467): stderr 2023/11/09 17:31:31 gRPC Server listening at 127.0.0.1:35467 | |
[90m5:31PM[0m [33mDBG[0m GRPC Service Ready | |
[90m5:31PM[0m [33mDBG[0m GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:<nil>} sizeCache:0 unknownFields:[] Model:/models/luna-ai-llama2 ContextSize:512 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:false NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:1 LibrarySearchPath:/tmp/localai/backend_data/backend-assets/gpt4all RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0} | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:35467): stderr whisper_init_from_file_no_state: loading model from '/models/luna-ai-llama2' | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:35467): stderr whisper_model_load: loading model | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:35467): stderr whisper_model_load: invalid model data (bad magic) | |
[90m5:31PM[0m [33mDBG[0m GRPC(luna-ai-llama2-127.0.0.1:35467): stderr whisper_init_no_state: failed to load model | |
[90m5:31PM[0m [33mDBG[0m [whisper] Fails: could not load model: rpc error: code = Unknown desc = unable to load model | |
[90m5:31PM[0m [33mDBG[0m [stablediffusion] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model stablediffusion from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model stablediffusion: {backendString:stablediffusion modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m [stablediffusion] Fails: grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/stablediffusion. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS | |
[90m5:31PM[0m [33mDBG[0m [piper] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model piper from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading model in memory from file: /models/luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m Loading GRPC Model piper: {backendString:piper modelFile:luna-ai-llama2 threads:1 assetDir:/tmp/localai/backend_data context:0xc00003e0b0 gRPCOptions:0xc0006c40a0 externalBackends:map[huggingface-embeddings:/build/extra/grpc/huggingface/huggingface.py]} | |
[90m5:31PM[0m [33mDBG[0m [piper] Fails: grpc process not found: /tmp/localai/backend_data/backend-assets/grpc/piper. some backends(stablediffusion, tts) require LocalAI compiled with GO_TAGS | |
[90m5:31PM[0m [33mDBG[0m [/build/extra/grpc/huggingface/huggingface.py] Attempting to load | |
[90m5:31PM[0m [33mDBG[0m Loading model /build/extra/grpc/huggingface/huggingface.py from luna-ai-llama2 | |
[90m5:31PM[0m [33mDBG[0m [/build/extra/grpc/huggingface/huggingface.py] Fails: backend unsupported: /build/extra/grpc/huggingface/huggingface.py | |
[172.17.0.1]:58330 [91m 500 [0m - [92m POST [0m /v1/chat/completions | |
[127.0.0.1]:34174 [92m 200 [0m - [96m GET [0m /readyz |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment