Created
October 28, 2023 05:51
-
-
Save AnttiRae/dc5c09f40d2474caa959155a88b7426d to your computer and use it in GitHub Desktop.
llama-gpt-amd - /run.sh --model 7b --with-rocm
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
❯ ./run.sh --model 7b --with-rocm | |
[+] Building 227.3s (31/31) FINISHED docker:default | |
=> [llama-gpt-api-rocm-ggml internal] load build definition from ggml.Dockerfile 0.2s | |
=> => transferring dockerfile: 920B 0.0s | |
=> [llama-gpt-api-rocm-ggml internal] load .dockerignore 0.2s | |
=> => transferring context: 2B 0.0s | |
=> [llama-gpt-ui internal] load build definition from Dockerfile 0.3s | |
=> => transferring dockerfile: 918B 0.0s | |
=> [llama-gpt-ui internal] load .dockerignore 0.2s | |
=> => transferring context: 141B 0.0s | |
=> [llama-gpt-api-rocm-ggml internal] load metadata for docker.io/rocm/dev-ubuntu-22.04:latest 0.0s | |
=> [llama-gpt-api-rocm-ggml 1/5] FROM docker.io/rocm/dev-ubuntu-22.04 0.3s | |
=> [llama-gpt-api-rocm-ggml internal] load build context 0.3s | |
=> => transferring context: 3.60kB 0.0s | |
=> [llama-gpt-ui internal] load metadata for ghcr.io/ufoscout/docker-compose-wait:latest 1.5s | |
=> [llama-gpt-ui internal] load metadata for docker.io/library/node:19-alpine 1.9s | |
=> [llama-gpt-api-rocm-ggml 2/5] RUN apt-get update && apt-get upgrade -y && apt-get install -y git build-essential python3 python3-pip gcc wget rocm-dev rocm-libs rocblas hipblas && mkdir 167.0s | |
=> [llama-gpt-ui auth] library/node:pull token for registry-1.docker.io 0.0s | |
=> [llama-gpt-ui production 1/9] FROM docker.io/library/node:19-alpine@sha256:8ec543d4795e2e85af924a24f8acb039792ae9fe8a42ad5b4bf4c277ab34b62e 11.9s | |
=> => resolve docker.io/library/node:19-alpine@sha256:8ec543d4795e2e85af924a24f8acb039792ae9fe8a42ad5b4bf4c277ab34b62e 0.1s | |
=> => sha256:8ec543d4795e2e85af924a24f8acb039792ae9fe8a42ad5b4bf4c277ab34b62e 1.43kB / 1.43kB 0.0s | |
=> => sha256:d0ba7111bc031323ce2706f8e424afc868db289ba40ff55b05561cf59c123be1 1.16kB / 1.16kB 0.0s | |
=> => sha256:e2a8cc97f817417787050d381376568c494547f9af9decfca6463dee6db6561c 6.73kB / 6.73kB 0.0s | |
=> => sha256:8a49fdb3b6a5ff2bd8ec6a86c05b2922a0f7454579ecc07637e94dfd1d0639b6 3.40MB / 3.40MB 1.4s | |
=> => sha256:1197750296b3abe1d21ffbb3d3ea76df5ba887cf82c8e3284d267cbb2aa1724a 48.15MB / 48.15MB 9.5s | |
=> => sha256:f352bc07f19b43a8678cc8c8efe162ccb6193ead7af6dd366639a01402d1819e 2.34MB / 2.34MB 3.0s | |
=> => extracting sha256:8a49fdb3b6a5ff2bd8ec6a86c05b2922a0f7454579ecc07637e94dfd1d0639b6 0.1s | |
=> => sha256:47be83a79857fb67c4d144471b8301ae6fb874971bfaa60d12dc97ea1355cffe 449B / 449B 2.6s | |
=> => extracting sha256:1197750296b3abe1d21ffbb3d3ea76df5ba887cf82c8e3284d267cbb2aa1724a 1.4s | |
=> => extracting sha256:f352bc07f19b43a8678cc8c8efe162ccb6193ead7af6dd366639a01402d1819e 0.0s | |
=> => extracting sha256:47be83a79857fb67c4d144471b8301ae6fb874971bfaa60d12dc97ea1355cffe 0.0s | |
=> [llama-gpt-ui] FROM ghcr.io/ufoscout/docker-compose-wait:latest@sha256:ee1b58447dcf9ae2aaf84e5904ffc00ed5a983bf986535b19aeb6f2d4a7ceb8a 3.3s | |
=> => resolve ghcr.io/ufoscout/docker-compose-wait:latest@sha256:ee1b58447dcf9ae2aaf84e5904ffc00ed5a983bf986535b19aeb6f2d4a7ceb8a 0.1s | |
=> => sha256:ee1b58447dcf9ae2aaf84e5904ffc00ed5a983bf986535b19aeb6f2d4a7ceb8a 2.38kB / 2.38kB 0.0s | |
=> => sha256:2ebf0b77a3d10f528b6f924d31a14c49dba25184fbac0f4b6f08c9e26050e3ee 480B / 480B 0.0s | |
=> => sha256:f50c9bd54f7d6e3caefb9a608e6815bfdb1b2a4b6400d940acc09a744b8d63c2 1.08kB / 1.08kB 0.0s | |
=> => sha256:ca20d452e2dc7814422d2b5d30b52f34889adac67dec57a3fd50fa4fac8a242c 257.97kB / 257.97kB 3.0s | |
=> => extracting sha256:ca20d452e2dc7814422d2b5d30b52f34889adac67dec57a3fd50fa4fac8a242c 0.0s | |
=> [llama-gpt-ui internal] load build context 0.2s | |
=> => transferring context: 1.30MB 0.1s | |
=> [llama-gpt-ui production 2/9] WORKDIR /app 0.1s | |
=> [llama-gpt-ui base 3/3] COPY package*.json ./ 0.2s | |
=> [llama-gpt-ui dependencies 1/1] RUN npm ci 27.2s | |
=> [llama-gpt-ui build 1/2] COPY . . 0.2s | |
=> [llama-gpt-ui build 2/2] RUN npm run build 27.0s | |
=> [llama-gpt-ui production 3/9] COPY --from=dependencies /app/node_modules ./node_modules 5.7s | |
=> [llama-gpt-ui production 4/9] COPY --from=build /app/.next ./.next 0.2s | |
=> [llama-gpt-ui production 5/9] COPY --from=build /app/public ./public 0.1s | |
=> [llama-gpt-ui production 6/9] COPY --from=build /app/package*.json ./ 0.7s | |
=> [llama-gpt-ui production 7/9] COPY --from=build /app/next.config.js ./next.config.js 0.1s | |
=> [llama-gpt-ui production 8/9] COPY --from=build /app/next-i18next.config.js ./next-i18next.config.js 0.1s | |
=> [llama-gpt-ui production 9/9] COPY --from=ghcr.io/ufoscout/docker-compose-wait:latest /wait /wait 0.1s | |
=> [llama-gpt-ui] exporting to image 5.7s | |
=> => exporting layers 5.7s | |
=> => writing image sha256:be40e19cca3000a22f5a06b616da972ab09af4708bbe17dd70fe97489e4186fd 0.0s | |
=> => naming to docker.io/library/llama-gpt-amd-llama-gpt-ui 0.0s | |
=> [llama-gpt-api-rocm-ggml 3/5] COPY . . 0.2s | |
=> [llama-gpt-api-rocm-ggml 4/5] RUN python3 -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi uvicorn sse-starlette pydantic-settings 9.6s | |
=> [llama-gpt-api-rocm-ggml 5/5] RUN CMAKE_ARGS="-DLLAMA_HIPBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python==0.1.78 16.0s | |
=> [llama-gpt-api-rocm-ggml] exporting to image 33.8s | |
=> => exporting layers 33.8s | |
=> => writing image sha256:1c6663a2c01288118b12e9af305f8a53528365845197c355c92b5f72b3f835e5 0.0s | |
=> => naming to docker.io/library/llama-gpt-amd-llama-gpt-api-rocm-ggml 0.0s | |
[+] Running 3/3 | |
✔ Network llama-gpt-amd_default Created 0.1s | |
✔ Container llama-gpt-amd-llama-gpt-ui-1 Created 0.1s | |
✔ Container llama-gpt-amd-llama-gpt-api-rocm-ggml-1 Created 0.1s | |
Attaching to llama-gpt-amd-llama-gpt-api-rocm-ggml-1, llama-gpt-amd-llama-gpt-ui-1 | |
llama-gpt-amd-llama-gpt-ui-1 | [INFO wait] -------------------------------------------------------- | |
llama-gpt-amd-llama-gpt-ui-1 | [INFO wait] docker-compose-wait 2.12.1 | |
llama-gpt-amd-llama-gpt-ui-1 | [INFO wait] --------------------------- | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] Starting with configuration: | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Hosts to be waiting for: [llama-gpt-api-rocm-ggml:8000] | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Paths to be waiting for: [] | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Timeout before failure: 3600 seconds | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - TCP connection timeout before retry: 5 seconds | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Sleeping time before checking for hosts/paths availability: 0 seconds | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Sleeping time once all hosts/paths are available: 0 seconds | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] - Sleeping time between retries: 1 seconds | |
llama-gpt-amd-llama-gpt-ui-1 | [DEBUG wait] -------------------------------------------------------- | |
llama-gpt-amd-llama-gpt-ui-1 | [INFO wait] Checking availability of host [llama-gpt-api-rocm-ggml:8000] | |
Error response from daemon: could not select device driver "amdgpu" with capabilities: [[gpu]] |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment