Skip to content

Instantly share code, notes, and snippets.

View vilaca's full-sized avatar
🎯
Focusing

João Vilaça vilaca

🎯
Focusing
View GitHub Profile
@adrienbrault
adrienbrault / llama2-mac-gpu.sh
Last active April 22, 2024 08:47
Run Llama-2-13B-chat locally on your M1/M2 Mac with GPU inference. Uses 10GB RAM. UPDATE: see https://twitter.com/simonw/status/1691495807319674880?s=20
# Clone llama.cpp
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
# Build it
make clean
LLAMA_METAL=1 make
# Download model
export MODEL=llama-2-13b-chat.ggmlv3.q4_0.bin
@blockloop
blockloop / docker-compose.yaml
Last active April 23, 2022 02:41
Thanos Receive example with docker-compose
version: "3"
networks:
default:
driver: bridge
ipam:
config:
- subnet: 172.16.57.0/24
services:
@steve-chavez
steve-chavez / map.sql
Last active January 31, 2022 02:25
Create a OSM map from a pg function with PostgREST
-- From
-- + http://duspviz.mit.edu/tutorials/intro-postgis.php
-- + http://duspviz.mit.edu/web-map-workshop/leaflet_nodejs_postgis/
CREATE TABLE coffee_shops
(
id serial NOT NULL,
name character varying(50),
address character varying(50),
city character varying(50),
state character varying(50),
@O1ahmad
O1ahmad / Alertmanager CLI options
Created January 5, 2020 23:02
Alertmanager command-line interface (CLI) options
$ ./alertmanager -h [14/1934]
usage: alertmanager [<flags>]
Flags:
-h, --help Show context-sensitive help (also try --help-long and --help-man).
--config.file="alertmanager.yml"
Alertmanager configuration file name.