Skip to content

Instantly share code, notes, and snippets.

View jonahgeorge's full-sized avatar

Jonah George jonahgeorge

View GitHub Profile
@jonahgeorge
jonahgeorge / .gitignore
Last active January 30, 2016 19:24 — forked from Lewuathe/shift.py
Optimal assignment of GTAs to Courses via ILP
__pycache__
{
"firstNames": [
"Crimson",
"Cinnamon",
"Silent",
"Sheik",
"Dreamy",
"Wondering",
"Spicy",
"Catchy",
@jonahgeorge
jonahgeorge / playground.rs
Created May 13, 2018 17:09 — forked from rust-play/playground.rs
Code shared from the Rust Playground
use std::collections::HashMap;
use std::fmt::Debug;
pub trait StaticBuilder {
fn build(&self) -> Box<HasName>;
}
struct UserBuilder;
impl StaticBuilder for UserBuilder {
@jonahgeorge
jonahgeorge / semantic-commit-messages.md
Created September 27, 2021 16:13 — forked from joshbuchea/semantic-commit-messages.md
Semantic Commit Messages

Semantic Commit Messages

See how a minor change to your commit message style can make you a better programmer.

Format: <type>(<scope>): <subject>

<scope> is optional

Example

@jonahgeorge
jonahgeorge / llama-7b-m1.md
Created April 4, 2023 05:20 — forked from cedrickchee/llama-7b-m1.md
4 Steps in Running LLaMA-7B on a M1 MacBook with `llama.cpp`

4 Steps in Running LLaMA-7B on a M1 MacBook

The large language models usability

The problem with large language models is that you can’t run these locally on your laptop. Thanks to Georgi Gerganov and his llama.cpp project, it is now possible to run Meta’s LLaMA on a single computer without a dedicated GPU.

Running LLaMA

There are multiple steps involved in running LLaMA locally on a M1 Mac after downloading the model weights.