NOTE: commands and UI are deprecated
- Negative Engineering
- What is workflow orchestration?
- Introduction to Prefect 2.0
- First Prefect flow and Basics
import sys | |
from awsglue.transforms import * | |
from awsglue.utils import getResolvedOptions | |
from pyspark.context import SparkContext | |
from awsglue.context import GlueContext | |
from awsglue.job import Job | |
## @params: [JOB_NAME] | |
args = getResolvedOptions(sys.argv, ['JOB_NAME']) | |
bucketpathparam = getResolvedOptions(sys.argv, ['s3_path']) |
The problem with large language models is that you can’t run these locally on your laptop. Thanks to Georgi Gerganov and his llama.cpp project, it is now possible to run Meta’s LLaMA on a single computer without a dedicated GPU.
There are multiple steps involved in running LLaMA locally on a M1 Mac after downloading the model weights.
Let's talk to an Alpaca-7B model using LangChain with a conversational chain and a memory window.
Install python packages using pip. Note that you need to install HuggingFace Transformers from source (GitHub) currently.
$ pip install git+https://github.com/huggingface/transformers
llava
and mistral
BUT it would reload the models every time I switched model requests./usr/share/ollama/.ollama