Skip to content

Instantly share code, notes, and snippets.

@slopp
slopp / README.md
Last active November 19, 2025 18:18
Dynamic pipeline that invokes k8 ops

Dynamic Pipeline

This example shows the psuedo-code for a Dagster pipeline that:

  1. Accepts the path to a raw dataset as a string
  2. Runs a step to break the raw dataset into partitions
  3. For each partition, the pipeline runs a series of two processing steps. Each processing step is a call out to a Docker container to run supplying the partition key as an input argument. The partitions are run together in parallel before being collected in a final processing step that operates on all the partitions.

To run the pipeline:

@slopp
slopp / penguins.csv
Created March 31, 2021 20:07
Palmer Penguins Dataset as CSV
rowid species island bill_length_mm bill_depth_mm flipper_length_mm body_mass_g sex year
1 Adelie Torgersen 39.1 18.7 181 3750 male 2007
2 Adelie Torgersen 39.5 17.4 186 3800 female 2007
3 Adelie Torgersen 40.3 18 195 3250 female 2007
4 Adelie Torgersen NA NA NA NA NA 2007
5 Adelie Torgersen 36.7 19.3 193 3450 female 2007
6 Adelie Torgersen 39.3 20.6 190 3650 male 2007
7 Adelie Torgersen 38.9 17.8 181 3625 female 2007
8 Adelie Torgersen 39.2 19.6 195 4675 male 2007
9 Adelie Torgersen 34.1 18.1 193 3475 NA 2007
@slopp
slopp / credits.sql
Created September 28, 2023 21:01
Credits Dagster OSS
with events as (select distinct
DATE_FORMAT(timestamp, '%Y-%m') as event_month,
dagster_event_type,
coalesce(run_id, '||', step_key) as step_id,
count(1) as credits
from event_logs
where dagster_event_type = 'STEP_START'

Dynamo

Executive Summary

This report provides a comprehensive comparison of NVIDIA Dynamo with other frameworks, including vLLM and NVIDIA Triton Server, for large language model (LLM) inference workloads. The key findings highlight Dynamo's superior performance in multi-GPU setups, achieving higher throughput and lower latency compared to vLLM. Additionally, Dynamo's disaggregated serving approach offers more flexible performance tuning, especially for large models and variable workload conditions. In comparison to NVIDIA Triton Server, Dynamo is optimized for low-latency generative AI/LLM workloads, while Triton Server excels in multi-model inference serving. The report also delves into Dynamo's technical architecture, which is designed to accelerate inference workloads through components such as disaggregated serving, smart routing, and distributed KV cache management. Benchmarking methodologies and key performance metrics are also discussed, emphasizing the importance of standardized evaluation for

@slopp
slopp / handout.Rmd
Created October 27, 2017 12:51
LaTex + RMD
---
title: \textbf{USING RSTUDIO WITH TERADATA \\ \large RStudio makes it easy to access and analyze your data with R}
geometry: margin=0.6in
output:
pdf_document:
fig_caption: false
pandoc_args: [
"-V", "classoption=twocolumn"
]
---
@slopp
slopp / users_by_role.py
Created July 3, 2023 17:25
Get Users by Role
from gql import Client, gql
from gql.transport.requests import RequestsHTTPTransport
import os
import pandas as pd
from datetime import datetime, timedelta
USER_GRANTS_QUERY = """
query UsersByRole {
usersOrError {
@slopp
slopp / README.md
Created January 2, 2025 22:37
LangGraph Exploration

A modification of the LangChain SQL Q&A tutorial https://python.langchain.com/docs/tutorials/sql_qa/.

The changes are:

  • uses Pydantic to type the state and inputs/outputs
  • uses duckDB on the palmerpenguin dataset
  • uses a Nvidia NIM for the LLM
  • instead of a sequence write_query -> run_query -> gen_answer, this graph adds a LLM that checks the write_query output for validity and to see if it answers the question, leading to a more dynamic graph that looks like this:

graph

@slopp
slopp / README.md
Last active January 2, 2025 22:32
Simple LLM agent to recommend fake coffee shops via tool calling
@slopp
slopp / README.md
Last active January 2, 2025 22:32
Code for creating a RAG chatbot based on theradavist
@slopp
slopp / README.md
Created December 31, 2024 00:10
AI Coffee Shop Streamlit App

This simple streamlit app uses the Google Maps and Places API, along with a hosted Nvidia NIM wrapper of the Llama model, to help you find coffee shops near an address.

Screenshot 2024-12-30 at 5 03 38 PM

To run

  1. Install dependencies