Feel free to contact me at robert.balicki@gmail.com or tweet at me @statisticsftw
This is a rough outline of how we utilize next.js and S3/Cloudfront. Hope it helps!
It assumes some knowledge of AWS.
# MIT No Attribution | |
# | |
# Copyright 2022 Ben Kehoe | |
# | |
# Permission is hereby granted, free of charge, to any person obtaining a copy of this | |
# software and associated documentation files (the "Software"), to deal in the Software | |
# without restriction, including without limitation the rights to use, copy, modify, | |
# merge, publish, distribute, sublicense, and/or sell copies of the Software, and to | |
# permit persons to whom the Software is furnished to do so. | |
# |
/* | |
Adapted from https://github.com/sindresorhus/github-markdown-css | |
The MIT License (MIT) | |
Copyright (c) Sindre Sorhus <sindresorhus@gmail.com> (sindresorhus.com) | |
Permission is hereby granted, free of charge, to any person obtaining a copy | |
of this software and associated documentation files (the "Software"), to deal | |
in the Software without restriction, including without limitation the rights |
import os | |
import requests | |
PLAYLIST_URI = os.environ['PLAYLIST_URI'] | |
SPOTIFY_TOKEN = os.environ['SPOTIFY_TOKEN'] # from default api login or more easily from chrome network inspecting calls directed to api.spotify.com | |
url = "https://api.spotify.com/v1/playlists/{}".format(PLAYLIST_URI) | |
querystring = {"type": "track,episode", "market": "from_token"} | |
headers = { |
# Adapted from https://unix.stackexchange.com/questions/60577/concatenate-multiple-files-with-same-header#170692 | |
# `head -1` gets first line of the file of the file and stashes them as the header row into `all.txt` | |
# `tail -n +2 -q *.csv >> all.txt` grabs every csv file from the second row down and stashes them into `all.txt` | |
# `all.txt` is a csv file but we use the txt extension to avoid it being captured in the `*.csv` glob. | |
# You could also output to a csv file by having your input files share a naming convention such as `file-001.csv` and glob on `file*.csv` | |
# But renaming `all.txt` to `all.csv` is sometimes easier than worrying about a naming convention and txt and csvs are the same thing | |
head -1 my-csv-01.csv > all.txt; tail -n +2 -q *.csv >> all.txt |
package com.operative.pipelinetracker.controller; | |
import java.util.HashMap; | |
import java.util.Map; | |
import java.util.Properties; | |
import org.apache.kafka.clients.consumer.KafkaConsumer; | |
import org.apache.kafka.clients.consumer.OffsetAndMetadata; | |
import org.apache.kafka.common.TopicPartition; | |
import org.apache.kafka.common.serialization.StringDeserializer; |
# .net core, kubernetes tool stack on MacOS | |
https://www.jetbrains.com/rider/ | |
http://www.mono-project.com/download/stable/ | |
https://www.microsoft.com/net/download/thank-you/dotnet-sdk-2.1.300-macos-x64-installer | |
https://www.microsoft.com/net/download/macos | |
https://github.com/kubernetes/minikube/releases | |
https://kubernetes.io/docs/tasks/tools/install-kubectl/ | |
# dotnet core docker | |
https://hub.docker.com/r/microsoft/dotnet/ |
Feel free to contact me at robert.balicki@gmail.com or tweet at me @statisticsftw
This is a rough outline of how we utilize next.js and S3/Cloudfront. Hope it helps!
It assumes some knowledge of AWS.
package main | |
import ( | |
"context" | |
"flag" | |
"fmt" | |
"log" | |
"net/http" | |
"os" | |
"os/signal" |
Kafka 0.11.0.0 (Confluent 3.3.0) added support to manipulate offsets for a consumer group via cli kafka-consumer-groups
command.
kafka-consumer-groups --bootstrap-server <kafkahost:port> --group <group_id> --describe
Note the values under "CURRENT-OFFSET" and "LOG-END-OFFSET". "CURRENT-OFFSET" is the offset where this consumer group is currently at in each of the partitions.
license: gpl-3.0 | |
height: 480 | |
border: no |