I hereby claim:
- I am posquit0 on github.
- I am posquit0 (https://keybase.io/posquit0) on keybase.
- I have a public key ASAWooySOb8CkZYPKNDR9_Elv28KtIHwyquzeptBXdN0jQo
To claim this, I am signing this object:
#!/usr/bin/env python | |
# encoding: utf-8 | |
from Queue import Queue | |
from SocketServer import BaseRequestHandler, ThreadingTCPServer | |
from datetime import datetime | |
import threading | |
ThreadingTCPServer.allow_reuse_address = True |
# Add the Oracle JDK Repos | |
UBUNTU_VERSION=precise | |
echo "deb http://ppa.launchpad.net/webupd8team/java/ubuntu ${UBUNTU_VERSION} main" | tee /etc/apt/sources.list.d/webupd8team-java.list | |
echo "deb-src http://ppa.launchpad.net/webupd8team/java/ubuntu ${UBUNTU_VERSION} main" | tee -a /etc/apt/sources.list.d/webupd8team-java.list | |
apt-key adv --keyserver keyserver.ubuntu.com --recv-keys EEA14886 | |
DEBIAN_FRONTEND="noninteractive" apt-get update | |
# Accept the Oracle License | |
echo "oracle-java7-installer shared/accepted-oracle-license-v1-1 boolean true" > /tmp/oracle-license-debconf | |
/usr/bin/debconf-set-selections /tmp/oracle-license-debconf |
I hereby claim:
To claim this, I am signing this object:
apt-get install python2.7-dev python2.7 | |
apt-get build-dep gdb | |
apt-get source gdb | |
sed -i -E "s|python3|/usr/bin/python2.7|" debian/rules | |
dpkg-buildpackage -uc -us -j8 | |
dpkg-install ../*.deb |
This script can be used to feed collectd with cpu and memory usage statistics for running docker containers using the collectd exec
plugin.
This script will report the used and cached memory as well as the user and system cpu usage by inspecting the appropriate cgroup stat file for each running container.
This script is intented to be executed by collectd on a host with running docker containers. To use, simply configure the exec
plugin in collectd to execute the collectd-docker.sh
script. You may need to adjust the script to match your particulars, such as the mount location for cgroup.
public AWSCredentialsProvider getDefaultCredentials() { | |
final String profileName = AwsProfileNameLoader.INSTANCE.loadProfileName(); | |
final AllProfiles allProfiles = new AllProfiles(Stream.concat( | |
BasicProfileConfigLoader.INSTANCE.loadProfiles( | |
AwsProfileFileLocationProvider.DEFAULT_CONFIG_LOCATION_PROVIDER.getLocation()).getProfiles().values().stream(), | |
BasicProfileConfigLoader.INSTANCE.loadProfiles( | |
AwsProfileFileLocationProvider.DEFAULT_CREDENTIALS_LOCATION_PROVIDER.getLocation()).getProfiles().values().stream()) | |
.map(profile -> new BasicProfile(profile.getProfileName().replaceFirst("^profile ", ""), profile.getProperties())) | |
.collect(Collectors.toMap(profile -> profile.getProfileName(), profile -> profile, |
This gist will drive you through creating a Docker 1.12 Swarm cluster (with Swarm mode) on AWS infrastructure.
You need a few things already prepared in order to get started. You need at least Docker 1.12 set up. I was using the stable version of Docker for mac for preparing this guide.
$ docker --version
Docker version 1.12.0, build 8eab29e
You also need Docker machine installed.
I have been an aggressive Kubernetes evangelist over the last few years. It has been the hammer with which I have approached almost all my deployments, and the one tool I have mentioned (shoved down clients throats) in almost all my foremost communications with clients, and it was my go to choice when I was mocking my first startup (saharacluster.com).
A few weeks ago Docker 1.13 was released and I was tasked with replicating a client's Kubernetes deployment on Swarm, more specifically testing running compose on Swarm.
And it was a dream!
All our apps were already dockerised and all I had to do was make a few modificatons to an existing compose file that I had used for testing before prior said deployment on Kubernetes.
And, with the ease with which I was able to expose our endpoints, manage volumes, handle networking, deploy and tear down the setup. I in all honesty see no reason to not use Swarm. No mission-critical feature, or incredibly convenient really nice to have feature in Kubernetes that I'm go