##Git mergetool.
$ git config --global merge.tool=vimdiff
##vimdiff key mappings.
connection = Faraday::Connection.new('http://example.com') do |builder| | |
builder.request :url_encoded # for POST/PUT params | |
builder.adapter :net_http | |
end | |
# same as above, short form: | |
connection = Faraday.new 'http://example.com' | |
# GET | |
connection.get '/posts' |
#!/bin/bash | |
# This tool can be used to sync down Red Hat based packages from RHN using only Red Hat shipped tools | |
# Brian "Red Beard" Harrington <brian@dead-city.org> | |
# Copyright 2013 | |
# | |
# To satisfy the pre-reqs for this script install the following two rpms: | |
# yum-utils | |
# createrepo_c (in RHEL 8 createrepo and createrepo_c have been combined) | |
# See https://github.com/rpm-software-management/createrepo_c#differences-in-behavior-between-createrepo_c-and-createrepo |
Note: total experiment and hack, looks nasty, could be awesome:
kitchen.local.yml
into $HOME/.kitchen/config.yml
brew install polipo
, with Ubuntu: apt-get install polipo
)polipo-start
and polipo-console
somewhere useful (perhaps $HOME/bin
?)# a 5 minute disk thrashing benchmark | |
# generates equal amounts of random read and write IO on every drive | |
# will generate metrics for each drive | |
[global] | |
ioengine=libaio | |
direct=1 | |
unified_rw_reporting=1 | |
rw=randrw | |
time_based=1 | |
runtime=300s |
Producer | |
Setup | |
bin/kafka-topics.sh --zookeeper esv4-hcl197.grid.linkedin.com:2181 --create --topic test-rep-one --partitions 6 --replication-factor 1 | |
bin/kafka-topics.sh --zookeeper esv4-hcl197.grid.linkedin.com:2181 --create --topic test --partitions 6 --replication-factor 3 | |
Single thread, no replication | |
bin/kafka-run-class.sh org.apache.kafka.clients.tools.ProducerPerformance test7 50000000 100 -1 acks=1 bootstrap.servers=esv4-hcl198.grid.linkedin.com:9092 buffer.memory=67108864 batch.size=8196 |