Attention: the list was moved to
https://github.com/dypsilon/frontend-dev-bookmarks
This page is not maintained anymore, please update your bookmarks.
Attention: the list was moved to
https://github.com/dypsilon/frontend-dev-bookmarks
This page is not maintained anymore, please update your bookmarks.
# -*- coding: utf-8 -*- | |
#!/usr/bin/env python | |
import sys | |
class Polyomino(object): | |
def __init__(self, iterable): | |
self.squares = tuple(sorted(iterable)) | |
def __repr__(self): |
TODO | |
implement security measures | |
git config | |
config files | |
full sublimetext config | |
set up openvpn | |
rdesktop and network drive to terra | |
set up evolution | |
RStudio |
$ python get_logs.py rds-db-instance-name rds.log aws-access-key aws-secret-key
$ ./pgbadger -p '%t:%r:%u@%d:[%p]:' ./rds.log
$ open out.html
Here I will go over how to setup Logstash, Kibana, Redis, and Elasticsearch in an EC2 environment behind a public Load Balancer. The setup I'll be doing will have: | |
1) One server for Redis to act as the broker/buffer to receive logs. | |
2) One server using Logstash receive logs from Redis and parse/index them over to Elasticsearch. | |
3) One server for Elasticsearch to receive logs and Kibana to view them in a browser. | |
4) One server to send the logs using logstash. | |
5) One public Load Balancer. | |
This may seem like a lot but follow these steps and you'll get the hang of it :) | |
What you will need: |
I have moved this over to the Tech Interview Cheat Sheet Repo and has been expanded and even has code challenges you can run and practice against!
\
Section "Extensions" | |
Option "XVideo" "Disable" | |
EndSection | |
Section "Device" | |
Identifier "Intel Graphics" | |
Driver "intel" | |
Option "AccelMethod" "sna" | |
Option "TearFree" "true" | |
Option "DRI" "true" |
Using perf:
$ perf record -g binary
$ perf script | stackcollapse-perf.pl | rust-unmangle | flamegraph.pl > flame.svg
NOTE: See @GabrielMajeri's comments below about the
-g
option.