Skip to content

Instantly share code, notes, and snippets.

View Sathiyarajan's full-sized avatar
:octocat:
I may be slow to respond.

sathyadev Sathiyarajan

:octocat:
I may be slow to respond.
View GitHub Profile
@Sathiyarajan
Sathiyarajan / urllib2_upload.py
Created September 28, 2017 06:56 — forked from zhenyi2697/urllib2_upload.py
Python: urllib2 upload file demo
#!/usr/bin/python
# demo from here: http://pymotw.com/2/urllib2/index.html#uploading-files
import itertools
import mimetools
import mimetypes
from cStringIO import StringIO
import urllib
import urllib2

Notes while learning Docker

These notes are derived from my efforts in working through The Docker Book: Containerization is the new virtualization written by James Turnbull. I highly recommend getting a copy of Jame's book which is an $8 dollar e-book on Amazon.

The following notes are my own and simply an interpretation of what James has written. Caveat lector.

Ensuring Docker is ready

$ docker info
@Sathiyarajan
Sathiyarajan / pep20_by_example.py
Created October 31, 2017 13:08 — forked from evandrix/pep20_by_example.py
PEP 20 (The Zen of Python) by example
#!/usr/bin/env python
"""
=====================================
PEP 20 (The Zen of Python) by example
=====================================
Usage: %prog
:Author: Hunter Blanks, hblanks@artifex.org / hblanks@monetate.com
@Sathiyarajan
Sathiyarajan / docker_kill.sh
Created November 12, 2017 03:19 — forked from evanscottgray/docker_kill.sh
kill all docker containers at once...
docker ps | awk {' print $1 '} | tail -n+2 > tmp.txt; for line in $(cat tmp.txt); do docker kill $line; done; rm tmp.txt
@Sathiyarajan
Sathiyarajan / Sample.java
Created January 3, 2018 18:24 — forked from Saberko/Sample.java
spark hbase integration
package utils;
import org.apache.commons.cli.Options;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellScanner;
import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result;
@Sathiyarajan
Sathiyarajan / HBaseOp.java
Created January 3, 2018 18:24 — forked from Saberko/HBaseOp.java
HBase api curd操作
package learn;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.*;
import org.apache.hadoop.hbase.util.Bytes;
@Sathiyarajan
Sathiyarajan / SparkHbaseIntegraton.java
Created January 3, 2018 18:25 — forked from Saberko/SparkHbaseIntegraton.java
spark读取、写入hbase操作
package learn;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.filter.CompareFilter;
import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
@Sathiyarajan
Sathiyarajan / ansible-summary.md
Created February 2, 2018 10:52 — forked from andreicristianpetcu/ansible-summary.md
This is an ANSIBLE Cheat Sheet from Jon Warbrick

An Ansible summary

Jon Warbrick, July 2014, V3.2 (for Ansible 1.7)

Configuration file

intro_configuration.html

First one found from of

@Sathiyarajan
Sathiyarajan / SparkSQL with json example
Created February 2, 2018 10:53 — forked from tmcgrath/SparkSQL with json example
SparkSQL with json file example
// download sample json
http://bit.ly/2gY39Ay
// start spark-shell in same directory as where customers.json
// is downloaded
val customers = sqlContext.jsonFile("customers.json")
// register a temp table
customers.registerTempTable("customers")
@Sathiyarajan
Sathiyarajan / Cassandra Spark SQL
Created February 2, 2018 10:53 — forked from tmcgrath/Cassandra Spark SQL
SparkSQL with Cassandra from Killrweather processed data
//If you want to run a local cluster
//start-master.sh
//start-slave.sh <your-master-url>
// connect to spark-shell and load cassandra connector jar OR
~/Development/spark-1.6.3-bin-hadoop2.6/bin/spark-shell --packages datastax:spark-cassandra-connector:1.6.0-s_2.10
// connect to spark-shell to spark cluster and load cassandra connector jar
~/Development/spark-1.6.3-bin-hadoop2.6/bin/spark-shell --master <your-master-url> --packages datastax:spark-cassandra-connector:1.6.0-s_2.10