Mañana realizaremos una serie de ejercicios sobre el dataset del Titanic de Kaggle, una pagina de competiciones de ciencia de datos.
Es interesante que te traigas descargado los archivos CSV.
package spark.example | |
import org.apache.spark.SparkContext | |
import org.apache.spark.SparkContext._ | |
import org.apache.spark.SparkConf | |
object SparkGrep { | |
def main(args: Array[String]) { | |
if (args.length < 3) { | |
System.err.println("Usage: SparkGrep <host> <input_file> <match_term>") |
sudo apt-get remove \ | |
aisleriot \ | |
brltty \ | |
duplicity \ | |
empathy \ | |
empathy-common \ | |
example-content \ | |
gnome-accessibility-themes \ | |
gnome-contacts \ | |
gnome-mahjongg \ |
# Basic commands | |
:Git [args] # does what you'd expect | |
all of your `~/.gitconfig` aliases are available. | |
:Git! [args] # same as before, dumping output to a tmp file | |
Moving inside a repo. |
Mañana realizaremos una serie de ejercicios sobre el dataset del Titanic de Kaggle, una pagina de competiciones de ciencia de datos.
Es interesante que te traigas descargado los archivos CSV.
compiler: ghc-8.2.0.20170404 | |
compiler-check: match-exact | |
resolver: lts-8.8 | |
setup-info: | |
ghc: | |
linux64: | |
8.2.0.20170404: | |
url: https://downloads.haskell.org/~ghc/8.2.1-rc1/ghc-8.2.0.20170404-x86_64-deb8-linux.tar.xz | |
content-length: 138846620 | |
sha1: 3fd303b5753b095ce020131405dcf43199e6b447 |
#!/usr/bin/env stack | |
{- stack | |
--resolver lts-6.27 | |
--install-ghc | |
runghc | |
--package containers | |
-} | |
{-# LANGUAGE ScopedTypeVariables #-} | |
{-# LANGUAGE PolyKinds #-} | |
{-# LANGUAGE DataKinds #-} |
Taken from Deedle in 10 minutes, adapted to Analyze
import Analyze
import Analyze.Frame as Frame
import Analyze.Series as Series
import qualified Date.Time.Calendar -- from the 'time' package
I hereby claim:
To claim this, I am signing this object:
I hereby claim:
To claim this, I am signing this object:
I hereby claim:
To claim this, I am signing this object: