library(BayesianFirstAid)
debugonce(bayes.t.test) # turns off debugging for function after first run
bayes.t.test(co2)
library(ggplot2) | |
library(dplyr) # easier data wrangling | |
library(viridis) # colour blind friendly palette, works in B&W also | |
library(Interpol.T) # will generate a large dataset on initial load | |
library(lubridate) # for easy date manipulation | |
library(ggExtra) # because remembering ggplot theme options is beyond me | |
library(tidyr) | |
data<- data(Trentino_hourly_T,package = "Interpol.T") |
require(quantmod) | |
require(ggplot2) | |
require(reshape2) | |
require(plyr) | |
require(scales) | |
# Download some Data, e.g. the CBOE VIX | |
getSymbols("^VIX",src="yahoo") | |
# Make a dataframe |
#### VPIN calculation ######################################################### | |
#install.packages('fasttime',repos='http://www.rforge.net/') | |
require(data.table); require(fasttime); require(plyr) | |
# Assuming TAQ data is arranged in 1 year stock csv files | |
stock=fread('/TAQ_data.csv'); stock=stock[,1:3,with=FALSE] | |
setnames(stock,colnames(stock),c('DateTime','Price','Volume')); | |
stock[,DateTime:=paste(paste(substr(DateTime,1,4),substr(DateTime,5,6), | |
substr(DateTime,7,8),sep='-'),substr(DateTime,10,17))] | |
setkey(stock,DateTime); | |
stock=as.xts(stock[,2:3,with=FALSE],unique=FALSE, |
- macOS 10.15.5
- tmux 3.1b
macOS has ncurses version 5.7 which does not ship the terminfo description for tmux. There're two ways that can help you to solve this problem.
Instead of tmux-256color
, use screen-256color
which comes with system. Place this command into ~/.tmux.conf
or ~/.config/tmux/tmux.conf
(for version 3.1 and later):
<?xml version="1.0"?> | |
<root> | |
<item> | |
<name>For Topre REALFORCE 87u Users</name> | |
<devicevendordef> | |
<vendorname>TopreCorporation</vendorname> | |
<vendorid>0x0853</vendorid> | |
</devicevendordef> | |
<item> | |
<name>Fix FN + F1..F12 and functional keys</name> |
- General Background and Overview
- Probabilistic Data Structures for Web Analytics and Data Mining : A great overview of the space of probabilistic data structures and how they are used in approximation algorithm implementation.
- Models and Issues in Data Stream Systems
- Philippe Flajolet’s contribution to streaming algorithms : A presentation by Jérémie Lumbroso that visits some of the hostorical perspectives and how it all began with Flajolet
- Approximate Frequency Counts over Data Streams by Gurmeet Singh Manku & Rajeev Motwani : One of the early papers on the subject.
- [Methods for Finding Frequent Items in Data Streams](http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.187.9800&rep
The response to my first few posts has been much larger than I’d imagined and I’d like to thank everyone for the encouragement.
If you’re interested in building a trading system I recommend first reading my previous post on general ideas to keep in mind.
My first really technical post will be on how to build a limit order book, probably the single most important component of a trading system. Because the data structure chosen to represent the limit order book will be the primary source of market information for trading models, it is important to make it both absolutely correct and extremely fast.
To give some idea of the data volumes, the Nasdaq TotalView ITCH feed, which is every event in every instrument traded on the Nasdaq, can have data rates of 20+ gigabytes/day with spikes of 3 megabytes/second or more. The individual messages average about 20 bytes each so this means handling
; /usr/local/bin/nasm -f macho 32.asm && ld -macosx_version_min 10.7.0 -o 32 32.o && ./32 | |
global start | |
section .text | |
start: | |
push dword msg.len | |
push dword msg | |
push dword 1 | |
mov eax, 4 |