Assumed you have files with extension .ebk
stored by Kies.
Install Ruby 2.3 or later
require 'concurrent' | |
require 'benchmark' | |
require 'logger' | |
module Logging # Credits: https://stackoverflow.com/a/6768164/10118318 | |
# This is the magical bit that gets mixed into your classes | |
def logger | |
Logging.logger | |
end |
/* | |
Simple GTK+ pipe viewer | |
Contributors: Romiras | |
Based on paned.c | |
LICENSE: GNU GPLv3 | |
*/ | |
#include <stdio.h> | |
#include <unistd.h> | |
#include <gtk/gtk.h> |
# Concurrently classify items by type and batch them with multiple workers | |
require 'logger' | |
NUM_WORKERS = 10 | |
NUM_ITEMS = 12 | |
NUM_ITERATIONS = 5000 | |
Item = Struct.new(:id, :type) |
split --verbose -l1000 dataset-bulk-30k.ndjson bulk. | |
for f in bulk.??; do echo $f; curl -i -X POST localhost:9200/_bulk -H "Content-Type: application/x-ndjson" --data-binary @$f; done |
# Reads .jsonl file and transforms it into file ready for bulk import to ES. | |
require 'json' | |
File.open(ARGV[1], 'wb') {|f| | |
ARGF.each_line { |line| | |
a = JSON.parse(line) | |
out = { | |
index: { | |
"_id": a["_id"], |
require 'concurrent' | |
require 'benchmark' | |
max = 20 # number of tasks to process | |
n_workers = 4 # number of concurrent workers | |
results = [] # | |
lambda = -> (iter, i) { results << ('%04d' % i); d = rand(0.005)+0.001; puts("Iteration #{iter}. sleep %.3f" % d); sleep(d) } | |
workers = Array.new(n_workers, lambda) | |
iter = 0 |
"<!--" | |
"><!--" | |
"||[];" | |
"--><!--" | |
"--></" | |
"----" | |
"!--<" | |
"//--></" | |
"//-->" | |
"...</" |
source "https://rubygems.org" | |
gem 'eventmachine' | |
gem 'em-http-request' |
require 'active_support/all' # mb_chars | |
def simple_fuzzy_match(s1, s2) | |
levenshtein_distance( normalize_str(s1), normalize_str(s2) ) < 2 | |
end | |
def normalize_str(s) | |
s. | |
mb_chars. # convert to multibyte string (ActiveSupport::Multibyte::Chars) - required in Ruby version below 2.4 | |
downcase. # lower case for all characters |