In this lab, we will learn to:
- Configure Solr to store indexes in HDFS
- Create a solr cluster of 2 solr instances running on port 8983 and 8984
- Index documents in HDFS using the Hadoop connectors
- Use Solr to search documents
#find jar containing class org.apache.hive.jdbc.HiveStatement | |
find / -iname '*.jar' | xargs -i bash -c "jar -tvf {} | tr / . | grep org.apache.hive.jdbc.HiveStatement && echo {}" |
This lab is part of a 'Sql on Hadoop' webinar. The recording and slides can be found here
How/when to use Hive vs Phoenix vs SparkSQL
{"paragraphs":[{"title":"Create HAWQ table and generate data series","text":"%psql.sql\ndrop table if exists tt;\ncreate table tt (i int);\ninsert into tt select generate_series(1,1000000);","dateUpdated":"Jun 7, 2016 7:12:59 PM","config":{"colWidth":12,"editorMode":"ace/mode/scala","title":true,"graph":{"mode":"table","height":300,"optionOpen":false,"keys":[],"values":[],"groups":[],"scatter":{}},"enabled":true},"settings":{"params":{},"forms":{}},"jobName":"paragraph_1465351928921_1500086996","id":"20160603-083343_921281900","result":{"code":"SUCCESS","type":"TABLE","msg":"Update Count\n0\n","comment":"","msgTable":[[{"value":"0"}]],"columnNames":[{"name":"Update Count","index":0,"aggr":"sum"}],"rows":[["0"]]},"dateCreated":"Jun 7, 2016 7:12:08 PM","status":"FINISHED","progressUpdateIntervalMs":500,"$$hashKey":"object:863","dateFinished":"Jun 7, 2016 7:13:00 PM","dateStarted":"Jun 7, 2016 7:12:59 PM","focus":true},{"title":"Calculate average of subset of data","text":"%psql.sql\nselect avg(i) from tt where |
########################################################################################################################### | |
##HDB on HDP sandbox setup script | |
###Pre-reqs: | |
#- Laptop with at least 10-12 GB RAM (mine has 16 GB) | |
#- ISO image of Centos 6.7 or later downloaded from [here](http://isoredirect.centos.org/centos/6/isos/x86_64/). | |
# - In my case, I used CentOS-6.7-x86_64-bin-DVD1.iso. | |
##### Setup Centos 6.7 or later on VM | |
#- Start a CentOS VM using ISO |
sudo git clone https://github.com/hortonworks-gallery/ambari-vnc-service.git /var/lib/ambari-server/resources/stacks/HDP/2.3/services/VNCSERVER
#download toolkit
wget https://hipchat.hortonworks.com/files/1/2055/bT1LbKB8SS26X9t/nifi-toolkit-1.0.0-SNAPSHOT-bin.zip
#create nifi certs dir under ambari-server resources dir
mkdir /var/lib/ambari-server/resources/host_scripts/nifi-certs
#generate certs using toolkit into ambari-server resources dir
Goals:
Notes:
build=447 | |
tee /etc/yum.repos.d/HDF.repo > /dev/null << EOF | |
[HDF-2.0] | |
name=HDF-2.0 | |
baseurl=http://s3.amazonaws.com/dev.hortonworks.com/HDF/centos6/2.x/BUILDS/2.0.0.0-$build | |
gpgcheck=0 | |
path=/ | |
enabled=1 | |
EOF |