Shantanu's Blog

Corporate Consultant

May 23, 2016

 

load data into elastic using logstash

# install logstash

wget https://download.elastic.co/logstash/logstash/packages/centos/logstash-2.3.2-1.noarch.rpm

rpm -iUh logstash-2.3.2-1.noarch.rpm

# test logstash can generate json output

bin/logstash -e 'input {file {path => "/var/log/cron" start_position => beginning} } output {stdout { codec => json } } '

# push data to elastic
cd /opt/logstash/
bin/logstash -e 'input {file {path => "/var/log/cron" start_position => beginning} } output { elasticsearch { hosts => "https://search-xxxx-demo-75hhndin3fhmovqanbqwfimu4q.us-east-1.es.amazonaws.com"} } '

Labels: , , , ,


May 22, 2016

 

Backing up data only containers

The advantage of data only container is that you can easily take backup of the data. For e.g. in this case the tutum/mysql:5.5 image is used to store data. The data container is committed and then pushed to docker hub. Since this data my contain important information it is better to keep it as private repository.

[root@ip-172-31-15-118 mylamp]# docker ps
CONTAINER ID        IMAGE                 COMMAND             CREATED             STATUS              PORTS                                                  NAMES
99526bfa2f0f        mylamp_web            "/run.sh"           11 minutes ago      Up 11 minutes       0.0.0.0:80->80/tcp, 3306/tcp, 0.0.0.0:8000->8000/tcp   mylamp_web_1
c7950eeab8e9        tutum/mysql:5.5       "/run.sh"           11 minutes ago      Up 11 minutes       0.0.0.0:3306->3306/tcp                                 mylamp_mysql_1
3802f7c35133        tutum/ubuntu:trusty   "/run.sh"           11 minutes ago      Up 11 minutes       0.0.0.0:24->22/tcp                                     mylamp_crmdbvol1_1

[root@ip-172-31-15-118 mylamp]# docker commit 3802f7c35133 shantanuo/todel
f8a3d4f0376bcc0eb07f0bde592e1e4a58f7d8d8d82eb7bcb8d1522324725fe3

[root@ip-172-31-15-118 mylamp]# docker login
Username: XXX
Password:
Email: XXX@gmail.com
WARNING: login credentials saved in /root/.docker/config.json
Login Succeeded

[root@ip-172-31-15-118 mylamp]# docker push shantanuo/todel
The push refers to a repository [docker.io/shantanuo/todel] (len: 1)
f8a3d4f0376b: Pushed
ebe73f29e6e1: Pushed
latest: digest: sha256:f1fda8ef1ac49ed013131177c9213716a3cb761671e0ff43315ed06422213f7a size: 28444

Labels: ,


May 21, 2016

 

decouple application using docker compose

# install packages
yum install -y git mysql-server docker

curl -L https://github.com/docker/compose/releases/download/1.7.1/docker-compose-`uname -s`-`uname -m` > /usr/local/bin/docker-compose
chmod +x /usr/local/bin/docker-compose

/etc/init.d/docker start

# create and copy public key to github to clone private repo

ssh-keygen -t rsa -b 4096 -C "user@gmail.com"

cat ~/.ssh/id_rsa.pub


# clone your private repository
git clone git@github.com:shantanuo/xxx.git

# create your custom my.cnf config file

mkdir -p /my/custom/
vi /my/custom/my.cnf
[mysqld]
sql_mode=''

# start docker containers as per yml file

/usr/local/bin/docker-compose up -d

# restore mysql data into container from host:
mysqladmin -h localhost -P 3306 --protocol=tcp -u root -ppasswd create livebox

mysql -h localhost -P 3306 --protocol=tcp -u root -ppasswd livebox < livebox.sql

# access mysql container from another container:

mysql -h mysql -uroot -ppasswd
_____

## use container ID of tutum/mysql:5.5 for e.g.
docker logs c7950eeab8e9
    mysql -uadmin -pxmFShXB1Asgn -h127.0.0.1

# use the password to connect to 127.0.0.1 and execute commands:
mysql> grant all on *.* to 'root'@'%' identified by 'passwd' with grant option;
Query OK, 0 rows affected (0.00 sec)

# data only container to be used for mysql data
docker run -d -v /var/lib/mysql --name db_vol1 -p 23:22 tutum/ubuntu:trusty
docker run -d --volumes-from db_vol1 -p 3306:3306 tutum/mysql:5.5

Labels: , , , , , ,


May 16, 2016

 

List all databases on elasticsearch server

Here is the python code that will list all indices from elastic.
It will return the current date if the data is being partitioned daily or hourly.

import elasticsearch
es = elasticsearch.Elasticsearch('https://search-demo-75hhndin3fhmovqanbqwfimu4q.us-east-1.es.amazonaws.com')
indices=es.indices.get_aliases().keys()
sorted(indices)[-1:]

We can also use curl to get the same list...

# curl 'search-demo-75hhndin3fhmovqanbqwfimu4q.us-east-1.es.amazonaws.com/_cat/indices?v'
health status index            pri rep docs.count docs.deleted store.size pri.store.size
yellow open   blank-struts2      5   1          0            0       575b           575b
yellow open   vq-2016-05-02-07   5   1          1            0      3.4kb          3.4kb
yellow open   vq-2016-05-02-09   5   1         94            0     41.6kb         41.6kb

Labels: , ,


May 14, 2016

 

Python multiprocessing made easy

This is how a linear download would happen that will take time because files are downloaded one after other. This process can be improved using multiprocessing module. But using that module is complicated.
import os
import requests

def worker(url):
    fname = os.path.basename(url)
    print fname, url
    print "\n"
    r = requests.get(url)
    with open(fname, "wb") as f:
        f.write(r.content)
def run(urls):
    for url in urls:
        print url
        worker(url)
       
urls = ["http://www.irs.gov/pub/irs-pdf/f1040.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040a.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040ez.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040sb.pdf" ]
run(urls)

I am not even showing how multiprocessing would do this because here is a new module called deco!
I need to add only 3 lines to the code. One is import and 2 are decorators!
from deco import concurrent, synchronized
import os
import requests

@concurrent
def worker(url):
    fname = os.path.basename(url)
    print fname, url
    print "\n"
    r = requests.get(url)
    with open(fname, "wb") as f:
        f.write(r.content)
@synchronized
def run(urls):
    for url in urls:
        print url
        worker(url)
       
urls = ["http://www.irs.gov/pub/irs-pdf/f1040.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040a.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040ez.pdf",
            "http://www.irs.gov/pub/irs-pdf/f1040sb.pdf" ]
run(urls)

pip install deco is available from here...
https://github.com/alex-sherman/deco

Labels: ,


Archives

June 2001   July 2001   January 2003   May 2003   September 2003   October 2003   December 2003   January 2004   February 2004   March 2004   April 2004   May 2004   June 2004   July 2004   August 2004   September 2004   October 2004   November 2004   December 2004   January 2005   February 2005   March 2005   April 2005   May 2005   June 2005   July 2005   August 2005   September 2005   October 2005   November 2005   December 2005   January 2006   February 2006   March 2006   April 2006   May 2006   June 2006   July 2006   August 2006   September 2006   October 2006   November 2006   December 2006   January 2007   February 2007   March 2007   April 2007   June 2007   July 2007   August 2007   September 2007   October 2007   November 2007   December 2007   January 2008   February 2008   March 2008   April 2008   July 2008   August 2008   September 2008   October 2008   November 2008   December 2008   January 2009   February 2009   March 2009   April 2009   May 2009   June 2009   July 2009   August 2009   September 2009   October 2009   November 2009   December 2009   January 2010   February 2010   March 2010   April 2010   May 2010   June 2010   July 2010   August 2010   September 2010   October 2010   November 2010   December 2010   January 2011   February 2011   March 2011   April 2011   May 2011   June 2011   July 2011   August 2011   September 2011   October 2011   November 2011   December 2011   January 2012   February 2012   March 2012   April 2012   May 2012   June 2012   July 2012   August 2012   October 2012   November 2012   December 2012   January 2013   February 2013   March 2013   April 2013   May 2013   June 2013   July 2013   September 2013   October 2013   January 2014   March 2014   April 2014   May 2014   July 2014   August 2014   September 2014   October 2014   November 2014   December 2014   January 2015   February 2015   March 2015   April 2015   May 2015   June 2015   July 2015   August 2015   September 2015   January 2016   February 2016   March 2016   April 2016   May 2016  

This page is powered by Blogger. Isn't yours?