Big-Data - Hadoop Multi-Node Cluster over AWS using Ansible

Apache Framework — To Do Big Data Computing and Storage through Distributed Approach. Big Size files are Stripped out in some block sizes and Stores at different Data Storage Nodes using HDFS protocol. Benefit of the tool is that I/O process become faster…

Created by Akanksha

🤔 What Exactly Big Data is?

🤔 How to solve this Big Data Management Challenge?

Hadoop

source : IBM

Let me Direct you to the practical part now,

# mkdir hadoop-ws
# cd hadoop-ws
# mkdir roles
ansible.cfg
file format:
access_key: GUJGWDUYGUEWVVFEWGVFUYV
secret_key: huadub7635897^%&hdfqt57gvhg

Steps:

# mkdir role
# cd role
# ansible-galaxy init ec2
# ansible-galaxy init hadoop_master
# ansible-galaxy init hadoop_slave
# ansible-galaxy init hadoop_client
# cd role/hadoop_master/tasks
# vim main.yml
task.yml
# cd role/kube_master/vars
# vim main.yml
var.yml
task.yml
var.yml
core-site.xml
hdfs.xml
task.yml
var.yml
core-site.xml
hdfs-site.xml
task.yml
var.yml
core-site.xml
setup.yml
# ansible-playbook setup.yml -ask-vault-pass

Thanks for reading. Hope this blog have given you some valuable inputs!!

Technology enhancement take a journey of learning and exploring!! On a way to achieve and Follow my own star!!