Big-Data - Hadoop Multi-Node Cluster over AWS using Ansible

Apache Framework — To Do Big Data Computing and Storage through Distributed Approach. Big Size files are Stripped out in some block sizes and Stores at different Data Storage Nodes using HDFS protocol. Benefit of the tool is that I/O process become faster…

Created by Akanksha

🤔 What Exactly Big Data is?

🤔 How to solve this Big Data Management Challenge?

Hadoop

source : IBM

Let me Direct you to the practical part now,

# mkdir hadoop-ws
# cd hadoop-ws
# mkdir roles
ansible.cfg
  • First:
file format:
access_key: GUJGWDUYGUEWVVFEWGVFUYV
secret_key: huadub7635897^%&hdfqt57gvhg
  • Second

Steps:

# mkdir role
# cd role
# ansible-galaxy init ec2
# ansible-galaxy init hadoop_master
# ansible-galaxy init hadoop_slave
# ansible-galaxy init hadoop_client
# cd role/hadoop_master/tasks
# vim main.yml
task.yml
# cd role/kube_master/vars
# vim main.yml
var.yml
task.yml
var.yml
core-site.xml
hdfs.xml
task.yml
var.yml
core-site.xml
hdfs-site.xml
task.yml
var.yml
core-site.xml
setup.yml
# ansible-playbook setup.yml -ask-vault-pass

Thanks for reading. Hope this blog have given you some valuable inputs!!

--

--

Platform Engineer | Kubernetes | Docker | Terraform | Helm | AWS | Azure | Groovy | Jenkins | Git, GitHub | Sonar | NMAP and other Scan and Monitoring tool

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Akanksha Singh

Platform Engineer | Kubernetes | Docker | Terraform | Helm | AWS | Azure | Groovy | Jenkins | Git, GitHub | Sonar | NMAP and other Scan and Monitoring tool