Skip to main content

How to learn Hadoop


First of all, I want to tell you. here all content you will get  practical only. for the theoretical part, you can follow edureka videos and Durgasoft videos. once you complete some videos related to map-reduce and Hadoop ecosystem..after that you can follow me...
you need to go step by step:-
1. What is big data
2. What is the need of Hadoop
3. How Hadoop works.
4. Tools in Hadoop ecosystem.



If you are a beginner then first you need some configuration in your system which is:
1: RAM:-
            Minimum RAM size should be 8gb otherwise you will get frustrated because of a slow system.
            if you go with 16gb RAM then it will be great. My opinion is to go with 16gb RAM.
2: Download VMWARE.https://my.vmware.com/en/web/vmware/info/slug/desktop_end_user_computing/vmware_workstation_pro/14_0
3. Download Cloudera ->https://www.cloudera.com/downloads/cdh/5-14-0.html. install it in VMWARE without installing os in WMWARE.
4. Download Eclipse for Map Reduce program

Click here to know about Big Data in Business?





Comments

Popular posts from this blog

sqoop

----------for database connect-------- mysql -u root -pcloudera ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- >>sqoop import --connect jdbc://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test01 >>sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test02 -m 10 >>sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test03 --split-by customer_id --fields-terminated-by '|' -m 10 >>sqoop list-databases --connect jdbc:mysql://localhost/retail_db --username root --password cloudera >>sqoop list-tables --connect jdbc:mysql://localhost/retail_db --username root --password cloudera -----------

important

create database veer; use veer; create table item_details(item_id int,item_name varchar(30),item_price int,item_color varchar(10)) row format delimited fields terminated by',' lines terminated by '\n' stored as textfile; hadoop fs -put /home/cloudera/sk/data/item.txt /user/hive/warehouse/veer.db/item select count(*) from item_details; select * from item_details order by item_id desc; select item_color,count(*) from item_details group by item_color; ---------------------------31-01-2018---------------------------------- create table flight_nt(s_id int,s_name varchar(30),code varchar(20),state varchar(40)) row format delimited fields terminated by',' lines terminated by '\n' stored as textfile; create table flightdata_part (s_id int, s_name varchar(30), code varchar(20) ) PARTITIONED BY (state VARCHAR(64)) row format delimited fields terminated by ',' lines terminated by '\n'  stored as textfile; LOAD DATA LOCAL INPATH '/local/home/srv