Exp_1
Exp_1
1
2
3
4
Infrastructure-as-a-service (IaaS)
•With IaaS, you rent IT infrastructure—servers and virtual machines (VMs), storage,
networks, operating systems—from a cloud provider on a pay-as-you-go basis.
• PaaS is designed to make it easier for developers to quickly create web or mobile apps,
without worrying about setting up or managing the underlying infrastructure of servers,
storage, network and databases needed for development.
5
Software as a service (SaaS)
•Software-as-a-service (SaaS) is a services for delivering software applications
over the Internet, on demand and typically on a subscription basis.
•With SaaS, cloud providers host and manage the software application and
underlying infrastructure and handle any maintenance, like software upgrades.
6
VMware Workstation
• VMware Workstation is a hosted hypervisor that runs on x64
versions of Windows and Linux operating systems.
• It enables users to set up virtual machines on a single physical
machine, and use them simultaneously along with the actual
machine.
7
REQUIREMENTS
• Minimum of 2 Gb ram
8
Installing of Hadoop Single Node Cluster Configuration .
9
STEP-1
SOFTWARES REQUIRED
• Hadoop 1.2.0-bin.tar.gz
• Jdk 7u67-linux-i586.tar.gz
Link for JDK
http://www.oracle.com/technetwork/java/javase/downloads/in
dex.html
Link of Hadoop :
http://mirror.fibergrid.in/apache/hadoop/common/hadoop-1.2.
0/
10
STEP-2
Untar the software
• Open the terminal window
• Type ls command
11
Step 3
Bashrc configurations
Open the terminal type the command
sudo gedit ~/.bashrc
export JAVA_HOME=/home/user/Desktop/jdk1.7.0_67
export PATH=$PATH:$JAVA_HOME/bin
export HADOOP_HOME=/home/user/Desktop/hadoop-1.2.0
export PATH=$PATH:$HADOOP_HOME/bin
Open new terminal and check for java version and Hadoop version.
12
Step-4
Hadoop Configuration Files
1. Core-site.xml
(Configuration setting for hadoop core, such as I/O setting that are
common to HDFS and Mapreduce)
2.Hdfs-Site.xml
13
3. Mapred-site.xml
(configuration setting for MapReduce daemons: the job tracker and the
task tracker)
4. Hadoop-Env_sh
14
• Open the new terminal window
• Go to the hadoop folder
• cd Desktop
• cd Hadoop (Press tab) for path
• cd conf
• Type ls
• All the above mentioned files are listed
15
• Go to terminal and type
16
• Go to terminal and type
sudo gedit hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
<name>dfs.name.dir</name>
<value>/home/user/Desktop/hadoop-1.2.0/name/data</value>
</property>
</configuration>
17
• Go to terminal and type
18
• Go to terminal and type
export JAVA_HOME=/home/user/Desktop/jdk1.7.0_67
19
STEP-5
SSh Configuration
• sudo apt-get install ssh
• ssh-keygen -t rsa -P ""
Sharing the public key with host
20
STEP 6
Formatting the Name Node
• $ hadoop namenode –format
21
STEP-7&8
Starting Hadoop Daemons
Open the terminal window and type
$ start-all.sh
and type
$ jps
22
• user@ubuntuvm:~$ start-all.sh
• Warning: $HADOOP_HOME is deprecated.
• starting namenode, logging to
/home/user/Downloads/hadoop-1.2.0/libexec/../logs/hadoop-user-
namenode-ubuntuvm.out
• localhost: datanode running as process 2978.
• localhost: secondarynamenode running as process 3123.
• jobtracker running as process 3204.
• localhost: tasktracker running as process 3342.
• user@ubuntuvm:~$ jps
• 4020 Jps
• 3342 TaskTracker
• 3204 JobTracker
• 3123 SecondaryNameNode
• 3606 NameNode
• 2978 DataNode
23
24