1CC Notes
1CC Notes
UNIT - I
SYSTEMS MODELING,CLUSTERING AND VIRTUALIZATION
b) P2P Networks : In a P2P system, every node acts as both a client and
a server, providing part of the system resources. Peer machines are
simply client computers connected to the Internet. All client machines act
autonomously to join or leave the system freely. This implies that no
master-slave relationship exists among the peers. No central coordination
or central database is needed. The system is self-organizing with
distributed control.
➢ Para-Virtualization Architecture
When the x86 processor is virtualized, a virtualization layer is
inserted between the hardware and the OS. According to the x86
ring definition, the virtualization layer should also be installed at
Ring 0. Different instructions at Ring 0 may cause some problems. In
Figure 3.8, we show that para-virtualization replaces non-
virtualizable instructions with hyper calls that communicate directly
with the hypervisor or VMM. However, when the guest OS kernel is
modified for virtualization, it can no longer run on the hardware
directly. Compared with full virtualization, para-virtualization is
relatively easy and more practical. The main problem in full
virtualization is its low performance in binary translation. To speed
up binary translation is difficult. Therefore, many virtualization
products employ the para-virtualization architecture. The popular
Xen, KVM, and VMware ESX are good examples
3. Virtualization of CPU
Virtualization of the CPU refers to the process of creating a virtual
version of a physical CPU, allowing multiple virtual machines (VMs) to
share and use the same underlying hardware resources. This concept is
central to modern cloud computing, data centers, and server
management, enabling efficient use of physical hardware.
• Hard Disk Drives (HDD): IBM introduced the first commercial HDD
in 1956, the IBM 350, with a storage capacity of 5 MB spread across
50 platters. This marked the beginning of disk-based storage
technology, which allowed direct access to data.
5.1.3. Early Semiconductor Memory (1970s)
• Core Memory: Before the dominance of modern semiconductor
memory, core memory (magnetic cores) was a common technology.
Each bit was stored on a magnetic core.
2. Storage models
Storage models describe the layout of a data structure in a physical
storage; a data model captures the most important logical aspects of a
data structure in a database. Physical storage can be a local disk, a
removable media, or storage accessible via the network.
Two abstract models of storage are commonly used: cell storage and
journal storage. Cell storage assumes that the storage consists of cells of
the same size and that each object fits in one cell. This model reflects the
physical organization of several storage media; the primary memory of a
computer is organized as an array of memory cells and a secondary
storage device, e.g., a disk, is organized in sectors or blocks read and
written as a unit. Read/write coherence and before-or-after atomicity are
two highly desirable properties of any storage model and, in particular, of
cell storage
Parallel file systems are scalable, are capable of distributing files across a
large number of nodes, and provide a global naming space. In a parallel
data system, several I/O nodes serve data to all computational nodes
4. Databases
Most cloud applications are data-intensive, test the limitations of existing
cloud storage infrastructure, and demand database management systems
capable of supporting rapid application development and a short time-to-
market. Cloud applications require low latency, scalability, high
availability, and demand a consistent view of data. These requirements
cannot be satisfied simultaneously by existing database models; for
example, relational databases are easy to use for application
development but do not scale well.
As its name implies, the NoSQL model does not support SQL as a query
language and may not guarantee the ACID, Atomicity, Consistency,
Isolation, and Durability properties of traditional databases. It usually
guarantees an eventual consistency for transactions limited to a single
data item.
5. Distributed file systems
A Distributed File System (DFS) is a file system that is distributed on multiple file
servers or multiple locations. It allows programs to access or store isolated files as
they do with the local ones, allowing programmers to access files from any
network or computer. In this article, we will discuss everything about Distributed
File System.
DFS (Distributed File System) is a technology that allows you to group shared
folders located on different servers into one or more logically structured
namespaces. The main purpose of the Distributed File System (DFS) is to allows
users of physically distributed systems to share their data and resources by using
a Common File System. A collection of workstations and mainframes connected by
a Local Area Network (LAN) is a configuration on Distributed File System. A DFS
is executed as a part of the operating system. In DFS, a namespace is created and
this process is transparent for the clients.
Components of DFS