Document from Politecnico Di Milano about Computing Infrastructures. The Pdf, a summary of a university course, details hardware and software infrastructures, including data centers, servers, and virtualisation concepts for Computer Science students.
See more50 Pages


Unlock the full PDF for free
Sign up to get full access to the document and start transforming it with AI.
Computing Infrastructures Summary of the course by prof. Manuel Roveri Federica Laudizi - Federico Liuzzi POLITECNICO DI MILANOComputing Infrastructures | Federica Laudizi - Federico Liuzzi 1
What is a computing infrastructure? It is a technological infrastructure that provides hardware and software for computation to other systems and services.
A computing infrastructure may be formed of various layers, each layer being formed by computing units of diverse computational power and/or power usage. The lower the layer, the closer we are to the physical world we are measuring and/or interacting with.
RAM Data Center 50x - 100x Edge Computing Systems PC 0.0001x - 0.0005x 0.05x - 0.1x 20x - 50x 100x - 1000x Computation Speedup Embedded PCs Embedded Devices Internet of Things 0.01x-0.05x 0.1x - 0.5x Each layer (or type of computing unit) has its own characteristics.
Data centers are big conglomerates of interconnected machines working together, using this kind of solution has the advantage of being overall less expensive, having high storage and computational capacity with higher reliability. However, it requires being constantly connected to the internet (with high bandwidth!), there is higher latency and power consumption, together with privacy and security issues.
Moving lower along the layers, we find ourselves in the realm of edge computers. They provide high computational capacity, higher security levels and lower latencies. They too require relatively a lot of power and an internet connection.
Cloud Servers Edge Computing (EC) HOmi O O Vehicle-2-Vehicle Smart Devices Intelligent Monitoring Systems
Embedded PCs are low power and low performance units, but still capable of running complex OSes such as Linux. They allow for a pervasive implementation of a system, high performances for the buck, high availability of hardware and large support communities around each board. They will still need a substantial amount of power to work (around the tens of watt) and often some hardware design is needed.
IoT devices are highly pervasive, super low cost and low power but are affected by substantial limitations in computing ability, memory and thus being difficult to program.
BUILDING AND INFRASTRUCTURE SERVERS NETWORKING SYSTEMS COOLING STORAGE SUPPLY POWER RECOVERY FAILURE In the last few decades, computing and storage have moved from PC-like clients to smaller, often mobile, devices, combined with large internet services. Traditional enterprises are also shifting to this new paradigm of cloud computing.
This shift comes with an advantage to vendors, it enables them to deliver products in SaaS fashion, allowing for faster application development, easier improvements, and fixes of the software, together with more robust deployment due to the use of few well tested configurations.
Some workloads require so much computing capability that they are a more natural fit in datacenter (and not in client-side computing), these can vary from web search services and machine learning. (As an example, it would take 355 years to train GPT-3 on the fastest GPU on the market).
The trends toward server-side computing and widespread internet services created a new class of computing systems: Warehouse-scale computers (WSCs). In one of these systems, a program may be a whole internet service, consisting of tens or more individual programs that interact in complex ways.
Data centers are buildings where multiple servers and communication units are co-located because of their common environmental requirements and physical security needs, and for ease of maintenance.
Traditional data centers typically host a large number of small- or medium- sized applications, each application is running on a dedicated hardware infrastructure that is de-coupled and protected from pag. 6Computing Infrastructures | Federica Laudizi - Federico Liuzzi other systems in the same facility, because of this, applications tend not to communicate each other.
These data centers host hardware and software for multiple organizational units or even different companies.
WSCs, instead, belong to a single organization, use a homogeneous hardware and system software platform, and share a common systems' management layers.
WSCs run a smaller number of large applications (or internet services). The common resource management infrastructure allows significant deployment flexibility.
Traditional Datacenters Warehouse-Scale Computer Customer A Customer B Customer C Service Service Service Service Service Service Service Internet Internet Internet Internet Internet Provider V IT IT IT WSCs were initially designed for online data-intensive web workloads (such as managing Amazon's ecosystem), when the companies that developed these systems realized the power of these infrastructures, they started selling public clouds computing systems (e.g., Amazon, Google, Microsoft). Such public clouds do run many small applications, like a traditional data center but these applications rely on Virtual Machines (or Containers), and they access large, common services for block or database storage, load balancing, and so on, fitting very well within the WSC model.
These big data centers are often replicated in various physical locations to reduce user latency and improve serving throughput, nevertheless a single request will be processed within one data center.
The world is divided into Geographic Areas (GAS) defined by Geo-political boundaries (or country Region1 Region2 borders) or determined by data residency. In each GA >100 miles there are at least 2 Computing Regions (CRs). <2ms Customers see regions as the finer grain discretization of the infrastructure, this means that multiple DCs in the same region are not exposed.
Each region has a perimeter defined by the maximum latency to reach its DC.
Availability Zones (AZs) are finer grain location within a single computing region, they allow customers to run mission critical applications with high availability and fault tolerance to datacenter failures. Each availability zone is fault-isolated with redundant power, cooling, and networking.
Region AZ1 AZ2 AZ3 AZ4 Customer 1 No multi AZs Customer 2 Multi AZs Customer 3 Multi AZs Customer A Customer A Customer B Customer C Internal IT customer A Provider IT pag. 7
pag. 1
pag. 2
pag. 3
pag. 4