UNIT-1 Computing Paradigms
Introduction ⮚What is computing? ● The process of utilizing computer technology to complete a task. ● Computing may involve computer hardware and/or software, but must involve some form of a computer system. ⮚What is paradigms? ● A style or a way of doing something. ● A set of practices to be followed to accomplish a task. ⮚Computing Paradigms ● In the domain of computing, there are many different standard practices being followed based on inventions and technological advancements.
Computing Paradigms ●The various computing paradigms: ~ High performance computing, ~ Parallel computing ~ Cluster computing, ~ Distributed computing ~ Grid computing, ~ Cloud computing, ~ Bio-computing, ~ Mobile computing, ~ Quantum computing, ~ Optical computing, ~ Nano computing.
High Performance Computing
High Performance Computing ● In high-performance computing systems, a pool of processors are connected with other resources like memory, storage, and input and output devices, and the deployed software is enabled to run in the entire system of connected components. ● The processor machines can be of homogeneous or heterogeneous type. ● The legacy meaning of high-performance computing (HPC) is the supercomputers; however, it is not true in present-day computing scenarios.
● Thus, examples of HPC include a small cluster of desktop computers or personal computers (PCs) to the fastest supercomputers. ● HPC systems are normally found in those applications where it is required to use or solve scientific problems. ● Most of the time, the challenge in working with these kinds of problems is to perform suitable simulation study, and this can be accomplished by HPC without any difficulty. ● Scientific examples such as protein folding in molecular biology and studies on developing models and applications based on nuclear fusion are worth noting as potential applications for HPC.
High Performance Computing
High Performance Computing ●HPC Use In Africa And How Different Industries Can Take Advantage of It
High Performance Computing UCL to host National High Performance Computing Hub for Materials Science A UCL-led consortium has been awarded to establish a new national High Performance Computing (HPC) facility for the Materials and Molecular Modelling community.
High Performance Computing Global High Performance Computing market
High Performance Computing
Parallel Computing
Parallel Computing ●Serial Computing: ●Traditionally, software has been written for serial computation: ○A problem is broken into a discrete series of instructions ○Instructions are executed sequentially one after another ○Executed on a single processor ○Only one instruction may execute at any moment in time
Parallel Computing ●Serial Computing:
Parallel Computing ●Serial Computing: ●For Example-
Parallel Computing ●Parallel Computing: ●In the simplest sense, parallel computing is the simultaneous use of multiple compute resources to solve a computational problem: ○A problem is broken into discrete parts that can be solved concurrently ○Each part is further broken down to a series of instructions ○Instructions from each part execute simultaneously on different processors ○An overall control/coordination mechanism is employed
Parallel Computing
Parallel Computing ●For Example-
Parallel Computing ●The computational problem should be able to: ○Be broken apart into discrete pieces of work that can be solved simultaneously; ○Execute multiple program instructions at any moment in time; ○Be solved in less time with multiple compute resources than with a single compute resource. ●The compute resources are typically: ○A single computer with multiple processors/cores ○An arbitrary number of such computers connected by a network
This figure is showing the majority of the world's large parallel computers (supercomputers) are clusters of hardware produced by a handful of (mostly) well known vendors.
Distributed Computing
Distributed Computing ● Distributed computing is also a computing system that consists of multiple computers or processor machines connected through a network, which can be homogeneous or heterogeneous, but run as a single system. ● The connectivity can be such that the CPUs in a distributed system can be physically close together and connected by a local network, or they can be geographically distant and connected by a wide area network. ● The heterogeneity in a distributed system supports any number of possible configurations in the processor machines, such as mainframes, PCs, workstations, and minicomputers.
● The goal of distributed computing is to make such a network work as a single computer. ● Distributed computing systems are advantageous over centralized systems, because there is a support for the following characteristic features: 1. Scalability: It is the ability of the system to be easily expanded by adding more machines as needed, and vice versa, without affecting the existing setup. 2. Redundancy or replication: Here, several machines can provide the same services, so that even if one is unavailable (or failed), work does not stop because other similar computing supports will be available.
Distributed Computing Centralized ● A system with centralized multiprocessor parallel architecture. In the late 1980 s Centralized systems have been progressively replaced by distributed systems. Characteristics of centralized system ● Non autonomous components ● usually homogeneous technology ● Multiple users share the same resources at all time ● single point of control ● single point of failure
Distributed Computing ● Set of tightly coupled programs executing on one or more computers which are interconnected through a network and coordinating their actions. These programs know about one another and carry out tasks that none could carry out in isolation ●Characteristics of distributed system ● Autonomous components ● Mostly build using heterogeneous technology ● System components may be used exclusively ● Concurrent processes can execute ● Scalability- possibility of adding new hosts ● openness- easily extended and modified ● Heterogeneity-supports various H/W S/w platforms ● Resource sharing- H/w, S/W and data ● fault tolerance- ability to function correctly even if faults occur
Distributed Computing
Distributed Computing ● the following are the different application of the distributed system. ● Global positioning System ● World Wide Web ● Air Traffic Control System ● Automated Banking System ● In the World Wide Web application the data or application were distributed on the several number of the heterogeneous computer system, but for the end user or the browser it seems to be a single system from which user got the information. ● The multiple number of computer working concurrently and perform the resource sharing in the World Wide Web.
Cluster Computing
Cluster Computing ● A cluster computing system consists of a set of the same or similar type of processor machines connected using a dedicated network infrastructure. ● All processor machines share resources such as a common home directory and have a software such as a message passing interface (MPI) implementation installed to allow programs to be run across all nodes simultaneously. ● This is also a kind of HPC category. ● The individual computers in a cluster can be referred to as nodes.
● The reason to realize a cluster as HPC is due to the fact that the individual nodes can work together to solve a problem larger than any computer can easily solve. ● And, the nodes need to communicate with one another in order to work cooperatively and meaningfully together to solve the problem in hand. ● If we have processor machines of heterogeneous types in a cluster, this kind of clusters become a subtype and still mostly are in the experimental or research stage.
Cluster Computing ● A computer cluster help to solve complex operations more efficiently with much faster processing speed, better data integrity than a single computer and they only used for mission-critical applications. ● The Clustering methods have identified as- HPC IAAS, HPC PAAS, that are more expensive and difficult to setup and maintain than a single computer. ● A computer cluster defined as the addition of processes for delivering large-scale processing to reduce downtime and larger storage capacity as compared to other desktop workstation or computer.
Cluster Computing ● Some of the critical Applications of Cluster Computers are Google Search Engine, Petroleum Reservoir Simulation, Earthquake Simulation, Weather Forecasting. ● Cluster Can be classified into two category Open and Close Cluster. ● Open Cluster: All nodes in Open Cluster are needed IPs, and that are accessible through internet/web, that cause more security concern. Close Cluster: On the other hand Close Cluster are hide behind the gateway node and provide better security.
Cluster Computing ● Types of Cluster computing ● 1. Load-balancing clusters: As the name implies, This system is used to distribute workload across multiple computers. That system ● distributes the processing load as possible across a cluster of computers. 2. High availability (HA) clusters: A high availability clusters (HA cluster) are the bunch of computers that can reliably utilise for ● redundant operations in the event of nodes failure in Cluster computing. 3. High performance (HP) clusters: This computer networking methodology use supercomputers and Cluster computing to solve advanced computation problems.
Cluster Computing Advantages of using Cluster computing ● 1. Cost efficiency: In a Cluster computing Cost efficiency is the ratio of cost to output, that is the connecting group of the computer as computer cluster much cheaper as compared to mainframe computers. 2. Processing speed: The Processing speed of computer cluster is the same as a mainframe computer. 3. Expandability: The best benefit of Cluster Computing is that it can be expanded easily by adding the additional desktop workstation to the system. 4. High availability of resources: If any node fails in a computer cluster, another node within the cluster continue to provide uninterrupted processing. When a mainframe system fails, the entire system fails.
Cluster Computing Architecture
Grid Computing
Grid Computing ● The computing resources in most of the organizations are underutilized but are necessary for certain operations. ● The idea of grid computing is to make use of such non utilized computing power by the needy organizations, and thereby the return on investment (ROI) on computing investments can be increased. ● Thus, grid computing is a network of computing or processor machines managed with a kind of software such as middleware, in order to access and use the resources remotely. ● The managing activity of grid resources through the middleware is called grid services.
● Grid services provide access control, security, access to data including digital libraries and databases, and access to large-scale interactive and long-term storage facilities. ● Grid computing is more popular due to the following reasons: ~ Its ability to make use of unused computing power, and thus, it is a cost-effective solution (reducing investments, only recurring costs) ~ As a way to solve problems in line with any HPC-based application ~ Enables heterogeneous resources of computers to work cooperatively and collaboratively to solve a scientific problem ● Researchers associate the term grid to the way electricity is distributed in municipal areas for the common man.
Cloud Computing
Cloud Computing ● The computing trend moved towards cloud from the concept of grid computing, particularly when large computing resources are required to solve a single problem, using the ideas of computing power as a utility and other allied concepts. ● However, the potential difference between grid and cloud is that grid computing supports leveraging several computers in parallel to solve a particular application, while cloud computing supports leveraging multiple resources, including computing resources, to deliver a unified service to the end user.
● In cloud computing, the IT and business resources, such as servers, storage, network, applications, and processes, can be dynamically provisioned to the user needs and workload. ● In addition, while a cloud can provision and support a grid, a cloud can also support non-grid environments, such as a three-tier web architecture running on traditional or Web 2.0 applications.
BioComputing
Biocomputing ● Biocomputing systems use the concepts of biologically derived or simulated molecules (or models) that perform computational processes in order to solve a problem. ● The biologically derived models aid in structuring the computer programs that become part of the application. ● Biocomputing provides the theoretical background and practical tools for scientists to explore proteins and DNA. ● DNA and proteins are nature’s building blocks, but these building blocks are not exactly used as bricks.
● The function of the final molecule rather strongly depends on the order of these blocks. ● Thus, the bio computing scientist works on inventing the order suitable for various applications mimicking biology. ● Bio computing shall, therefore, lead to a better understanding of life and the molecular causes of certain diseases.
Mobile Computing
Mobile Computing ● In mobile computing, the processing (or computing) elements are small (i.e., handheld devices) and the communication between various resources is taking place using wireless media. ● Mobile communication for voice applications (e.g., cellular phone) is widely established throughout the world and witnesses a very rapid growth in all its dimensions including the increase in the number of subscribers of various cellular networks. ● An extension of this technology is the ability to send and receive data across various cellular networks using small devices such as smart phones.
● There can be numerous applications based on this technology; ⮚for example, video call or conferencing is one of the important applications that people prefer to use in place of existing voice (only) communications on mobile phones. ● Mobile computing–based applications are becoming very important and rapidly evolving with various technological advancements as it allows users to transmit data from remote locations to other remote or fixed locations.
MOBILE COMPUTING DEVICES
PRINCIPLES OF MOBILE COMPUTING
LIMITATIONS OF MOBILE COMPUTING
Quantum Computing
Quantum Computing ● Quantum computing is the area of study focused on developing computer technology based on the principles of quantum theory, which explains the nature and behavior of energy and matter on the quantum (atomic and subatomic) level. ● Quantum computing is an as-of-yet theoretical computing model that uses a very different form of data handling to perform calculations. ● The emergence of quantum computing is based on a new kind of data unit that could be called non-binary, as it has more than two possible values. ● Manufacturers of computing systems say that there is a limit for cramming more and more transistors into smaller and smaller spaces of integrated circuits (ICs) and thereby doubling the processing power about every 18 months.
● This problem will have to be overcome by a new quantum computing–based solution, wherein the dependence is on quantum information, the rules that govern the subatomic world. ● Quantum computers are millions of times faster than even our most powerful supercomputers today. ● Since quantum computing works differently on the most fundamental level than the current technology, and although there are working prototypes, these systems have not so far proved to be alternatives to today’s silicon-based machines.
Quantum Computing ● Classical Vs Quantum computer
Applications Some of the applications of Quantum Computing: Machine Learning, Computational Chemistry, Financial Portfolio Optimization, Logistics and Scheduling, Drug Design, Cyber Security, Code breaking, (Circuit, Software, and System Fault Simulation).
Quantum Computing
IBM Simulates a 56-Qubit Machine
Optical Computing
Optical Computing ● Optical computing system uses the photons in visible light or infrared beams, rather than electric current, to perform digital computations. ● An electric current flows at only about 10% of the speed of light. ● This limits the rate at which data can be exchanged over long distances and is one of the factors that led to the evolution of optical fiber. ● By applying some of the advantages of visible and/or IR networks at the device and component scale, a computer can be developed that can perform operations 10 or more times faster than a conventional electronic computer.
Nano Computing
Nano computing
Nano computing ● Nano computing refers to computing systems that are constructed from nanoscale components. ● The silicon transistors in traditional computers may be replaced by transistors based on carbon nanotubes. ● The successful realization of nanocomputers relates to the scale and integration of these nanotubes or components. ● The issues of scale relate to the dimensions of the components; they are, at most, a few nanometers in at least two dimensions. ● The issues of integration of the components are twofold: first, the manufacture of complex arbitrary patterns may be economically infeasible, and second, nanocomputers may include massive quantities of devices.
Cloud Computing-UNIT 1 claud computing basics
Cloud Computing-UNIT 1 claud computing basics
Cloud Computing-UNIT 1 claud computing basics
Cloud Computing-UNIT 1 claud computing basics

Cloud Computing-UNIT 1 claud computing basics

  • 1.
  • 2.
    Introduction ⮚What is computing? ●The process of utilizing computer technology to complete a task. ● Computing may involve computer hardware and/or software, but must involve some form of a computer system. ⮚What is paradigms? ● A style or a way of doing something. ● A set of practices to be followed to accomplish a task. ⮚Computing Paradigms ● In the domain of computing, there are many different standard practices being followed based on inventions and technological advancements.
  • 3.
    Computing Paradigms ●The variouscomputing paradigms: ~ High performance computing, ~ Parallel computing ~ Cluster computing, ~ Distributed computing ~ Grid computing, ~ Cloud computing, ~ Bio-computing, ~ Mobile computing, ~ Quantum computing, ~ Optical computing, ~ Nano computing.
  • 4.
  • 5.
    High Performance Computing ●In high-performance computing systems, a pool of processors are connected with other resources like memory, storage, and input and output devices, and the deployed software is enabled to run in the entire system of connected components. ● The processor machines can be of homogeneous or heterogeneous type. ● The legacy meaning of high-performance computing (HPC) is the supercomputers; however, it is not true in present-day computing scenarios.
  • 6.
    ● Thus, examplesof HPC include a small cluster of desktop computers or personal computers (PCs) to the fastest supercomputers. ● HPC systems are normally found in those applications where it is required to use or solve scientific problems. ● Most of the time, the challenge in working with these kinds of problems is to perform suitable simulation study, and this can be accomplished by HPC without any difficulty. ● Scientific examples such as protein folding in molecular biology and studies on developing models and applications based on nuclear fusion are worth noting as potential applications for HPC.
  • 7.
  • 8.
    High Performance Computing ●HPCUse In Africa And How Different Industries Can Take Advantage of It
  • 9.
    High Performance Computing UCLto host National High Performance Computing Hub for Materials Science A UCL-led consortium has been awarded to establish a new national High Performance Computing (HPC) facility for the Materials and Molecular Modelling community.
  • 10.
    High Performance Computing GlobalHigh Performance Computing market
  • 11.
  • 12.
  • 13.
    Parallel Computing ●Serial Computing: ●Traditionally,software has been written for serial computation: ○A problem is broken into a discrete series of instructions ○Instructions are executed sequentially one after another ○Executed on a single processor ○Only one instruction may execute at any moment in time
  • 14.
  • 15.
  • 16.
    Parallel Computing ●Parallel Computing: ●Inthe simplest sense, parallel computing is the simultaneous use of multiple compute resources to solve a computational problem: ○A problem is broken into discrete parts that can be solved concurrently ○Each part is further broken down to a series of instructions ○Instructions from each part execute simultaneously on different processors ○An overall control/coordination mechanism is employed
  • 17.
  • 18.
  • 19.
    Parallel Computing ●The computationalproblem should be able to: ○Be broken apart into discrete pieces of work that can be solved simultaneously; ○Execute multiple program instructions at any moment in time; ○Be solved in less time with multiple compute resources than with a single compute resource. ●The compute resources are typically: ○A single computer with multiple processors/cores ○An arbitrary number of such computers connected by a network
  • 20.
    This figure isshowing the majority of the world's large parallel computers (supercomputers) are clusters of hardware produced by a handful of (mostly) well known vendors.
  • 21.
  • 22.
    Distributed Computing ● Distributedcomputing is also a computing system that consists of multiple computers or processor machines connected through a network, which can be homogeneous or heterogeneous, but run as a single system. ● The connectivity can be such that the CPUs in a distributed system can be physically close together and connected by a local network, or they can be geographically distant and connected by a wide area network. ● The heterogeneity in a distributed system supports any number of possible configurations in the processor machines, such as mainframes, PCs, workstations, and minicomputers.
  • 23.
    ● The goalof distributed computing is to make such a network work as a single computer. ● Distributed computing systems are advantageous over centralized systems, because there is a support for the following characteristic features: 1. Scalability: It is the ability of the system to be easily expanded by adding more machines as needed, and vice versa, without affecting the existing setup. 2. Redundancy or replication: Here, several machines can provide the same services, so that even if one is unavailable (or failed), work does not stop because other similar computing supports will be available.
  • 24.
    Distributed Computing Centralized ● Asystem with centralized multiprocessor parallel architecture. In the late 1980 s Centralized systems have been progressively replaced by distributed systems. Characteristics of centralized system ● Non autonomous components ● usually homogeneous technology ● Multiple users share the same resources at all time ● single point of control ● single point of failure
  • 25.
    Distributed Computing ● Setof tightly coupled programs executing on one or more computers which are interconnected through a network and coordinating their actions. These programs know about one another and carry out tasks that none could carry out in isolation ●Characteristics of distributed system ● Autonomous components ● Mostly build using heterogeneous technology ● System components may be used exclusively ● Concurrent processes can execute ● Scalability- possibility of adding new hosts ● openness- easily extended and modified ● Heterogeneity-supports various H/W S/w platforms ● Resource sharing- H/w, S/W and data ● fault tolerance- ability to function correctly even if faults occur
  • 27.
  • 28.
    Distributed Computing ● thefollowing are the different application of the distributed system. ● Global positioning System ● World Wide Web ● Air Traffic Control System ● Automated Banking System ● In the World Wide Web application the data or application were distributed on the several number of the heterogeneous computer system, but for the end user or the browser it seems to be a single system from which user got the information. ● The multiple number of computer working concurrently and perform the resource sharing in the World Wide Web.
  • 29.
  • 30.
    Cluster Computing ● Acluster computing system consists of a set of the same or similar type of processor machines connected using a dedicated network infrastructure. ● All processor machines share resources such as a common home directory and have a software such as a message passing interface (MPI) implementation installed to allow programs to be run across all nodes simultaneously. ● This is also a kind of HPC category. ● The individual computers in a cluster can be referred to as nodes.
  • 31.
    ● The reasonto realize a cluster as HPC is due to the fact that the individual nodes can work together to solve a problem larger than any computer can easily solve. ● And, the nodes need to communicate with one another in order to work cooperatively and meaningfully together to solve the problem in hand. ● If we have processor machines of heterogeneous types in a cluster, this kind of clusters become a subtype and still mostly are in the experimental or research stage.
  • 32.
    Cluster Computing ● Acomputer cluster help to solve complex operations more efficiently with much faster processing speed, better data integrity than a single computer and they only used for mission-critical applications. ● The Clustering methods have identified as- HPC IAAS, HPC PAAS, that are more expensive and difficult to setup and maintain than a single computer. ● A computer cluster defined as the addition of processes for delivering large-scale processing to reduce downtime and larger storage capacity as compared to other desktop workstation or computer.
  • 33.
    Cluster Computing ● Someof the critical Applications of Cluster Computers are Google Search Engine, Petroleum Reservoir Simulation, Earthquake Simulation, Weather Forecasting. ● Cluster Can be classified into two category Open and Close Cluster. ● Open Cluster: All nodes in Open Cluster are needed IPs, and that are accessible through internet/web, that cause more security concern. Close Cluster: On the other hand Close Cluster are hide behind the gateway node and provide better security.
  • 34.
    Cluster Computing ● Typesof Cluster computing ● 1. Load-balancing clusters: As the name implies, This system is used to distribute workload across multiple computers. That system ● distributes the processing load as possible across a cluster of computers. 2. High availability (HA) clusters: A high availability clusters (HA cluster) are the bunch of computers that can reliably utilise for ● redundant operations in the event of nodes failure in Cluster computing. 3. High performance (HP) clusters: This computer networking methodology use supercomputers and Cluster computing to solve advanced computation problems.
  • 35.
    Cluster Computing Advantages ofusing Cluster computing ● 1. Cost efficiency: In a Cluster computing Cost efficiency is the ratio of cost to output, that is the connecting group of the computer as computer cluster much cheaper as compared to mainframe computers. 2. Processing speed: The Processing speed of computer cluster is the same as a mainframe computer. 3. Expandability: The best benefit of Cluster Computing is that it can be expanded easily by adding the additional desktop workstation to the system. 4. High availability of resources: If any node fails in a computer cluster, another node within the cluster continue to provide uninterrupted processing. When a mainframe system fails, the entire system fails.
  • 37.
  • 38.
  • 39.
    Grid Computing ● Thecomputing resources in most of the organizations are underutilized but are necessary for certain operations. ● The idea of grid computing is to make use of such non utilized computing power by the needy organizations, and thereby the return on investment (ROI) on computing investments can be increased. ● Thus, grid computing is a network of computing or processor machines managed with a kind of software such as middleware, in order to access and use the resources remotely. ● The managing activity of grid resources through the middleware is called grid services.
  • 40.
    ● Grid servicesprovide access control, security, access to data including digital libraries and databases, and access to large-scale interactive and long-term storage facilities. ● Grid computing is more popular due to the following reasons: ~ Its ability to make use of unused computing power, and thus, it is a cost-effective solution (reducing investments, only recurring costs) ~ As a way to solve problems in line with any HPC-based application ~ Enables heterogeneous resources of computers to work cooperatively and collaboratively to solve a scientific problem ● Researchers associate the term grid to the way electricity is distributed in municipal areas for the common man.
  • 45.
  • 46.
    Cloud Computing ● Thecomputing trend moved towards cloud from the concept of grid computing, particularly when large computing resources are required to solve a single problem, using the ideas of computing power as a utility and other allied concepts. ● However, the potential difference between grid and cloud is that grid computing supports leveraging several computers in parallel to solve a particular application, while cloud computing supports leveraging multiple resources, including computing resources, to deliver a unified service to the end user.
  • 47.
    ● In cloudcomputing, the IT and business resources, such as servers, storage, network, applications, and processes, can be dynamically provisioned to the user needs and workload. ● In addition, while a cloud can provision and support a grid, a cloud can also support non-grid environments, such as a three-tier web architecture running on traditional or Web 2.0 applications.
  • 50.
  • 51.
    Biocomputing ● Biocomputing systemsuse the concepts of biologically derived or simulated molecules (or models) that perform computational processes in order to solve a problem. ● The biologically derived models aid in structuring the computer programs that become part of the application. ● Biocomputing provides the theoretical background and practical tools for scientists to explore proteins and DNA. ● DNA and proteins are nature’s building blocks, but these building blocks are not exactly used as bricks.
  • 52.
    ● The functionof the final molecule rather strongly depends on the order of these blocks. ● Thus, the bio computing scientist works on inventing the order suitable for various applications mimicking biology. ● Bio computing shall, therefore, lead to a better understanding of life and the molecular causes of certain diseases.
  • 54.
  • 55.
    Mobile Computing ● Inmobile computing, the processing (or computing) elements are small (i.e., handheld devices) and the communication between various resources is taking place using wireless media. ● Mobile communication for voice applications (e.g., cellular phone) is widely established throughout the world and witnesses a very rapid growth in all its dimensions including the increase in the number of subscribers of various cellular networks. ● An extension of this technology is the ability to send and receive data across various cellular networks using small devices such as smart phones.
  • 56.
    ● There canbe numerous applications based on this technology; ⮚for example, video call or conferencing is one of the important applications that people prefer to use in place of existing voice (only) communications on mobile phones. ● Mobile computing–based applications are becoming very important and rapidly evolving with various technological advancements as it allows users to transmit data from remote locations to other remote or fixed locations.
  • 57.
  • 58.
  • 59.
  • 60.
  • 61.
    Quantum Computing ● Quantumcomputing is the area of study focused on developing computer technology based on the principles of quantum theory, which explains the nature and behavior of energy and matter on the quantum (atomic and subatomic) level. ● Quantum computing is an as-of-yet theoretical computing model that uses a very different form of data handling to perform calculations. ● The emergence of quantum computing is based on a new kind of data unit that could be called non-binary, as it has more than two possible values. ● Manufacturers of computing systems say that there is a limit for cramming more and more transistors into smaller and smaller spaces of integrated circuits (ICs) and thereby doubling the processing power about every 18 months.
  • 62.
    ● This problemwill have to be overcome by a new quantum computing–based solution, wherein the dependence is on quantum information, the rules that govern the subatomic world. ● Quantum computers are millions of times faster than even our most powerful supercomputers today. ● Since quantum computing works differently on the most fundamental level than the current technology, and although there are working prototypes, these systems have not so far proved to be alternatives to today’s silicon-based machines.
  • 63.
  • 64.
    Applications Some of theapplications of Quantum Computing: Machine Learning, Computational Chemistry, Financial Portfolio Optimization, Logistics and Scheduling, Drug Design, Cyber Security, Code breaking, (Circuit, Software, and System Fault Simulation).
  • 65.
  • 66.
    IBM Simulates a56-Qubit Machine
  • 67.
  • 68.
    Optical Computing ● Opticalcomputing system uses the photons in visible light or infrared beams, rather than electric current, to perform digital computations. ● An electric current flows at only about 10% of the speed of light. ● This limits the rate at which data can be exchanged over long distances and is one of the factors that led to the evolution of optical fiber. ● By applying some of the advantages of visible and/or IR networks at the device and component scale, a computer can be developed that can perform operations 10 or more times faster than a conventional electronic computer.
  • 75.
  • 76.
  • 77.
    Nano computing ● Nanocomputing refers to computing systems that are constructed from nanoscale components. ● The silicon transistors in traditional computers may be replaced by transistors based on carbon nanotubes. ● The successful realization of nanocomputers relates to the scale and integration of these nanotubes or components. ● The issues of scale relate to the dimensions of the components; they are, at most, a few nanometers in at least two dimensions. ● The issues of integration of the components are twofold: first, the manufacture of complex arbitrary patterns may be economically infeasible, and second, nanocomputers may include massive quantities of devices.