Category Archives: Cloud Servers
Has Arm Discovered the Ecosystem Keys? – The Next Platform
Arm server development is a reality and a growing one at that. Not just from a performance point of view but also, perhaps more important, from an ecosystem view.
Be it the Marvell ThunderX2 processor or the Ampere eMAG Skylark processor, the hyperscale, cloud, enterprise ecosystems are willing to adopt these new processors to further improve their TCO or dollars/core.
The all-important ecosystem is catching up with Arm, which is key to the momentum necessary to make the Arm servers a sustainable reality. With AWS launching their version of Arm instances i.e. Graviton processors, theres the much needed push to make the software ecosystem more widely acceptable in the industry. Not just that, AWS even announced bare-metal offerinings for EC2 A1 instances.
Slowly but steadily, Arm has also made a mark for itself in high performance computing, something we expect to see in full force at this years Supercomputing Conference. Arm has the most traction in terms of deployments and software development in HPC in the United States, Europe and Japan with each region leading the way along different trajectories to deploy systems based on the Arm architecture for their supercomputers.
All of this has taken time and extended development, of course. The first wave of Arm based servers came in 2010 until 2014 and were more experimental in nature than real production systems.
The first 64-bit Arm design i.e. the ARMv8-A was introduced in 2011 and since then the Arm server ecosystem have seen lots of ups and downs. ZTSystems, in November 2010 had launched a 1U Data Center Arm server based on Cortex-A9 cores (32-bit) which was supposed to be energy efficient and a denser solution compared to Intel Servers. Then came Calxeda with their version of 32-bit Arm servers i.e. the EnergyCore-ECX-1000 which did not see adoption and Calxeda eventually went defunct in 2013. In 2011 AppliedMicro launched the X-Gene 1 processor followed by X-Gene 2 in 2014. Samsung, Cavium (now Marvell) and AMD came up with their versions of Arm processors which tried to penetrate the server market but could not generate tangible interest among the end-users to adopt these technologies.
Arm servers have undergone a transformation in terms of development and early signs of this were seen in a semi-secret project within Broadcom which was taking shape in the form of Project Vulcan. The idea was to develop a world class 64-bit serious Arm server to take on Intel in the HPC and cloud market.
In late 2016, when Avago gave up on Broadcoms ambitions to develop a first class Arm server, Cavium jumped in and brought the Vulcan IP and Team on-board and fully funded the Vulcan project, re-christened as Cavium ThunderX2 now, Marvell ThunderX2. In more ways than one, the ThunderX2 is a serious contender to Intel and AMD in the HPC, hyperscale and cloud businesses.
To make things better for the Arm ecosystem, in 2017, a brand new company, Ampere Computing bought the X-Gene assets and re-introduced the X-Gene processor as the Ampere eMAG processor. It needs to be mentioned that Qualcomm tried its hand at building a true Data Center Arm Server Centriq based on the Falkor Architecture and given Qualcomms standing, with time, it could have made their data center server project a success. However, for reasons unknown to many, they chose to significantly disinvest and many personnel from Qualcomms Centriq project were hired by Ampere Computing in Raleigh. Huawei has a very compelling Arm Server offering in the Kunpeng 920, which is a 7-nm, 64 core CPU.
Figure 1: Diverse Arm architectures (source)
The question many have is whether the Arm server ecosystem is mature enough to be excited about?
The ecosystem has come a long way to become a stable one. However, it has many miles to go to reach the same level as x86. Given this momentum, it would not be surprising if the likes of Google, Facebook, Tencent etc. are actively experimenting with Arm platforms. Amazon and Microsoft have already invested in Arm platforms in their respective clouds i.e. AWS & Azure.
Figure 2: Commits to Linux GitHub repository for x86 vs. arm64 as of 13th November, 2019
The contributions towards enabling aarch64 for Linux operating system have steadily increased since 2012 while the growth rate for x86 has not been as consistent. These are good indications that the Arm ecosystem is here to stay and growing.
An ongoing debate among software engineers is whether to implement a business logic in a monolithic architecture or take the same logic and break it down into multiple pieces. There is a growing trend of organizations moving to a Microservices architecture for various reasons be it unit testing, ease of deployment, server performance among many others. Also, microservices based architecture are relatively easy to scale compared to a monolith. Linaro, Arm and Arm Server Manufacturers are leading this charge. Also, Packet is providing the developer community a platform to develop and sustain the ecosystem.
If theres one area where Arm servers have taken the biggest strides, it is definitely be High Performance Computing (HPC). The Arm ecosystem for HPC is also the most developed compared to Arms progress in cloud datacenters.
The momentum for Arm in HPC was driven by many centers, but Dr. Simon McIntosh-Smith and the University of Bristol and Cray hosting the 1st Isambard Hackathon to optimize HPC applications for ThunderX2 based servers back in November 2017 at Bristol. This was promptly followed up by a 2nd Isambard Hackathon in March 2018.
Most of the HPC applications compile and run out of the box for Arm based servers with Arm compilers, GCC, OpenMPI, OpenMP support.
I participated in both representing Cavium Inc, assisting developers, architects and engineers optimize their codes/applications for ThunderX2 Processors. Collectively, we optimized key HPC applications like NAMD, UM-NEMO, OpenFOAM, NWCHEM, CASTEP, etc. and compared to Intel CPU Architectures like Broadwell and Skylake. Prof Smith and team did a detailed study identifying the opportunities and benefits of Arm Servers with regards to the incumbent Intel servers with compelling performance per dollar for the Arm-based servers.
Figure 3: Cray-Isambard performance comparison on mini-apps
Figure 4: Cray-Isambard performance comparison on key Archer applications
Figure 5: Cavium Inc. published HPC Performance comparison vs. Intel Skylake CPUs (2017)
This was a significant movement that Arm servers needed in the HPC space. The two Isambard hackathons also fast-tracked the Arm HPC development with Arm optimizing their compilers as well as Math libraries in collaboration with Arm server manufacturers like Cavium Inc (now Marvell Semiconductors). There is tremendous movement in the Arm HPC Performance Libraries optimization world. Arm has invested in optimizing GEMM, SVE, spMM, spMV and FFT libraries in collaboration with developers and Silicon manufacturers like Marvell. The Arm Allinea Studio has successfully established itself as a go-to tool for Arm server Workload Analysis, similar to what VTune would be for Intel.
Another major milestone was the Vanguard Astra Arm based supercomputer at Sandia National Laboratories powered by DoE, Cavium and HPE. This is the first Arm based supercomputer to make the TOP500 list at 156th position as of June 2019 and 198th rank in the November 2019 rankings. The building blocks are HPE Apollo 70 platforms, Marvell ThunderX2 CPUs with 4xEDR Infiniband interconnect. The Astra Supercomputer is made up of 2592 compute servers i.e. 145k cores and 663 TB memory. US DoE is making a concerted effort to invest in diverse as well as future proof technologies such as Arm, in its path towards achieving exascale computing.
Figure 6: Astra, the Arm based supercomputer debuted on the TOP500 list in November 2018
Europe and Asia are taking huge strides in deploying Arm based clusters and systems for HPC and Research. Be it Monte-Carlo, Isambard or CINECA-E4 projects in Europe or Japans Arm based Fugaku supercomputer, its just the beginning of a new era of Arm in HPC. Cray is betting big with the A64FX Arm chip built by Fujitsu. The A64FX prototype is number one on the Green500 list and 160th on the Top500 list..
HPC workloads tend to be highly parallelizable in nature, and Arm CPUs provide an opportunity to leverage lots of cores at reasonable price points. Further, having competition in the CPU market benefits all buyers, not just HPC shops, to negotiate the best resources for their workloads.
Marvell is a pioneer in more ways than one in introducing the Arm server ecosystem to the hyperscale world with Marvell and Microsoft partnering on ThunderX2 platforms for Azure. Oracle has invested $40 Million in Ampere Computing, which is home to the ARMv8 eMAG processor. Oracle also has plans to massively expand their datacenter footprint in the coming months and this investment in Ampere could mean potential deployment of eMAG processors in Oracle Data Centers.
In the recent past, theres been a slew of announcements regarding enhancements to the Arm ecosystem. VMware announced 64-bit support Arm Support. In an official announcement, DDN announced professional support for Lustre on Arm servers in 2018 In mid 2019 at ISC, AMI announced firmware support for the Marvell ThunderX2 Arm based servers in March 2019.
NVIDIA announced CUDA support for Arm at ISC19 and backed it up with a major announcement of introducing a reference design to enable organizations to build GPU-accelerated Arm based servers, which is a big shift towards enabling Arm to be successful in the HPC and accelerated computing segment. Imagine a system with power efficient Arm based CPUs with GPUs for training and AI ASICs for inference. Machine Learning & Artificial Intelligence pose interesting opportunities & the collaboration with NVIDIA will enable this segment for Arm based solutions.
Like Intel, AMD and Arm, Ampere Computing too has created a developer program for developers to build and expand their Cloud Ecosystem. This will enable further and faster integration of Arm servers in the hyperscale and datacenter world in a much more open and collaborative way.
While the ecosystem still needs more time to grow and mature, it is steadily moving towards that nirvana of It just works. With the emergence of Arm in the computer architecture world along with RISC-V and many other semiconductor start-ups, its only a matter of time until aarch64 is the new normal like x86. That is what the community is all striving towards.
Once the developers are convinced that their software stack just works on Arm Servers, it would be a big win for the Arm Server ecosystem, and I for one am willing to make the bold claim that for many workloads especially HPC It just works
About the Author
Indraneil Gokhale is a Performance Engineer and leads the Hardware Engineering team at Box Inc. Indraneil has previously worked at Cavium (now Marvell), Uber and Intel. Indraneil has experience in optimizing HPC applications and workloads for x86 and aarch64 architectures. He has published white papers, book chapters on optimizing the Weather Research and Forecasting (WRF) application. Indraneil holds a Masters Degree in Electrical Engineering from Auburn University, USA and a Bachelors Degree in EEE from Jawaharlal Nehru Technological University, Hyderabad, India.
See the original post here:
Has Arm Discovered the Ecosystem Keys? - The Next Platform
SME disaster recovery made easy with cloud, hybrid and HCI – ComputerWeekly.com
Disaster recovery(DR) must be reliable, speedy and economical. These are the basic requirements for all businesses, and small to medium-sized enterprises (SMEs) are no exception. But, for smaller firms, cost considerations will be at or near the top of the IT managers list.
Organisations often view disaster recovery as little more than insurance, or as IDC analyst Phil Goodwin puts it, an expense that is likely to have little payback.
Large firms in highly regulated markets such as financial services invest heavily in disaster recovery and business continuity, not least because they are mandated to do so.
RBS, for example, was fined 56m for an IT failure in 2012.
Smaller firms might not have the budget for DR, however. Or they might choose not to pay for it, hoping they can ride out any incident with backups and hard work. This is short-sighted.
The Uptime Institutes 2018 Global Data Center Surveyfound that 31% of businesses experienced downtime that caused severe damage, but that 80% of incidents could have been prevented.
Operating disaster recovery on a pay-as-you-go model, combined with faster data transfers over the public internet, is transforming DR options for SMEs
Disaster recovery is becoming cheaper, simpler and more effective, through the growth of cloud-based services.
Firms no longer need to invest in dedicated or duplicate hardware, remote datacentres and the skilled staff to maintain them. Cloud or hybrid cloud technologies allow smaller companies to outsource much of the technical side of DR provision and to move to an on-demand model.
Operating disaster recovery on a pay-as-you-go model, combined with faster data transfers over the public internet, is transforming DR options for SMEs.
For smaller firms, conventional backup and disaster recovery has long meant saving data to tape or other removable media and storing it off-site. In case of a disaster the firm would need to source new hardware and restore data and applications.
Larger or better-resourced businesses, or those depending heavily on their data, will have invested in automatic off-site data replication and even standby servers. Others contracted with specialist suppliers to manage DR.
In the past, DR meant having a redundant location that was either always-on, a hot DR facility, or the use of shared resources that would be configured and set up when needed, says Roy Illsley, a distinguished analyst at Ovum who covers infrastructure solutions. These solutions are typically expensive or bring long recovery times.
Some businesses have moved to ad-hoc cloud backup solutions, including consumer-grade online storage. But this is still labour-intensive.
Small and medium-sized enterprises often deem disaster recovery too expensive or peripheral to core operations to fully invest in, warns Mark Wass, a director at business continuity supplier Sungard.
They often settle for a DIY, cloud-based platform approach, which they believe to be cheaper, and then assign responsibility for managing it to the office manager, he says.
Although this avoids reliance on on-site backups, it is still a manual process. And, Wass says, assuming staff can rely on cloud-based storage and access to a personal device to work on is risky.
Fortunately, the market for online disaster recovery is growing, and so is the choice of suppliers. Companies such as Veeam, Rubrik, Commvault, Cohesity and Nutanix provide options for businesses that need an off-the-shelf solution.
Cloud-based DR is available either directly from suppliers, or through managed service providers and IT integrators.
OGL Computer is a managed service provider with 1,200 UK SME customers. It offers a choice of cloud recovery and cloud-based data replication, as well as dedicated recovery options for VMware and Hyper-V. Recovery allows firms to restore their key applications within 24 hours cloud-based replication provides recovery within seconds.
But, as OGLs enterprise solutions architect, Steve Bennett, says, there are also customers who need, or prefer, an on-premise solution.
The availability of cloud services is an obvious way backup is becoming easier for SMEs, says Bryan Betts, analyst at Freeform Dynamics. But if they cant or dont want to use cloud, theres also the appliance option, either physical or virtual.
The availability of cloud services is an obvious way backup is becoming easier for SMEs. But if they cant or dont want to use cloud, theres also the appliance option, either physical or virtual Bryan Betts, Freeform Dynamics
Add in the availability of flash storage and modern software that provides user self-service, and you can get a box that not only takes care of all or almost all your backups, but also lets users themselves quickly sort out most of their data recovery needs.
Changes in the way businesses buy technology are also affecting disaster recovery.
Hyper-converged infrastructure (HCI) started out as a way to make it easier to deploy virtual machines. But because hyper-converged systems usually include their own storage, they lend themselves to disaster recovery too. On-premise HCI systems can be out of the reach of SMEs, but the implementation of cloud-based HCI makes it much more accessible.
An example is Cohesitys Clusters, but Nutanix and Rubrik use similar technologies. Businesses with the in-house expertise can also look at replication to enterprise public cloud providers, such as Amazon Web Services, Microsoft Azure and Google Cloud Platform.
Cloud DR isnt [as] neatly packaged but is more flexible, says Peter Groucutt, managing director of backup provider Databarracks. You can choose your replication software, such as ASR, Zerto or Veeam, and choose your cloud platform. For day-to-day replication you can keep cloud resources to a minimum but have the resources available to scale up as much as you need when you invoke DR.
Another factor in favour of a bespoke approach is the growing need to back up cloud-based applications. The appliance-to-offsite-backup or appliance-to-cloud route is tried and tested, but data backups from cloud-based applications are largely best kept in the cloud.
IT managers should check their disaster recovery plan includes cloud applications. More software-as-a-service (SaaS) providers are including application availability and data protection options with their products.
No amount of technology, though, will protect a business if it fails to work. This means having a disaster recovery plan, a robust testing regimeand a plan to deal with human factors, from the availability of technical experts to senior managements ability to act under pressure.
You must have a robust process that has clear rules on when a switch to disaster recovery is required and when it is not Roy Illsley, Ovum
For SMEs, the main point to consider is ease of use in setting up the replication, and then how the DR capability can be tested and verified it is operating and fit for purpose, says Ovums Illsley. The area where most SMEs fail is in the process for invoking the DR plan. You must have a robust process that has clear rules on when a switch to DR is required and when it is not.
Freeform Dynamics Betts goes further. The absolutely key requirement for DR is to make sure you can recover, he warns. Test it often enough to make sure not just that your backup process is reliable, but that you can rebuild a working system from it within the specified recovery time objective and recovery point objective.
Read more:
SME disaster recovery made easy with cloud, hybrid and HCI - ComputerWeekly.com
AllSaints tech team explains tools behind its cloud journey – Essential Retail
Members of the AllSaints technology team have explained the companys five-year journey to becoming a cloud-based business, describing it as the biggest infrastructure change in its history.
Since 2014, AllSaints has used the capabilities of data management specialist Pythian, API codifier Terraform, and an array of Google Cloud tools to help create a new infrastructure. During the project, the retailer has halved the number of servers it runs from 60 to 30, and it says it is in a stronger position to flex its online business to meet consumers ever-changing demands.
As previously reported by Essential Retail, AllSaints uses Google G Suite for internal communications, which the retailers executive consultant for digital and technology, John Bovill, said was a deliberate move to get staff familiar with using the cloud.
Soon after, AllSaints migrated its infrastructure to Google Cloud with support from Pythian. Andy Dean, technical operations manager at AllSaints, said 60 individual services moved onto the new technology set-up.
The interdependencies between them meant that it made more sense to move them all at once, and that took a lot of planning, he noted, adding that the development team needed to change to Cloud SQL technology and undertake significant re-coding.
It was the biggest infrastructure change wed made in the history of the company, so one of our goals was that nobody noticed the change, explained Dean.
One result of that change is AllSaints now utilises Google Compute Engine and Google Kubernetes Engine autoscaling to meet the needs of additional online traffic at peak times. It said this means it no longer has to rely on additional servers and has contributed to a 50% reduction in infrastructure costs.
The team now monitors web performance through Google Clouds Stackdriver, while Googles wider network provides disaster recovery services which replaces a previous reliance on a single data centre.
In-house development
AllSaints develops its customer-facing services in-house, including electronic point-of-sale systems and its mobile app, and the move towards a microservice architecture gave the development team a chance to change their way of working. They can now build a continuous integration/continuous deployment (CI/CD) pipeline to automate the software delivery process, using Jenkins on Google Cloud and Terraform.
Before, we couldnt confidently say a bug was fixed until we actually tested it in production now we can deploy code in test environments that exactly mimic production, noted Dean.
The improved CI/CD pipeline means we can update our services every day, with a shorter lifespan on bugs, and minimal disruption. That makes us more responsive to customer needs, more proactive. And that's exactly what were trying to achieve.
AllSaints said online page speeds and conversions have increased since rolling out the new infrastructure. Next up, the retailer is preparing to deploy Istio to connect and monitor its microservices model, and it is planning to explore ways of leveragingdata within its organisation via Google Cloud tool, BigQuery.
Strategically we are looking to maximise our usage of Google Cloud, driving this and associated technologies to provide the best possible AllSaints experience for our customers, explains Bovill, who having overseen the latter stages of the AllSaints cloud journey is preparing to leave the retailer at the end of this year.
See the rest here:
AllSaints tech team explains tools behind its cloud journey - Essential Retail
Kubernetes – taming the cloud – TechRadar
When you want to use Linux to provide services to a business, those services will need to be secure, resilient and scalable. Nice words, but what do we mean by them?
Secure means that users can access to the data they require, be that read-only access or write access. At the same time, no data is exposed to any party thats not authorised to see it. Security is deceptive: you can think you have everything protected only to find out later that there are holes. Designing in security from the start of a project is far easier than trying to retrofit it later.
Resilient means your services tolerate failures within the infrastructure. A failure might be a server disk controller that can no longer access any disks, rendering the data unreachable. Or the failure might be a network switch that no longer enables two or more systems to communicate. In this context, a single point of failure or SPOF is a failure that adversely affects service availability. A resilient infrastructure is one with no SPOFs.
Scalable describes the ability of systems to handle spikes of demand gracefully. It also dictates how easily changes may be made to systems. For example, adding a new user, increasing the storage capacity or moving an infrastructure from Amazon Web Services to Google Cloud or even moving it in-house.
As soon as your infrastructure expands beyond one server, there are lots of options for increasing the security, resilience and scalability. Well look at how these problems have been solved traditionally, and what new technology is available that changes the face of big application computing.
(Image credit: Future)
Enjoying what you're reading? Want more Linux and open source? We can deliver, literally! Subscribe to Linux Format today at a bargain price. You can get print issues, digital editions or why not both? We deliver to your door worldwide for a simple yearly fee. So make your life better and easier, subscribe now!
To understand whats possible today, its helpful to look at how technology projects have been traditionally implemented. Back in the olden days that is, more than 10 years ago businesses would buy or lease hardware to run all the components of their applications. Even relatively simple applications, such as a WordPress website, have multiple components. In the case of WordPress, a MySQL database is needed along with a web server, such as Apache, and a way of handling PHP code. So, theyd build a server, set up Apache, PHP and MySQL, install WordPress and off theyd go.
By and large, that worked. It worked well enough that there are still a huge number of servers configured in exactly that way today. But it wasnt perfect, and two of the bigger problems were resilience and scalability.
Lack of resilience meant that any significant issue on the server would result in a loss of service. Clearly a catastrophic failure would mean no website, but there was also no room to carry out scheduled maintenance without impacting the website. Even installing and activating a routine security update for Apache would necessitate a few seconds outage for the website.
The resilience problem was largely solved by building high availability clusters. The principle was to have two servers running the website, configured such that the failure of either one didnt result in the website being down. The service being provided was resilient even if the individual servers were not.
Part of the power of Kubernetes is the abstraction it offers. From a developers perspective, they develop the application to run in a Docker container. Docker doesnt care whether its running on Windows, Linux or some other operating system. That same Docker container can be taken from the developers MacBook and run under Kubernetes without any modification.
The Kubernetes installation itself can be a single machine. Of course, a lot of the benefits of Kubernetes wont be available: there will be no auto-scaling; theres an obvious single point of failure, and so on. As a proof of concept in a test environment, though, it works.
Once youre ready for production, you can run in-house or on a Cloud provider such as AWS or Google Cloud. The Cloud providers have some built-in services that assist in running Kubernetes, but none of are hard requirements. If you want to move between Google, Amazon and your own infrastructure, you set up Kubernetes and move across. None of your applications have to change in any way.
And where is Linux? Kubernetes runs on Linux, but the operating system is invisible to the applications. This is a significant step in the maturity and usability of IT infrastructures.
The scalability problem is a bit trickier. Lets say your WordPress site gets 1,000 visitors a month. One day, your business is mentioned on Radio 4 or breakfast TV. Suddenly, you get more than a months worth of visitors in 20 minutes. Weve all heard stories of websites crashing, and thats typically why: a lack of scalability.
The two servers that helped with resilience could manage a higher workload than one server alone could, but thats still limited. Youd be paying for two servers 100 per cent of the time and most of the time both were working perfectly. Its likely that one alone could run your site. Then John Humphrys mentions your business on Today and youd need 10 servers to handle the load but only for a few hours.
The better solution to both the resilience and scalability problem was cloud computing. Set up a server instance or two the little servers that run your applications on Amazon Web Services (AWS) or Google Cloud, and if one of the instances failed for some reason, it would automatically be restarted. Set up auto-scaling correctly and when Mr Humphrys causes the workload on your web server instances to rapidly rise, additional server instances are automatically started to share the workload. Later, as interest dies down, those additional instances are stopped, and you only pay for what you use. Perfect or is it?
Whilst the cloud solution is much more flexible than the traditional standalone server, there are still issues. Updating all the running cloud instances isnt straightforward. Developing for the cloud has challenges too: the laptop your developers are using may be similar to the cloud instance, but its not the same. If you commit to AWS, migrating to Google Cloud is a complex undertaking. And suppose, for whatever reason, you simply dont want to hand over your computing to Amazon, Google or Microsoft?
Containers have emerged as a means to wrap applications with all of their dependencies up into a single package that can be run anywhere. Containers, such as Docker, can run on your developers laptops in the same way as they run on your cloud instances, but managing a fleet of containers becomes increasingly challenging as the number of containers grows.
The answer is container orchestration. This is a significant shift in focus. Before, we made sure we had enough servers, be they physical or virtual, to ensure we could service the workload. Using the cloud providers autoscaling helped, but we were still dealing with instances. We had to configure load balancers, firewalls, data storage and more manually. With container orchestration, all of that (and much more) is taken care of. We specify the results we require and our container orchestration tools fulfil our requirements. We specify what we want done, rather than how we want it done.
Kubernetes (ku-ber-net-eez) is the leading container orchestration tool today, and it came from Google. If anyone knows how to run huge-scale IT infrastructures, Google does. The origin of Kubernetes is Borg, an internal Google project thats still used to run most of Googles applications including its search engine, Gmail, Google Maps and more. Borg was a secret until Google published a paper about it in 2015, but the paper made it very apparent that Borg was the principal inspiration behind Kubernetes.
Borg is a system that manages computational resources in Googles data centres and keeps Googles applications, both production and otherwise, running despite hardware failure, resource exhaustion or other issues occurring that might otherwise have caused an outage. It does this by carefully monitoring the thousands of nodes that make up a Borg cell and the containers running on them, and starting or stopping containers as required in response to problems or fluctuations in load.
Kubernetes itself was born out of Googles GIFEE (Googles Infrastructure For Everyone Else) initiative, and was designed to be a friendlier version of Borg that could be useful outside Google. It was donated to the Linux Foundation in 2015 through the formation of the Cloud Native Computing Foundation (CNCF).
Kubernetes provides a system whereby you declare your containerised applications and services, and it makes sure your applications run according to those declarations. If your programs require external resources, such as storage or load balancers, Kubernetes can provision those automatically. It can scale your applications up or down to keep up with changes in load, and can even scale your whole cluster when required. Your programs components dont even need to know where theyre running: Kubernetes provides internal naming services to applications so that they can connect to wp_mysql and be automatically connected to the correct resource.
The end result is a platform that can be used to run your applications on any infrastructure, from a single machine through an on-premise rack of systems to cloud-based fleets of virtual machines running on any major cloud provider, all using the same containers and configuration. Kubernetes is provider-agnostic: run it wherever you want.
Kubernetes is a powerful tool, and is necessarily complex. Before we get into an overview, we need to introduce some terms used within Kubernetes. Containers run single applications, as discussed above, and are grouped into pods. A pod is a group of closely linked containers that are deployed together on the same host and share some resources. The containers within a pod work as a team: theyll perform related functions, such as an application container and a logging container with specific settings for the application.
Four key Kubernetes components are the API Server, the Scheduler, the Controller Manager and a distributed configuration database called etcd. The API Server is at the heart of Kubernetes, and acts as the primary endpoint for all management requests. These may be generated by a variety of sources including other Kubernetes components, such as the scheduler, administrators via command-line or web-based dashboards, and containerised applications themselves. It validates requests and updates data stored in etcd.
The Scheduler determines which nodes the various pods will run on, taking into account constraints such as resource requirements, any hardware or software constraints, workload, deadlines and more.
The Controller Manager monitors the state of the cluster, and will try to start or stop pods as necessarily, via the API Server, to bring the cluster to the desired state. It also manages some internal connections and security features.
Each node runs a Kubelet process, which communicates with the API server and manages containers generally using Docker and Kube-Proxy, which handles network proxying and load balancing within the cluster.
The etcd distributed database system derives its name from the /etc folder on Linux systems, which is used to hold system configuration information, plus the suffix d, often used to denote a daemon process. The goals of etcd are to store key-value data in a distributed, consistent and fault-tolerant way.
The API server keeps all its state data in etcd and can run many instances concurrently. The scheduler and controller manager can only have one active instance but uses a lease system to determine which running instance is the master. All this means that Kubernetes can run as a Highly Available system with no single points of failure.
So how do we use those components in practice? What follows is an example of setting up a WordPress website using Kubernetes. If you wanted to do this for real, then youd probably use a predefined recipe called a helm chart. They are available for a number of common applications, but here well look at some of the steps necessary to get a WordPress site up and running on Kubernetes.
The first task is to define a password for MySQL:
kubectl will talk to the API Server, which will validate the command and then store the password in etcd. Our services are defined in YAML files, and now we need some persistent storage for the MySQL database.
The specification should be mostly self-explanatory. The name and labels fields are used to refer to this storage from other parts of Kubernetes, in this case our WordPress container.
Once weve defined the storage, we can define a MySQL instance, pointing it to the predefined storage. Thats followed by defining the database itself. We give that database a name and label for easy reference within Kubernetes.
Now we need another container to run WordPress. Part of the container deployment specification is:
The strategy type Recreate means that if any of the code comprising the application changes, then running instances will be deleted and recreated. Other options include being able to cycle new instances in and removing existing instances, one by one, enabling the service to continue running during deployment of an update. Finally, we declare a service for WordPress itself, comprising the PHP code and Apache. Part of the YAML file declaring this is:
Note the last line, defining service type as LoadBalancer. That instructs Kubernetes to make the service available outside of Kubernetes. Without that line, this would merely be an internal Kubernetes only service. And thats it. Kubernetes will now use those YAML files as a declaration of what is required, and will set up pods, connections, storage and so on as required to get the cluster into the desired state.
This has necessarily been only a high-level overview of Kubernetes, and many details and features of the system have been omitted. Weve glossed over autoscaling (both pods and the nodes that make up a cluster), cron jobs (starting containers according to a schedule), Ingress (HTTP load balancing, rewriting and SSL offloading), RBAC (role-based access controls), network policies (firewalling), and much more. Kubernetes is extremely flexible and extremely powerful: for any new IT infrastructure, it must be a serious contender.
If youre not familiar with Docker start here: https://docs.docker.com/get-started.
Theres an interactive, tutorial on deploying and scaling an app here: https://kubernetes.io/docs/tutorials/kubernetes-basics.
And see https://kubernetes.io/docs/setup/scratch for how to build a cluster.
You can play with a free Kubernetes cluster at https://tryk8s.com.
Finally, you can pore over a long, technical paper with an excellent overview of Googles use of Borg and how that influenced the design of Kubernetes here: https://storage.googleapis.com/pub-tools-public-publication-data/pdf/43438.pdf.
Find out more about Tiger Computing.
(Image credit: Future)
Enjoying what you're reading? Want more Linux and open source? We can deliver, literally! Subscribe to Linux Format today at a bargain price. You can get print issues, digital editions or why not both? We deliver to your door worldwide for a simple yearly fee. So make your life better and easier, subscribe now!
Here is the original post:
Kubernetes - taming the cloud - TechRadar
State of the Cloud, November 2019 – Cloudwards
Hello and welcome to this latest State of the Cloud, our monthly column where we go over the biggest stories from the cloud and tech industries. October was an eventful month, but November moved at an almost breakneck speed, so strap in as we review all the goings on in our niche.
Subscribe to our monthly newsletter for updates on reviews, articles and investigations.
Of course, our usual suspects of Facebook and Libra, its supremely dodgy currency, will make an appearance, but were not just throwing our usual stones from our glass house. Also checking in are the streaming wars and some shorter stories. First, though, well be talking about a pretty serious security breach affecting one of our favorite VPN providers, NordVPN.
Early in the month, news came to light that a NordVPN server had been breached back in March 2018. The attack was possible due to an error made by the data center operator, which the VPN is no longer in business with, but no records were exposed.
All in all, its not the biggest deal, really, considering cybercrime can hit anybody and at any time, but it does raise the issue of transparency.
After all, why did we not find out till October 2019, almost 18 months later, that this attack had taken place? Anybody grounded in the realities of online business knows attacks happen and occasionally the people behind them will gain access, but as long as you encrypt all your data and, like in the case of NordVPN, dont keep user logs, theres no harm done, usually.
Though the service did quickly come out with a statement answering these questions, as well as an explanation of what exactly happened, we have decided to ding the service a little in our NordVPN review, if only just a few points. That said, we still have faith in NordVPN, though this episode does remind us all that nothing is completely safe online.
On a less serious note, the streaming wars heated up ahead of the launch of Disney+ in the U.S. on November 12, with launches all over the world following soon after. If youd like to know more about the service or would like to figure out how to access it from anywhere, check out our guide on how to watch Disney+.
The new behemoth on the scene offering the full Star Wars and Marvel universe experiences, as well as everything Disney has ever produced is a serious threat to existing channels. Somehow this hasnt deterred anybody from doing what theyre doing, though, and in some cases even launching new competitors.
However, much as we said in the August edition of this column, all that this is likely to do is make piracy more attractive. Disney has effectively stolen away a huge chunk of Netflixs library and is setting up a killer deal with Hulu that will package it with Disney+, but Netflix still has enough to offer viewers, so its unlikely people will give it up.
Having subscriptions to both Netflix and Disney+ will cost you roughly $20 per month, which should be doable for most. However, if you add any other entertainment subscriptions or start thinking in annual terms, then many consumers might decide its all costing too much.
While these large corporations are battling over market share, they might very well find the market is shrinking simply because people are busy torrenting.
Mark Zuckerberg became a meme after his testimony on Capitol Hill in May 2018, thanks to his robotic performance and evasive answers, not to mention his excessive drinking of water. Since then, the boy wonder hasnt gotten much in the way of media training, it seems, though several U.S. politicians have sharpened up their tech knowledge.
This was evident during The Zucks hearing on the Libra, his would-be digital currency. To cut a long story short, it seems that nobody within Facebook has any real idea of how Libra would work, what the risks are to consumers (and their privacy) and what exactly the company hopes to achieve with its introduction.
As you would expect, many backers of the Zuck Buck quickly abandoned ship after the massive public outcry following the hearing.
Adding to Facebooks woes and yes, you may picture us rubbing our hands in barely suppressed glee a massive class action lawsuit filed against it over its misuse of facial recognition technology has received the go-ahead from a judge, a federal one, no less.
The price tag for this could be as high as $35 billion, plus whatever court costs, potentially stymying Zuckerbergs steamroller for the foreseeable future. We can only hope.
To finish up our regular coverage, theres yet another chapter being written in the saga of whether or not Western countries should let Huawei build (part of) nascent 5G networks.
As we talked about in our June edition, the only companies capable right now of delivering on the superfast data connections are from the Middle Kingdom, and all those come with the very real threat of spying by the Chinese intelligence services.
Thus, the EU had a risk assessment commissioned on what we can and cannot expect when its member states let Huawei work on its 5G networks. Though its a long, nuanced report, what it boils down to is that the risks are manageable, provided that the right precautions are taken. In turn, Germany has decided to open the doors to Huawei, though not completely.
However, this completely ignores the problem that any time China is given any way in, it exploits this opening mercilessly; for examples, just see whats going on in its nearby waters or the brutal suppression of Hong Kong.
The problem lies in the fact that rational Western politicians see China as a reasonable partner, which it very much is not. Letting Huawei work on telecommunications is just sowing the wind.
To see the damage that could possibly be wrought by somebody with access to countrywide computer systems, look no further than Georgia the Caucasian nation, rather than the U.S. state. A massive cyberattack pretty much laid the entire country out, causing untold damage and likely seriously freaking people out.
However, according to at least one Chinese official, its actually the West that is to blame for all these problems, thanks to the Cold War mentality that hinders mutual trust in cyberspace. We have a feeling Chinese censorship and its human rights abuses might be partially to blame for that, too, though.
In better news, Google has banned a whole mess of payday loan apps from the Play Store, to the chagrin of said loan sharks. Anything that makes one of the most predatory industries upset is a good move in our book, so all praise to Mountain View in this case.
In other Google-related news, the company has unveiled a quantum computing breakthrough that, according to people a lot smarter than we are, is pretty great, but not as huge as announced. Well be honest, all we could think of is whether or not it could run The Witcher 3 on full settings on a big screen.
With that, well leave you for this month. November is shaping up to be another exciting month, and were already looking forward to telling you all about it in December. For now, we wish you luck with the darkening days in the Northern Hemisphere and much enjoyment of the spring in the Southern.
Subscribe to our monthly newsletter for updates on reviews, articles and investigations.
Did we miss any important news from October? Or did we wildly misinterpret any facts? Let us know in the comments below. Thank you for reading.
Read the original here:
State of the Cloud, November 2019 - Cloudwards
Inspur steps up with Innovative Liquid Cooling Technology at SC19 – insideHPC
Inspur is showcasing new HPC systems with Natural Circulation Evaporative Cooling technology this week at SC19. Inspur combines high-density computing servers with natural circulation evaporative cooling technology, which is more reliable, energy-saving, and easier to deploy than other liquid cooling solutions.
Currently, HPC is in urgent need of cooling technologies with high efficiency and lower energy consumption. However, the cost, safety, deployment and maintenance challenges of cooling solutions are big concerns of users. Liquid cooling technologies, compared to traditional air cooling counterparts, boast prominent advantages in heat dissipation efficiency, energy utilization, and other aspects, and have seen rapid growth in recent years. Nevertheless, liquid cooling technologies are still subject to challenges. For example, immersion and spray-type liquid cooling technologies, despite their more efficient heat dissipation, require to have the IT components continuously observed to maintain the functionality and reliability because of direct contact with the coolant. Moreover, excessive usage of coolant pushes the limit on machine room load-bearing capacity, and raises operation and maintenance costs. Plate-type water cooling technology, though not exposed to heating elements directly, uses uninsulated water as the coolant which, once leaked, will cause lethal damage to HPC systems, giving rise to safety hazards.
Inspur collaborated with the Institute of Electrical Engineering of Chinese Academy of Sciences (IEECAS), combining Inspurs leading supercomputing servers with IEECASs natural circulation evaporative cooling technology to achieve an efficient, reliable and energy saving liquid cooling HPC system. The system is equipped with Inspurs high-density server i24 which can support 4 high-performance two-socket computing nodes in 2U and the natural circulation evaporative cooling suite developed by IEECAS. The natural circulation evaporation method requires no circulating pumps which are necessary in traditional plate-type water cooling but vulnerable and energy consuming and enables automatic control over condenser fans, eliminating manual operation for over 90% of the time. This further reduces cooling overheads and lowers the PUE values of data centers to below 1.1 for green and energy-efficient operation. The non-corrosive insulating cooling liquid protects IT devices from damage in the event of leakages, greatly improving safety. In addition, the entire cooling system is compact in size and easy to deploy and maintain, with less demanding requirements on machine rooms.
The cooling system has already been successfully deployed in a large science project and is performing as expected.
Inspur is a leading provider of data center infrastructure, cloud computing, and AI solutions, ranking among the worlds top 3 server manufacturers. Through engineering and innovation, Inspur delivers cutting-edge computing hardware design and extensive product offerings to address important technology arenas like open computing, cloud data center, AI and deep learning. Performance-optimized and purpose-built, our world-class solutions empower customers to tackle specific workloads and real-world challenges.
See our complete coverage of SC19
Check out our insideHPC Events Calendar
See the article here:
Inspur steps up with Innovative Liquid Cooling Technology at SC19 - insideHPC
How to ditch the cloud and move to do-it-yourself NAS instead – The Age
What is NAS?
NAS is essentially a bunch of hard drives that connects to your home network, powered and housed by a small computer, enabling a centralised file storage system you can access from anywhere.
Traditionally a NAS box required a bit of know-how to get running, but manufacturers have made great strides in this area to the point that almost anyone can set up a powerful network storage solution that is more capable and flexible than a cloud storage service.
Synology's DS218J is a powerful two-bay NAS box at an entry level price of around $230.
Since you're buying the NAS box and requisite hard drives outright, there's more of an upfront investment. But it works out cheaper in the long run as there are no ongoing monthly fees. Cloud storage is akin to renting a place for your data to live, whereas NAS is more like owning your own home, giving you complete control and ownership. Boxes are designed to run 24/7, but generally don't consume a lot of power.
A two-bay NAS box can be picked up from as low as $200. Filling those slots with two 1TB hard drives will set you back another $100, so in total you're looking at $300. By comparison, a Google One plan with 2TB storage will set you back $125 a year. I invested in a more expensive five bay Synology DS1019+, and filled up the hard drive slots as and when I needed more storage over time. More drive bays also give you better redundancy, as you can mirror data so you won't lose any if one or two drives fail.
Loading
Keep in mind that NAS boxes can do more than just store and share your photos. The likes of Synology and QNAP have an extensive app ecosystem that extends the functionality well beyond the bounds of traditional network storage.
I use mine as a media server so the family can easily stream movies and music stored on the NAS to any connected smartphone, tablet, PC or streaming box, in addition to serving as a PVR for recording major sporting events on free-to-air television. I also use it to drive and monitor my home security cameras.
There are a number of companies that make NAS boxes, including QNAP, Western Digital and Seagate, but Synology's DiskStation line is far and away the best in the industry when it comes to ease of use, stability and features.
For example making your NAS accessible from outside of your home network usually involves setting up port forwarding rules or other complicated network settings. But Synology's QuickConnect feature bypasses this by allowing you to assign a simple URL or ID for access.
The DS1019+ is a five-bay NAS box that supports 4K transcoding expansion bays for even more storage, at around $1000.
It's also the only NAS system that can match Google Photos in terms of smarts. Synology's Moments app, which runs on top of the company's Diskstation Manager operating system, analyses all your photos and puts them into sensible albums for you, making it much easier to find the photo you're looking for.
It uses facial recognition to group photos with similar faces, and scene recognition so you can search based on things that are in the picture.
Loading
Unlike Google Photos, Synology does all this without ever collecting any user data or sending a single photo to the public cloud.
You can also set the app to automatically upload any new photos from your phone to your NAS.
Another strong point for NAS compared to cloud storage services is speed. Cloud services are limited by your internet speeds and the bandwidth of their servers, whereas NAS utilises the speed of your local home network which is significantly faster.
Of course it's always wise to keep a backup offsite of all your important files in case there's a fire or burglary. Synology has multiple options for doing this, backing up data stored on the NAS to a public cloud service like Google Drive, OneDrive or Dropbox.
The benefit here is that Synology will encrypt your data before it is uploaded, so your data can't be trawled and won't be compromised if the cloud service is hacked or breached.
Some NAS boxes allow you to sync an encrypted backup of your storage to the cloud.
Synology also offers its own private cloud option called Synology C2 Backup, with the basic plan costing between $16 and $100 a year depending on your needs.
Another option is to invest in a secondary Synology NAS offsite and have files synchronise over the internet. I personally go the manual route; plugging a USB drive into the NAS on a monthly basis to back up my most precious data, namely my collection of family photos and videos.
Krishan is a multi-award-winning Australian technology journalist.
Here is the original post:
How to ditch the cloud and move to do-it-yourself NAS instead - The Age
What is the Cloud – Definition | Microsoft Azure
The definition for the cloud can seem murky, but essentially, its a term used to describe a global network of servers, each with a unique function. The cloud is not a physical entity, but instead is a vast network of remote servers around the globe which are hooked together and meant to operate as a single ecosystem. These servers are designed to either store and manage data, run applications, or deliver content or a service such as streaming videos, web mail, office productivity software, or social media. Instead of accessing files and data from a local or personal computer, you are accessing them online from any Internet-capable devicethe information will be available anywhere you go and anytime you need it.
Businesses use four different methods to deploy cloud resources. There is a public cloud that shares resources and offers services to the public over the Internet, a private cloud that isnt shared and offers services over a private internal network typically hosted on-premises, a hybrid cloud that shares services between public and private clouds depending on their purpose, and a community cloud that shares resources only between organizations, such as with government institutions.
What is cloud services? – Definition from WhatIs.com
The term cloud services is a broad category that encompasses the myriad IT resources provided over the internet. The expression may also be used to describe professional services that support the selection, deployment and ongoing management of various cloud-based resources.
The first sense of cloud services covers a wide range of resources that a service provider delivers to customers via the internet, which, in this context, has broadly become known as the cloud. Characteristics of cloud services include self-provisioning and elasticity; that is, customers can provision services on an on-demand basis and shut them down when no longer necessary. In addition, customers typically subscribe to cloud services, under a monthly billing arrangement, for example, rather than pay for software licenses and supporting server and network infrastructure upfront. In many transactions, this approach makes a cloud-based technology an operational expense, rather than a capital expense. From a management standpoint, cloud-based technology lets organizations access software, storage, compute and other IT infrastructure elements without the burden of maintaining and upgrading them.
The usage of cloud services has become closely associated with common cloud offerings, such as software as a service (SaaS), platform as a service (PaaS) and infrastructure as a service (IaaS).
SaaS is a software distribution model in which applications are hosted by a vendor or service provider and made available to customers over a network, typically the internet. Examples include G Suite -- formerly Google Apps -- Microsoft Office 365, Salesforce and Workday.
PaaS refers to the delivery of operating systems and associated services over the internet without downloads or installation. The approach lets customers create and deploy applications without having to invest in the underlying infrastructure. Examples include Amazon Web Services' Elastic Beanstalk, Microsoft Azure -- which refers to its PaaS offering as Cloud Services -- and Salesforce's App Cloud.
IaaS involves outsourcing the equipment used to support operations, including storage, hardware, servers and networking components, all of which are made accessible over a network. Examples include Amazon Web Services, IBM Bluemix and Microsoft Azure. SaaS, PaaS and IaaS are sometimes referred to collectively as theSPI model.
Cloud services that a service provider offers to multiple customers through the internet are referred to as public cloud services. The SaaS, PaaS and IaaS providers noted above may all be said to be providing public cloud-based services.
Private cloud services, in contrast, are not made generally available to individual or corporate users or subscribers. Private cloud-based services use technologies and approaches associated with public clouds, such as virtualization and self-service. But private cloud services run on an organization's own infrastructure and are dedicated to internal users, rather than multiple, external customers.
The second sense of cloud services involves professional services that enable customers to deploy the various types of cloud services. Consulting firms, systems integrators and other channel partners may offer such services to help their clients adopt cloud-based technology.
In this context, cloud services might include any or all of the following offerings: cloud-readiness assessment, application rationalization, migration, deployment, customization, private and public cloud integration -- hybrid clouds -- and ongoing management. Companies specializing in cloud services have become an attractive acquisition target for large IT services providers -- Accenture, IBM and Wipro, for instance -- that seek expertise in cloud consulting and deployment.
Cloud services are sometimes deemed synonymous with web services. The two fields, although related, are not identical. A web service provides a way for applications or computers to communicate with each over the World Wide Web. So, web services are generally associated with machine-to-machine communications, while cloud services are generally associated with scenarios in which individuals or corporate customers consume the service -- users accessing office productivity tools via a SaaS-based application, for example.
Some web services, however, may be closely intertwined with cloud services and their delivery to individuals and organizations. Cloud services, for instance, often use RESTful web services, which are based on representational state transfer (REST) technology. REST is viewed as providing open and well-defined interfaces for application and infrastructure services.
See also: XaaS (anything as a service)
The rest is here:
What is cloud services? - Definition from WhatIs.com
What is cloud server? – Definition from WhatIs.com
A cloud server is a hosted, and typically virtual, compute server that is accessed by users over a network. Cloud servers are intended to provide the same functions, support the same operating systems (OSes) and applications, and offer performance characteristics similar to traditional physical servers that run in a local data center. Cloud servers are often referred to as virtual servers, virtual private servers or virtual platforms.
An enterprise can choose from several types of cloud servers. Three primary models include:
Public cloud servers: The most common expression of a cloud server is a virtual machine (VM) -- or compute "instance" -- that a public cloud provider hosts on its own infrastructure, and delivers to users across the internet using a web-based interface or console. This model is broadly known as infrastructure as a service (IaaS). Common examples of cloud servers include Amazon Elastic Compute Cloud instances, Azure instances and Google Compute Engine instances.
Private cloud servers: A cloud server may also be a compute instance within an on-premises private cloud. In this case, an enterprise delivers the cloud server to internal users across a local area network, and, in some cases, also to external users across the internet. The primary difference between a hosted public cloud server and a private cloud server is that the latter exists within an organization's own infrastructure, where a public cloud server is owned and operated outside of the organization.
Dedicated cloud servers: In addition to virtual cloud servers, cloud providers can also supply physical cloud servers, also known as bare-metal servers, which essentially dedicate a cloud provider's physical server to a user. These dedicated cloud servers also called dedicated instances -- are typically used when an organization must deploy a custom virtualization layer, or mitigate the performance and security concerns that often accompany a multi-tenant cloud server.
Cloud servers are available in a wide array of compute options, with varying amounts of processors and memory resources. This enables a user to select an instance type that best fits the needs of a specific workload. For example, a smaller Amazon EC2 instance might offer one virtual CPU and 2 GB of memory, while a larger Amazon EC2 instance provides 96 virtual CPUs and 384 GB of memory. In addition, it is possible to find cloud server instances that are tailored to unique workload requirements, such as compute-optimized instances that include more processors relative to the amount of memory.
While it's common for traditional physical servers to include some storage, most public cloud servers do not include storage resources. Instead, cloud providers typically offer storage as a separate cloud service, such as Amazon Simple Storage Service and Google Cloud Storage. A user provisions and associates storage instances with cloud servers to hold content, such as VM images and application data.
The choice to use a cloud server will depend on the needs of the organization and its specific application and workload requirements. Some potential benefits and drawbacks include:
Ease of use: One of the biggest benefits of cloud servers is that a user can provision them in a matter of minutes. With a public cloud server, an organization does not need to worry about server installation, maintenance or other tasks that come with ownership of a physical server.
Globalization: Public cloud servers can "globalize" workloads. With a traditional centralized data center, users can still access workloads globally, but network latency and disruptions can reduce performance for geographically distant users. By hosting duplicate instances of a workload in different global regions, users can benefit from faster and often more reliable access.
Cost: Public cloud servers follow a pay-as-you-go pricing model. Compared to a traditional physical server, this can save an organization money, particularly for workloads that only need to run temporarily or are used infrequently. Cloud servers are often used in such temporary use cases, such as software development and testing, as well as where high scalability is important. However, depending on the amount of use, the long-term and full-time cost of cloud servers can become more expensive than owning the server outright. In addition, regulatory obligations and corporate governance standards may prohibit organizations from using cloud servers and storing data in different geographic regions.
Performance: Because cloud severs are typically multi-tenant environments, and a user has no direct control over those servers' physical location, a VM may be adversely impacted by excessive storage or network demands of other cloud servers on the same hardware. This is often referred to as the "noisy neighbor" issue. Dedicated or bare-metal cloud servers can help a user avoid this problem.
Outages and resilience: Cloud servers are subject to periodic and unpredictable service outages, usually due to a fault within the provider's environment or an unexpected network disruption. For this reason, and because a user has no control over a cloud provider's infrastructure, some organizations choose to keep mission-critical workloads within their local data center rather than the public cloud. Also, there is no inherent high availability or redundancy in public clouds. Users that require greater availability for a workload must deliberately architect that availability into the workload.
Continued here:
What is cloud server? - Definition from WhatIs.com