Connect with us

Security

Yes, there are security ramifications to serverless computing

Published

on

At least one in five organizations, 21%, have implemented serverless computing as part of their cloud-based infrastructure. That’s the finding of a recent survey of 108 IT managers conducted by Datamation. Another 39% are planning or considering serverless resources. 


Photo: Joe McKendrick

The question is, will serverless computing soon gain critical mass, used by a majority of enterprises? Along with this, what are the ramifications for security? 

Existing on-premises systems and applications — you can call some of them “legacy” — still require more traditional care and feeding. Even existing cloud-based applications are still structured around the more serverful mode of development and delivery. 

That’s what many enterprises are dealing with now — loads of traditional applications to manage even while they begin a transition to serverless mode. Again, even if applications or systems are in the cloud, that still is closer to traditional IT than serverless on the continuum, says Marc Feghali, founder and VP product management for Attivo Networks. “Traditional IT architectures use a server infrastructure, that requires managing the systems and services required for an application to function,” he says. It doesn’t matter if the servers happen to be on-premises or cloud-based. “The application must always be running, and the organization must spin up other instances of the application to handle more load which tends to be resource-intensive.”  

Serverless architecture goes much deeper than traditional cloud arrangements, which are still modeled on the serverful model. Serverless, Feghali says, is more granular, “focusing instead on having the infrastructure provided by a third party, with the organization only providing the code for the applications broken down into functions that are hosted by the third party. This allows the application to scale based on function usage. It’s more cost-effective since the third-party charges for how often the application uses the function, instead of having the application running all the time.”

How should the existing or legacy architecture be phased out? Is it an instant cut over, or should it be a more gradual migration? Feghali urges a gradual migration, paying close attention to security requirements. “There are specific use cases that will still require existing legacy architecture,” and serverless computing “is constrained by performance requirements, resource limits, and security concerns,” Feghali points out. The advantage serverless offers is that it “excels at reducing costs for compute. That being said, where feasible, one should gradually migrate over to serverless infrastructure to make sure it can handle the application requirements before phasing out the legacy infrastructure.”   

Importantly, a serverless architecture calls for looking at security in new ways, says Feghali, “With the new service or solution, security frameworks need to be evaluated to see what new gaps and risks will present themselves. They will then need to reassess their controls and processes to refine them to address these new risk models.”

Security protocols and processes differ in a serverless environment. Namely, with the use of serverless computing, an enterprise’s attack surface widens. “The attack surface is much larger as attackers can leverage every component of the application as an entry point,” Feghali says, which includes “the application layer, code, dependencies, configurations and any cloud resources their application requires to run properly. There is no OS to worry about securing, but there is no way to install endpoint or network-level detection solutions such as antivirus or [intrusion protection or prevention systems[. This lack of visibility allows attackers to remain undetected as they leverage vulnerable functions for their attacks, whether to steal data or compromise certificates, keys, and credentials to access the organization.”

At this point, introducing the security measures needed to better protect serverless environments may add more cost and overhead, according to a study out of the University of California at Berkeley, led by Eric Jonas. “Serverless computing reshuffles security responsibilities, shifting many of them from the cloud user to the cloud provider without fundamentally changing them,” their report states. “However, serverless computing must also grapple with the risks inherent in both application disaggregation multi-tenant resource sharing.”

One approach to securing serverless is “oblivious algorithms,” the UC Berkeley team continues. “The tendency to decompose serverless applications into many small functions exacerbates this security exposure. While the primary security concern is from external attackers, the network patterns can be protected from employees by adopting oblivious algorithms. Unfortunately, these tend to have high overhead.” 

Physical isolation of serverless resources and functions is another approach — but this, of course, comes with premium pricing from cloud providers. Jonas and his team also see possibilities with generating very rapid instances of serverless functions. “The challenge in providing function-level sandboxing is to maintain a short startup time without caching the execution environments in a way that shares state between repeated function invocations. One possibility would be to locally snapshot the instances so that each function can start from clean state.” 

Feghali’s firm, Attivio Networks, focuses on adoption of “deception technologies” intended to provide greater visibility across the various components in a serverless stack, “as a way to understand when security controls are not working as they should, detect attacks that have by-passed them, and for notification of policy violations by insiders, suppliers, or external threat actors.”  

The bottom line is handing over the keys of the server stack to a third-party cloud provider doesn’t mean outsourcing security as well. Security needs to remain the enterprise customer’s responsibility, because it is they who will need to answer in the event of a breach.

Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Security

Key Criteria for Evaluating Unified Endpoint Management

Published

on

Endpoint management is one of the most significant challenges in the enterprise today. An increasingly large percentage of our workforce is distributed and demands flexibility to work wherever they want, whenever they want. We must respond by giving them access to the services they require to do their jobs effectively. The alternative is that we, as a business, will suffer, lose good people, and become less competitive. However, we must achieve this essential access while maintaining security and control of our business’s data assets.

An appropriate endpoint management strategy is key to addressing these issues. Our approach should be holistic and unified, bringing together control of devices, management of applications, security of data, and access controls.

Unified endpoint management (UEM) is the approach to meeting this challenge. It has evolved from traditionally disparate solutions for endpoint management, application delivery, and security into a single platform. This single platform delivers a consistent end-user experience across all devices, applications, and locations while maintaining security and control of data assets. The leading solutions allow us to enroll devices easily into our control, provide support, and ensure constituency and compliance while managing access to our applications and data.

This GigaOM Key Criteria Report describes UEM solutions and identifies key criteria and evaluation metrics for selecting such a solution. The corresponding GigaOm Radar Report identifies vendors and products that excel in this sector. Together, these reports give decision-makers an overview of the market to help them evaluate existing platforms and decide where to invest.

How to Read this Report

This GigaOm report is one of a series of documents that helps IT organizations assess competing solutions in the context of well-defined features and criteria. For a fuller understanding consider reviewing the following reports:

Key Criteria report: A detailed market sector analysis that assesses the impact that key product features and criteria have on top-line solution characteristics—such as scalability, performance, and TCO—that drive purchase decisions.

GigaOm Radar report: A forward-looking analysis that plots the relative value and progression of vendor solutions along multiple axes based on strategy and execution. The Radar report includes a breakdown of each vendor’s offering in the sector.

Solution Profile: An in-depth vendor analysis that builds on the framework developed in the Key Criteria and Radar reports to assess a company’s engagement within a technology sector. This analysis includes forward-looking guidance around both strategy and product.

The post Key Criteria for Evaluating Unified Endpoint Management appeared first on Gigaom.

Continue Reading

Security

Data Storage for Ever Changing Business Needs

Published

on

Join GigaOm analyst Enrico Signoretti and CTERA CTO Aron Brand in this one-hour live webinar as they explore file storage trends and dynamics through the lens of IT infrastructure modernization projects.

The file and cloud experts will discuss the limitations of traditional NAS architectures in today’s corporate environments and how organizations are implementing distributed cloud file storage to solve remote collaboration, ransomware protection, and unstructured data growth challenges.

Signoretti and Brand will also examine the recently published GigaOm Radar for Distributed Cloud File Storage, in which CTERA was named the leader. They will review the report’s key criteria and evaluation metrics for choosing a distributed cloud file storage platform, helping IT leaders to understand which vendors are most aligned to their needs today as well as 12-18 months down the road.

The post Data Storage for Ever Changing Business Needs appeared first on Gigaom.

Continue Reading

Security

High Performance Application Security Testing – Cloud WAF Security Platforms

Published

on

This free 1-hour webinar from GigaOm Research features analyst Jake Dolezal and will focus on comparing Web Application Firewall (WAF) security platforms in an enterprise with high performance needs.

This webinar will discuss web application security mechanisms deployed in the cloud. The cloud enables enterprises to differentiate and innovate with microservices at a rapid pace. However, the cloud is just as vulnerable, if not more so, to attacks and breaches as on-premises APIs and apps are. Our focus is specifically on approaches to securing apps, APIs, and microservices that are tuned for high performance and availability. We define “high performance” as companies that experience workloads of more than 1,000 transactions per second (tps) and require a maximum latency below 30 milliseconds across the landscape.

In this webinar, we will reveal the performance tests of security mechanisms on NGINX, AWS, and Azure, specifically: ModSecurity, NGINX App Protect WAF, AWS Web Application Firewall (WAF), and Azure WAF.

Register now to join GigaOm and NGINX for this free expert webinar.

The post High Performance Application Security Testing – Cloud WAF Security Platforms appeared first on Gigaom.

Continue Reading

Trending