Category Archives: Cloud Servers
How these 3 Companies Leverage the Hybrid Cloud – TechGenix
The Covid-19 pandemic sparked a boom in user and operational requirements that could only be fulfilled by cloud-native solutions. To increase operating efficiency and cost-savings, businesses began looking to shift the bulk of their operations to the cloud. However, the way forward was not always immediately apparent, since the shift to hybrid cloud often needs specialized support and resources. Since each business has its own unique existing infrastructure, there was a need for tailored solutions for migrating apps and services from on-premises architecture to the hybrid cloud.
Several companies have managed to make the transition to hybrid cloud successfully, with minimal downtime or disruption in services, and their strategies can offer businesses today a way forward. Lets take a look at three companies that efficiently leverage the hybrid cloud for their business operations.
Global travel company Hotelbeds, headquartered in Spain, connects nearly 180,000 travel service providers worldwide to around 60,000 different travel sellers through their advanced online platform. As the worlds leading bedbank, they need to ensure that their services are running smoothly so that travel providers have seamless access to distribution channels and sources of revenue.
Image Source: Pixabay
Hotelbeds initially built their infrastructure by tightly coupling their legacy on-premises architecture with a single cloud provider. While this approach worked for them for a while, they found themselves facing rising costs and realized that they were missing out on the benefits of a distributed architecture. As the cloud computing landscape evolved to support more hybrid and multi-cloud strategies, Hotelbeds decided that they, too, needed to jump on the bandwagon and develop their own hybrid cloud strategy that leveraged the best offerings among cloud service providers.
Hotelbeds partnered with Rancher, an open-source multi-cluster orchestration platform that delivers Kubernetes-as-a-service (KaaS) in order to make this transition. With Rancher handling deployments and workload distribution, Hotelbeds began to run Kubernetes clusters in production in Amazon Web Services (AWS) and Google Cloud Platform (GCP). This gave them greater flexibility in managing and scaling their workloads, while also allowing them to offer more reliable and performance available services.
In addition to dodging vendor lock-in, Hotelbeds also managed to reduce cloud migration time by 90% and deployment time by 80%. Through their new hybrid cloud strategy, the company found that they were able to scale and deploy their workloads in just minutes, guaranteeing more reliable service and increasing their operating efficiency.
Similar to Hotelbeds, Easybook is in the travel industry, and is the largest online transport booking site in the Southeast Asian region, having sold over 5,000,000 tickets since it was established in 2005 in Singapore. Easybook connects travellers to ticketing services for buses, trains, and ferries as well as car rental services and local tours, while also providing Software-as-a-Service (SaaS) back-office solutions to the local transportation industry.
Easybooks IT infrastructure was originally located in the public cloud, which allowed them to quickly scale their services up to meet demand. As the company expanded its user base and grew into new markets, the public cloud approach no longer became cost effective. In order to continue its expansion while controlling costs, Easybook started looking toward a hybrid cloud solution that allowed it to continue maintaining its core business applications in Microsoft Azure.
Easybook enlisted the services of Equinix, a digital infrastructure company, and InfoFabrica, a hybrid and multi-cloud consulting organization, to move their infrastructure from the public cloud to a hybrid cloud architecture. They managed to accomplish this in less than three months with zero disruptions to service or downtime. InfoFabricas network and database engineers helped Easybook set up their new servers at Equinixs IBX (International Business Exchange) colocation data centers and install their apps, databases, and web portal software, while also offering facilities management services. This freed up Easybook to focus its staff and resources on developing their business strategy.
Since their transition to the hybrid cloud, Easybook has managed to cut costs by 30 to 40 percent and leverage the scalability and security of direct access to Azure cloud from within the colocation data center. This keeps their data secure while ensuring high application performance. Equinix also enables Easybooks access to the offerings of several cloud service providers through its software-defined interconnection platform, Equinix Cloud Exchange (ECX) Fabric.
BP is a British multinational oil and gas giant with operations in nearly 80 countries around the world, employing over 70,000 people. BP also has a network of mega data centers around the globe, which they began to rethink as global cloud service providers entered the market and revolutionized the way companies ran their operations. Realizing that they would never be able to match up to the scale, the scope, the resilience, and the flexibility of public cloud platforms, BP decided to transition out of data center management and shift their IT architecture to the cloud, specifically to Microsoft Azure.
Image Source: Pixabay
The first resource that BP shifted to Azure was a Microsoft SharePoint Server deployment. This experiment made them realize that rather than just transferring on-premises behaviors to the cloud, they also needed to adopt a cloud-native approach in order to reap the full benefits of a cloud-centric infrastructure. As BP shifted to a cloud-first strategy, they realized that they were able to free up resources to focus on innovation at the cutting-edge to deliver business value. This was because they were no longer investing enormous sums of capital in developing the infrastructure to support a new project before starting the project.
The reason why BP hasnt migrated completely to the public cloud is because stringent data protection laws in certain countries require them to maintain some resources on-premises in their data centers. The hybrid cloud approach then meets compliance requirements while ensuring flexibility, scalability, and geographic coverage. BPs hybrid cloud strategy appears to be proceeding according to plan, and it may soon accomplish its goal of decommissioning its mega data centers.
BP has leveraged several of Azures offerings to support their move to hybrid cloud. They used Azure Stack to extend cloud functionality to remote regions, networking service Azure ExpressRoute to connect on-premises data centers to Azure, Azure Active Directory for identity and access management, Azure Security Center to manage their resources in Azure, and Azure Front Door Service to quickly deliver globally distributed application through a single, highly secure entry point.
The case studies explored in this article highlight an important point about adopting a hybrid cloud strategy: no two companies will share an identical hybrid cloud strategy. Whether or not you need a hybrid cloud strategy and the kind of approach you need to take depends on your business strategy, your current IT infrastructure, the apps and services your company uses, your user base, and the direction your company is heading in.
And this is not a transition you will need to make alone. As illustrated in this article, there are several tools, platforms, and organizations that can make the shift to hybrid cloud a seamless and productive experience, while ensuring that your customers are provided consistent and reliable services.
You may employ a KaaS platform as in the case of Hotelbeds, outsource your network and database engineering needs to a digital infrastructure and/or hybrid cloud company like Easybook did, or employ your own teams and resources to make the most of public cloud offerings as BP did.
Make sure to conduct a thorough analysis of your business needs and assess the offerings in the market before you make the switch over from public cloud or on-premises architecture to a hybrid cloud setup. If not, you may miss out on the flexibility, scalability, reliability, and high performance that an effective hybrid cloud strategy can offer you.
Featured Image Source: Piqsels
Post Views:47
Read more here:
How these 3 Companies Leverage the Hybrid Cloud - TechGenix
Multi-cloud security doesn’t have to be complicated, just consistent – IT-Online
As organisations in every industry shift infrastructure and services to the cloud by means of a multi-cloud strategy, their business assets, software and applications become distributed across several cloud-hosting environments.
By Kumar Vaibhav, solution architect at In2IT
Despite the many business benefits including agility, flexibility, competitive pricing, scalability, and reliance, to list a few there are several hurdles that must be addressed when adopting cloud across the business. It can be particularly tricky securing a plethora of clouds due to a lack of visibility across services and providers.
With multiple clouds comes multiple layers of risk, such as an increased attack surface, improper user management, constantly shifting workloads, DevOps and automation, all of which can get complicated.
Multiple cloud benefits
However, cloud security shouldnt be as complicated as it has become. Despite cloud having been around for more than a decade, there is still this perception that it is new technology, which makes people uncomfortable.
Cloud is many things, including scalable, reliable and cost-effective, but its no longer new. While on-premise security and own data centres is what most organisations think they need to secure their digital assets, the reality is that this is no longer sustainable its time-consuming and cost-intensive to operate and manage, particularly in comparison to the cloud.
Security must meet in the middle
So how does cloud security compare to on-premise security? Essentially, there isnt that much difference. Its easy to think that on-premise is more secure because one has direct control over all the servers, systems and data living in that data centre.
However, its important to remember when moving to the cloud that all cloud service providers, like Microsoft, Bing and Amazon all have their own security measures in place. The main concern that businesses have when it comes to moving data to the cloud is that theyre uncertain where it will live, but realistically, its possible to have the same controls in the cloud as with on-premise security.
The two go hand-in-hand and security in the cloud is a responsibility that must be shared between the cloud service provider and the customer, depending on the service theyre using.
The service provider has to ensure (in line with the SLA) that customer data is safe in their cloud, while the customer has to ensure everything in their cloud up to the point where it onramps to the service provider is secured and that their users are properly managed.
Users are the weakest link in security
Proper user management is particularly important now that the workforce is split between working at home, in the office or out in the field as 80% to 90% of all cyber breaches or attacks happen because of users.
Whether it is users being tricked into giving out credentials, or credentials being compromised by exploiting vulnerabilities, the effect is the same, making it critical to implement and utilise Multi-Factor Authentication (MFA) as part of a stringent Identity Management Program.
Password sniffing or spoofing is easy, and there are thousands of ways that attackers can gain unauthorised access to data, but having MFA drastically reduces the chances of getting defrauded from the inside. In addition to MFA, its necessary to have a proper access control program in place. Role-based access is one of the most important keys to preventing data leaks.
Here, its important that not everyone gets the same level of access, and specific users must be granted only the permissions necessary to fulfil their job description.
Countering the DevOps risk
Securing web-based applications to ensure theyre not used as attack vectors is as simple as proper testing. One of the main problems with the DevOps approach thats becoming increasingly popular because of the agility it enables is that the fast pace of work can lead to an increase in coding mistakes, which can result in undetected bugs and errors.
Attackers can exploit these coding mistakes to gain access to digital assets. To counter this risk, it is necessary to pay more attention to thorough vulnerability testing on the web app continuously while following best practices for maps. Although penetration testing can be expensive, this cost needs to be evaluated against the real possibility that a single breach can cause untold damage, both reputational and financial.
Protecting against network threats and vulnerabilities in the cloud isnt much different to securing web apps, and its important to ensure that all applications and operating systems are up to date in terms of security patches, along with proper access control through a firewall and a secure perimeter.
Access must be on a needs basis only, and when vulnerabilities are detected, these must be addressed as soon as possible. In the case of virtual machines, its important to have the appropriate security controls and to pay particular attention to endpoint hygiene. Theres no point in having antivirus protection, or a firewall if its incorrectly configured, malfunctioning or not reporting properly.
Visibility through simplification
Secure Access Service Edge (SASE), as defined by Gartner, can make a difference here.
SASE is a security framework specifying that security and network connectivity technologies should come together in a single cloud-delivered platform to enable rapid, secure cloud transformation. In addition to providing a singular point through which services are delivered to the client, this also streamlines network access and security measures, while eliminating operational complexity by reducing the number of vendors involved and helping to protect the business from third-party vulnerability.
This plays a massive role in achieving visibility and transparency in cloud environments, along with the fact that public cloud providers generally have their own compliance requirements to meet such as ISO 20 001, PCI, DSS and HIPAA all of which can be passed onto the customer.
Secure the data wherever it goes
Ultimately, the most effective approach to securing anything in the cloud will be one that focuses on securing data both in transit and in motion.
Asset protection is important, and visibility is critical given the scalability and flexibility of the cloud. Endpoint protection is required to secure servers or workstations or any machine in the cloud, along with operational security which ensures that when any changes are made, these occur without accidentally opening system loopholes.
Monitoring is just as vital, along with vulnerability and penetration testing. Finally, to ensure security and continuity, businesses should avoid putting all their eggs into a single cloud basket. Using multiple clouds ensures that if one goes down, theres another ready to take its place and ensure security through business continuity.
Related
Read more:
Multi-cloud security doesn't have to be complicated, just consistent - IT-Online
Podcast: why the future of data management sits in the cloud – Central Banking
It is no longeracceptablefor central banks to rely on daily batch processing within their data management frameworks,says Henrik Crone, deputy chief executive of Skysparc.
[Central banks] need to have a data lake, which constantly streams updates to their data analytics platforms, Crone tells Central Banking.
One way to achieve real-time data updates is to migrate from on-premises data centres to cloud-hosted centres.
Historically, central banks have used on-premises data centres which means their IT infrastructure is onsite. These centres include everything from servers that support email to the network hardware connecting them to support infrastructure.
Over the years, however, a larger proportion of central banks are switching to the cloud. Within a cloud data centre, a central bank leases infrastructure from a third party like Amazon Web Service or Microsoft and accesses data centre resources over the internet.
Adopting cloud, Crone says during the latest CB on Air Partners in Focus episode, not only allows central banks access to data from wherever they are in the world, but also allows them to scale up their data platforms.
You can programmatically increase the power of their querying. They can find their business data insights within data platforms using the power of the cloud the storage is also endless, he says.
Having infinite data storage allows central banks to aggregate historical data with real-time data warehousing.
Central banks would also reduce their data maintenance costs should they make the move to a cloud-hosted data network.
When you have traditional frameworks, you have to ensure there is an administrator and partitioning tables, all this technical mumbo jumbo regarding the maintenance, says Crone.
Data partitioning occurs to allow data to be distributed across sites and accessed seperately in order to improve query performance and increase the manageability of an on-premises database. In the cloud, there is zero maintenance, Crone says.
Index
00:00 Introductions
00:50 From data warehouses to the cloud
03:10 Cloud technology post-pandemic
08:44 Overcoming analytics challenges
11:27 Future data trends
Link:
Podcast: why the future of data management sits in the cloud - Central Banking
Streaming Analytics Market worth $50.1 billion by 2026 – Exclusive Report by MarketsandMarkets – Yahoo Finance
CHICAGO, Jan. 12, 2022 /PRNewswire/ -- According to a research report "Streaming Analytics Market with COVID-19 Impact Analysis, by Component, Application (Supply Chain Management, Sales & Marketing, and Fraud Detection), Industry Vertical, Deployment Mode, Organization Size, and Region - Global Forecast to 2026", published by MarketsandMarkets, the streaming analytics size is projected to grow from USD 15.4 billion in 2021 to 50.1 USD billion in 2026, at a Compound Annual Growth Rate (CAGR) of 26.5% during the forecast period.
MarketsandMarkets Logo
The Streaming analytics industry is driven increased digitalization and emerging technologies such as big data, IoT, and AI to drive the market growth. However, Strategic shift toward real-time accurate forecasts and rising data connectivity through hybrid and multi-cloud environments further contributes to the growth of the Streaming Analytics Market.
Browse in-depth TOC on "Streaming Analytics Market"
348 Tables 61 Figures 342 Pages
Download PDF Brochure: https://www.marketsandmarkets.com/pdfdownloadNew.asp?id=64196229
Based on the Component, software segment to account for a larger market size during the forecast period
The Streaming Analytics Market has been segmented by two components: software and services. The deployment of streaming analytics has witnessed an increase in adoption, as serves a variety of purposes, such as fraud detection and risk management. The growing adoption of streaming analytics across all major verticals, such as BFSI, Telecommunication and IT, Retail and eCommerce, Healthcare and Life Sciences, Manufacturing, Government, Energy and Utilities, Transportation and Logistics, Media and Entertainment, Other Verticals (travel & hospitality and education).
Based on deployment mode, cloud segment to grow at a higher CAGR during the forecast period
Cloud computing refers to the storage, management, and processing of data via networks of remote servers, which are typically accessed via the Internet. According to Statista, cloud computing would generate more than USD 300 billion in revenue in 2020 as a component of IT services. At the same time, PwC shows the COVID-19 crisis has accelerated the cloud transition even further as per data during the first quarter of 2020, cloud spending increased by 37% to USD 29 billion. The increasing generation of data leads to various challenges for several organizations. These challenges include storage, privacy, and affordability.
Story continues
Based on application, Location Intelligence segment to hold the largest market size during the forecast period
The Streaming Analytics Market based on application is segmented into Fraud Detection, Sales and Marketing, Predictive Asset Management, Risk Management, Network Management and Optimization, Location Intelligence, Supply Chain Management, Other Applications (product innovation and customer management). Streaming analytics combines geospatial, graph, and business analytics into a single platform purpose-built for performance and scale. GIS creates maps that can be accessed through a mobile app or software service. These maps incorporate imagery, coordinates, and spreadsheets, among other data layers that use spatial location. Streaming analytics can be used to examine the data, which can subsequently be released via an app or other user access point. Location intelligence (LI), which could be defined as the successor to GIS, is being driven in part by advances in streaming analytics. It is a significant advancement over GIS since it supports real-time data streams and large datasets, as well as new methods for evaluating stream data.
Request Sample Pages: https://www.marketsandmarkets.com/requestsampleNew.asp?id=64196229
Based on cloud type, Hybrid Cloud segment to grow at a higher CAGR during the forecast period
A hybrid cloud is a cloud computing environment that is a mix of both public cloud and private cloud. It helps organizations enhance their data centers by deploying data on a multi-cloud platform. Various benefits, such as agility, scalability, and cost optimization features, are boosting the adoption of hybrid cloud analytics solutions in the global cloud analytics market. Enterprises are adopting hybrid cloud as it helps them overcome complexities related to the traditional IT environments.
Based on vertical, the Energy and Utilities segment is expected to grow at a higher CAGR during the forecast period
Streaming analytics are gaining acceptance among all verticals to improve profitability and reduce overall costs. The major verticals adopting streaming analytics software are BFSI, Telecommunication and IT, Retail and eCommerce, Healthcare and Life Sciences, Manufacturing, Government, Energy and Utilities, Transportation and Logistics, Media and Entertainment, Other Verticals (travel & hospitality and education). Energy and utilities, by vertical segment, is expected to grow at a higher CAGR during the forecast period
Based on organization size, SMEs segment to grow at the highest CAGR during the forecast period
The SMEs are organizations with an employee strength of less than 1,000. SMEs are also implementing streaming analytics software that enable better orientation in the complex building providing streaming-based analytics and tracking functionalities. SMEs are also implementing streaming analytics, which enables companies to adopt the analytics of data streaming and sensor data from grids to provide real-time insights on operational performance. The adoption of new technologies tailored to streaming analytics environments has helped companies identify the broad risk areas under various functional units, such as supply chain operations for product delivery.
Speak to Our Analyst: https://www.marketsandmarkets.com/speaktoanalystNew.asp?id=64196229
North America to hold the largest market size during the forecast period
North America is estimated to account for the largest market share during the forecast period. In North America, sales and marketing and location intelligence are considered highly effective by most organizations and verticals. On the other hand, Europe is gradually incorporating these advanced solutions within its enterprises. APAC is witnessing a substantial rise in the adoption of streaming analytics owing to the increasing digitalization and rising demand for centrally managed systems.
Major Streaming Analytics Market vendors include IBM (US), Google (US), Oracle (US), Microsoft (US), SAS (US), SAP (Germany), Cloudera (US), Teradata (US), TIBCO (US), AWS (US), Software AG (Germany), Informatica (US), Impetus (US), HPE (US), Intel (US), Iguazio (Israel), Conviva (US), Axonize (Israel), Adobe (US), Altair (US), Mphasis (India), Striim (US), INTECO (Canada), WSO2 (US), SQLstream (US), EsperTech (US), Materialize (US), StarTree(US), Crosser (Sweden), Quix (UK), Lenses (UK), BangDB (India), Imply (US), Coralogix (US), Ververica (US), StreamSets (US). These market players have adopted various growth strategies, such as partnerships, collaborations, and new product launches, to expand have been the most adopted strategies by major players from 2019 to 2021, which helped companies innovate their offerings and broaden their customer base.
Browse Adjacent Markets: Analytics Market Research Reports & Consulting
Related Reports:
Video Analytics Market with COVID-19 Impact, by Component, Application (Intrusion Management, Incident Detection, People/Crowd Counting, Traffic Monitoring), Deployment Model (On-premises and Cloud), Type, Vertical, and Region - Global Forecast to 2026
Big Data Market by Component, Deployment Mode, Organization Size, Business Function (Operations, Finance, and Marketing and Sales), Industry Vertical (BFSI, Manufacturing, and Healthcare and Life Sciences), and Region - Global Forecast to 2025
About MarketsandMarkets
MarketsandMarkets provides quantified B2B research on 30,000 high growth niche opportunities/threats which will impact 70% to 80% of worldwide companies' revenues. Currently servicing 7500 customers worldwide including 80% of global Fortune 1000 companies as clients. Almost 75,000 top officers across eight industries worldwide approach MarketsandMarkets for their painpoints around revenues decisions.
Our 850 fulltime analyst and SMEs at MarketsandMarkets are tracking global high growth markets following the "Growth Engagement Model GEM". The GEM aims at proactive collaboration with the clients to identify new opportunities, identify most important customers, write "Attack, avoid and defend" strategies, identify sources of incremental revenues for both the company and its competitors. MarketsandMarkets now coming up with 1,500 MicroQuadrants (Positioning top players across leaders, emerging companies, innovators, strategic players) annually in high growth emerging segments. MarketsandMarkets is determined to benefit more than 10,000 companies this year for their revenue planning and help them take their innovations/disruptions early to the market by providing them research ahead of the curve.
MarketsandMarkets's flagship competitive intelligence and market research platform, "Knowledge Store" connects over 200,000 markets and entire value chains for deeper understanding of the unmet insights along with market sizing and forecasts of niche markets.
Contact:
Mr. Aashish MehraMarketsandMarkets INC.630 Dundee RoadSuite 430Northbrook, IL 60062USA: +1-888-600-6441 Email: sales@marketsandmarkets.com Research Insight: https://www.marketsandmarkets.com/ResearchInsight/streaming-analytics-market.asp Visit Our Website: https://www.marketsandmarkets.com Content Source: https://www.marketsandmarkets.com/PressReleases/streaming-analytics.asp
Cision
View original content:https://www.prnewswire.com/news-releases/streaming-analytics-market-worth-50-1-billion-by-2026--exclusive-report-by-marketsandmarkets-301459292.html
SOURCE MarketsandMarkets
Go here to read the rest:
Streaming Analytics Market worth $50.1 billion by 2026 - Exclusive Report by MarketsandMarkets - Yahoo Finance
Growing Technical Advancements in DevOps Technologies and Their Rising Demand for Optimizing Business Operations to Drive the Global DevOps Market by…
The DevOps market is projected to flourish immensely by 2027 due to growing technical advancements in DevOps technologies and their increasing need in business optimization. The cloud deployment sub-segment is expected to be the most lucrative. Market in the North America region is anticipated to witness better growth opportunities.
New York, USA, Jan. 12, 2022 (GLOBE NEWSWIRE) -- According to the report published by Research Dive, the global DevOps market is estimated to garner a revenue of $23,362.8 million by 2027 and grow at a healthy CAGR of 22.9% over the forecast period from 2020-2027. The comprehensive report offers a concise outline of the DevOps markets present framework including chief facets of the market such as growth factors, hindrances, restraints and several opportunities during the estimated period of 2020-2027. The report also provides all the market figures to help new entrants analyze the market easily.
BIGGEST DISCOUNT EVER (Offer Limited Period Only)
End of Season Discount: Flat 20% OFF Read-Only Access Starting at $2999 Individual User (Single User) at $4560 Multi-User Access at $8700 TO $6960 Business User Access at $10700 TO $8560
Download FREE Sample Report of the Global DevOps Market: https://www.researchdive.com/download-sample/2801
Dynamics of the DevOps Market:
Drivers: Increasing demand for advanced technologies to run business operations smoothly is the main factor expected to drive the growth of the global DevOps market during the forecast timeframe. In addition, growing need for fast and constant application delivery systems among businesses is projected to further boost the market growth by 2027.
Opportunities: Continuous involvement of innovative technologies like Artificial Intelligence (AI) and Machine Learning (ML) to offer reliable DevOps platforms and solutions is estimated to offer ample growth opportunities for the DevOps market by 2027.
Restraints: Exorbitant costs of advanced DevOps technologies is the main factor predicted to restraint the market growth during the forecast period.
Story continues
Covid-19 Impact on the DevOps Market:
The onset of the Covid-19 pandemic has disrupted several businesses and markets, however, the global DevOps market remained unaffected and witnessed immense growth. Strict lockdowns and mobility restrictions imposed by governments worldwide led to growing demand for implementing advanced cloud systems and platforms to enhance business growth. In addition, enterprises began switching to digital transformation services to restart their businesses and emphasize on secured IT infrastructures. These factors boosted the global DevOps market growth during the analysis timeframe.
Check out How COVID-19 impacts the Global DevOps Market: https://www.researchdive.com/connect-to-analyst/2801
Segments of the DevOps Market:
The report has fragmented the DevOps market into a few segments based on solution, deployment, end-user, and region.
Solution: Monitoring and Performance Management Sub-segment to be Most Dominant
By solution, the monitoring and performance management solution sub-segment is expected to hold a dominant market share and register a revenue of $6,410.3 million by 2027. This dominance is due to wide utilization of DevOps tools for full stack operations and infrastructure monitoring and performance management. This includes databases, cloud networks, app and web servers, and others. In addition, continuous monitoring of customer behavior is essential for gaining timely insights on response times for client satisfaction. These factors are estimated to drive the growth of the sub-segment during the analysis timeframe.
Check out all Information and communication technology & media Industry Reports: https://www.researchdive.com/information-and-communication-technology-and-media
Deployment: Cloud Deployment to be Most Lucrative
By deployment, the cloud type sub-segment accounted for $2,944.2 million in 2019 and is projected to witness significant growth rate throughout the markets analysis period. Advantages of DevOps cloud-based platforms like easy access to files at any time, less costs for deployment, lower testing and operation costs, etc. are predicted to boost the growth of the DevOps markets sub-segment by 2027. In addition, rising demand for software automations among enterprises is estimated to further propel the sub-segments growth during the forecast period.
End-user: Small and Medium Enterprises Sub-segment to be Most Profitable
By end-user, the small and medium enterprises sub-segment of the global DevOps market is estimated to have a significant market size and grow at a stable rate throughout the analysis years. Wide adoption of DevOps solutions for software optimization and development services to reinforce business is expected to boost the sub-segments growth by 2027. Moreover, benefits of DevOps like faster testing, designing, etc. is expected to further accelerate the sub-segments growth during the forecast timeframe.
Access Varied Market Reports Bearing Extensive Analysis of the Market Situation, Updated With The Impact of COVID-19: https://www.researchdive.com/covid-19-insights
Region: North America Region to Witness Better Growth Opportunities
By region, the market in the North America region is anticipated to have the highest market share due to the existence of technically advanced economies like the US and Canada. Moreover, the US is a significant hub for emerging technologies that encourage enterprises to adopt DevOps platforms and services. These factors are predicted to boost the market growth in the North America region by 2027.
Key Players of the DevOps Market
1. Alphabet2. Hewlett Packard Enterprise Development, LP3. IBM4. Amazon Web Services5. Broadcom6. Microsoft7. Cigniti8. Oracle9. Alibaba Group Holding Limited10. Micro Focus
These players are working on building strategies such as product enhancement, merger and acquisition, partnerships and collaborations to assist the market development.
For instance, in December 2021, Stacklet Inc., a commercial cloud governance platform, announced its decision to collaborate in compliance-as-a-code platform that automatically groups related notifications to later route them to correct stakeholders and integrate combine with the existing workflows. This will help businesses comply with a wide array of mandates by adopting DevOps solutions.
The report also sums up many crucial facets of the DevOps market including financial performance of the key players, SWOT analysis, product portfolio, and newest strategic developments. Click Here to Get Absolute Top Companies Development Strategies Summary Report.
TRENDING REPORTS WITH COVID-19 IMPACT ANALYSIS
Cognitive Cloud Computing Market: https://www.researchdive.com/2800/cognitive-cloud-computing-market
Quantum Computing Market: https://www.researchdive.com/8332/quantum-computing-market
Communication Intelligence (COMINT) Market: https://www.researchdive.com/8420/communication-intelligence-comint-market
View original post here:
Growing Technical Advancements in DevOps Technologies and Their Rising Demand for Optimizing Business Operations to Drive the Global DevOps Market by...
Data Center Market to Grow by USD 519.34 Bn | Adoption of Multi-cloud and Network Upgrades to Support 5G will Drive Growth | Technavio – PRNewswire
Major Five Data Center Companies:
Get lifetime access to our Technavio Insights. Subscribe now to our most popular "Lite Plan" billed annually at USD 3000. View 3 reports monthly and Download 3 Reports Annually!
Data Center Market Component Outlook (Revenue, USD bn, 2020-2025)
The data center market share growth by the IT infrastructure segment will be significant during the forecast period. Datacenter information technology (IT) infrastructure comprises all IT equipment, infrastructure, and solutions required to set up or scale a data center. The use of such IT infrastructure is rising due to the increase in demand for computing power and storage to support the growth in global data traffic. Enterprises globally have embraced cloud technologies and are moving their data from on-premises data centers to cloud-based data centers. This trend is expected to continue during the forecast period to create the demand for servers and storage infrastructure, and other IT equipment.
Data Center Market Geography Outlook (Revenue, USD bn, 2020-2025)
The data center market is expected to be dominated by North America during the forecast period. The US is the key country for the data centers market in the region. Market growth in this region will be slower than the growth of the market in Europe, APAC, and South America. The increasing adoption of cloud services by enterprises across industries in the US and Canada will drive the data center market growth in North America during the forecast period.
Grab an Exclusive Sample Reportfor more information about the contribution of each segment of the data center market
Related Reports:
Data Center Market Scope
Report Coverage
Details
Page number
120
Base year
2020
Forecast period
2021-2025
Growth momentum & CAGR
Accelerate at a CAGR of 21%
Market growth 2021-2025
USD 519.34 billion
Market structure
Fragmented
YoY growth (%)
18.30
Regional analysis
North America, APAC, Europe, South America, and MEA
Performing market contribution
North America at 35%
Key consumer countries
US, China, UK, Australia, Japan, and Germany
Competitive landscape
Leading companies, Competitive strategies, Consumer engagement scope
Companies profiled
Alphabet Inc., Amazon.com Inc., Cisco Systems Inc., Equinix Inc., Microsoft Corp., NTT DATA Corp., Oracle Corp., SAP SE, Huawei Investment & Holding Co. Ltd., and International Business Machines Corp.
Market Dynamics
Parent market analysis, Market growth inducers and obstacles, Fast-growing and slow-growing segment analysis, COVID-19 impact and future consumer dynamics, Market condition analysis for the forecast period
Customization purview
If our report has not included the data that you are looking for, you can reach out to our analysts and get segments customized.
About Technavio
Technavio is a leading global technology research and advisory company. Their research and analysis focuson emerging market trends and provide actionable insights to help businesses identify market opportunities and develop effective strategies to optimize their market positions.
With over 500 specialized analysts, Technavio's report library consists of more than 17,000 reports and counting, covering 800 technologies, spanning across 50 countries. Their client base consists of enterprises of all sizes, including more than 100 Fortune 500 companies. This growing client base relies on Technavio's comprehensive coverage, extensive research, and actionable market insights to identify opportunities in existing and potential markets and assess their competitive positions within changing market scenarios.
Contacts
Technavio ResearchJesse MaidaMedia & Marketing ExecutiveUS: +1 844 364 1100UK: +44 203 893 3200Email:[emailprotected]Website:www.technavio.com/
SOURCE Technavio
Go here to read the rest:
Data Center Market to Grow by USD 519.34 Bn | Adoption of Multi-cloud and Network Upgrades to Support 5G will Drive Growth | Technavio - PRNewswire
From 1920s to 2020s: Get ready for a new Roaring Twenties – Big Think
On New Years Eve of 2019, revelers gathered around the globe to ring in a new decade. Many jubilantly attended Roaring Twenties parties, adorned in elegant evening wear, cloche and Panama hats, and knickerbockers, harkening back to an exciting, culturally vibrant era of economic prosperity. But whatever veiled hopes partygoers had for a booming future soon met jarring realities: a once-in-a-century pandemic, global lockdowns, an economic recession, and widespread civil unrest stemming from an incident of police brutality. The Roaring 2020s were not to be, it seemed.
Take heart:Mark P. Mills, a physicist, senior fellow at the Manhattan Institute, faculty fellow at Northwestern University, and a partner in Montrose Lane, an energy-tech venture fund, is out to rekindle our collectively dashed hopes. In his new book,The Cloud Revolution: How the Convergence of New Technologies Will Unleash the Next Economic Boom and a Roaring 2020s, Mills convincingly argues with verve, vitality, and most importantly evidence, that humanity is about to take a great step forward in the coming decade. And unlike the first Roaring Twenties, these wont need to end with a Great Depression.
In the opening pages, Mills reminds us that the original Roaring Twenties didnt start off so auspiciously, either. In fact, separated by a century, our situation seems eerily similar. The 1918 flu pandemic ran well into 1920, triggering a severe U.S. recession that lasted through summer 1921. Violent riots and political instability were also prevalent. Yet from this pit of public despair, Americans pulled themselves out. Propelled byremarkable advancementsin mass production, medicine, electrification, communications via telephone and radio, movies, automobiles, and aviation, the United States saw its GDPriseby an astounding 43% between 1921 and 1929.
Were now in store for a similar wave of growth, Mills says. What comes next will likely be more consequential than the comparable technological flourishing that began in the 1920s. We will again see a boost to the economys productivity, which always increases overall wealth. The rising tide does lift all boats. The future will repeat a central pattern of the past. The 25 percent in the near future will live like the 5 percent today, and the future 5 percent will live like todays 1 percent, and so on.
The key driver of this collective boon will be the Cloud, Mills says, along with the knowledge and technologies it spawns. The most basicdefinitionof this nebulous term is software and services run on computer servers in data centers accessed via the Internet. But the Cloud is much more than just Netflix, Google Drive, and Apple iCloud. It democratizes technology as never before, connecting everyone and everything, allowing unprecedented gathering and splicing of information. Cloud data centers the new tech eras digital cathedrals, as Mills calls them now occupy vastly more space than skyscrapers, formerly the hubs for business and innovation. The most powerful Cloud supercomputers are now 3 million times more powerful than the top machine of 1990, and they already fuel the discovery of new materials and medicines, as well as powering machine-centered manufacturing.
The Cloud, Mills argues, brings together the three foundational spheres for technological revolution: the means for gathering and propagating information, the means (machines) of production, and the class of materials available to do everything.
Mills offers copious examples of recent advancements in these three spheres. He lists numerous discoveries currently relegated to esoteric academic journals; any of these innovations could profoundly affect our everyday lives soon without us really noticing.
Those effects take center stage in the final section of Mills book. The arenas of work, health, education, entertainment, and science will all profoundly change in the near future, he writes. Services and manufacturing will be robotified, but doomsday predictions of mass unemployment will not happen. Luddites, who fret that more-efficient machines will replace them in their jobs, have always been wrong. Efficiency doesnt eliminate jobs in the long-term; it drives job growth. The reason productivity expands wealth is that its the only means for liberating human time, the most precious commodity in the universe, Mills writes. When humans have more free time, they educate themselves, innovate, and create new industries such as, for example, entertainment industries in cloud-based gaming or the metaverse. Moreover, the Cloud will enable more people around the world to obtain more education.
At the same time, big data, stored and analyzed in the Cloud, will lead to new breakthroughs in health and science. Research that once took months or years in fields like medicine and materials science can now be done in mere hours or days with AI algorithms available via Cloud supercomputing.
We do in fact live in a time of a new normal, Mills writes. But instead of our future being one of perennial slow growth and technological stagnation, it will be just the opposite.
What could get in the way of this optimistic view? What could spiral a new Roaring Twenties down into a second Great Depression? Mills sees China and climate change as the biggest dangers, primarily because they are linked with thinking, business practices, and government policies that encourage controlling or even limiting growth and innovation.
Whether societies ultimately prosper by taking advantage of what technology offers depends on having a culture that encourages free thinking and risk taking, and it depends on a government that allows markets to operate, even when some outcomes are less than ideal, Mills writes.
The Cloud Revolutionis indeed a liberating book. Mills succeeds in infusing readers with refreshing, evidence-based optimism for the future. Heres to the new Roaring Twenties.
This article originally appeared on RealClearScience and is reprinted with permission from RealClearWire.
Read more here:
From 1920s to 2020s: Get ready for a new Roaring Twenties - Big Think
Post Pandemic: Cloud Adoption Needs to Be Accelerated – APN News
Published on January 12, 2022
Authored by Nilesh Gupta, Head of CloudFirst & Edge Services, 3i Infotech
We had a lot to learn around us from last two years. We became even more close to our family and friends, celebrated new heroes outside our work environments and of course WFH became the most used abbreviated word in English for many enterprises, the impact of Covid-19 has brought the benefits and limitations of their IT landscape into the open. Some were ready to ensure there is no impact at all, but many struggled and learnt the hard way. As companies look ahead to a post-pandemic era, many will need to consider how to meet unique needs and priorities with potentially limited resources with WFH being the new norm.
In my many discussions with various stakeholders at these companies many are still overly reliant on legacy infrastructure and business models that still needs to be addressed to ensure any kind of acceleration to cloud journey. While it is especially important, they need to develop a proactive strategy that explores the use of cloud (hybrid or multi) and how to transition away from legacy systems towards modern technologies to embrace the digital podium, it is equally important for a partner to ensure the adoption acceleration is seamless and successful in the end.
It is important to engage with a cloud lifecycle management services partner for existing and new enterprises to transform their technology to the cloud faster and without risk. Today, industry-leading automation delivers 10x acceleration with complete ease, governance, efficiency, and fine-grained control of multi-cloud environments based on customers business and IT ecosystem. Efficient transformational partners use their global delivery capabilities and strong partnerships with established doyens in the space such as Oracle (OCI), Amazon (AWS) and Microsoft (Azure) to deliver cloud computing services that help organizations enjoy substantial reduction in CapEx across hardware, licensing, and software. With the increased agility and flexibility offered by cloud computing services, your organization will be able to cut OpEx costs like cooling and power supply, and build affordable, tested business continuity and disaster recovery plans. A suitable service provider needs to provide a wide choice of engagement models shared, hybrid, SLA-based, and project-based, that help organisations to experiment and innovate.
Broadly speaking there are three dimensions to cloud adoption journey:
1. Discover the Apps & IT Infrastructure: In this phase, there is a 360O analysis of the infrastructure and application to map the dependencies and existing workload performance. We will need to ensure there is full hierarchical view of the application and its dependencies that will help ease easy migration, asset management, and any compliance needs. So, answers to some of the typical questions like below will be addressed during the deep discovery phase:
It is in this phase that as a partner with a single comprehensive cloud management platform will help build a roadmap for the readiness, prepare an automated discovery plan for the business environment, create a set of cloud transformation sprints, and then formulate the structured road map to address the subsequent sprints to comprehend what should be considered for cloud transformation.
2. Accelerate Cloud Migration: Typically, in cloud migration phase, it is known to be fraught with complications and tedious tasks. The objective is to facilitate a seamless migration of the existing on-prem workloads to hybrid or multi-cloud environments with an aim to provide uninterrupted services to end-users. Be it migration of monolithic applications into a hosted cloud environment or housing them in containers, the virtual resources need to be available in minutes as well as manage deployments across multiple cloud-providers such as public, private and hybrid at the same time.
It is in this phase, the transformational partner, with a 360O overview of the application, and the cloud management platform, identifies the required virtual infrastructure (OS, supporting software, servers, network) and get the target cloud environment ready. This simple step accelerates the migration and achieves faster time-to-market and optimised cost to migrate.
3. Continuous Optimization and Management: One major concern post cloud adoption is to accommodate for high availability of the applications round the clock. What is needed is to have a full-stack visibility across the required platforms by factoring in unforeseen demands and failure scenarios. Multi-cloud application monitoring is rendered in a detailed and effective method by gathering cloud-specific insights. Artificial Intelligence for IT operations is the critical change in basic assumptions that is required to manage these digital transformation issues.
While ensuring round the clock availability of cloud environment, it is equally important to optimize cloud costs. One needs to ensure the reporting of cloud usage at the application and services level, dynamic scaling up and down of cloud resources, and intelligent rebalancing of workloads are available on real time basis. It helps you stay ahead of the cloud expenses by providing comprehensive chargeback / reports on resource consumption and enables to cut costs by making recommendations based on the actual usage. Be it public, hybrid or multi-cloud, periodic analysis of cloud usage and associated costs needs to be conducted to prevent overspending.
It is important to note that a comprehensive single cloud management platform helps customers broadly in these three dimensions and in accelerating the cloud adoption journey seamlessly.
See the original post here:
Post Pandemic: Cloud Adoption Needs to Be Accelerated - APN News
2 Growth Stocks That Could Double Your Money in 5 Years – Motley Fool
Just because the S&P 500 index is sitting close to all-time highs doesn't mean there aren't good stock deals to be found out there. You just have to look in the right places.
One industry that is becoming increasingly important to our economy and will surely continue to breed some long-term investment winners is the semiconductor industry. The chip market is forecast to see sales rise by 8.8% to reach $601 billion in 2022, according to the Semiconductor Industry Association.The ongoing buildout of data centers and other high-performance applications is a big tailwind for leading component suppliers.
Image source: Getty Images.
That tailwind is expected to help two best-of-breed semiconductor companies potentially double in value by 2027 -- Taiwan Semiconductor Manufacturing (NYSE:TSM) and Micron Technology (NASDAQ:MU). Indeed, both stocks have attracted the attention of top investment managers lately, including Mohnish Pabrai of Pabrai Investments and Seth Klarman of Baupost Group.
Here's why these two tech stocks have room to run.
TSMC is the largest dedicated semiconductor foundry in the world.It makes the capital investment in factories and manufacturing chips so that its customers can focus on what they do best, which is product design. Three notable customers are Advanced Micro Devices, Nvidia, and Intel.
There are two reasons to consider buying and holding TSMC stock for the long term. First, it's got a great track record of delivering market-beating returns to shareholders. Over the last 15 years, the stock has returned over 1,000% to investors, which thrashes the 235% gain over the same timeframe from the S&P 500 index.It generated $10 billion in free cash flow over the last year and paid out 90% of that in dividends to shareholders.
Second, what used to be a highly cyclical semiconductor industry is beginning to see narrower peaks and troughs in demand.TSMC's revenue growth has accelerated over the last two years,and management is increasing its capital investment as it sees a period of higher structural growth.
TSMC reported third-quarter revenue growth ( in U.S. dollars) of 22.6% over the year-ago period.As CFO Jen-Chau Huang said on the Q3 earnings call, "We are witnessing a structural increase in underlying semiconductor demand underpinned by the industry megatrends of 5G-related and [high-performance computing] applications."The latter would involve the expansion in data center capacity, which doubled over the last four years.
The stock is not cheap, trading at a price-to-earnings (P/E) ratio of 33,but a more sustained demand trend in the industry should support this valuation. Analysts expect TSMC to grow earnings per share at an annualized rate of 15% over the next five years, which is consistent with industry prospects and the company's past rate of earnings growth.Assuming the stock is still trading around the same P/E, that would deliver a double to investors, not counting the extra 1.47% return from the current dividend yield.
Image source: Getty Images.
Micron is a leader in making memory and storage components for everything that matters in tech: consumer PCs, TVs, cars, cloud servers, mobile phones, and the Internet of Things, among others. The two main products Micron is known for are dynamic random-access memory (DRAM) modules and non-volatile memory (NAND) products used in solid-state drives (SSDs).
The stock fell in 2021 over near-term demand in the PC, mobile, and cloud markets, in addition to uncertainty about near-term pricing trends for DRAM products, which have a history of wild swings based on supply/demand.The stock has since rebounded after Micron reported better-than-expected earnings that negated these concerns.
However, Micron's growth over the last five years shows a business riding similar secular demand trends as TSMC. Revenue has doubled to $29 billion, and free cash flow has grown faster, reaching $3.9 billion.While temporary bouts of weak memory pricing can cause lumpy revenue results, the long-term trend has supported growing demand for Micron's products that are essential in the new data economy. Through all the swings in memory pricing, the top line has increased 253% over the last 10 years.
Micron just launched its first DDR5 memory chip, which offers 50% faster data transfer speeds. Along with that release, it also introduced its high-performance 16 GB/16GBpsGDDR6 memory for AMD's RX 6000 graphics processors designed for gaming.Demand for DDR5 is significantly exceeding supply due to non-memory component shortages, but CEO Sanjay Mehrotra expects bit shipments of this product to "grow to meaningful levels in the second half of calendar 2022."
The shares trade at a P/E of 14. That would be a fair valuation for a low-growth, cyclical company, but that description doesn't completely fit Micron's businessconsidering the long-term opportunities across data centers, 5G, cloud computing, AI servers, and a growing gaming industry. As investors recognize Micron more as a growth business, the valuation multiple should expand and potentially lead to a double over the next five years.
This article represents the opinion of the writer, who may disagree with the official recommendation position of a Motley Fool premium advisory service. Were motley! Questioning an investing thesis -- even one of our own -- helps us all think critically about investing and make decisions that help us become smarter, happier, and richer.
Follow this link:
2 Growth Stocks That Could Double Your Money in 5 Years - Motley Fool
The Future of Records & Compliance With Optimere CEO ICYMI – Government Technology
Communication channels have evolved over time, and this has put pressure on governments to remain agile with their communications strategies.
New platforms, an increase in public interest and the necessary expansion of regulatory frameworks require state and local agencies to monitor their communications closer than ever, including on social media and elsewhere online.
In this episode, the In Case You Missed It crew put a button on the latest GovTech 100 news and interviewed Ray Carey, CEO of Optimere, about how public records are changing, what his company is doing to help agencies remain in compliance and where government communications could go in the future.
The following interview was lightly edited for clarity and brevity:
A: Ive been involved in technology for 25 years, for my whole career as a banker and investor and manager. Im now the lucky CEO and janitor of Optimere. And I came to Optimere as the CEO of ArchiveSocial.
This is my third assignment in the CEO and janitor role. My last company was called Amplify, and I helped some of the worlds largest brands like McDonalds or General Mills or Estee Lauder harness the data in their customer communications. And that data was moving from traditional communications, think email and phone, over to social. So that should sound familiar to a powerhouse company.
Now [Amplify has] about 1,000 employees and continues to sort of help those global folks. But ArchiveSocial allowed me to get back to kind of a mission-driven company, which was cool.
Before that I had a company called Neo Nova and we were helping roll out rural broadband from Alaska to Key West and, you know, it felt really good to tell my mom that I was helping the rural broadband program. So Amplify was great, but its nice to get back to a company that, you know, you brag to your mom about.
Q: Optimere was recently formed out of ArchiveSocial, NextRequest and Monsido as a public records powerhouse. Can you share more about the vision for this new company?
A: Our mission here is to try to help government communicators build trust in digital communications. All of our communicators know they do their job because great communication is totally key to having a great community. If you dont have great communication, you have a hard time building that community and engaging the citizens in the way that you want.
But its really hard to do. There are so many rules and regulations that our customers government communicators are subject to, it can sometimes be paralyzing. I just want to talk about, you know, the issue of the day, the landslide that happened or whats going on with COVID, but Im paralyzed by all these regulations around data retention, or [Americans with Disabilities Act] ADA or [Freedom of Information Act] FOIA requests or data privacy laws. And so thats why we brought together these three great companies, so that our customers can focus on the message and we focus on them.
Q: How does Optimere work with government agencies right now to modernize the way they manage public records and communications?
A: Back to this idea that we help them build trust, because theres all these regulations and rules. And theres a reason those rules are there. In my mind its to drive compliance. And too often they drive liability.
Its kind of like saying we have speed limits for safety, not so that we can give people tickets. And so we believe it takes sort of a holistic approach to your communications social-web workflow. So thats ArchiveSocial-Monsido-NextRequest: social-web workflow.
And the issue in social, the biggest issue that people have, is one of record capture. So unlike other forms of communication where I send it to you the chain of custody is really clear with an email, per se, you have it the chain of custody on social media is really murky. It sits on servers, cloud servers, as weve talked about at Facebook, at Twitter, and if its gone, its gone. Oftentimes, they wont give it to you, even if youre the White House.
Weve looked in the archive and 1 in 12 records is edited or deleted within a year in the archive. And so if you dont have a real-time version of those, you are missing a fundamental record of your government communication. So we work with folks on social to make sure that they have a permanent record of the communication that they have.
On the web, theres lots of issues on the web, you have to comply with your policy. But the biggest issue that were working with folks (about) on the web is accessibility. Nobody would build a government building and in the parking lot not have parking spots for folks that have placards and need access. But we build websites that are not accessible to millions of folks all the time. And so we have to have that same mindset that what we do in the physical world, we have to be doing that in the digital world.
And if you get those two right youre creating this massive volume of digital records and you have to be able to produce those things without bankrupting your agency. And so we work with folks on automating their workflows. We work with them to handle time-intensive tasks like redaction. And then one thing that weve seen, particularly in the public records production spaces, is theres lots of, we call them oopsies, where people produce confidential information. And so we scan all these records ahead of time and give them risk scores to keep you from accidentally maybe producing something that has Social Security numbers or, you know, private citizen data. And unfortunately, that happens at times. So thats one of the great things about using technology to correct for what can be human process and human process errors.
Q: Now, weve seen a general trend that citizens are having shifting expectations, and they have greater expectations now in the public sector. No longer do they say, Oh, thats government, you know, theyre old school. They have much greater expectations on government to deliver services differently. And so those old ways of doing things are no longer accepted. How are you seeing this in the way that this is impacting agencies and how they should think about the record and communication aspects of what they do?
A: Yeah, I think in many ways government is very unique and different. But in some ways, its the same as the rest of industry. And people that want to communicate with their citizens, with their customers, with their constituents, you have to communicate in ... the place where your citizens live. And thats increasingly Facebook and Twitter and Pinterest and social media. You have to communicate, particularly in a post-COVID world, in a digital, sort of accessible way. And that means on the web.
Were seeing this massive growth in social media communications. One of our largest folks in the archive is the [Centers for Disease Control and Prevention] CDC. So just think about how much the CDC was putting out on Facebook two years ago, and whats going on right now.
If I look back three years ago, our average public agency had about four accounts. So think one agency, four accounts that might be two contributors on Facebook and Twitter. Right now, thats tripled, thats 12 accounts on average.
The number of records per agency over the last three years is up 300 percent. That archive today has around 3,500 agencies in it, and it just crossed over half a billion records. So what is happening is government is engaging its citizens where they live. Thats major.
Another big change were seeing is in the records space. Transparency has always been at the core of public records law, sunshine laws, you know, (the) Freedom of Information Act. Were seeing citizens want transparency in the record production process, not just for the records: Where does my request sit? How many days? Whos approving this?
So you need a system fundamentally to be able to not only be transparent as a government, but to be transparent in the way you are being transparent, to sort of stretch the analogy. Were certainly seeing those two big, big trends, with citizens demanding that from their public officials.
Q: I want to come back to something you mentioned earlier, and thats accessibility. How is Optimere approaching making records and communications more accessible to the country?
A: Each of our solutions can go together, but they each solve a very specific problem. And the problem on the web is you need something to monitor your sites and domain. So that is Monsido. There is more than one demand letter per day in a public agency threatening lawsuits for inaccessible websites, right? So it is a place that theres a significant risk factor with financial implications for folks.
But if you get past that money thing, it is just the right thing to do. Everybody communicates. Everybody wants to communicate in a way that is accessible to all folks. I havent met one of our customers whos trying to build websites that cant be digested by people with disabilities.
And so we are monitoring the websites across the board. If you look at your big agencies, they may have hundreds of people producing content on websites, hundreds of folks, and one misstep can really get them into trouble. So you need some automation to monitor those sites and domains. Are they complying with the policy? Are they complying with [Web Content Accessibility Guidelines]?
A little example of this is every single image needs to have whats called alt text that describes what that image is. So somebody whos visually impaired, their reader can read that website. So we scan the website for all of these potential violations, these potential risk factors. And then they can be scored, prioritized, and then dealt with.
Q: Im glad that you mentioned the theme of compliance, because thats something that continues to evolve. What are some of the macro trends that youre seeing impact record and communication compliance, that agencies need to take note of? And then second, how are you responding to these changes?
A: Theres more records. Theres more lawsuits. Theres more of these demand letters around accessibility. But theres one new threat vector that started coming up about six or seven months ago that I really wanted to draw peoples attention to. And that is the issue of social media blocking, and viewpoint discrimination.
Weve been working with folks for years to archive their social media. But we started seeing this pop up where agencies were being sued because somebody was blocking users and didnt have a good reason why they were blocking users. And recently there was one in Colorado state: $25,000 for a Senate president. A city in Rhode Island: $7,000, for blocked comments. Were seeing dozens of these every single month.
And if you looked at the policy, most every agency says we dont block anybody. I looked in the archive and 80 percent of public agencies are blocking folks, and no ones writing down why. And were starting to see records requests asking to show everyone that was blocked on Facebook. And were going to go through and say there are eight specific sort of exclusions to free speech. And you better have a reason why you blocked those folks, that is a sort of constitutionally valid reason, or youre going to have an issue with viewpoint discrimination.
So specific compliance with free speech and First Amendment violations is something really new and active that we see growing. And so we built into the system, on the back of the APIs for Facebook and Twitter, the ability to see everyone who youve blocked and then create metadata to write down the reason for this specific violation.
So when someone comes to you and says you blocked me, you can say either No, I didnt definitively, or you can say I did, because you violated our policy in this specific way. Heres everyone else we blocked. And it wasnt because they were of a political affiliation, or they said something about the mayors spouse that the mayor didnt like.
And so this is a new real threat vector for folks, because our communicators and our [public information officers] PIOs are usually doing a great job. They have people that get access to their social media that may not be as trained. They might be an elected official that goes a little haywire and puts their agency at risk.
Q: Whats on your radar for the year ahead? Whats next for Optimere?
A: So weve spent a lot of time in data privacy. We have a pretty big operation in Copenhagen, in the UK and [General Data Protection Regulation] GDPR. And data privacy, and how you handle consumer information, is very robust.
That is coming with CCNA in California. Theres laws in Vermont. Raise your hand if youve been assaulted by cookie banners popping up on websites in the last six months. That is directly related to law and regulation changes around data privacy.
Weve just introduced a product called ConsentManager, which is the first accessibility-driven cookie banner. Imagine you spend all this time in your website, making sure your website is compliant and accessible, and you get a third-party cookie banner that doesnt have that same rigor that the rest of your website has.
So I think 2022 is going to be a year of continued growth in data privacy, both in regulations and in solutions from us.
We have, in alpha, a new product called ArchiveChat. We are seeing modern collaboration communications, like Slack, and others get used in government. And that has the same duty of care that email and social media has. And so were working diligently on that.
And then, you know, were growing. We add about 150 public agencies a month. We added 1,500 public agency customers last year alone across the three companies. And so we need to hire optimists (thats the official word for an employee at Optimere). And we got a bunch of hiring to do this year. I love it.
Follow Optimere here: LinkedIn | Twitter | Facebook | Instagram
Until then, watch the previous episode announcing the GovTech 100 with special guests.
In Case You Missed It is Government Technologys weekly news roundup and interview live show featuring e.Republic* Chief Innovation Officer Dustin Haisler, Deputy Chief Innovation Officer Joe Morris and Gov Tech Assistant News Editor Jed Pressgrove as they bring their analysis and insight to the weeks most important stories in state and local government.
Follow along live each Friday at 12 p.m. PST on LinkedIn and YouTube.
*e.Republic is Government Technologys parent company.
Read the rest here:
The Future of Records & Compliance With Optimere CEO ICYMI - Government Technology