Learn how BLDD Architects developed IT security that supports all of its servers and mix of clients in a way that’s invisible to its end users.
Listen to this podcast discussion on how Texmark, with support from HPE and HPE channel partner CB Technologies, has been combining the refinery of the future approach with the best of OT, IT, and IoT technology solutions to deliver data-driven insights that promote safety, efficiency, and unparalleled sustained operations.
Companies large and small are changing in order to innovate faster, provide better customer experiences, and achieve greater cost efficiencies. British philosopher Alan Watts has a suggestion for dealing with this type of disruption: “The only way to make sense out of change is to plunge into it, move with it, and join the dance.”
Sounds simple, right? It is not.
Many businesses are dancing straight into the arms of public cloud because it enables them to meet time-to-market deadlines by scaling quickly and easily. Yet, others find that certain workloads are not appropriate for this type of Tango due to cost, performance, compliance, security, or complexity issues. And a growing number of enterprises are looking for a mix of IT deployments to attain ideal results. In order to adjust quickly to changing business needs, IT wants the flexibility to place some applications in the public cloud and others in a private cloud on-premises – sort of like choosing to enjoy both hip-hop and ballet.
Transforming the traditional
As organizations try to select the best deployment options, they are finding that cloud is no longer a destination; instead, it is new way of doing business that focuses on speed, scalability, simplicity, and economics. This type of business model allows cloud architects to distribute workloads across a mix of on-premises and public clouds. No matter where IT places the workload, everyone in the enterprise expects fast service delivery, operational simplicity, and optimization of costs.
If this scenario sounds too good to be true, it actually is…for the moment.
IT is struggling to achieve this type of cloud transformation due to a number of constraints typically found in data centers. Most people acknowledge that much of today’s data center infrastructure is slow, complex, and manual, which means that IT can’t properly deliver the services needed for a modern, cloud-based deployment model. Yet, the challenge is actually much bigger – it involves legacy thinking, which can be harder to change than technology.
Out with the old way of thinking … in with the new
Many developers in the past routinely used a type of waterfall model for project management, where the project leaders define the project at the start, and then it goes through a number of sequential phases during its lifecycle. This model has its roots in engineering where a physical design was a critical part of the project and any changes to that design were costly. Changes occurred infrequently and all at once. IT operations was comfortable with this process, because the old way of thinking believed that if the frequency of change is reduced, risk is also reduced.
Modern developers have discovered that the opposite can be true. If something goes wrong with a massive change, it could very well bring down the entire company. Therefore, the new way of thinking is to implement small changes much more frequently. That way, if something fails, it is a small failure – and the team can quickly change course without causing major problems.
A transformed data center needs a new mindset that embraces an agile set of principles, similar to how application developers work – delivering and accepting project changes in short duration phases called sprints. During each sprint, continuous change is encouraged, creating a more agile and flexible environment. And failure is allowed, because that is when learning – and adjustment – occurs.
Another big change involves capital spending and total cost of ownership. The old thinking involved inflexible consumption models that forced the organization to pay for everything up front. Again, IT believed that this model was less risky because they knew the costs upfront and could accurately plan accordingly.
Yet this model can be more risky because it is not agile; IT could not increase infrastructure for a short duration during a critical need, and then dial it back down when the need no longer existed. Today’s new way of thinking about IT infrastructure involves a flexible, as-a-service consumption model, where customers only pay for what they use when they use it.
Creating a perfectly choreographed experience across your enterprise
Hewlett Packard Enterprise (HPE) is working to solve your legacy thinking challenges in the data center and in the public cloud. Cloud Technology Partners (CTP), a Hewlett Packard Enterprise company, will help your team learn the mindset changes your business needs to succeed in a digital transformation and the steps you need to make toward a truly hybrid model.
HPE is also creating a perfectly choreographed series of solutions that will quickly modernize your data center and public cloud infrastructure footprint. With the help of HPE’s industry experts and innovative infrastructure, you can quickly turn your legacy data center into a hybrid cloud experience that combines modern technologies and software-defined infrastructure such as composable infrastructure, hyperconvergence, infrastructure management, and multi-cloud management.
A new hybrid cloud operating model built for speed, agility, and cost optimization is upon us. Make sure you have the right partner to “plunge into it, move with it, and join the dance.”
Advisory services at Cloud Technology Partners can help you understand how to take advantage of today’s new, modern multi-cloud technology. To learn more about how composable infrastructure can power your digital transformation, click here. Visit HPE OneView and HPE OneSphere to read how to simplify and automate infrastructure and multi-cloud management.
About Gary Thome
Gary Thome is the Vice President and Chief Technologist for the Software-Defined and Cloud Group at Hewlett Packard Enterprise (HPE). He is responsible for the technical and architectural directions of converged datacenter products and technologies.
To read more articles from Gary, check out the HPE Shifting to Software-Defined blog.
“A world with millions of clouds distributed everywhere - that's the future as we see it.” – HPE CEO Antonio Neri
When cloud computing first began disrupting traditional IT over 10 years ago, who would have imagined millions of clouds would soon follow? According to industry experts, that is exactly where the industry is heading. The next wave of digital disruption will store and analyze data at the edge and in the cloud instantly, compliments of millions of clouds distributed everywhere.
To cope with this tsunami of widely distributed data, businesses will need to go beyond on-premises environments and multi-cloud deployments. They must connect a hybrid system that stretches from the edge to the cloud and everywhere in-between. A recent report from 451 Research, From Edge to Cloud, Managing the Next Wave of IT Disruption, explains this new reality.
8 Essential Steps for Managing Edge-to-Cloud
The report details 8 essentials businesses need to consider as they enter the next wave of IT disruption.
1. Proactive cloud strategy
Organizations everywhere are pursuing a proactive hybrid cloud and multi-cloud strategy, balancing performance, cost, and compliance. At the same time, they are meeting specific needs of applications and workloads. All of this takes planning, along with time and skills – which are in short supply in today’s fast-paced, competitive environment. Organizations must seek ways to unify access to multiple clouds and simplify management.
2. Modernize and automate
Traditional, manual-laden IT processes will become outdated, as orchestration and automation tools transform the data center. Hyperconvergence and composability are providing the agility of public cloud through software-defined strategies, which increases automation and saves time.
3. Take out the complexity
An ideal hybrid IT environment must be simple and quick to deploy and manage -- and capable of seamlessly bridging multiple workloads across traditional, private, and public cloud infrastructure. A hybrid cloud management platform must allow IT administrators or business managers to view all available infrastructure resources without requiring detailed knowledge of the underlying hardware.
4. Future-proof for emerging technologies
Hybrid IT must support not only OS, virtualization, and popular cloud options that businesses are using, but also fast-growing new alternatives. These include bare-metal and container platforms, along with extensions to the architecture, such as the distributed edge. Unified APIs will help with the integration of existing apps, making everything easier to manage.
5. Deliver everything as a service
Enterprises that want to optimize resources are moving toward deploying everything as a service. Software-defined and hybrid cloud management help to integrate off-premises services with workloads that need to stay on-premises.
6. Deal with the data and gains insights faster
As data explodes from the edge to the cloud, software-defined services and hybrid cloud data management will become vital. Organizations will need to decide where to generate data, how to analyze it quickly, and what actions to take based on their analysis.
7. Control spending and utilization
Public cloud providers are expanding their portfolios to provide more options, which include more pricing models, increased instance sizes, smaller time increments, better reporting, and competitive pricing. Because the price of cloud is falling only marginally, providers differentiate themselves by offering flexibility in procurement and products. Yet, as more choice is offered, complexity also increases, driving the need for hybrid cloud management solutions.
8. Extend to the edge
Edge computing marks the beginning of a massive increase in a vast infrastructure of endpoints that will be part of tomorrow’s IT. Moving data centers such as cars, airplanes, trains, robots, and drones will increase rapidly. Enterprise customers need to invest now by integrating their private and public cloud resources with an eye toward expanding to a highly distributed infrastructure in the future.
A world with millions of clouds distributed everywhere will soon become commonplace. While the rest of the world is moving toward the cloud, multitudes of smart endpoints are starting to force computing closer to the edge. Analytics, edge processing, artificial intelligence, and machine learning are also on the rise. Combining cloud and hybrid IT models with edge computing—all tied together with a multi-cloud management platform—is an important milestone to combat the next wave of IT disruption.
Read the full report from 451 Research, From Edge to Cloud, Managing the Next Wave of IT Disruption. Learn more about hybrid cloud management here.
Gary Thome is the Vice President and Chief Technologist for the Software-Defined and Cloud Group at Hewlett Packard Enterprise (HPE). He is responsible for the technical and architectural directions of converged datacenter products and technologies which include HPE OneSphere – multi-cloud management, HPE SimpliVity – Hyperconverged Infrastructure, HPE Synergy – Composable Infrastructure and HPE OneView – Integrated Management.
To read more articles from Gary, check out the HPE Shifting to Software-Defined blog.
Get ready to meet the team of data scientists behind HPE InfoSight. And learn how we’re leveraging this AI-driven analytics platform to improve the customer, support, and sales experience.
In 2012, InfoSight was launched with the intention of leveraging telemetry data to identify, predict, and solve customer issues with storage arrays. This in turn would allow our support team to consist primarily of Level 3 engineers.
The bold vision was put in place by our Chief Data Scientist at the time. In the early days of Nimble Storage, the data science team knew data would be valuable and had the foresight to begin collecting DNA about the hardware being deployed in the field. InfoSight enabled the team to leverage multiple years of detailed performance data pertaining to hard drives, solid state drives, fans, CPUs, power supplies, and network cards. This trove of data became the backbone of InfoSight and the data science team.
In the years since the launch of InfoSIght, the data science team has played an integral role in differentiating our storage arrays from our competitors. Leveraging installed base data allowed the team to programmatically open, remediate, and close cases on our customers’ behalf. As the scope of data increased, the data science team has been able to create more sophisticated models and tools that not only enhance the customer experience but also the support and sales experiences too. Additionally, the data used by the data science team has expanded into the stack leveraging sensor data and configuration data from virtual machines connected to the various HPE platform.
The lessons learned from our multi-petabyte analytics platform has directly improved the following experiences:
InfoSight data science has been providing customers with the ability to view capacity and performance predictions based on specific workloads allowing them to avoid troublesome situations that could lead to poor performance. Various resource models have been created to help a customer identify periods of saturation, cache latency, and CPU usage.
Configuration data also plays an important role in the analytics provided to customers. Individual volumes can also become problematic so we closely monitor overprovisioning on the customer array. When analyzing all the data for a particular customer we are able to confidently use our models based on real-world data to make recommendations to our customers on upgrade needs to ensure they continue receiving the performance they expect from our hardware.
The InfoSight data science team has worked closely with our support organization since the beginning to provide insights into the install base performance of similar customer configurations. The team partners with support to help create signatures that are used for proactive case monitoring. As we feed our support data into our data lake we are able to look for similar patterns across various models of hardware or software releases to refine rules that are currently in place.
InfoSight has also been a key tool in the support investigation process. It allows a customer to gain insights into the operation of their HPE equipment while viewing the same information a technical support engineer will see.
The team has created tools which are used in the sales cycle to help right-size a customer environment leveraging our real-world installed base knowledge. These sizing models are sophisticated in nature and take into account known performance metrics, sensor data, and latency. They allow users to input various criteria that matters to the customer and weigh cost versu performance—while providing recommendations that are based not just on marketing data but on known performance for particular workloads, array models, or disk types.
Historically. we have also been able to leverage our recommendation engine to generate leads and opportunities the allow the sales team to help serve our customers.
Get ready to get to know today’s InfoSight data science team
The HPE InfoSight data science team has a wide breadth of experience, ranging from recent grads to 20+ years. The team has received multiple patents for the work in the data space. The data scientists work closely with our dedicated data engineers, support team, and product specialists. In the coming weeks you can check this space for blogs written by the individual data scientists discussing the technology stack, research approach, real-world findings, and formula usage.
ABOUT THE AUTHOR
I am Marketing Manager for HPE InfoSight. Before that, I led the marketing for the deep learning startup, Nervana Systems, up until its acquisition by Intel in 2016.
Discover how a joint data storage solution from HPE and Cohesity eliminates secondary storage silos and restores sanity to application chaos.
Without question, secondary storage has become a critical issue for IT decision makers. A previous blog discusses how disjointed secondary storage can cause mass data fragmentation, which can slow down the business and potentially raise compliance risk.
A poor secondary data storage solution can certainly create inefficient workflows, but that's just the beginning. Unmanaged secondary storage sprawl also results in:
Increased costs due to loss of economies of scale
Increased security risk, since you may not have a clear picture of what data assets you have and who has access to them
Increased compliance risk if you're holding regulated data without a crystal-clear understanding of what you have
Increased risk of poor decision making, as copies of data can float around with no clear indication of which one is the most current or accurate
Increased storage capacity utilization as various secondary storage silos lose the capability to bring aggregated data reduction technologies to bear on the ongoing capacity challenge
It's clear that secondary storage has become a primary problem. What's needed is a solution that can eliminate the need for so many of these silos.
Now, via a partnership between two enterprise IT firms—HPE and Cohesity, iPhone-like simplicity has come to handily solve the secondary storage dilemma, bringing with it some incredible outcomes.
A data storage solution for secondary storage
This solution: HPE Solutions for Cohesity DataPlaltform. Cohesity, a market-leading purveyor of a hyperconverged secondary storage software stack, provides software that runs atop HPE Apollo and ProLiant servers. This combination handily solves the mass data fragmentation problem. With Cohesity, your data is no longer massively siloed, copies of data are reduced or eliminated, and you reduce or halt the spread of data to every corner of your organization.
Why is this so important? In research performed by VansonBourne, 98 percent of respondents indicated that their secondary storage needs will grow in the next 18 months, with more than one-half saying that their storage needs will grow between 25 and 75 percent per year.
Fixing this without the right tools isn't easy. In fact, 26 percent of respondents said that they would rather quit their jobs than be tasked with fixing their company's secondary storage problems without the right tools in place.
Hyperconverged secondary storage
You may be wondering how this HPE/Cohesity mashup can help you solve the storage growth issue, since you'll still have a lot of data to contend with. What Cohesity brings to the picture is a software solution that conglomerates all of an organization's secondary storage assets under one managed umbrella, instantly imbuing visibility into what used to be something of a black box. Where the real magic comes in, however, is through Cohesity's global deduplication, compression, and erasure coding features. These capabilities allow organizations to grow in a far more sustainable way than they can with a smattering of point solutions.
This transition to a converged secondary storage architecture can't happen soon enough. Today, 48 percent of survey respondents say that their secondary storage sprawl means that their IT team spends 30-to-100 percent of their time cleaning up. Almost all respondents—98 percent—say that it's getting worse.
The HPE Solutions for Cohesity DataPlatform are data storage solutions that provide a broad set of capabilities, including:
The previously mentioned data reduction services, which slow your organization's need to constantly procure storage for secondary needs. (Bear in mind that the HPE Cohesity System isn't designed for primary workloads. For those, you should turn your attention to HPE Nimble Storage and HPE 3PAR solutions, which bring all-flash power to the storage equation.)
Data protection capabilities for workloads that operate in the secondary arena.
Full support for object file types. No longer do you need a separate solution to manage object-centric application needs. Cohesity can handle that for you as a part of its integrated software.
Test/dev support. For many, test and development environments were just fragments sitting on their own. With Cohesity, all of your test and dev activities can be managed via the integrated platform and then easily promoted into production once you're ready. This provides a seamless DevOps experience for your organization.
Data indexing so that you always know what you have and where you have it.
Encryption of your data both at rest and in flight.
If you're looking for a solution that's hyperconverged onto a single platform, the joint HPE and Cohesity solution is exactly what the CIO ordered. You can collapse a number of previously disparate secondary-centric services into one. And it's not limited to an on-premises approach.
Instant hybrid cloud
The HPE Solutions for Cohesity DataPlatfrom also helps you accelerate your hybrid cloud journey. The Cohesity software optionally leverages the public cloud for myriad purposes, including as a backup target, a target for long-term data retention and archive, a storage tier, disaster recovery, and test/dev purposes.
All of the features described earlier, including data reduction, global indexing, and security, apply to the public cloud side of the Cohesity equation, too. The result is a seamless experience between your HPE-driven on-premises Cohesity solution and your favorite public cloud, whether that's AWS, Azure, Google, or another Cohesity cloud partner.
Ensuring a solid hardware foundation
With great software comes a need for great hardware from a trusted partner. Cohesity is a hyperconverged solution, meaning that the underlying hardware needs to work in concert with the solution—not as a hindrance. Even with a conglomeration of secondary data sources, secondary storage needs will still grow, and these data sources have and will continue to have importance to the organization. The only way that such a solution works is if it's running on hardware that is stable, fast, easily managed, and scalable. HPE brings that in spades with HPE Solutions for Cohesity DataPlatfrom.
A solution is only as good as the support behind it. The HPE and Cohesity solution is a full member of HPE's global supply chain, helping customers around the world reduce deployment risk and complexity and unlocking a fully optimized deployment and support experience. For those that are uncomfortable taking the plunge on their own, the HPE and Cohesity solution can be configured and deployed by HPE Pointnextconsulting services.
As you consider your secondary application needs, look no further than the hyperconverged secondary storage solution from HPE and Cohesity to allow you to simplify your environment by conglomerating workloads into a single, scalable hybrid environment.
Meet Around the Storage Block blogger Scott D. Lowe, CEO and Lead Analyst for ActualTech Media.
Since 1994, Scott has helped organizations of all stripes solve critical technology challenges. He has served in a variety of technical roles, spent ten years as a CIO, and has spent another ten as a strategic IT consultant in higher education. Today, his company helps educate IT pros and decision makers and brings IT consumers together with the right enterprise IT solutions to help them propel their businesses forward.
APRIL 9, 2019 • BLOG POST • PHIL DAVIS, PRESIDENT OF HYBRID IT & CHIEF SALES OFFICER
HPE partners with Google Cloud and Nutanix to provide customers greater choice and agility
Organizations everywhere are on a hybrid cloud expedition. One that can be complex but one that we, at HPE, aim to simplify no matter which path our customers choose.
In recent years, we’ve taken many steps to accelerate customers’ hybrid cloud journeys. We’ve simplified our own organizational structure, creating a Hybrid IT business group that is integrated and easier for customers to navigate – because all infrastructure, software, and services capabilities are under one roof.
Plus, we have made extensive additions to our suite of products and services in order to give customers the choice and flexibility they crave for a consistent and optimal experience across public and private clouds. One of those is HPE SimpliVity, a key component of our Composable Cloud portfolio which enables customers to streamline IT operations with a fast, uncomplicated and efficient hyperconverged infrastructure (HCI) platform – and at a fraction of the cost.
And today, we’re further expanding our commitment to deliver the options and the experiences our customers desire for hybrid cloud by strategically aligning with two powerful industry players. Each is designed to extend our fast-growing and ever-evolving HPE GreenLake ecosystem.
To extend the HPE Composable Cloud portfolio, HPE and Google Cloud have entered into a strategic partnership to deliver hybrid cloud solutions that accelerate innovation and expand choice and agility for customers. The partnership will provide customers with a consistent experience across public cloud and on premises environments.
As an initial part of this strategic agreement, HPE will offer two validated designs for Google Kubernetes Engine (GKE) based on HPE SimpliVity hyperconverged offering and HPE Nimble Storage with HPE ProLiant offering. In addition, HPE will offer these solutions as a service through HPE GreenLake, HPE’s fully managed consumption offering. Customers that choose this offering can run applications as a Service in the Google Kubernetes Engine environment on premises and benefit from the same container-based design across their hybrid cloud.
In addition to that, we are partnering with Nutanix to deliver an integrated hybrid cloud as-a-service solution. The offer -- which leverages Nutanix’s Enterprise Cloud OS delivered through the HPE GreenLake as-a-service solution – will provide customers with a fully HPE-managed hybrid cloud.
As part of the agreement, Nutanix is also expanding platform choice to its customers and will enable its channel partners to directly sell HPE ProLiant DX and HPE Apollo DX servers combined with Nutanix’s Enterprise Cloud OS software, so that customers can purchase an integrated, turnkey appliance with built-in intelligence and security.
We believe our offering with Nutanix is an attractive choice that will reduce cost and complexity by offering a fully managed hybrid cloud infrastructure delivered as-a-service, to be deployed in customer data centers or in a customer’s co-location facility.
With tools like these, customers have the wherewithal to build hybrid and private clouds and transform their existing applications. They can provision workloads of all types, across virtualization, containers and bare metal, in minutes. And as a result, they will spend less time managing infrastructure and more time creating value-added services at a fraction of past operational costs.
Some might ask why we are pursuing such unique and collaborative offerings with other providers. But to me, the reason is obvious: For certain customers, in certain situations, we can offer stronger composable solutions, together. It’s all about delivering what the customer wants and needs, and we don’t mind sharing – while still continuing to compete vigorously across our entire portfolio of solutions.
The transformation to hybrid cloud can be difficult and confusing. With bold, customer-centric, and simplified initiatives – such as today’s two groundbreaking announcements with Google and Nutanix – HPE further strengthens our position as the strategic hybrid cloud transformation partner for any enterprise.
And in doing so, the pathway to hybrid cloud just became a lot easier to travel.