
Today’s IT infrastructure and operations teams are readying for the intersection of multiple technology trends that will impact their ability to deliver scalable, agile, and resilient platforms.
Among those trends discussed at the recent Gartner IT Infrastructure, Operations, and Cloud Strategies conference, three stand out:
Read on for more key takeaways from the conference, which drew some 5,000 attendees and nearly 130 vendors.
Generative AI uses machine learning and AI foundation models to generate new content, product designs, and business processes, among other examples, which resemble but don’t repeat the original data.
Listen to enough analysts and you’ll come away thinking generative AI will be the biggest industry disrupter since the internet. While a “distrust and verify” approach to GenAI’s outputs is warranted, generative AI’s role as a strategic innovation for I&O means more companies will want to try GenAI initiatives.
According to Gartner’s 2024 CIO and Technology Executive Survey, which a Gartner I&O conference keynote highlighted:
80% of CIOs and tech leaders plan full GenAI adoption within three years.
Before adopting generative AI, companies must decide whether they want to buy or build GenAI solutions in their environment. Those decisions include:
Gartner recommends starting small, then iterate by testing AI functionality in a proof of concept before moving into production.
Mark McCully, Director of Infrastructure & Operations Modernization Engineering at Sayers, says:
“Companies may want to kick the tires on GenAI, build a chatbot, and expand from there, based on what they think will drive the most business value. It probably will be less expensive initially in the public cloud, rather than purchasing a lot of compute and storage, GPUs, and related requirements.”
Other considerations in your GenAI approach include power consumption and cooling. Next-generation data centers have to handle the power and scale needed to run GenAI, which requires heavy-duty GPU’s and DPU’s (data processing units). As companies spin up more AI-related workloads, those computing processes generate more heat and require liquid cooling solutions.
Many companies will likely start their GenAI journey in the public cloud using technologies such as Microsoft Copilot, OpenAI, or Hugging Face for access to large language models.
Expected to play a large role in GenAI adoption is Azure Stack, a portfolio of products that extend Microsoft Azure services and capabilities from the data center to edge locations and remote offices. McCully says:
“Azure is a good vehicle to try GenAI initiatives through a consumption model, without having to outlay a lot of capital or opex expenses to bring things you will need on premise such as GPUs and liquid cooling.”
Platform engineering incorporates key tenants such as automation, observability, and self-service into your infrastructure environment so your software development team can deliver stable and secure applications faster.
“By 2026, 30% of enterprises will automate more than half of their network activities, an increase from less than 10% of enterprises in early 2023.”
The growth of AI and the Internet of Things will fuel not only unstructured data growth but also edge computing, where processing occurs closer to data sources for more efficient IoT and AI applications.
According to Gartner: Modernize Data and Analytics Capabilities (gartner.com)
By 2025, more than 50% of enterprise-managed data will be created and processed outside the data center or cloud.
Several industries will push the boundaries of edge computing. That’s because much of their data isn’t processed or held in the main data center, but rather at edge locations such as branch offices.
Those industries and use case examples include manufacturing (system automation), retail (immersive e-commerce), healthcare (heart monitors), finance (high-frequency trading), and energy (real-time grid adjustments based on consumption).
Public clouds such as Microsoft Azure provide edge capabilities with IoT enablement options, while Azure Stack offers on-prem cloud-like services.
With a solution like Azure Stack, you can bring specific data sets or applications back down from the cloud to on-premise locations. This avoids the egress charges associated with taking data out of the cloud, and also moves your data closer to your end users who need to consume data from your applications. McCully says:
“While you might be trying to put as much as you can in the cloud, there are going to be certain workloads or edge use cases where it makes sense to use something like Azure Stack to keep the data or the application more local and avoid some egress fees or latency issues.”
The growth of data from AI and IoT brings more concerns about data security as well as application security. Your infrastructure platform engineering initiatives should include conversations about how you embed security into your application development and deployment.
For securing the edge, options include software-defined wide area networking (SD-WAN), secure access service edge (SASE), and cyber-physical systems security solutions.
Questions? Contact us at Sayers today for help in choosing the right technology solutions for your business.