logo
GBM Joins Red Hat Partner Practice Accelerator

GBM Joins Red Hat Partner Practice Accelerator

TECHx3 days ago

Home » Emerging technologies » Cloud Computing » GBM Joins Red Hat Partner Practice Accelerator
Gulf Business Machines (GBM), an end-to-end digital solutions provider, has announced its participation in the Red Hat Partner Practice Accelerator. GBM is the first company in the MEA and GCC region to join this global initiative, which focuses on application development.
Red Hat's Partner Practice Accelerator recognizes partners that demonstrate advanced technical expertise, service delivery excellence, and hold Red Hat certifications. GBM qualified by meeting rigorous training and validation standards.
With this recognition, GBM is now positioned as a trusted expert in application modernization using Red Hat OpenShift. This hybrid cloud application platform, powered by Kubernetes, enables GBM to design, implement, and configure enterprise-grade solutions tailored to customers' needs.
Additionally, GBM gains access to enhanced Red Hat resources and co-delivery opportunities. These include full lifecycle implementation services from design to deployment and scalable, future-ready cloud solutions aligned with business goals. GBM is the first in MEA/GCC to join the Red Hat initiative.
The partnership strengthens GBM's hybrid cloud and Kubernetes expertise.
Customers benefit from GBM's certified services and global delivery model.
This milestone reinforces GBM's commitment to innovation, operational excellence, and customer success through Red Hat technologies. Red Hat Partner Practice Accelerator is part of Red Hat's global engagement model designed to drive co-creation and hybrid cloud advancement.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

GBM Joins Red Hat Partner Practice Accelerator
GBM Joins Red Hat Partner Practice Accelerator

TECHx

time3 days ago

  • TECHx

GBM Joins Red Hat Partner Practice Accelerator

Home » Emerging technologies » Cloud Computing » GBM Joins Red Hat Partner Practice Accelerator Gulf Business Machines (GBM), an end-to-end digital solutions provider, has announced its participation in the Red Hat Partner Practice Accelerator. GBM is the first company in the MEA and GCC region to join this global initiative, which focuses on application development. Red Hat's Partner Practice Accelerator recognizes partners that demonstrate advanced technical expertise, service delivery excellence, and hold Red Hat certifications. GBM qualified by meeting rigorous training and validation standards. With this recognition, GBM is now positioned as a trusted expert in application modernization using Red Hat OpenShift. This hybrid cloud application platform, powered by Kubernetes, enables GBM to design, implement, and configure enterprise-grade solutions tailored to customers' needs. Additionally, GBM gains access to enhanced Red Hat resources and co-delivery opportunities. These include full lifecycle implementation services from design to deployment and scalable, future-ready cloud solutions aligned with business goals. GBM is the first in MEA/GCC to join the Red Hat initiative. The partnership strengthens GBM's hybrid cloud and Kubernetes expertise. Customers benefit from GBM's certified services and global delivery model. This milestone reinforces GBM's commitment to innovation, operational excellence, and customer success through Red Hat technologies. Red Hat Partner Practice Accelerator is part of Red Hat's global engagement model designed to drive co-creation and hybrid cloud advancement.

GBM becomes region's first company to join Red Hat partner practice accelerator
GBM becomes region's first company to join Red Hat partner practice accelerator

Zawya

time4 days ago

  • Zawya

GBM becomes region's first company to join Red Hat partner practice accelerator

GBM will now have access to enhanced Red Hat resources and co-delivery opportunities Dubai, United Arab Emirates: Gulf Business Machines (GBM), a leading end-to-end digital solutions provider, has become the first company in the MEA/GCC region to join Red Hat Partner Practice Accelerator specializing in application development. This global initiative by Red Hat recognizes a select group of partners who have demonstrated advanced technical expertise and proven service delivery practices by achieving Red Hat professional training credentials, certifications, and partner services validation. With its participation, GBM is a trusted expert in application modernization using Red Hat OpenShift, the industry's leading hybrid cloud application platform powered by Kubernetes, capable of architecting, implementing, and configuring tailored enterprise-grade solutions that drive impactful business outcomes. GBM will now have access to enhanced Red Hat resources and co-delivery opportunities to support customers with end-to-end implementation services, from design to deployment, as well as scalable, future-ready solutions that align with their business goals. This distinction reaffirms GBM's commitment to innovation, operational excellence, and long-term success for its customers adopting Red Hat technologies. Combining its deep local roots with global standards, GBM is positioned to help shape the future of application development across the region. Red Hat Partner Practice Accelerator is part of Red Hat's evolving partner engagement model that implements a globally unified approach to collaboration. The initiative provides partners with simplified paths to co-create, innovate, and deliver solutions and services to support customers on their hybrid cloud journeys. About Gulf Business Machines (GBM) With more than 35 years of experience, 7 offices and over 1500 employees across the region, Gulf Business Machines (GBM) is a leading end-to-end digital solutions provider, offering the region's broadest portfolio, including industry-leading digital infrastructure, digital business solutions, security and services. We have nurtured partnerships since 1990 with the world's leading technology companies and invested in a talented, skilled workforce to implement solutions that cater to customer's specific, complex and diverse business needs. Some of our strategic partners in the Gulf include IBM as their sole distributor throughout the GCC (excluding Saudi Arabia and selected IBM product and services), Cisco as a Gold Partner (the highest level of certification at Cisco), and VMware as a Premier Partner (the highest partner status within VMware).

F5 and NVIDIA to meet the needs of accelerated computing and AI
F5 and NVIDIA to meet the needs of accelerated computing and AI

Tahawul Tech

time5 days ago

  • Tahawul Tech

F5 and NVIDIA to meet the needs of accelerated computing and AI

F5 has announced new capabilities for F5 BIG-IP Next for Kubernetes accelerated with NVIDIA BlueField-3 DPUs and the NVIDIA DOCA software framework, underscored by customer Sesterce's validation deployment. Sesterce is a leading European operator specialising in next-generation infrastructures and sovereign AI, designed to meet the needs of accelerated computing and artificial intelligence. Extending the F5 Application Delivery and Security Platform, BIG-IP Next for Kubernetes running natively on NVIDIA BlueField-3 DPUs delivers high-performance traffic management and security for large-scale AI infrastructure, unlocking greater efficiency, control, and performance for AI applications. In tandem with the compelling performance advantages announced along with general availability earlier this year, Sesterce has successfully completed validation of the F5 and NVIDIA solution across a number of key capabilities, including the following areas: – Enhanced performance, multi-tenancy, and security to meet cloud-grade expectations, initially showing a 20% improvement in GPU utilisation. – Integration with NVIDIA Dynamo and KV Cache Manager to reduce latency for the reasoning of large language model (LLM) inference systems and optimisation of GPUs and memory resources. – Smart LLM routing on BlueField DPUs, running effectively with NVIDIA NIM microservices for workloads requiring multiple models, providing customers the best of all available models. – Scaling and securing Model Context Protocol (MCP) including reverse proxy capabilities and protections for more scalable and secure LLMs, enabling customers to swiftly and safely utilise the power of MCP servers. – Powerful data programmability with robust F5 iRules capabilities, allowing rapid customisation to support AI applications and evolving security requirements. 'Integration between F5 and NVIDIA was enticing even before we conducted any tests', said Youssef El Manssouri, CEO and Co-Founder at Sesterce. 'Our results underline the benefits of F5's dynamic load balancing with high-volume Kubernetes ingress and egress in AI environments. This approach empowers us to more efficiently distribute traffic and optimise the use of our GPUs while allowing us to bring additional and unique value to our customers. We are pleased to see F5's support for a growing number of NVIDIA use cases, including enhanced multi-tenancy, and we look forward to additional innovation between the companies in supporting next-generation AI infrastructure'. Highlights of new solution capabilities include: LLM Routing and Dynamic Load Balancing with BIG-IP Next for Kubernetes With this collaborative solution, simple AI-related tasks can be routed to less expensive, lightweight LLMs in supporting generative AI while reserving advanced models for complex queries. This level of customisable intelligence also enables routing functions to leverage domain-specific LLMs, improving output quality and significantly enhancing customer experiences. F5's advanced traffic management ensures queries are sent to the most suitable LLM, lowering latency and improving time to first token. 'Enterprises are increasingly deploying multiple LLMs to power advanced AI experiences—but routing and classifying LLM traffic can be compute-heavy, degrading performance and user experience', said Kunal Anand, Chief Innovation Officer at F5. 'By programming routing logic directly on NVIDIA BlueField-3 DPUs, F5 BIG-IP Next for Kubernetes is the most efficient approach for delivering and securing LLM traffic. This is just the beginning. Our platform unlocks new possibilities for AI infrastructure, and we're excited to deepen co-innovation with NVIDIA as enterprise AI continues to scale'. Optimizing GPUs for Distributed AI Inference at Scale with NVIDIA Dynamo and KV Cache Integration Earlier this year, NVIDIA Dynamo was introduced, providing a supplementary framework for deploying generative AI and reasoning models in large-scale distributed environments. NVIDIA Dynamo streamlines the complexity of running AI inference in distributed environments by orchestrating tasks like scheduling, routing, and memory management to ensure seamless operation under dynamic workloads. Offloading specific operations from CPUs to BlueField DPUs is one of the core benefits of the combined F5 and NVIDIA solution. With F5, the Dynamo KV Cache Manager feature can intelligently route requests based on capacity, using Key-Value (KV) caching to accelerate generative AI use cases by speeding up processes based on retaining information from previous operations (rather than requiring resource-intensive recomputation). From an infrastructure perspective, organisations storing and reusing KV cache data can do so at a fraction of the cost of using GPU memory for this purpose. 'BIG-IP Next for Kubernetes accelerated with NVIDIA BlueField-3 DPUs gives enterprises and service providers a single point of control for efficiently routing traffic to AI factories to optimize GPU efficiency and to accelerate AI traffic for data ingestion, model training, inference, RAG, and agentic AI,' said Ash Bhalgat, Senior Director of AI Networking and Security Solutions, Ecosystem and Marketing at NVIDIA. 'In addition, F5's support for multi-tenancy and enhanced programmability with iRules continue to provide a platform that is well-suited for continued integration and feature additions such as support for NVIDIA Dynamo Distributed KV Cache Manager'. Improved Protection for MCP Servers with F5 and NVIDIA Model Context Protocol (MCP) is an open protocol developed by Anthropic that standardizes how applications provide context to LLMs. Deploying the combined F5 and NVIDIA solution in front of MCP servers allows F5 technology to serve as a reverse proxy, bolstering security capabilities for MCP solutions and the LLMs they support. In addition, the full data programmability enabled by F5 iRules promotes rapid adaptation and resilience for fast-evolving AI protocol requirements, as well as additional protection against emerging cybersecurity risks. 'Organisations implementing agentic AI are increasingly relying on MCP deployments to improve the security and performance of LLMs', said Greg Schoeny, SVP, Global Service Provider at World Wide Technology. 'By bringing advanced traffic management and security to extensive Kubernetes environments, F5 and NVIDIA are delivering integrated AI feature sets—along with programmability and automation capabilities—that we aren't seeing elsewhere in the industry right now'. F5 BIG-IP Next for Kubernetes deployed on NVIDIA BlueField-3 DPUs is generally available now. For additional technology details and deployment benefits, go to and visit the companies at NVIDIA GTC Paris, part of this week's VivaTech 2025 event. Further details can also be found in a companion blog from F5. Image Credit: F5

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store