Many enterprises today are moving from generative AI (gen AI) experimentation to production, deployment and scaling. Code generation and modernization are now among the top enterprise use cases that offer a clear path to value creation, cost reduction and return on investment (ROI).

IBM® Granite™ is a family of enterprise-grade models developed by IBM Research® with rigorous data governance and regulatory compliance. Granite currently supports multilingual language and code modalities. And as of the NVIDIA AI Summit in Taiwan this month, IBM Granite code models, 8b-code-instruct and 34b-code-instruct, are available as NVIDIA-hosted NIM inference microservices on the NVIDIA API catalog.

These models are optimized for higher throughput and performance, powered by NVIDIA NIM. The new availability of these models comes after IBM’s announcement of its collaboration with NVIDIA to drive enterprise gen AI adoption by pairing NVIDIA AI Enterprise software and accelerated computing with industry solutions from IBM Consulting®.

Enterprise decision makers are facing the challenge of scaling gen AI faster while mitigating foundation model-related risks. They are looking for truly enterprise-grade foundation models and software capabilities to bring trusted, performant and cost-effective generative AI to key business workflows and processes.

Based on HumanEvalPack evaluation, Granite code models can outperform some models that are even twice their size. In the evaluation, no single model could perform at a high level at generation, fixing and explanation—apart from Granite. The family of models was also recognized by the Stanford Transparency Index as one of the most transparent modes in the industry, with a perfect score in several categories designed to measure how open models really are. In fact, since the recognition, IBM has taken further steps to enhance Granite’s transparency by releasing Granite code models into open source, aimed at making coding as easy as possible for the developer community. Granite models are trained on 116 programming languages, including Python, JavaScript, Java, Go, C++ and Rust.

Granite models are available in a curated foundation models library part of the IBM watsonx™ data and AI platform, open-source platforms—Hugging Face, GitHub, watsonx.ai™ and RHEL AI (the new foundation model platform from Red Hat®)—and now on the NVIDIA API catalog, which makes coding easy and accessible for as many developers as possible.

IBM Granite code models on the API catalog are planned to be offered as downloadable NIM inference microservices—designed to simplify and accelerate the deployment of AI models across GPU-accelerated workstations, data center and cloud platforms. The flexibility to deploy on your preferred infrastructure ensures your data is private and secured. Containerized for easy deployment, NVIDIA NIM microservices deliver superior throughput to power more responses on the same infrastructure and support industry-standard APIs that can be easily incorporated into existing workflows.

Furthermore, as part of the NVIDIA AI Enterprise software platform, self-hosted NIM models include ongoing security updates and are backed by enterprise-grade support. Developers can access NIM for free-to-start testing of IBM Granite code models at scale and build a proof of concept (POC) by connecting applications to the NVIDIA-hosted API endpoint running on a fully accelerated stack.

By bringing Granite code models to the NVIDIA API catalog, IBM is enabling enterprises to easily use industry-leading models for trusted code generation and translation, GPU infrastructure, and inference management software capabilities for price-performance optimization. More Granite models will soon be available on the NVIDIA API catalog as IBM and NVIDIA continue to expand their collaboration.

Explore what Granite and NVIDIA can do for you

More from Artificial intelligence

Putting AI to work in finance: Using generative AI for transformational change

2 min read - Finance leaders are no strangers to the complexities and challenges that come with driving business growth. From navigating the intricacies of enterprise-wide digitization to adapting to shifting customer spending habits, the responsibilities of a CFO have never been more multifaceted. Amidst this complexity lies an opportunity. CFOs can harness the transformative power of generative AI (gen AI) to revolutionize finance operations and unlock new levels of efficiency, accuracy and insights. Generative AI is a game-changing technology that promises to reshape…

AI that’s ready for business starts with data that’s ready for AI

6 min read - By 2026, over 80% of enterprises will deploy AI APIs or generative AI applications. AI models and the data on which they're trained and fine-tuned can elevate applications from generic to impactful, offering tangible value to customers and businesses. For example, the Master’s generative AI-driven golf fan experience uses real-time and historical data to provide insights and commentary for over 20,000 video clips. The quality and quantity of data can make or break AI success, and organizations that effectively harness…

Applying generative AI to revolutionize telco network operations 

5 min read - Generative AI is shaping the future of telecommunications network operations. The potential applications for enhancing network operations include predicting the values of key performance indicators (KPIs), forecasting traffic congestion, enabling the move to prescriptive analytics, providing design advisory services and acting as network operations center (NOC) assistants.   In addition to these capabilities, generative AI can revolutionize drive tests, optimize network resource allocation, automate fault detection, optimize truck rolls and enhance customer experience through personalized services. Operators and suppliers are…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters