ZDNETIBM is zooming along with new open-source Granite Large Language Models (LLM) releases every few months. Granite 3.1 is the latest generation model, building upon the success of Granite 3.0. The model offers enhanced capabilities and performance optimized for business applications.Also: Gemini Advanced users can now access Google’s most experimental modelThe family of Granite 3.1 models boasts an impressive 128K token context window, a substantial increase from their predecessors. This expansion allows the models to process and understand much larger amounts of text — equivalent to approximately 85,000 English words — enabling more comprehensive analysis and generation tasks. By comparison, OpenAI’s ChatGPT 3, which ignited the AI revolution, could handle only 2,000 tokens.Outperforming the competitionBig Blue claims its new Granite 8B Instruct model outperforms its rivals, such as Google Gemma 2, Meta Llama 3.1, and Qwen 2.5, on HuggingFace’s OpenLLM Leaderboard benchmarks.Also: Want generative AI LLMs integrated with your business data? You need RAGThe Granite 3.1 family includes dense models and Mixture of Experts (MoE) variants. IBM states its Granite 2B and 8B models are text-only dense LLMs trained on over 12 trillion data tokens. The dense models are designed to support tool-based use cases and for retrieval augmented generation (RAG), streamlining code generation, translation, and bug fixing.The MoE models are trained on over 10 trillion tokens of data. IBM claims these models are ideal for deployment in on-device applications with low latency. More