Cerence AI and Arm have joined forces to enhance CaLLM Edge, an embedded small language model, using Arm’s Kleidi software library. This collaboration focuses on optimizing CPU and GPU performance for real-time language processing at the edge, emphasizing the significance of edge computing and generative AI in the automotive sector.
Cerence AI has partnered with semiconductor manufacturer, Arm to enhance its embedded small language model (SLM), CaLLM Edge, using Arm’s Kleidi software library.
The collaboration aims to optimize CPU and GPU performance for real-time language processing at the edge, improving speed, efficiency, and privacy highlighting the growing importance of edge computing and generative AI in the automotive industry.
Arm’s Kleidi technology accelerates machine learning and neural network operations on Arm-based devices, addressing the challenges of limited compute power in vehicles. CaLLM Edge operates fully on Arm-based chipsets, enabling advanced in-car AI capabilities without relying on cloud connectivity.
“We are excited to partner with Arm to take CaLLM Edge to the next level, setting new standards for performance and efficiency in edge computing in the car,” says Nils Schanz, EVP, Product & Technology, Cerence AI. “By combining our expertise in AI-powered language models with Arm’s innovative library, we are continuing our journey to create a new era of voice-first experiences and next-generation AI applications in the automotive space, empowering consumers with smarter, faster, and more responsive in-car assistants.”
This partnership supports automakers in delivering smarter, faster, and more responsive AI-powered user experiences for drivers and setting new standards for in-car AI applications, enhancing safety and connectivity.
Related
AI/ML | ARM | automotive AI | Cerence AI | edge AI | edge computing | embedded LLS