LLM acceleration: Apple cooperates with Nvidia

The ReDrafter software is designed to significantly speed up the execution of large language models on Nvidia GPUs. The tool is open source.

listen Print view
Apple Intelligence logo and icon

Apple Intelligence logo and icon: Apple has some catching up to do when it comes to AI.

(Image: Apple)

3 min. read

Apple has launched a project in collaboration with Nvidia to accelerate inferencing in large language models (LLMs), which are used to establish connections between tokens, among other things. During inferencing, AI accelerators execute AI algorithms that have already been trained.

In November, the company published an open-source software called Recurrent Drafter, or ReDrafter for short, in a paper including code on GitHub. Nivida itself is already using ReDrafter in its in-house TensorRT LLM framework, as the AI giant announced in a blog post. It is a "novel, speculative decoding technique" that helps developers to "significantly accelerate" workload performance on Nvidia GPU chips.

According to Apple, ReDrafter and TensorRT-LLM succeed in accelerating token generation by 2.7 times per second (in so-called greedy decoding). According to Apple, this was tested on a production model with several tens of billions of parameters. "The benchmark results show that this technology could significantly reduce the latency perceived by users." At the same time, performance and power are saved.

Videos by heise

According to Nvidia, speculative decoding is a process in which LLM inferencing is accelerated by generating multiple tokens in parallel. "Smaller 'design' modules are used to predict future tokens, which are then checked by the main model." With this method, the output quality is as good as before, "while response times are significantly reduced, especially with low traffic". This makes better use of the available resources.

Apple emphasizes that in parallel to its work in the server area with Nvidia GPUs, it is also working on accelerating LLM inference on Apple Silicon devices. Like its competitors from Meta or OpenAI, the iPhone company apparently relies heavily on Nvidia technology when training its own LLMs. The rest of the industry should therefore also benefit from the work of the AI team. With open source models, ReDrafter is said to have succeeded in being up to 3.5 tokens faster per generation step. This has surpassed the performance of earlier speculative decoding methods.

The latest version of the TensorRT LLM framework contains both the necessary drafting and validation logic in a single engine, writes Nvidia. This minimizes overhead. The collaboration with Apple has made TensorRT-LLM "more powerful and flexible".

Empfohlener redaktioneller Inhalt

Mit Ihrer Zustimmung wird hier ein externer Preisvergleich (heise Preisvergleich) geladen.

Ich bin damit einverstanden, dass mir externe Inhalte angezeigt werden. Damit können personenbezogene Daten an Drittplattformen (heise Preisvergleich) übermittelt werden. Mehr dazu in unserer Datenschutzerklärung.

(bsc)

Don't miss any news – follow us on Facebook, LinkedIn or Mastodon.

This article was originally published in German. It was translated with technical assistance and editorially reviewed before publication.