The Larger the Better? Improved LLM Code-Generation via Budget Reallocation.

Michael Hassid, Tal Remez, Jonas Gehring, Roy Schwartz, Yossi Adi

Research output: Contribution to conferencePaperpeer-review

Abstract

It is a common belief that large language models (LLMs) are better than smaller-sized ones. However, larger models also require significantly more time and compute during inference. This begs the question: what happens when both models operate under the same budget? (e.g., compute, run-time). To address this question, we analyze code generation LLMs of various sizes and make comparisons such as running a 70B model once vs. generating five outputs from a 13B model. We consider a standard unit-test setup, which can be used to select the correct output from the smaller model. Our findings reveal that the repeated use of smaller models can yield consistent improvements, with gains of up to 15% across five tasks. On the other hand, in scenarios where unit-tests are unavailable, a ranking-based selection of candidates from the smaller model falls short of the performance of a single output from larger ones. Our results highlight the potential of using smaller models instead of larger ones, and the importance of studying approaches for ranking LLM outputs.
Original languageEnglish
Pages1-16
Number of pages16
StatePublished - 10 Jul 2024
Event1st Conference On Language Modeling, COLM 2024. - University of Pennsylvania, Philadelphia, United States
Duration: 7 Oct 20249 Oct 2024
Conference number: 1
https://2024.colmweb.org/

Conference

Conference1st Conference On Language Modeling, COLM 2024.
Country/TerritoryUnited States
CityPhiladelphia
Period7/10/249/10/24
Internet address

Keywords

  • Compute efficient LMs
  • code-generation
  • budget constraint
  • model size

Fingerprint

Dive into the research topics of 'The Larger the Better? Improved LLM Code-Generation via Budget Reallocation.'. Together they form a unique fingerprint.

Cite this