April 16, 2024, 4:51 a.m. | Matthew DeLorenzo, Vasudev Gohil, Jeyavijayan Rajendran

cs.CL updates on arXiv.org arxiv.org

arXiv:2404.08806v1 Announce Type: new
Abstract: Large Language Models (LLMs) have proved effective and efficient in generating code, leading to their utilization within the hardware design process. Prior works evaluating LLMs' abilities for register transfer level code generation solely focus on functional correctness. However, the creativity associated with these LLMs, or the ability to generate novel and unique solutions, is a metric not as well understood, in part due to the challenge of quantifying this quality.
To address this research gap, …

abstract arxiv code code generation creativity cs.cl design focus functional hardware however language language models large language large language models llm llms prior process transfer type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Engineer - New Graduate

@ Applied Materials | Milan,ITA

Lead Machine Learning Scientist

@ Biogen | Cambridge, MA, United States