Aug 20, 2024 · In this work, we evaluate new commercial and open-source models of varying sizes against an improved VerilogEval benchmark suite.
Aug 20, 2024 · The enhanced VerilogEval benchmark provides a more robust framework for evaluating the performance of large-language models (LLMs) on digital ...
Aug 20, 2024 · We enhance VerilogEval's infrastructure and dataset by automatically classifying failures, introduce new prompts for supporting in-context ...
This work enhances VerilogEval's infrastructure and dataset by automatically classifying failures, introduce new prompts for supporting in-context learning ...
Aug 20, 2024 · The study aimed to improve the evaluation of large-language models (LLMs) in hardware code generation, specifically focusing on Verilog.
Aug 20, 2024 · The paper revisits the VerilogEval benchmark to evaluate newer large language models (LLMs) and in-context learning techniques for generating ...
Aug 21, 2024 · VerilogEval evaluates an LLM's ability to generate RTL code. In this new major version, we've added several enhancements: • A new spec-to-RTL ...
Aug 21, 2024 · Revisiting VerilogEval: Newer LLMs, In-Context Learning, and Specification-to-RTL Tasks paper: https://arxiv.org/abs/2408.11053.
To address this gap, the open-source VerilogEval benchmark was released in 2023, providing a consistent evaluation framework for LLMs on code completion tasks.