Verifying Hardware CWEs in RTL Designs Generated by GenAI


A new technical paper titled “All Artificial, Less Intelligence: GenAI through the Lens of Formal Verification” was published by researchers at Infineon Technologies.

“Modern hardware designs have grown increasingly efficient and complex. However, they are often susceptible to Common Weakness Enumerations (CWEs). This paper is focused on the formal verification of CWEs in a dataset of hardware designs written in SystemVerilog from Regenerative Artificial Intelligence (AI) powered by Large Language Models (LLMs). We applied formal verification to categorize each hardware design as vulnerable or CWE-free. This dataset was generated by 4 different LLMs and features a unique set of designs for each of the 10 CWEs we target in our paper. We have associated the identified vulnerabilities with CWE numbers for a dataset of 60,000 generated SystemVerilog Register Transfer Level (RTL) code. It was also found that most LLMs are not aware of any hardware CWEs; hence they are usually not considered when generating the hardware code. Our study reveals that approximately 60% of the hardware designs generated by LLMs are prone to CWEs, posing potential safety and security risks. The dataset could be ideal for training LLMs and Machine Learning (ML) algorithms to abstain from generating CWE-prone hardware designs.”

Find the technical paper here. Published March 2024.

Gadde, Deepak Narayan, Aman Kumar, Thomas Nalapat, Evgenii Rezunov and Fabio Cappellini. “All Artificial, Less Intelligence: GenAI through the Lens of Formal Verification.” (2024).arXiv:2403.16750v1

Leave a Reply

(Note: This name will be displayed publicly)