Logarithm-Approximate Floating-Point Multiplier for Hardware-efficient Inference in Probabilistic Circuits

Research output: Contribution to conferencePaperScientificpeer-review


Machine learning models are increasingly being deployed onto edge devices, for example, for smart sensing, reinforcing the need for reliable and effi- cient modeling families that can perform a variety of tasks in an uncertain world (e.g., classification, outlier detection) without re-deploying the model. Probabilistic circuits (PCs) offer a promising avenue for such scenarios as they support efficient and exact computation of various probabilistic inference tasks by design, in addition to having a sparse structure. A critical challenge towards hardware acceleration of PCs on edge devices is the high computational cost associated with mul- tiplications in the model. In this work, we propose the first approximate computing framework for energy-efficient PC computation. For this, we leverage addition-as-int approximate multipliers, which are significantly more energy-efficient than regular floating-point multipliers, while preserving computation accuracy. We analyze the expected approximation error and show through hardware simulation results that our approach leads to a significant reduction in energy consumption with low approximation error and provides a remedy for hardware acceleration of general-purpose probabilistic models.
Original languageEnglish
Number of pages6
Publication statusPublished - 13 Jul 2023
MoE publication typeNot Eligible
EventWorkshop on Tractable Probabilistic Modeling - Pittsburgh, Pittsburgh, United States
Duration: 4 Aug 20234 Aug 2023
Conference number: 6


WorkshopWorkshop on Tractable Probabilistic Modeling
Abbreviated titleTPM
Country/TerritoryUnited States
Internet address


Dive into the research topics of 'Logarithm-Approximate Floating-Point Multiplier for Hardware-efficient Inference in Probabilistic Circuits'. Together they form a unique fingerprint.

Cite this