Agenda
MSc ME Thesis Presentation
- Tuesday, 29 November 2022
- 15:00-15:40
- HB17.140
A New Logarithmic Quantization Technique and Corresponding Processing Element Design for CNN Accelerators
Longxing Jiang
Convolutional Neural Networks (CNN) have become a popular solution for computer vision problems. However, due to the high data volumes and intensive computation involved in CNNs, deploying CNNs on low-power hardware systems is still challenging. The power consumption of CNNs can be prohibitive in the most common implementation platforms: CPUs and GPUs. Therefore, hardware accelerators that can exploit CNN parallelism and methods to reduce the computation burden or memory requirements are still hot research topics. Quantization is one of these methods.
One suitable quantization strategy for low-power deployments is logarithmic quantization.
Logarithmic quantization for Convolutional Neural Networks (CNN): a) fits well typical weights and activation distributions, and b) allows the replacement of the multiplication operation by a shift operation that can be implemented with fewer hardware resources. In this thesis, a new quantization method named Jumping Log Quantization (JLQ) is proposed. The key idea of JLQ is to extend the quantization range, by adding a coefficient parameter ”s” in the power of two exponents (2sx+i ).
This quantization strategy skips some values from the standard logarithmic quantization. In addition, a small hardware-friendly optimization called weight de-zeroing is proposed in this work. Zero-valued weights that cannot be performed by a single shift operation are all replaced with logarithmic weights to reduce hardware resources with little accuracy loss.
To implement the Multiply-And-Accumulate (MAC) operation (needed to compute convolutions) when the weights are JLQ-ed and dezeroed, a new Processing Element (PE) have been developed. This new PE uses a modified barrel shifter that can efficiently avoid the skipped values.
Resource utilization, area, and power consumption of the new PE standing alone and in a systolic array prototype are reported. The results show that JLQ performs better than other state-of-the-art logarithmic quantization methods when the bit width of the operands becomes very small.
Agenda
- Wed, 11 Mar 2026
- 17:30
- Aula Senaatszaal
PhD Thesis Defence
Simin Zhu
Towards Robust Radar Perception in Autonomous Vehicles: Deep Learning Methods for Motion Estimation, Radar Calibration, and Scene Segmentation
- Thu, 30 Apr 2026
- 12:30
- Aula Senaatszaal
PhD Thesis Defence
Yanbin He
Kronecker Compressed Sensing With Structured Sparsity
Algorithms, guarantees, and applications
- Thu, 21 May 2026
- 10:00
- Aula Senaatszaal
PhD Thesis Defence
Yanbo Wang
Compositional Generative Models: for Generalizable Scene Generation and Understanding
building intelligent agents with the flexible, systematic compositional imagination characteristic of human cognition