Özet
Long Short Term Memory (LSTM) networks as one of the most used Recurrent Neural Networks (RNN) structures offer high accuracy for sequence learning tasks. However, it is challenging to offer low latency and high throughput while satisfying the low power constraints at the same time for computationally expensive LSTM operations. This work offers a two-pronged approach to accelerate inference in RNN networks. First, linear quantization technique is applied to reduce the complexity of operations, power consumption and required memory resources. Then, a new activation implementation method is proposed, called lookupx, to accelerate sigmoid function computation during inference. It is shown that lowering precision to 4-bit integer numbers for inputs causes only 2% accuracy loss and the lookupx activation methodology has 1.9x better performance and 50x lower power consumption while decreasing the required chip area 1.2x compared to integer domain activation functions with the same accuracy result.
Orijinal dil | İngilizce |
---|---|
Ana bilgisayar yayını başlığı | ICECS 2023 - 2023 30th IEEE International Conference on Electronics, Circuits and Systems |
Ana bilgisayar yayını alt yazısı | Technosapiens for Saving Humanity |
Yayınlayan | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Elektronik) | 9798350326499 |
DOI'lar | |
Yayın durumu | Yayınlandı - 2023 |
Etkinlik | 30th IEEE International Conference on Electronics, Circuits and Systems, ICECS 2023 - Istanbul, Turkey Süre: 4 Ara 2023 → 7 Ara 2023 |
Yayın serisi
Adı | ICECS 2023 - 2023 30th IEEE International Conference on Electronics, Circuits and Systems: Technosapiens for Saving Humanity |
---|
???event.eventtypes.event.conference???
???event.eventtypes.event.conference??? | 30th IEEE International Conference on Electronics, Circuits and Systems, ICECS 2023 |
---|---|
Ülke/Bölge | Turkey |
Şehir | Istanbul |
Periyot | 4/12/23 → 7/12/23 |
Bibliyografik not
Publisher Copyright:© 2023 IEEE.