Solventum
Global Technology Park
Devarabisanahalli, Varthur Hobli, Outer Ring Road
Bengaluru, Karnataka, India - 560103
Email:
kunal [dot] banerjee [at] solventum [dot] com
[CV]
About
I have recently joined Solventum as the founding member of the AI Science team in India, where we focus on applying AI — particularly Generative AI — to solve complex problems in the healthcare domain for global markets, with a strong emphasis on the EMEA region.
Prior to this, I spent nearly 5.5 years at Walmart as a Principal Data Scientist, working on a wide range of initiatives including text extraction from images, anomaly prediction in production channels, and the introduction of sequential testing within the A/B experimentation framework. I also led the development of the world’s first super-resolution technique optimized for improving text legibility in low-resolution images. Until my departure, I was responsible for the implementation and monitoring of company-wide Responsible GenAI. Through this role, I gained extensive experience across Computer Vision, Natural Language Processing, Recommender Systems, A/B Testing, AutoML, and AI Governance.
Earlier, I was a Research Scientist at Intel Labs, where my research focused on kernel optimization of deep learning workloads on Intel architectures. My work on optimized implementations of convolution (Winograd), RNNs, LSTMs, and GRUs is available in open-source libraries such as, LIBXSMM and Intel MKL-DNN. These libraries have been adopted across major frameworks and products including PyTorch, TensorFlow, Caffe, MS CNTK, Apache MXNet, Chainer, and OpenVINO, enabling significant performance improvements on Intel silicon. My research interests also include low-precision deep neural networks. In collaboration with colleagues at Intel Labs, I co-developed Ternary Residual Networks, which use 8-bit activations and 2-bit weights (with residual edges where required). I am also a co-author of the foundational paper on the BFLOAT16 datatype, a collaborative effort between Intel and Facebook. Additionally, I contributed to Intel’s deep learning training accelerator as part of the Intel Artificial Intelligence Products Group.
Before joining Intel, I earned my PhD in Computer Science from IIT Kharagpur. My doctoral research was supported by multiple fellowships, and my dissertation received the Best PhD Thesis Award from several venues, including VLSI Design, ISVLSI, and the India Electronics & Semiconductor Association (IESA).
Verification of Code Motion Techniques using Value Propagation. Kunal Banerjee, Chandan Karfa, Dipankar Sarkar, Chittaranjan Mandal.
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (TCAD), vol. 33, no. 8, 2014, pp: 1180-1193.
Harnessing Deep Learning via a Single Building Block.
Evangelos Georganas, Kunal Banerjee, Dhiraj Kalamkar, Sasikanth Avancha, Anand Venkat, Michael Anderson, Greg Henry, Hans Pabst, Alexander Heinecke.
International Parallel & Distributed Processing Symposium (IPDPS), May 2020, pp: 222-233. [arXiv]
(Preliminary version accepted as research poster in SuperComputing 2019.)
Reliability Evaluation of Compressed Deep Learning Models.
Brunno F. Goldstein, Sudarshan Srinivasan, Dipankar Das, Kunal Banerjee, Leandro Santiago, Victor C. Ferreira, Alexandre S. Nery, Sandip Kundu, Felipe M. G. Franca.
Latin American Symposium on Circuits and Systems (LASCAS), February 2020, pp: 1-5.
Training Google Neural Machine Translation on an Intel CPU Cluster.
Dhiraj Kalamkar, Kunal Banerjee, Sudarshan Srinivasan, Srinivas Sridharan, Evangelos Georganas, Mikhail E. Smorkalov, Cong Xu, Alexander Heinecke.
International Conference on Cluster Computing (CLUSTER), September 2019, pp: 1-10.
Anatomy Of High-Performance Deep Learning Convolutions On SIMD Architectures.
Evangelos Georganas, Sasikanth Avancha, Kunal Banerjee, Dhiraj Kalamkar, Greg Henry, Hans Pabst, Alexander Heinecke.
International Conference for High Performance Computing, Networking, Storage, and Analysis (SuperComputing), November 2018, pp: 66:1-66:12. [arXiv]
K-TanH: Hardware Efficient Activations For Deep Learning.
Abhisek Kundu, Alexander Heinecke, Dhiraj Kalamkar, Sudarshan Srinivasan, Eric C. Qin, Naveen K. Mellempudi, Dipankar Das, Kunal Banerjee, Bharat Kaul, Pradeep Dubey.
Preprint on arXiv, September 2019, arXiv:1909.07729.
A Study of BFLOAT16 for Deep Learning Training.
Dhiraj Kalamkar, Dheevatsa Mudigere, Naveen Mellempudi, Dipankar Das, Kunal Banerjee, Sasikanth Avancha, Dharma Teja Vooturi, Nataraj Jammalamadaka, Jianyu Huang, Hector Yuen, Jiyan Yang, Jongsoo Park, Alexander Heinecke, Evangelos Georganas, Sudarshan Srinivasan, Abhisek Kundu, Misha Smelyanskiy, Bharat Kaul, Pradeep Dubey.
Preprint on arXiv, May 2019, arXiv:1905.12322.