Logo PTI Logo FedCSIS

Communication Papers of the 19th Conference on Computer Science and Intelligence Systems (FedCSIS)

Annals of Computer Science and Information Systems, Volume 41

Gradient Boosting Trees and Large Language Models for Tabular Data Few-Shot Learning

DOI: http://dx.doi.org/10.15439/2024F1407

Citation: Communication Papers of the 19th Conference on Computer Science and Intelligence Systems (FedCSIS), M. Bolanowski, M. Ganzha, L. Maciaszek, M. Paprzycki, D. Ślęzak (eds). ACSIS, Vol. 41, pages 5359 ()

Full text

Abstract. Large Language Models (LLM) have brought numerous of new applications to Machine Learning (ML). In the context of tabular data (TD), recent studies show that TabLLM is a very powerful mechanism for few-shot-learning (FSL) applications, even if gradient boosting decisions trees (GBDT) have historically dominated the TD field. In this work we demonstrate that although LLMs are a viable alternative, the evidence suggests that baselines used to gauge performance can be improved. We replicate public benchmarks and our methodology improves LightGBM by 290\%, this is mainly driven by forcing node splitting with few samples, a critical step in FSL with GBDT. Our results show an advantage to TabLLM for 8 or fewer shots, but as the number of samples increases GBDT provides competitive performance at a fraction of runtime. For other real-life applications with vast number of samples, we found FSL still useful to improve model diversity, and when combined with ExtraTrees provides strong resilience to overfitting, our proposal was validated in a ML competition setting ranking 1st place.

References

  1. Ravid Shwartz-Ziv and Amitai Armon, “Tabular data: Deep learning is not all you need,” Information Fusion, vol. 81, pp. 84–90, 2022.
  2. Ruoxi Wang, Bin Fu, Gang Fu, and Mingliang Wang, “Deep & cross network for ad click predictions,” 2017.
  3. Jinsung Yoon, Yao Zhang, James Jordon, and Mihaela van der Schaar, “Vime: Extending the success of self- and semi-supervised learning to tabular domain,” in Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, Eds. 2020, vol. 33, pp. 11033–11043, Curran Associates, Inc.
  4. Yixuan Zhang, Jialiang Tong, Ziyi Wang, and Fengqiang Gao, “Customer transaction fraud detection using xgboost model,” in 2020 International Conference on Computer Engineering and Application (ICCEA), 2020, pp. 554–558.
  5. Zifeng Wang and Suzhen Li, “Data-driven risk assessment on urban pipeline network based on a cluster model,” Reliability Engineering & System Safety, vol. 196, pp. 106781, 2020.
  6. Xi Fang, Weijie Xu, Fiona Anting Tan, Jiani Zhang, Ziqing Hu, Yanjun Qi, Scott Nickleach, Diego Socolinsky, Srinivasan Sengamedu, and Christos Faloutsos, “Large language models(llms) on tabular data: Prediction, generation, and understanding – a survey,” 2024.
  7. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin, “Attention is all you need,” CoRR, vol. abs/1706.03762, 2017.
  8. Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Weizhu Chen, “Deberta: Decoding-enhanced BERT with disentangled attention,” CoRR, vol. abs/2006.03654, 2020.
  9. Mingxing Tan and Quoc V. Le, “Efficientnet: Rethinking model scaling for convolutional neural networks,” CoRR, vol. abs/1905.11946, 2019.
  10. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” in 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. 2021, OpenReview.net.
  11. Jiashi Li, Xin Xia, Wei Li, Huixia Li, Xing Wang, Xuefeng Xiao, Rui Wang, Min Zheng, and Xin Pan, “Next-vit: Next generation vision transformer for efficient deployment in realistic industrial scenarios,” 2022.
  12. Vadim Borisov, Tobias Leemann, Kathrin Seßler, Johannes Haug, Martin Pawelczyk, and Gjergji Kasneci, “Deep neural networks and tabular data: A survey,” CoRR, vol. abs/2110.01889, 2021.
  13. Dugang Liu, Pengxiang Cheng, Hong Zhu, Xing Tang, Yanyu Chen, Xiaoting Wang, Weike Pan, Zhong Ming, and Xiuqiang He, “Diwift: Discovering instance-wise influential features for tabular data,” 2022.
  14. Tianqi Chen and Carlos Guestrin, “Xgboost: A scalable tree boosting system,” CoRR, vol. abs/1603.02754, 2016.
  15. Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu, “Lightgbm: A highly efficient gradient boosting decision tree,” in Advances in Neural Information Processing Systems, I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds. 2017, vol. 30, Curran Associates, Inc.
  16. Anna Veronika Dorogush, Andrey Gulin, Gleb Gusev, Nikita Kazeev, Liudmila Ostroumova Prokhorenkova, and Aleksandr Vorobev, “Fighting biases with dynamic boosting,” CoRR, vol. abs/1706.09516, 2017.
  17. Ravid Shwartz-Ziv and Amitai Armon, “Tabular data: Deep learning is not all you need,” 2021.
  18. Tomaso Poggio, Andrzej Banburski, and Qianli Liao, “Theoretical issues in deep networks,” Proceedings of the National Academy of Sciences, vol. 117, no. 48, pp. 30039–30045, 2020.
  19. Léo Grinsztajn, Edouard Oyallon, and Gaël Varoquaux, “Why do tree-based models still outperform deep learning on tabular data?,” 2022.
  20. Humza Naveed, Asad Ullah Khan, Shi Qiu, Muhammad Saqib, Saeed Anwar, Muhammad Usman, Naveed Akhtar, Nick Barnes, and Ajmal Mian, “A comprehensive overview of large language models,” 2024.
  21. Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou, “Chain-of-thought prompting elicits reasoning in large language models,” 2023.
  22. Sébastien Bubeck, Varun Chandrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, Harsha Nori, Hamid Palangi, Marco Tulio Ribeiro, and Yi Zhang, “Sparks of artificial general intelligence: Early experiments with gpt-4,” 2023.
  23. Xue Jiang, Yihong Dong, Lecheng Wang, Zheng Fang, Qiwei Shang, Ge Li, Zhi Jin, and Wenpin Jiao, “Self-planning code generation with large language models,” 2023.
  24. Ian J. Goodfellow, Yoshua Bengio, and Aaron Courville, Deep Learning, MIT Press, Cambridge, MA, USA, 2016, http://www.deeplearningbook.org.
  25. Vadim Borisov, Tobias Leemann, Kathrin Seßler, Johannes Haug, Martin Pawelczyk, and Gjergji Kasneci, “Deep neural networks and tabular data: A survey,” IEEE Transactions on Neural Networks and Learning Systems, p. 1–21, 2024.
  26. Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, Rohan Anil, Zakaria Haque, Lichan Hong, Vihan Jain, Xiaobing Liu, and Hemal Shah, “Wide and deep learning for recommender systems,” in Proceedings of the 1st Workshop on Deep Learning for Recommender Systems, New York, NY, USA, 2016, DLRS 2016, p. 7–10, Association for Computing Machinery.
  27. Huifeng Guo, Ruiming Tang, Yunming Ye, Zhenguo Li, Xiuqiang He, and Zhenhua Dong, “Deepfm an end-to-end wide and deep learning framework for ctr prediction,” 2018.
  28. Haoran Luo, Fan Cheng, Heng Yu, and Yuqi Yi, “Sdtr: Soft decision tree regressor for tabular data,” IEEE Access, vol. 9, pp. 55999–56011, 2021.
  29. Guolin Ke, Zhenhui Xu, Jia Zhang, Jiang Bian, and Tie-Yan Liu, “Deepgbm: A deep learning framework distilled by gbdt for online prediction tasks,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 2019, KDD ’19, p. 384–394, Association for Computing Machinery.
  30. Guolin Ke, Jia Zhang, Zhenhui Xu, Jiang Bian, and Tie-Yan Liu, “TabNN: A universal neural network solution for tabular data,” 2019.
  31. Sergei Ivanov and Liudmila Prokhorenkova, “Boost then convolve: Gradient boosting meets graph neural networks,” 2021.
  32. Sercan O. Arik and Tomas Pfister, “Tabnet: Attentive interpretable tabular learning,” 2019.
  33. Zifeng Wang and Jimeng Sun, “Transtab: Learning transferable tabular transformers across tables,” 2022.
  34. Xin Huang, Ashish Khetan, Milan Cvitkovic, and Zohar Karnin, “Tab-transformer: Tabular data modeling using contextual embeddings,” 2020
  35. Gowthami Somepalli, Micah Goldblum, Avi Schwarzschild, C. Bayan Bruss, and Tom Goldstein, “Saint: Improved neural networks for tabular data via row attention and contrastive pre-training,” 2021.
  36. Jannik Kossen, Neil Band, Clare Lyle, Aidan N. Gomez, Tom Rainforth, and Yarin Gal, “Self-attention between datapoints: Going beyond individual input-output pairs in deep learning,” 2022.
  37. Omurhan A. Soysal and Mehmet Serdar Guzel, “An introduction to zero-shot learning: An essential review,” in 2020 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), 2020, pp. 1–4.
  38. Ruiyu Wang, Zifeng Wang, and Jimeng Sun, “Unipredict: Large language models are universal tabular classifiers,” 2024.
  39. Sebastian Bordt, Harsha Nori, and Rich Caruana, “Elephants never forget: Testing language models for memorization of tabular data,” 2024.
  40. Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, and Tong Wang, “An interpretable model with globally consistent explanations for credit risk,” 2018.
  41. Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, Alina Oprea, and Colin Raffel, “Extracting training data from large language models,” 2021.
  42. Nicholas Carlini, Daphne Ippolito, Matthew Jagielski, Katherine Lee, Florian Tramer, and Chiyuan Zhang, “Quantifying memorization across neural language models,” 2023.
  43. Stefan Hegselmann, Alejandro Buendia, Hunter Lang, Monica Agrawal, Xiaoyi Jiang, and David Sontag, “Tabllm: Few-shot classification of tabular data with large language models,” 2023.
  44. Arlind Kadra, Marius Lindauer, Frank Hutter, and Josif Grabocka, “Well-tuned simple nets excel on tabular datasets,” 2021.
  45. Jack Smith, J. Everhart, W. Dickson, W. Knowler, and Richard Johannes, “Using the adap learning algorithm to forcast the onset of diabetes mellitus,” Proceedings - Annual Symposium on Computer Applications in Medical Care, vol. 10, 11 1988.
  46. Steinbrunn William Pfisterer Matthias Janosi, Andras and Robert Detrano, “Heart Disease,” UCI Machine Learning Repository, 1988, http://dx.doi.org/https://doi.org/10.24432/C52P4X.
  47. Noah Hollmann, Samuel Müller, Katharina Eggensperger, and Frank Hutter, “Tabpfn: A transformer that solves small tabular classification problems in a second,” 2023.
  48. Sebastian Stawicki Andrzej Janusz, Dominik Slezak and Mariusz Rosiak, “Data-driven risk assessment on urban pipeline network based on a cluster model,” Proceedings of the 24th International Workshop on Concurrency, Specification and Programming, 2015.
  49. Ivana T. Dragovic Ana M. Poledica Milica M. Zukanovic Andrzej Janusz Dominik Slezak Aleksandar M. Rakicevic, Pavle D. Milosevic, “Predicting stock trends using common financial indicators: A summary of fedcsis 2024 data science challenge held on knowledgepit.ai platform,” Proceedings of FedCSIS 2024, 2024.
  50. C. Huertas and Q. Zhao, “On the irrelevance of machine learning algorithms and the importance of relativity,” in 2023 IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Los Alamitos, CA, USA, jul 2023, pp. 16–21, IEEE Computer Society.
  51. Abien Fred Agarap, “Deep learning using rectified linear units (relu),” 2019.
  52. Diederik P. Kingma and Jimmy Ba, “Adam: A method for stochastic optimization,” 2017.