학술논문

FS-Boost: Communication-Efficient Federated Subtree-Based Gradient Boosting Decision Trees
Document Type
Conference
Source
2024 IEEE 21st Consumer Communications & Networking Conference (CCNC) Consumer Communications & Networking Conference (CCNC), 2024 IEEE 21st. :839-842 Jan, 2024
Subject
Communication, Networking and Broadcast Technologies
Computing and Processing
Robotics and Control Systems
Training
Costs
Federated learning
Distributed databases
Boosting
Numerical models
Decision trees
federated learning
gradient boosting decision trees
communication efficiency
edge computing
security
Language
ISSN
2331-9860
Abstract
Federated learning (FL) is a secure and distributed machine learning method in which clients learn cooperatively without disclosing private data to others. some decision tree-based FL have been proposed that employ gradient boosting decision trees (GBDT). However, previously proposed GBDT-based FL methods require sharing the entire decision tree, including the tree structure and leaf weights, for each synchronization step in training. This process inevitably results in extensive communication. To solve this problem, we propose FS-Boost-Federated Subtree-based GBDT, a horizontal FL method that reduces the communication cost. Suppose the maximum depth of the decision tree is set to d. In that case, a conventional GBDT trains only a single decision tree of depth d within one training round. In contrast, FS- Boost utilizes subtrees from depth 1 to d-1 generated in the learning process in addition to the whole trees. Sharing still-growing subtrees with other clients reduces the total amount of communication cost through accelerating model convergence. Our experiment results indicate that FS-Boost significantly reduced the communication cost by at least half in most cases while maintaining the accuracy.