Irsoy, O.Alpaydın, Ahmet İbrahim Ethem2023-08-042023-08-042022978-303123027-10302-9743http://hdl.handle.net/10679/8564https://doi.org/10.1007/978-3-031-23028-8_16In a budding tree, every node is part internal node and part leaf. This allows representing the tree in a continuous parameter space and training it with backpropagation, like a neural network. Unlike a traditional tree whose construction is composed of two distinct stages of growing and pruning, “bud” nodes grow into subtrees or are pruned back dynamically during learning. In this work, we extend the budding tree and propose the distributed tree where the children use different and independent splits; hence, multiple paths in a tree can be traversed at the same time. In a traditional tree, the learned representations are local, that is, activation makes a soft selection among all the root-to-leaf paths in a tree, but the ability to combine multiple paths of the distributed tree gives it the power of a distributed representation, as in a traditional perceptron layer. Our experimental results show that distributed trees perform comparably or better than budding and traditional hard trees.engrestrictedAccessDistributed decision treesconferenceObject1381315216200092758020001610.1007/978-3-031-23028-8_16Decision treesHierarchical mixture of expertsLocal vs distributed representations2-s2.0-85147848552