Talks | Parallel Computing
2024
Samuele Fonio Bruno Casella Oussama Harrak
Federated Adaboost for Survival Analysis Miscellaneous
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2nd Workshop on Advancements in Federated Learning (WAFL), 2024.
Abstract | Links | BibTeX | Tags: ai, epi, fl, icsc
@misc{23:ecmlpkdd:fedsurvboost,
title = {Federated Adaboost for Survival Analysis},
author = {Samuele Fonio Bruno Casella Oussama Harrak},
url = {https://datacloud.di.unito.it/index.php/s/DtXiQfne6BEC235},
year = {2024},
date = {2024-09-01},
address = {Vilnius, Lithuania},
abstract = {This work proposes FedSurvBoost, a federated learning pipeline for survival analysis based on the AdaBoost.F algorithm, which iteratively aggregates the best local weak hypotheses. Our method extends AdaBoost.F by removing the dependence on the number of classes coefficient from the computation of the weights of the best model. This makes it suitable for regression tasks, such as survival analysis. We show the effectiveness of our approach by comparing it with state-of-the-art methods, specifically developed for survival analysis problems, on two common survival datasets.},
howpublished = {European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2nd Workshop on Advancements in Federated Learning (WAFL)},
keywords = {ai, epi, fl, icsc},
pubstate = {published},
tppubtype = {misc}
}
2023
Marco Aldinucci
Experimenting with Systems for Decentralized Machine Learning Miscellaneous
NVidia GTC 2023, 2023.
Abstract | Links | BibTeX | Tags: across, admire, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa
@misc{23:gtc:fl,
title = {Experimenting with Systems for Decentralized Machine Learning},
author = {Marco Aldinucci},
url = {https://datacloud.di.unito.it/index.php/s/oyLt7xwkbKxz65c},
year = {2023},
date = {2023-03-01},
abstract = {Decentralized machine learning (DML) enables collaborative machine learning without centralized input data. Federated learning (FL) and edge inference (EI) are examples of DML. Collaboration naturally happens at the edge of a distributed system with inherently distributed data. While tools for DML are starting to flourish, much needs to be done to get more flexible and portable tools to experiment with novel techniques, non-fully connected topologies, multiple data domains, and asynchronous collaboration schemes. We'll present recent advances in DML, aiming to improve usability in data centers and, at the edge, to widen the class of models extending FL to non-DNN paradigms, to improve the accuracy of models controlling normalization and frequency of communications, and to boost data privacy though generative adversarial networks. Prerequisites: Intermediate understanding of machine learning methods and distributed & parallel computing.},
howpublished = {NVidia GTC 2023},
keywords = {across, admire, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa},
pubstate = {published},
tppubtype = {misc}
}
Marco Aldinucci
HPC4AI: The Research on AI beyond the public cloud Miscellaneous
CENTAI kick-off meeting, 2023.
Links | BibTeX | Tags: across, admire, brainteaser, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa
@misc{23:CENTAI:hpc4ai,
title = {HPC4AI: The Research on AI beyond the public cloud},
author = {Marco Aldinucci},
url = {https://datacloud.di.unito.it/index.php/s/PZXjPm8sfKTmTGb},
year = {2023},
date = {2023-03-01},
address = {Torino, Italy},
howpublished = {CENTAI kick-off meeting},
keywords = {across, admire, brainteaser, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa},
pubstate = {published},
tppubtype = {misc}
}
Bruno Casella, Samuele Fonio
Architecture-Based FedAvg for Vertical Federated Learning Miscellaneous
2023.
Abstract | Links | BibTeX | Tags: ai, epi, fl, icsc
@misc{23:casella:architecturalfedavgtalk,
title = {Architecture-Based FedAvg for Vertical Federated Learning},
author = {Bruno Casella and Samuele Fonio},
url = {https://datacloud.di.unito.it/index.php/s/kJQxnqG4d2ZSicK},
year = {2023},
date = {2023-01-01},
booktitle = {Proceedings of the 3rd Workshop on Distributed Machine Learning for the Intelligent Computing Continuum (DML-ICC), IEEE/ACM UCC 2023, Taormina, Italy, 4 December 2023},
abstract = {Federated Learning (FL) has emerged as a promising solution to address privacy concerns by collaboratively training Deep Learning (DL) models across distributed parties. This work proposes an architecture-based aggregation strategy in Vertical FL, where parties hold data with different attributes but shared instances. Our approach leverages the identical architectural parts, i.e. neural network layers, of different models to selectively aggregate weights, which is particularly relevant when collaborating with institutions holding different types of datasets, i.e., image, text, or tabular datasets. In a scenario where two entities train DL models, such as a Convolutional Neural Network (CNN) and a Multi-Layer Perceptron (MLP), our strategy computes the average only for architecturally identical segments. This preserves data-specific features learned from demographic and clinical data. We tested our approach on two clinical datasets, i.e., the COVID-CXR dataset and the ADNI study. Results show that our method achieves comparable results with the centralized scenario, in which all the data are collected in a single data lake, and benefits from FL generalizability. In particular, compared to the non-federated models, our proposed proof-of-concept model exhibits a slight performance loss on the COVID-CXR dataset (less than 8%), but outperforms ADNI models by up to 12%. Moreover, communication costs between training rounds are minimized by exchanging only the dense layer parameters.},
keywords = {ai, epi, fl, icsc},
pubstate = {published},
tppubtype = {misc}
}