Talks | Parallel Computing
2023
Marco Aldinucci
Experimenting with Systems for Decentralized Machine Learning Miscellaneous
NVidia GTC 2023, 2023.
Abstract | Links | BibTeX | Tags: across, admire, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa
@misc{23:gtc:fl,
title = {Experimenting with Systems for Decentralized Machine Learning},
author = {Marco Aldinucci},
url = {https://datacloud.di.unito.it/index.php/s/oyLt7xwkbKxz65c},
year = {2023},
date = {2023-03-01},
abstract = {Decentralized machine learning (DML) enables collaborative machine learning without centralized input data. Federated learning (FL) and edge inference (EI) are examples of DML. Collaboration naturally happens at the edge of a distributed system with inherently distributed data. While tools for DML are starting to flourish, much needs to be done to get more flexible and portable tools to experiment with novel techniques, non-fully connected topologies, multiple data domains, and asynchronous collaboration schemes. We'll present recent advances in DML, aiming to improve usability in data centers and, at the edge, to widen the class of models extending FL to non-DNN paradigms, to improve the accuracy of models controlling normalization and frequency of communications, and to boost data privacy though generative adversarial networks. Prerequisites: Intermediate understanding of machine learning methods and distributed & parallel computing.},
howpublished = {NVidia GTC 2023},
keywords = {across, admire, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa},
pubstate = {published},
tppubtype = {misc}
}
Decentralized machine learning (DML) enables collaborative machine learning without centralized input data. Federated learning (FL) and edge inference (EI) are examples of DML. Collaboration naturally happens at the edge of a distributed system with inherently distributed data. While tools for DML are starting to flourish, much needs to be done to get more flexible and portable tools to experiment with novel techniques, non-fully connected topologies, multiple data domains, and asynchronous collaboration schemes. We'll present recent advances in DML, aiming to improve usability in data centers and, at the edge, to widen the class of models extending FL to non-DNN paradigms, to improve the accuracy of models controlling normalization and frequency of communications, and to boost data privacy though generative adversarial networks. Prerequisites: Intermediate understanding of machine learning methods and distributed & parallel computing.
Marco Aldinucci
HPC4AI: The Research on AI beyond the public cloud Miscellaneous
CENTAI kick-off meeting, 2023.
Links | BibTeX | Tags: across, admire, brainteaser, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa
@misc{23:CENTAI:hpc4ai,
title = {HPC4AI: The Research on AI beyond the public cloud},
author = {Marco Aldinucci},
url = {https://datacloud.di.unito.it/index.php/s/PZXjPm8sfKTmTGb},
year = {2023},
date = {2023-03-01},
address = {Torino, Italy},
howpublished = {CENTAI kick-off meeting},
keywords = {across, admire, brainteaser, epi, eumaster4hpc, eupex, eupilot, hpc4ai, space, textarossa},
pubstate = {published},
tppubtype = {misc}
}
Bruno Casella, Samuele Fonio
Architecture-Based FedAvg for Vertical Federated Learning Miscellaneous
2023, (https://iris.unito.it/bitstream/2318/1949730/1/HALF_HVL_for_DML_ICC23___Taormina-2.pdf).
Abstract | Links | BibTeX | Tags: ai, epi, fl, icsc
@misc{23:casella:architecturalfedavg,
title = {Architecture-Based FedAvg for Vertical Federated Learning},
author = {Bruno Casella and Samuele Fonio},
url = {https://datacloud.di.unito.it/index.php/s/kJQxnqG4d2ZSicK},
doi = {10.1109/ICCVW60793.2023.00362},
year = {2023},
date = {2023-01-01},
booktitle = {Proceedings of the 3rd Workshop on Distributed Machine Learning for the Intelligent Computing Continuum (DML-ICC), IEEE/ACM UCC 2023, Taormina, Italy, 4 December 2023},
abstract = {Federated Learning (FL) has emerged as a promising solution to address privacy concerns by collaboratively training Deep Learning (DL) models across distributed parties. This work proposes an architecture-based aggregation strategy in Vertical FL, where parties hold data with different attributes but shared instances. Our approach leverages the identical architectural parts, i.e. neural network layers, of different models to selectively aggregate weights, which is particularly relevant when collaborating with institutions holding different types of datasets, i.e., image, text, or tabular datasets. In a scenario where two entities train DL models, such as a Convolutional Neural Network (CNN) and a Multi-Layer Perceptron (MLP), our strategy computes the average only for architecturally identical segments. This preserves data-specific features learned from demographic and clinical data. We tested our approach on two clinical datasets, i.e., the COVID-CXR dataset and the ADNI study. Results show that our method achieves comparable results with the centralized scenario, in which all the data are collected in a single data lake, and benefits from FL generalizability. In particular, compared to the non-federated models, our proposed proof-of-concept model exhibits a slight performance loss on the COVID-CXR dataset (less than 8%), but outperforms ADNI models by up to 12%. Moreover, communication costs between training rounds are minimized by exchanging only the dense layer parameters.},
note = {https://iris.unito.it/bitstream/2318/1949730/1/HALF_HVL_for_DML_ICC23___Taormina-2.pdf},
keywords = {ai, epi, fl, icsc},
pubstate = {published},
tppubtype = {misc}
}
Federated Learning (FL) has emerged as a promising solution to address privacy concerns by collaboratively training Deep Learning (DL) models across distributed parties. This work proposes an architecture-based aggregation strategy in Vertical FL, where parties hold data with different attributes but shared instances. Our approach leverages the identical architectural parts, i.e. neural network layers, of different models to selectively aggregate weights, which is particularly relevant when collaborating with institutions holding different types of datasets, i.e., image, text, or tabular datasets. In a scenario where two entities train DL models, such as a Convolutional Neural Network (CNN) and a Multi-Layer Perceptron (MLP), our strategy computes the average only for architecturally identical segments. This preserves data-specific features learned from demographic and clinical data. We tested our approach on two clinical datasets, i.e., the COVID-CXR dataset and the ADNI study. Results show that our method achieves comparable results with the centralized scenario, in which all the data are collected in a single data lake, and benefits from FL generalizability. In particular, compared to the non-federated models, our proposed proof-of-concept model exhibits a slight performance loss on the COVID-CXR dataset (less than 8%), but outperforms ADNI models by up to 12%. Moreover, communication costs between training rounds are minimized by exchanging only the dense layer parameters.