Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Proc. AISTATS, Fort Lauderdale, Florida, USA, Apr. 2017. • [Kairouz+, FTML2021] P. Kairouz et al., "Advances and open problems in federated learning", Foundations and Trends in Machine Learning: Vol. 14: No. 1–2, pp 1-210, 2021. • [Zinkevich+, NeurIPS2010] M. Zinkevich, M. Weimer, L. Li, and A. Smola, “Parallelized stochastic gradient descent,” in Proc. NeurIPS, Vancouver, Canada, 2010, pp. 1–9. • [Li+, MLSys2020] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” in Proc. MLSys 2020, virtual conference, 2020, pp. 429– 450. • [Lian+, NeurIPS2018] X. Lian, C. Zhang, H. Zhang, C.-J. Hsieh, W. Zhang, and J. Liu, “Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent,” in Proc. NeurIPS, Long Beach, CA, USA, Jan. 2018, pp. 5330– 5340. • [Lian+, ICML2018] X. Lian, W. Zhang, C. Zhang, and J. Liu, “Asynchronous decentralized parallel stochastic gradient descent,” in Proc. ICML, Stockholm, Sweden, 2018. • [Wang+, ICML Workshop2019] J. Wang and G. Joshi, “Cooperative SGD: A unified framework for the design and analysis of communication-efficient SGD algorithms,” in Proc. ICML Workshop, Long Beach, CA, USA, 2019. • [Sato+, TCCN2021] K. Sato and D. Sugimura, “Rate-adapted decentralized learning over wireless networks,” IEEE Trans. Cogn. Commun. Netw., vol. 7, no. 4, pp. 1412–1429, 2021. • [Nishio+, ICC2019] T. Nishio and R. Yonetani, “Client selection for federated learning with heterogeneous resources in mobile edge,” in Proc. IEEE ICC, Shanghai, China, May 2019, pp. 1–7. • [Yu+, IoTJ2022] L. Yu, R. Albelaihi, X. Sun, N. Ansari, and M. Devetsikiotis, “Jointly optimizing client selection and resource management in wireless federated learning for Internet of Things,” IEEE Internet Things J., vol. 9, no. 6, pp.4385-4395, March 2022.