Asynchronous stochastic optimization for sequence training of deep neural networks

G Heigold, E McDermott, V Vanhoucke… - … , Speech and Signal …, 2014 - ieeexplore.ieee.org
… This paper explores asynchronous stochastic optimization for sequence training of deep
neural networks. … We investigated asynchronous stochastic gradient descent for sequence …

[PDF][PDF] Asynchronous stochastic optimization for sequence training of deep neural networks: Towards big data

E McDermott, G Heigold, PJ Moreno… - … Annual Conference of …, 2014 - isca-archive.org
optimization criterion. Given the difficulty of guaranteeing high quality transcripts for large
training … Paper organization: The asynchronous optimization framework for sequence training

Asynchronous optimization methods for efficient training of deep neural networks with guarantees

V Kungurtsev, M Egan, B Chatterjee… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
… stochastic asynchronous optimization for … asynchronously updated by concurrent processes.
To this end, we use a probabilistic model which captures key features of real asynchronous

AMPNet: Asynchronous model-parallel training for dynamic neural networks

AL Gaunt, MA Johnson, M Riechert, D Tarlow… - arXiv preprint arXiv …, 2017 - arxiv.org
… 6) For example on the QM9 dataset [30, 27] our implementation of gated graph sequence
neural network (GGSNN) [21] on a 16 core CPU runs 9x faster than a (manually optimized) …

Neuro-distributed cognitive adaptive optimization for training neural networks in a parallel and asynchronous manner

P Michailidis, IT Michailidis, S Gkelios… - Integrated …, 2023 - journals.sagepub.com
… and asynchronous ND-CAO training, the algorithm is identified as an efficient scheme to
train neural networks in a novel gradient-independent, distributed, and asynchronous manner, …

APapo: An asynchronous parallel optimization method for DNN models

S Liu, T Ju - Future Generation Computer Systems, 2024 - Elsevier
… , extended training duration, and low equipment utilization in … optimization of large-scale
deep neural network (DNN) models, this paper proposes an asynchronous parallel optimization

Advances in asynchronous parallel and distributed optimization

M Assran, A Aytekin, HR Feyzmahdavian… - Proceedings of the …, 2020 - ieeexplore.ieee.org
… In Section V, we include a numerical example illustrating how asynchronous decentralized
algorithms may be used for training deep neural networks. We conclude, in Section VI, with a …

Asynchronous optimization for machine learning

R Leblond - 2018 - theses.hal.science
… We introduce new training algorithms for a class of neural networks targeted at sequential
on introducing new fast asynchronous parallel incremental optimization algorithms which can …

Asynchronous evolution of deep neural network architectures

J Liang, H Shahrzad, R Miikkulainen - Applied Soft Computing, 2024 - Elsevier
… Such variation is especially prominent in the training of deep neural network architectures, …
a new asynchronous EA called AES designed for complex problems such as optimizing the …

[PDF][PDF] Sequence discriminative distributed training of long short-term memory recurrent neural networks

H Sak, O Vinyals, G Heigold, A Senior, E McDermott… - entropy, 2014 - isca-archive.org
train the models in a distributed manner using asynchronous … We compare two sequence
discriminative criteria – maximum … order to investigate all the above, we use ASGD optimization