Tommaso Guerrini
Tommaso Guerrini

Reputation: 1718

Training neural networks in parallel, theoretical framework

I'm working with neural networks and currently exploring other ways than my cpu to accelerate the training process. Using Keras with Tensorflow on a GPU machine in AWS I was able to really speed up the process.

How does this parallelization work in theory? (I am not considering cluster parallelization)

I struggled to find a reference so I ask you for an insight into how it works.

Thank you in advance

Upvotes: 2

Views: 641

Answers (1)

Nat
Nat

Reputation: 1096

While I'm not familiar with the relevant implementations, training a neural network means optimizing that neural network's parameters, e.g. the weights of neural connections.

Common optimization methods include quasi-Newton methods, which are rate-limited by matrix math operations like matrix inversion. GPU's can greatly improve calculation speed here.

References:

Upvotes: 1

Related Questions