Indian Journal of Science and Technology
Year: 2016, Volume: 9, Issue: 19, Pages: 1-7
P. Chanthini* and K. Shyamala
*Author for correspondence
PG & Research Department of Computer Science, Dr. Ambedkar Government Arts College (Autonomous), University of Madras, Chennai-600 039, Tamil Nadu, India; [email protected]
Background: Neural networks are relatively crude computerized model based on the neural system of the human brain. The Complex problems may require sophisticated processing techniques to achieve practical speed. In human brain millions of neurons form a massively parallel information system. Method: A neural network is a parallel and distributed process. The parallel execution of the neural network is achieved up to the level of the training process. Parallelization approaches may work well on hardware implementations, a software package (SPANN), special purpose hardware and multicore CPUs through MPI. Findings: In order to achieve parallelism and speed up the training process, each neuron and full neural network is duplicated to multiple threads. In modern Microprocessor number of cores is rapidly increasing. So high-performance computing is a great challenge for the application developers. Improvements: Our future work directs to distribute neurons to multiple threads for parallel execution and duplicate the full neural network for parallel training.
Keywords: MPI, Multi-threaded and Multi-core System, OpenMP, Parallelization
Subscribe now for latest articles and news.