ON THE PROBLEM OF PREDICTING CALCULATION TIME NEEDED FOR NEURAL NETWORK EXECUTED BY MEANS OF GPU IN CASE OF CONVOLUTION NEURAL NETWORKS
( Pp. 47-51)
More about authors
Buryak Dmitry Yurievich
kand. fiz.-mat. nauk, starshiy inzhener-issledovatel
Branch of LG Electronics Popova Nina Nikolaevna kandidat fiziko-matematicheskih nauk; docent fakulteta vychislitelnoy matematiki i kibernetiki
Lomonosov Moscow State University
Branch of LG Electronics Popova Nina Nikolaevna kandidat fiziko-matematicheskih nauk; docent fakulteta vychislitelnoy matematiki i kibernetiki
Lomonosov Moscow State University
Abstract:
Computation performance of GPU devices has grown significantly in recent time. After CUDA architecture has appeared researchers could make active use of GPU devices in their work including nanotechnology area. However in many cases it is difficult to predict acceleration factor for an algorithm after its implementation by using GPU and consequently to estimate computational efficiency of this algorithm. Thus the task of computational performance prediction of an algorithm implemented using GPU is crucial. This work describes computational performance prediction model for algorithms based on artificial neural networks. Neural network depends on large amount of hyperparameters, which are defined on the architecture design stage, and affect its execution speed and results accuracy. A process of selecting these parameters values could take a long time. Application of prediction approaches allows to reduce time needed for the selection stage and to increase precision of hyperparameters' estimations.
How to Cite:
Buryak D.Y., Popova N.N., (2017), ON THE PROBLEM OF PREDICTING CALCULATION TIME NEEDED FOR NEURAL NETWORK EXECUTED BY MEANS OF GPU IN CASE OF CONVOLUTION NEURAL NETWORKS. Computational Nanotechnology, 2 => 47-51.
Reference list:
Amani A., Mohammadyani D. Artificial Neural Networks: Applications in Nanotechnology. Chapter in book: Artificial Neural Networks - Application. book edited by Chi Leung Patrick Hui, ISBN 978-953-307-188-6, Published: April 11, 2011.
Baghsorkhi S., Delahaye M., Gropp W., Wen-mei H.,W. Analytical performance prediction for evaluation and tuning of GPGPU applications. Workshop on Exploiting Parallelism using GPUs and other Hardware-Assisted Methods (EPHAM 09), In conjunction with The International Symposium on Code Generation and Optimization (CGO), 2009.
Fortune S., Wyllie J. Parallelism in Random Access Machines. In Proc. ACM STOC, 1978.
Gibbons P. B., Matias Y., Ramachandran V. The queue-read queue-write asynchronous pram model. In In Proc. of EURO-PAR, 1996.
G mez-Luna J., Gonz lez-Linares J.M., Benavides J.I., Guil N. Performance models for CUDA streams on NVIDIA GeForce series. Technical Report, University of M laga, 2011.
Guo P., Wang L. Accurate CUDA performance modeling for sparse matrix-vector multiplication. International Conference on High Performance Computing and Simulation (HPCS), 2012.
Hasan K.S., Chatterjee A., Radhakrishnan S., Antonio J.K. Performance Prediction Model and Analysis for Compute-Intensive Tasks on GPUs. In: Hsu CH., Shi X., Salapura V. (eds) Network and Parallel Computing. NPC 2014.
Hopf M., Ertl T. Hardware Accelerated Wavelet Transformations. In Proc. EG Symposium on Visualization, 2000.
Kothapalli K., Mukherjee R., Rehman M. S., Patidar, S., Narayanan, P.J., Srinathan K. A performance prediction model for the CUDA GPGPU platform. 2009 International Conference on High Performance Computing (HiPC), 2009.
LeCun Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. Handwritten digit recognition with a back-propagation network, Advanced in Neural Information Processing, 1990a, vol.2, 1990.
Luo Y., Duraiswami R. Canny Edge Detection on Nvidia CUDA. In Proc. of IEEE Computer Vision and Pattern Recognition, 2008.
Nguyen H. GPU Gems 3. Addison-Wesley Professional, 2007.
Sacha G. M., Varona P. Artificial intelligence in nanotechnology. Nanotechnology, 24 452002, 2013.
Valiant L. G. A Bridging Model for Parallel Computation. Comm. ACM 33, 8, 1990.
Vineet V., Narayanan P.J. CUDA cuts: Fast Graph Cuts on the GPU. In Proceedings of the CVPR Workshop on Visual Computer Vision on GPUs, 2008.
Baghsorkhi S., Delahaye M., Gropp W., Wen-mei H.,W. Analytical performance prediction for evaluation and tuning of GPGPU applications. Workshop on Exploiting Parallelism using GPUs and other Hardware-Assisted Methods (EPHAM 09), In conjunction with The International Symposium on Code Generation and Optimization (CGO), 2009.
Fortune S., Wyllie J. Parallelism in Random Access Machines. In Proc. ACM STOC, 1978.
Gibbons P. B., Matias Y., Ramachandran V. The queue-read queue-write asynchronous pram model. In In Proc. of EURO-PAR, 1996.
G mez-Luna J., Gonz lez-Linares J.M., Benavides J.I., Guil N. Performance models for CUDA streams on NVIDIA GeForce series. Technical Report, University of M laga, 2011.
Guo P., Wang L. Accurate CUDA performance modeling for sparse matrix-vector multiplication. International Conference on High Performance Computing and Simulation (HPCS), 2012.
Hasan K.S., Chatterjee A., Radhakrishnan S., Antonio J.K. Performance Prediction Model and Analysis for Compute-Intensive Tasks on GPUs. In: Hsu CH., Shi X., Salapura V. (eds) Network and Parallel Computing. NPC 2014.
Hopf M., Ertl T. Hardware Accelerated Wavelet Transformations. In Proc. EG Symposium on Visualization, 2000.
Kothapalli K., Mukherjee R., Rehman M. S., Patidar, S., Narayanan, P.J., Srinathan K. A performance prediction model for the CUDA GPGPU platform. 2009 International Conference on High Performance Computing (HiPC), 2009.
LeCun Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. Handwritten digit recognition with a back-propagation network, Advanced in Neural Information Processing, 1990a, vol.2, 1990.
Luo Y., Duraiswami R. Canny Edge Detection on Nvidia CUDA. In Proc. of IEEE Computer Vision and Pattern Recognition, 2008.
Nguyen H. GPU Gems 3. Addison-Wesley Professional, 2007.
Sacha G. M., Varona P. Artificial intelligence in nanotechnology. Nanotechnology, 24 452002, 2013.
Valiant L. G. A Bridging Model for Parallel Computation. Comm. ACM 33, 8, 1990.
Vineet V., Narayanan P.J. CUDA cuts: Fast Graph Cuts on the GPU. In Proceedings of the CVPR Workshop on Visual Computer Vision on GPUs, 2008.
Keywords:
Artificial neural network, convolutional neural networks, parallel computing, motherboard GPU, prediction of execution time.
Related Articles
Artificial intelligence and machine learning Pages: 9-18 DOI: 10.33693/2313-223X-2022-9-3-9-18 Issue №21873
Artificial Intelligence Elements for the Task of Determining the Position of the Vehicle in the Image
YOLO
computer vision
neural networks
convolutional neural networks
image recognition
Show more
1. Mathematical modeling Pages: 7-13 Issue №10450
PARALLEL SIMULATIONS OF ELECTRIC FIELDS IN MASS-SPECTROMETER TRAP FOR INCREASING OF IONS MASSES MEASUREMENTS ACCURACY
mathematical modeling
parallel computing
the mass spectrometer
the behavior of ion clouds
Show more
Cybersecurity Pages: 35-43 DOI: 10.33693/2313-223X-2023-10-3-35-43 Issue №23683
The Development of the Load Balancer and the Parallel Module for Managing Associatively Protected Map Databases
relational database
map data
associative steganography
database management
data protection
Show more
COMPUTER COMPLEXES AND INFORMATION TECHNOLOGIES Pages: 35-38 Issue №3497
SUPERCOMPUTER TRENDS
supercomputers
computational nanotechnology
parallel computing
Show more
Informatics and Information Processing Pages: 151-161 DOI: 10.33693/2313-223X-2024-11-1-151-161 Issue №95355
Development of a Web Application for Intelligent Analysis of Customer Reviews Using a Modified seq2seq Model with an Attention Mechanism
intelligent analysis
marketing
seq2seq model
artificial neural network
web application
Show more