Date: Nov 20, 2012 4:36 AM
Author: condor
Subject: GPU and Neural Network ... is this operation legitimate?

Hi, I need to compute many simple Neural Networks... thousands.

I would like to use a parallel for on the GPU.

The FIRST problem is that I am not able to compute the net on the GPU:

% Net creation
hiddenLayerSize = 10;
net = fitnet(hiddenLayerSize);
% Setup Division of Data for Training, Validation, Testing
net.divideParam.trainRatio = 70/100;
net.divideParam.valRatio = 15/100;
net.divideParam.testRatio = 15/100;
net.trainFcn = 'trainscg';
net.trainParam.showWindow = false;
net.trainParam.showCommandLine = false;
% Send input and target on the GPU
% Train
[net_new,~] = train(net,input_A,target_A)

I receive this error:
Error using network/train (line 293)
Number of samples (rows of gpuArrays) of data arguments do not match.

Is it possibile to compute the net on the GPU (I want that every core train a different net)? If yes, what am I doing wrong?