Search All of the Math Forum:

Views expressed in these public forums are not endorsed by NCTM or The Math Forum.

Notice: We are no longer accepting new posts, but the forums will continue to be readable.

Topic: Problem with 1-step ahead prediction in neural network
Replies: 9   Last Post: Oct 23, 2013 6:23 AM

 Messages: [ Previous | Next ]
 phuong Posts: 20 Registered: 8/23/13
Re: Problem with 1-step ahead prediction in neural network
Posted: Oct 19, 2013 2:26 PM

"Greg Heath" <heath@alumni.brown.edu> wrote in message <l3t37j\$bkp\$1@newscl01ah.mathworks.com>...
> "phuong" wrote in message <l3s2ha\$4qv\$1@newscl01ah.mathworks.com>...
> > Hi everybody,
> > I having a trouble with 1-step ahead of neural.
> > When I train network with fix parameter, I received another weight (IW,LW,b).
> > I know the reason is random intial weights. But why can we believe the predict result in 1-step if it alway changes for every train. May be the network not convergence. Because when it convergence, we just have only solution( or approximate solution). So is the network convergence?
> > All of things make the test result for 100 new predicted by neural network have many results, and some times different between so large.
> > Thank you very much.
> > Phuong

>
> The only problem is your assumption that there is only one solution. For any I-H-O network configuration with tansig hidden nodes there are (2^H)*H!-1 other nets that are equivalent. For the default value of H=10, there are (2^10)*factorial(10) = 3,715,891,200
> equivalent nets.
> 1. There are H! equivalent nets that only differ by the way they are ordered.
> 2. Since tansig is an odd function, for each of those orderings there are two equivalent
> nets that only differ by the polarity of the weights connected to one of the H hidden nodes.
>
> To make things worse, there can be local minima that are not global minima. The corresponding solutions range from excellent to very poor. Finally, there are other reasons
> (e.g., maximum mu in trainlm) that minimization searches fail.
>
> That is why I now use Ntrials = max(10,30/Ntst) random weight initializations for each candidate value of H.
>
> Hope this helps.
>
> Greg

Sorry, I don't understand your way. As i understand, you will train network Ntrials times, ok? And what is the next, compute the mean of result or what? Please help me more detail.
One more, i agree we have see H! net but i think the weight set is the same just change the oders, right? and If this idea right, i think the mse not change.
Thank you so much
Phuong

Date Subject Author
10/18/13 phuong
10/19/13 Greg Heath
10/19/13 phuong
10/19/13 Greg Heath
10/19/13 phuong
10/20/13 Greg Heath
10/21/13 phuong
10/21/13 Greg Heath
10/23/13 phuong
10/23/13 Greg Heath