Matrices as input and target for a neural network: "Number of inputs does not match net.numInputs."

2 views (last 30 days)
I'm attempting to train a neural network with a 425x12273 uint8 matrix as the network input, and a 1x12273 uint8 matrix as the network target. I have 12273 training examples and correct outputs stored in these matrices (the first and second ones, respectively). The neural network I've created takes in 425 inputs and ultimately outputs a single value.
When attempting to run (where X is 425x12273, T is 1x12273)
[net,tr] = train(net,X,T)
I get the error "Number of inputs does not match net.numInputs." How can I use these matrices as proper inputs to the network?
Note: I have tried
X = num2cell(X, 1)
T = num2cell(T, 1)
in order to get these matrices into cell array format; the error persists.

Answers (1)

MUHAMMED IRFAN
MUHAMMED IRFAN on 14 Jun 2018
Hello, Are you trying to attempt a classification problem ?? If so..
As I can understand , you are trying to train some 12273 elements , each of them are defined by their 425 features. You haven't mentioned in the question how many classes you want to classify them into .
You have just mentioned "correct outputs stored in these matrices (the first and second ones, respectively)". If you are trying to classify into 2 classes, your target should be og size 2 * 12273 instead of 1 * 12273 , each row corresponding to their respective classes.

Categories

Find more on Deep Learning Toolbox in Help Center and File Exchange

Products


Release

R2017b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!