266 views
1 votes
1 votes
You encounter a classification task, and after training your network on 20 samples, the training converges, but the training loss is remarkably high. You decide to train the same network on 10,000 examples to address this issue. Is your approach to fixing the problem correct?

A) Yes, increasing the amount of data will likely resolve the bias problem.

B) No, increasing the amount of data is unlikely to solve the bias problem.

C) Yes, increasing the amount of data will also reduce the bias of the model.

D) No, a better approach would be to keep the same model architecture and increase the learning rate.

1 Answer

1 votes
1 votes
No, increasing the amount of data is unlikely to solve the bias problem. The model is suffering from bias, and a more effective approach would be to decrease bias by, for example, adding more layers or learnable parameters. It's also possible that the training converged to a local optimum, so training longer, using a better optimizer, or restarting from a different initialization could be more fruitful.

Related questions

208
views
1 answers
1 votes
rajveer43 asked Jan 29
208 views
Suppose that you are training a neural network for classification, but you notice that the training loss is much lower than the validation loss. ... probability √Increase $L2$ regularization weightIncrease the size of each hidden layer
246
views
1 answers
0 votes
rajveer43 asked Jan 27
246 views
What is Error Analysis?(i) The process of analyzing the performance of a model through metrics such as precision, recall or F1-score.(ii) The process of ... iv) The process of identifying which parts of your model contributed to the error.
506
views
2 answers
1 votes
rajveer43 asked Jan 16
506 views
Suppose we are performing leave-one-out (LOO) validation and $10$-fold cross validation on a dataset of size $100, 000$ to pick between $4$ different ... that need to be trained for LOO validation versus $10$-fold cross validation?Answer:
268
views
2 answers
1 votes
rajveer43 asked Jan 29
268 views
Suppose a classifier predicts each possible class with equal probability. If there are 10 classes, what will the cross-entropy error be on a single example?$− log(10)$−0.1 log(1)$− log(0.1)$−10 log(0.1)$