## Should batch-normalization/dropout/activation-function layers be used after the last fully connected layer?

1

I am using the following architechture:

3*(fully connected -> batch normalization -> relu -> dropout) -> fully connected


Should I add the batch normalization -> relu -> dropout part after the last fully connected layer as well (the output is positive anyway, so the relu wouldn't hurt I suppose)?