Unconstrained Neural Network

Neural network with L1 regularization and weight pruning for sparse predictions

← Back to index

Model overview

Unconstrained neural network classifier using L1 regularization during training followed by post-training weight pruning to achieve sparsity. Unlike parameter-constrained models, this classifier uses ALL available tags as input features, relying on the combination of L1 regularization and pruning to eliminate unimportant connections.

Architecture

The network uses a single hidden layer with 256 neurons:

Training procedure

  1. Train with L1 regularization (λ=0.0001) to encourage small weights
  2. Apply weight pruning (threshold=0.01) to zero out small weights
  3. The pruned model achieves significant sparsity with minimal performance loss
5 CV folds
57.10% Mean accuracy
0.5556 Mean F1
0.210423 Mean p-adic loss
31,883 Avg non-zero params
96.1% Sparsity

Cross-validation results

FoldAccuracyF1P-adic loss (mean)Non-zero paramsSparsity
058.11%0.57080.20483331,68196.1%
155.96%0.54650.22774831,75096.1%
258.60%0.56770.19441831,94696.1%
357.04%0.55430.20587232,45396.0%
455.76%0.53860.21924231,58696.1%

Comparison with other models

The unconstrained neural network achieves the best p-adic loss among all models by using more parameters (after pruning), while the L1 regularization and pruning ensure that only the most important connections are retained. This demonstrates the tradeoff between model complexity and prediction accuracy.