ALS BiasedMF on ML100K

This page analyzes the hyperparameter tuning results for biased matrix factorization with ALS.

Parameter Search Space

Parameter Type Distribution Values
embedding_size Integer LogUniform 4 ≤ \(x\) ≤ 512
regularization.user Float LogUniform 1e-05 ≤ \(x\) ≤ 1
regularization.item Float LogUniform 1e-05 ≤ \(x\) ≤ 1
damping.user Float LogUniform 1e-12 ≤ \(x\) ≤ 100
damping.item Float LogUniform 1e-12 ≤ \(x\) ≤ 100

Final Result

Searching selected the following configuration:

{
    'embedding_size': 199,
    'regularization': {'user': 0.07695955941562233, 'item': 0.20416528660055916},
    'damping': {'user': 0.08134708424868575, 'item': 7.134891218000362e-08},
    'epochs': 7
}

With these metrics:

{
    'RBP': 0.0016333128715170513,
    'NDCG': 0.17844241206365657,
    'RecipRank': 0.018992777790004307,
    'RMSE': 0.8639979898614227,
    'TrainTask': '2efd7d36-8928-4e55-b227-6238048fa355',
    'TrainTime': 5.033226580999326,
    'TrainCPU': 23.041116,
    'timestamp': 1743028999,
    'checkpoint_dir_name': None,
    'done': True,
    'training_iteration': 7,
    'trial_id': 'a7114_00010',
    'date': '2025-03-26_18-43-19',
    'time_this_iter_s': 0.7188098430633545,
    'time_total_s': 5.61619234085083,
    'pid': 219279,
    'hostname': 'CCI-ws21',
    'node_ip': '10.248.127.152',
    'config': {
        'embedding_size': 199,
        'regularization': {'user': 0.07695955941562233, 'item': 0.20416528660055916},
        'damping': {'user': 0.08134708424868575, 'item': 7.134891218000362e-08},
        'epochs': 7
    },
    'time_since_restore': 5.61619234085083,
    'iterations_since_restore': 7,
    'experiment_tag': '10_item=0.0000,user=0.0813,embedding_size=199,item=0.2042,user=0.0770'
}

Parameter Analysis

Embedding Size

The embedding size is the hyperparameter that most affects the model’s fundamental logic, so let’s look at performance as a fufnction of it:

Learning Parameters

Iteration Completion

How many iterations, on average, did we complete?

How did the metric progress in the best result?

How did the metric progress in the longest results?