Parameter | Type | Distribution | Values |
---|---|---|---|
embedding_size | Integer | LogUniform | 4 ≤ \(x\) ≤ 512 |
regularization.user | Float | LogUniform | 1e-05 ≤ \(x\) ≤ 1 |
regularization.item | Float | LogUniform | 1e-05 ≤ \(x\) ≤ 1 |
damping.user | Float | LogUniform | 1e-12 ≤ \(x\) ≤ 100 |
damping.item | Float | LogUniform | 1e-12 ≤ \(x\) ≤ 100 |
ALS BiasedMF on ML100K
This page analyzes the hyperparameter tuning results for biased matrix factorization with ALS.
Parameter Search Space
Final Result
Searching selected the following configuration:
{ 'embedding_size': 199, 'regularization': {'user': 0.07695955941562233, 'item': 0.20416528660055916}, 'damping': {'user': 0.08134708424868575, 'item': 7.134891218000362e-08}, 'epochs': 7 }
With these metrics:
{ 'RBP': 0.0016333128715170513, 'NDCG': 0.17844241206365657, 'RecipRank': 0.018992777790004307, 'RMSE': 0.8639979898614227, 'TrainTask': '2efd7d36-8928-4e55-b227-6238048fa355', 'TrainTime': 5.033226580999326, 'TrainCPU': 23.041116, 'timestamp': 1743028999, 'checkpoint_dir_name': None, 'done': True, 'training_iteration': 7, 'trial_id': 'a7114_00010', 'date': '2025-03-26_18-43-19', 'time_this_iter_s': 0.7188098430633545, 'time_total_s': 5.61619234085083, 'pid': 219279, 'hostname': 'CCI-ws21', 'node_ip': '10.248.127.152', 'config': { 'embedding_size': 199, 'regularization': {'user': 0.07695955941562233, 'item': 0.20416528660055916}, 'damping': {'user': 0.08134708424868575, 'item': 7.134891218000362e-08}, 'epochs': 7 }, 'time_since_restore': 5.61619234085083, 'iterations_since_restore': 7, 'experiment_tag': '10_item=0.0000,user=0.0813,embedding_size=199,item=0.2042,user=0.0770' }
Parameter Analysis
Embedding Size
The embedding size is the hyperparameter that most affects the model’s fundamental logic, so let’s look at performance as a fufnction of it:
Learning Parameters
Iteration Completion
How many iterations, on average, did we complete?
How did the metric progress in the best result?
How did the metric progress in the longest results?