Parameter | Type | Distribution | Values | Selected |
---|---|---|---|---|
embedding_size | Integer | LogUniform | 4 ≤ \(x\) ≤ 512 | 182 |
regularization | Float | LogUniform | 0.0001 ≤ \(x\) ≤ 10 | 2.05 |
learning_rate | Float | LogUniform | 0.001 ≤ \(x\) ≤ 0.1 | 0.0381 |
reg_method | Categorical | Uniform | L2, AdamW | AdamW |
FlexMF Explicit
This page analyzes the hyperparameter tuning results for the FlexMF scorer in explicit-feedback mode (a biased matrix factorization model trained with PyTorch).
Parameter Search Space
Final Result
Searching selected the following configuration:
{ 'embedding_size': 182, 'regularization': 2.045564699262321, 'learning_rate': 0.03805132726783247, 'reg_method': 'AdamW', 'epochs': 4 }
With these metrics:
{ 'RBP': 0.02780012324782274, 'LogRBP': 0.2115251448513078, 'NDCG': 0.2129614179804314, 'RecipRank': 0.10860804526319909, 'RMSE': 0.9312385413697157, 'TrainTask': 'c545f414-37fa-4e5c-862b-a22da158fc6d', 'TrainTime': None, 'TrainCPU': None, 'max_epochs': 50, 'done': False, 'training_iteration': 4, 'trial_id': '0d832_00022', 'date': '2025-05-06_11-17-08', 'timestamp': 1746544628, 'time_this_iter_s': 0.7267158031463623, 'time_total_s': 3.288360357284546, 'pid': 475067, 'hostname': 'gracehopper1', 'node_ip': '192.168.225.60', 'config': { 'embedding_size': 182, 'regularization': 2.045564699262321, 'learning_rate': 0.03805132726783247, 'reg_method': 'AdamW', 'epochs': 4 }, 'time_since_restore': 3.288360357284546, 'iterations_since_restore': 4 }
Parameter Analysis
Embedding Size
The embedding size is the hyperparameter that most affects the model’s fundamental logic, so let’s look at performance as a fufnction of it:
Learning Parameters
Iteration Completion
How many iterations, on average, did we complete?
How did the metric progress in the best result?
How did the metric progress in the longest results?