FlexMF WARP

This page analyzes the hyperparameter tuning results for the FlexMF scorer in implicit-feedback mode with WARP loss.

Parameter Search Space

Parameter Type Distribution Values Selected
embedding_size Integer LogUniform 4 ≤ \(x\) ≤ 512 14
regularization Float LogUniform 0.0001 ≤ \(x\) ≤ 10 0.00536
learning_rate Float LogUniform 0.001 ≤ \(x\) ≤ 0.1 0.00513
reg_method Categorical Uniform L2, AdamW L2
item_bias Categorical Uniform True, False False

Final Result

Searching selected the following configuration:

{
    'embedding_size': 14,
    'regularization': 0.005362916165637716,
    'learning_rate': 0.005131215517492302,
    'reg_method': 'L2',
    'item_bias': False,
    'epochs': 23
}

With these metrics:

{
    'RBP': 0.23343363467501385,
    'NDCG': 0.4747794908475408,
    'RecipRank': 0.42092755669570436,
    'TrainTask': '1dcab606-c4b4-4c84-a4d1-999e6a0f4a08',
    'TrainTime': None,
    'TrainCPU': None,
    'max_epochs': 50,
    'done': False,
    'training_iteration': 23,
    'trial_id': '1f447_00090',
    'date': '2025-04-21_22-49-44',
    'timestamp': 1745290184,
    'time_this_iter_s': 10.31071925163269,
    'time_total_s': 522.4587495326996,
    'pid': 650652,
    'hostname': 'CCI-ws21',
    'node_ip': '10.248.127.152',
    'config': {
        'embedding_size': 14,
        'regularization': 0.005362916165637716,
        'learning_rate': 0.005131215517492302,
        'reg_method': 'L2',
        'item_bias': False,
        'epochs': 23
    },
    'time_since_restore': 288.4037449359894,
    'iterations_since_restore': 14
}

Parameter Analysis

Embedding Size

The embedding size is the hyperparameter that most affects the model’s fundamental logic, so let’s look at performance as a fufnction of it:

Learning Parameters

Iteration Completion

How many iterations, on average, did we complete?

How did the metric progress in the best result?

How did the metric progress in the longest results?