summaryrefslogtreecommitdiff
path: root/outputs/dev_run3/review_user_K4_d64_results.json
blob: 234a296887f3138eb8b4b88fe23a43dcc5a81f52 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
{
  "config": {
    "task": "review",
    "setting": "user",
    "dataset_config": "product_review_user",
    "model_name": "Qwen/Qwen2.5-1.5B-Instruct",
    "device": "cuda:1",
    "d": 64,
    "alpha": 0.1,
    "beta": 0.05,
    "lam": 0.0001,
    "lr": 0.05,
    "adapt_steps": 30,
    "max_grad_norm": 5.0,
    "basis_seed": 42,
    "K": 4,
    "max_new_tokens": 512,
    "temperature": 0.0,
    "top_p": 0.9,
    "num_eval": 200,
    "batch_size": 1,
    "blend_gamma": 0.5,
    "min_new_tokens": 128
  },
  "results": {
    "Base": {
      "rouge1": 0.25934513296875994,
      "rougeL": 0.12338481426247551,
      "meteor": 0.1469023840179831,
      "sfd": 1.4575752102695472,
      "num_examples": 200
    },
    "Prompt-All-K": {
      "rouge1": 0.2946653733430904,
      "rougeL": 0.139999532551422,
      "meteor": 0.19063956073455807,
      "sfd": 1.0439696689383455,
      "num_examples": 200
    },
    "BM25-Top1": {
      "rouge1": 0.29945906659382165,
      "rougeL": 0.13889418725618197,
      "meteor": 0.18864163482386093,
      "sfd": 1.0816142289529644,
      "num_examples": 200
    },
    "Uncond-Head": {
      "rouge1": 0.27471654695871556,
      "rougeL": 0.13993753615192442,
      "meteor": 0.14908919360059167,
      "sfd": 1.406899356051711,
      "num_examples": 200,
      "adapt_time": 6.94630047082901
    },
    "CVH": {
      "rouge1": 0.2669879308948325,
      "rougeL": 0.13772551245363476,
      "meteor": 0.1462595092740725,
      "sfd": 1.4302773361057746,
      "num_examples": 200,
      "adapt_time": 6.958050906658173
    }
  },
  "num_examples": 200
}