summaryrefslogtreecommitdiff
path: root/outputs/fair_audit/topic_user_K4_d64_N200_fair_results.json
blob: 436ad40803440b6b4dbb5ee1e896ff25140ce016 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
{
  "results": {
    "Base": {
      "rouge1": 0.29108420217791586,
      "rougeL": 0.11933661102378458,
      "meteor": 0.20440162670816078,
      "sfd_all": 1.57754139533291,
      "sfd_nolen": 0.968435638532869,
      "avg_len": 244.66,
      "feature_deltas": {
        "length": -0.89875,
        "avg_sent_len": -3.7377455393461467,
        "TTR": 0.06763033291033599,
        "newline_rate": 0.27872818568783064,
        "exclaim_rate": 0.1048047941518918,
        "first_person_rate": 0.0009034068211049077,
        "adj_adv_rate": -0.0037094728027600708,
        "sentiment_score": 0.002781973404286483
      }
    },
    "Prompt-All-K": {
      "rouge1": 0.25818662417404703,
      "rougeL": 0.12293156397427764,
      "meteor": 0.18520736264559315,
      "sfd_all": 1.4544789243060898,
      "sfd_nolen": 0.8730698626536205,
      "avg_len": 232.605,
      "feature_deltas": {
        "length": -17.15375,
        "avg_sent_len": -2.4913280395141304,
        "TTR": 0.026565419567024528,
        "newline_rate": 0.1756867598329655,
        "exclaim_rate": 0.03524532727422328,
        "first_person_rate": 0.001990369323950276,
        "adj_adv_rate": -0.007447287388785413,
        "sentiment_score": 0.0013350766977724307
      }
    },
    "BM25-Top1": {
      "rouge1": 0.2604153356020315,
      "rougeL": 0.1193909461818881,
      "meteor": 0.18568201388907532,
      "sfd_all": 1.4180388711957266,
      "sfd_nolen": 0.8918698274006386,
      "avg_len": 227.265,
      "feature_deltas": {
        "length": -25.24375,
        "avg_sent_len": -2.3812369921742818,
        "TTR": 0.04246840602123856,
        "newline_rate": 0.1523757212605453,
        "exclaim_rate": 0.06451831601647286,
        "first_person_rate": -0.0006676484724408885,
        "adj_adv_rate": -0.004714460870661791,
        "sentiment_score": 0.0019441456923959164
      }
    },
    "Uncond-Head": {
      "rouge1": 0.272253587033546,
      "rougeL": 0.1321174047956226,
      "meteor": 0.19203463885960004,
      "sfd_all": 1.6076602575522427,
      "sfd_nolen": 0.9548231613156074,
      "avg_len": 255.78,
      "feature_deltas": {
        "length": 14.12125,
        "avg_sent_len": -3.4104779329564514,
        "TTR": -0.1488780438134324,
        "newline_rate": 0.19151398235081096,
        "exclaim_rate": 0.054442049906536394,
        "first_person_rate": 0.007650724723829362,
        "adj_adv_rate": -0.014780960712514975,
        "sentiment_score": 0.0037896071184805994
      },
      "adapt_time": 5.503373540639878
    },
    "CVH": {
      "rouge1": 0.2738968089325739,
      "rougeL": 0.13034318090740352,
      "meteor": 0.19233542238408405,
      "sfd_all": 1.605802427600553,
      "sfd_nolen": 0.9598213964633208,
      "avg_len": 249.395,
      "feature_deltas": {
        "length": 4.51125,
        "avg_sent_len": -3.0078106411236507,
        "TTR": -0.137260791986814,
        "newline_rate": 0.18334382325488566,
        "exclaim_rate": 0.05609966650973971,
        "first_person_rate": 0.010012175202340347,
        "adj_adv_rate": -0.01552706206622135,
        "sentiment_score": 0.004098927650269633
      },
      "adapt_time": 5.494220098257065
    },
    "Uncond-SW": {
      "rouge1": 0.2728698688411594,
      "rougeL": 0.13082081008960608,
      "meteor": 0.19297967725738285,
      "sfd_all": 1.6217805873430902,
      "sfd_nolen": 0.9641445892199336,
      "avg_len": 253.305,
      "feature_deltas": {
        "length": 10.77125,
        "avg_sent_len": -3.0826858181641144,
        "TTR": -0.13979336041068025,
        "newline_rate": 0.20717424954482574,
        "exclaim_rate": 0.052789807352293626,
        "first_person_rate": 0.007670669350797363,
        "adj_adv_rate": -0.018823927161665116,
        "sentiment_score": 0.003211091078776497
      },
      "adapt_time": 2.707582322359085
    }
  },
  "num_examples": 200,
  "decode_policy": "greedy, min_new_tokens=128, max_new_tokens=512, blend_gamma=0.5"
}