summaryrefslogtreecommitdiff
path: root/outputs/fair_audit/topic_temporal_K4_d64_N200_fair_results.json
blob: 097b95c390c7e488ed0c50b85e520beb8faaf44b (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
{
  "results": {
    "Base": {
      "rouge1": 0.28984915072609035,
      "rougeL": 0.12250377793521261,
      "meteor": 0.21004758987715536,
      "sfd_all": 1.6643365277346691,
      "sfd_nolen": 1.0385348101589198,
      "avg_len": 246.335,
      "feature_deltas": {
        "length": -1.78,
        "avg_sent_len": -3.246577205494341,
        "TTR": 0.06897256967962712,
        "newline_rate": 0.4263220633386752,
        "exclaim_rate": 0.1120631034932264,
        "first_person_rate": -0.0037646675370617416,
        "adj_adv_rate": -0.00259664310238531,
        "sentiment_score": 0.002157300888416565
      }
    },
    "Prompt-All-K": {
      "rouge1": 0.2668110270292337,
      "rougeL": 0.1291186308652062,
      "meteor": 0.19818628603837213,
      "sfd_all": 1.410855843519239,
      "sfd_nolen": 0.8359962398843364,
      "avg_len": 249.06,
      "feature_deltas": {
        "length": -4.35,
        "avg_sent_len": -1.329287994749063,
        "TTR": 0.017625236539407726,
        "newline_rate": 0.20696549782868945,
        "exclaim_rate": 0.038863829209766554,
        "first_person_rate": -0.002770645129017844,
        "adj_adv_rate": -0.0005991015318490243,
        "sentiment_score": 0.0009170684119961202
      }
    },
    "BM25-Top1": {
      "rouge1": 0.2546033696273812,
      "rougeL": 0.11820224039247507,
      "meteor": 0.18746215667712504,
      "sfd_all": 1.4681828407377084,
      "sfd_nolen": 0.8656866959298877,
      "avg_len": 232.015,
      "feature_deltas": {
        "length": -21.19,
        "avg_sent_len": -2.2273077161985886,
        "TTR": 0.048089079588597955,
        "newline_rate": 0.2283326324239623,
        "exclaim_rate": 0.06476176816481617,
        "first_person_rate": -0.002062671240273185,
        "adj_adv_rate": -0.003105932780815531,
        "sentiment_score": 0.0026425405580024525
      }
    },
    "Uncond-Head": {
      "rouge1": 0.27929922437081617,
      "rougeL": 0.13416330840898946,
      "meteor": 0.20432220183006447,
      "sfd_all": 1.7457301119532245,
      "sfd_nolen": 1.0748504971524213,
      "avg_len": 259.19,
      "feature_deltas": {
        "length": 15.185,
        "avg_sent_len": -1.528666820505814,
        "TTR": -0.13456223126496966,
        "newline_rate": 0.3035233686677935,
        "exclaim_rate": 0.06546569935247971,
        "first_person_rate": 0.00608902419005909,
        "adj_adv_rate": -0.012713541636957098,
        "sentiment_score": 0.0032088299366138146
      },
      "adapt_time": 5.545737104415894
    },
    "CVH": {
      "rouge1": 0.2732668756486018,
      "rougeL": 0.1333292880216737,
      "meteor": 0.1994612023551069,
      "sfd_all": 1.6899249199714634,
      "sfd_nolen": 1.0363194964048348,
      "avg_len": 251.0,
      "feature_deltas": {
        "length": 7.57,
        "avg_sent_len": -1.636122177861851,
        "TTR": -0.14241579788497058,
        "newline_rate": 0.2739697954172092,
        "exclaim_rate": 0.06934072686287729,
        "first_person_rate": 0.005809860274156219,
        "adj_adv_rate": -0.0153621828929979,
        "sentiment_score": 0.004551798810441284
      },
      "adapt_time": 5.555131942033768
    },
    "Uncond-SW": {
      "rouge1": 0.27759121282072297,
      "rougeL": 0.13426233957013273,
      "meteor": 0.20317704007364315,
      "sfd_all": 1.7356393883366876,
      "sfd_nolen": 1.0575837665782766,
      "avg_len": 257.81,
      "feature_deltas": {
        "length": 15.08,
        "avg_sent_len": -1.5301018135923676,
        "TTR": -0.14273001338817937,
        "newline_rate": 0.3095342271756431,
        "exclaim_rate": 0.0598484461304457,
        "first_person_rate": 0.005052374635440565,
        "adj_adv_rate": -0.013400514550224818,
        "sentiment_score": 0.0029209458494426875
      },
      "adapt_time": 2.1236859560012817
    }
  },
  "num_examples": 200,
  "decode_policy": "greedy, min_new_tokens=128, max_new_tokens=512, blend_gamma=0.5"
}