1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
|
import json
import os
import random
# Source Files
FILE_ORIGINAL = "data/raw_datasets/labeled_full_dataset_batch.jsonl"
FILE_SYNTHESIS = "data/raw_datasets/synthesized_positives.jsonl"
# Output Files
OUTPUT_RAW = "data/finetune/preference_extractor_450k.jsonl"
def assemble_dataset():
os.makedirs(os.path.dirname(OUTPUT_RAW), exist_ok=True)
print("Assembling final dataset...")
records = []
# 1. Load Original (Pos + Neg)
print(f"Loading {FILE_ORIGINAL}...")
if os.path.exists(FILE_ORIGINAL):
with open(FILE_ORIGINAL, "r", encoding="utf-8") as f:
for line in f:
if line.strip():
item = json.loads(line)
# Standardize format: {"input": ..., "output": ...}
# Input is user query. Output is extracted JSON string.
query = item.get("original_query", "")
output_json = item.get("extracted_json", {"preferences": []})
# Ensure output is a string of JSON, minimal whitespace to save tokens
output_str = json.dumps(output_json, ensure_ascii=False)
records.append({
"input": query,
"output": output_str,
"source": item.get("source", "original")
})
else:
print(f"Warning: {FILE_ORIGINAL} missing!")
print(f"Loaded {len(records)} from original.")
# 2. Load Synthesis (Pos)
print(f"Loading {FILE_SYNTHESIS}...")
syn_count = 0
if os.path.exists(FILE_SYNTHESIS):
with open(FILE_SYNTHESIS, "r", encoding="utf-8") as f:
for line in f:
if line.strip():
item = json.loads(line)
query = item.get("original_query", "")
output_json = item.get("extracted_json", {"preferences": []})
output_str = json.dumps(output_json, ensure_ascii=False)
records.append({
"input": query,
"output": output_str,
"source": "synthesis"
})
syn_count += 1
else:
print(f"Warning: {FILE_SYNTHESIS} missing!")
print(f"Loaded {syn_count} from synthesis.")
# 3. Shuffle
print("Shuffling...")
random.shuffle(records)
# 4. Save
print(f"Saving {len(records)} records to {OUTPUT_RAW}...")
with open(OUTPUT_RAW, "w", encoding="utf-8") as f:
for r in records:
f.write(json.dumps(r, ensure_ascii=False) + "\n")
print("Done!")
print("\nTo upload to Hugging Face, run:")
print("huggingface-cli upload <repo_id> data/finetune/preference_extractor_450k.jsonl --repo-type dataset")
if __name__ == "__main__":
assemble_dataset()
|