blob: fe52bb473d8f7b1659a1b90a68d03404d78e91a0 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
|
import requests
import feedparser
def test_arxiv():
base_url = "http://export.arxiv.org/api/query"
# 时间段设为 2025-03-27 00:00 到 2025-03-29 00:00
# 注意: 论文在 3月27日 07:54Z 提交,应该在这个区间之内
search_query = (
"(all:bias+OR+all:fairness)"
"+AND+cat:cs.IR"
"+AND+submittedDate:[202503270000+TO+202503290000]"
)
params = {
"search_query": search_query,
"sortBy": "submittedDate",
"sortOrder": "descending",
"max_results": 100
}
print("[DEBUG] search_query =", search_query)
r = requests.get(base_url, params=params)
print("[DEBUG] Full URL =", r.url)
if r.status_code != 200:
print("[ERROR] HTTP Status:", r.status_code)
return
feed = feedparser.parse(r.content)
print("[DEBUG] Returned entries:", len(feed.entries))
# 打印出标题和发布时间供检查
for i, entry in enumerate(feed.entries, start=1):
print(f"{i}. Title: {entry.title} | updated: {entry.updated} | published: {entry.published}")
if __name__ == "__main__":
test_arxiv()
|