-
Notifications
You must be signed in to change notification settings - Fork 6
/
Copy pathrecreate.py
60 lines (49 loc) · 1.77 KB
/
recreate.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
import re
import ast
import argparse
import crawler
import citer
def set_args():
# 解析issue模版参数 --issue ${{ toJson(github.event.issue.body) }}
parser = argparse.ArgumentParser()
parser.add_argument("--issue", "-i", type=str, help="input issue", required=True)
args = parser.parse_args()
return args
def parse_issue(issue):
try:
issue = issue.replace("\\n", "").replace("\\r", "")
info = ast.literal_eval(issue)
print(info)
assert isinstance(info, list) and len(info) > 0
issue_dict = info[0]
assert "confs" in issue_dict and "year" in issue_dict and "filter" in issue_dict
#assert isinstance(info, list)
#assert len(info) > 0 and info[0].get("filter") and info[0].get("confs") and info[0].get("year")
except:
raise Exception("[-] Wrong input!")
return info
def run(confs_str, start_year, filter_str=''):
FILTERS = ['kddcup', 'w.html', 'lbr.html']
if filter_str:
FILTERS += filter_str.lower().split(' ')
try:
confs = confs_str.lower().split(' ')
except:
confs = []
start_year = int(start_year)
##dblp网站上的文章(daily推送)
crawler.run_all(confs=confs,filter_keywords=FILTERS,start_year=start_year,filename='results.json',threads=20)
##顶刊顶会
citer.run_all(confs=[conf + str(start_year) for conf in confs],mode='parallel')
def main():
###1.get issue template params
args = set_args()
###2.parse issue template params
info = parse_issue(args.issue)
assert len(info) == 1 # confs and year
item = info[0]
print(item)
###3.crawle conferences info
run(confs_str=item['confs'], start_year=item['year'],filter_str=item['filter'])
if __name__ == "__main__":
main()