-
Notifications
You must be signed in to change notification settings - Fork 8
/
Copy pathdoc-linter.py
executable file
·167 lines (148 loc) · 5.39 KB
/
doc-linter.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
#!/usr/bin/env python3
import re
import sys
import argparse
from functools import lru_cache
from dataclasses import dataclass
from pathlib import Path
from typing import Generator, Union
@dataclass
class Term():
name: str
value: str
@dataclass(repr=False)
class LintingError:
file: Path
line: int
column: int
message: str
fixed: bool = False
def __repr__(self) -> str:
fixed = " (fixed)" if self.fixed else ""
return f"{self.file}:{self.line}:{self.column}: {self.message}{fixed}"
DOCS_DIR = "docs"
RE_TERM = re.compile(r"^:(?P<term_name>[^:]+):\s*(?P<term_value>\S+.*)\s*$")
class Linter:
def __init__(self, start_path: Union[Path, str], auto_fix: bool, verbose: bool, direction: str = "from-terms"):
if not isinstance(start_path, Path):
start_path = Path(start_path)
self._start_path = start_path
self._auto_fix = auto_fix
self._verbose = verbose
self.errors: list[LintingError] = []
self.direction = direction
def verbose(self, message: str):
if self._verbose:
print(message)
@lru_cache(maxsize=0)
def load_terms(self, terms_file: Path):
self._terms = []
for line in terms_file.read_text().splitlines():
if match := RE_TERM.match(line):
if match.group("term_value").startswith("{") and match.group("term_value").endswith("}"):
# Reference to other term
continue
self._terms.append(Term(match.group("term_name"), match.group("term_value")))
def find_word(self, content: str, word: str, allow_part_of_word: bool = False) -> Generator[tuple[int, int, int, int], None, None]:
# Search from end of file backwards
end_search_idx = len(content)
while True:
start_idx = content.rfind(word, 0, end_search_idx)
if start_idx < 0:
break
end_search_idx = start_idx
end_idx = start_idx + len(word)
if (
not allow_part_of_word and
(content[start_idx-1] not in (" ", "\t", "\n") or content[start_idx + len(word)] not in (" ", "\t", "\n"))
):
# Part of a word
continue
line_start = content.rfind("\n", 0, start_idx)
line_number = content.count("\n", 0, start_idx) + 1
column_number = start_idx - line_start
yield start_idx, end_idx, line_number, column_number
@lru_cache(maxsize=1000)
def find_terms_file(self, start_path: Path) -> Path:
terms_file = None
while start_path.parent != start_path:
terms_file = start_path.joinpath("modules/common/partials/opsi_terms.adoc")
if terms_file.exists():
return terms_file.resolve()
start_path = start_path.parent
if not terms_file or not terms_file.exists():
raise FileNotFoundError(f"Terms file not found for path {start_path}")
def check_terms(self, file: Path, content: str) -> tuple[list[LintingError, str]]:
file = file.resolve()
errors = []
try:
terms_file = self.find_terms_file(file.parent)
except FileNotFoundError:
if self.verbose:
print(f"Skipping file {file} - no terms file found")
return [], content
if terms_file == file:
# Do not process the term file itself
return errors, content
self.load_terms(terms_file)
for term in self._terms:
if self.direction == "to-terms":
for start_idx, end_idx, line_number, column_number in self.find_word(content, term.value):
error = LintingError(file, line_number, column_number,f"Term {term.name!r} found")
if self._auto_fix:
content = f"{content[:start_idx]}{{{term.name}}}{content[end_idx:]}"
error.fixed = True
errors.append(error)
elif self.direction == "from-terms":
for start_idx, end_idx, line_number, column_number in self.find_word(
content,
f"{{{term.name}}}",
allow_part_of_word=True
):
error = LintingError(file, line_number, column_number,f"Term {term.name!r} found")
if self._auto_fix:
content = f"{content[:start_idx]}{term.value}{content[end_idx:]}"
error.fixed = True
errors.append(error)
return errors, content
def check_caution_used(self, file: Path, content: str) -> tuple[list[LintingError, str]]:
errors = []
for word in "[CAUTION]", "CAUTION:":
for start_idx, end_idx, line_number, column_number in self.find_word(content, word):
replace = word.replace("CAUTION", "WARNING")
error = LintingError(file, line_number, column_number,f"{word!r} used")
if self._auto_fix:
content = f"{content[:start_idx]}{replace}{content[end_idx:]}"
error.fixed = True
errors.append(error)
return errors, content
def run(self):
if self._start_path.is_file():
files = [self._start_path]
else:
files = list(self._start_path.rglob("*.asciidoc")) + list(self._start_path.rglob("*.adoc"))
for file in files:
self.verbose(f"Checking file: {file}")
content = file.read_text(encoding="utf-8")
orig_content = content
errors, content = self.check_terms(file, content)
self.errors.extend(errors)
errors, content = self.check_caution_used(file, content)
self.errors.extend(errors)
if orig_content != content:
file.write_text(content, encoding="utf-8")
def main():
# Get option --fix with argparse
parser = argparse.ArgumentParser()
parser.add_argument("--auto-fix", action="store_true")
parser.add_argument("--verbose", "-v", action="store_true")
parser.add_argument("path", nargs="?")
args = parser.parse_args()
linter = Linter(args.path or DOCS_DIR, args.auto_fix, args.verbose, direction="from-terms") # "to-terms"
linter.run()
for error in linter.errors:
print(error)
if [error for error in linter.errors if not error.fixed]:
sys.exit(1)
if __name__ == "__main__":
main()