v1.4.0: Major Skills Expansion - 75 Total Skills

This commit is contained in:
admin
2026-02-26 12:19:22 +04:00
Unverified
parent 029bad83b6
commit e2a7099273
392 changed files with 55962 additions and 2 deletions

View File

@@ -0,0 +1,11 @@
{
"name": "ultralytics-dev",
"version": "2.1.1",
"description": "Auto-formatting hooks for Python, JavaScript, Markdown, and Bash with Google-style docstrings and code quality checks.",
"author": {
"name": "Fatih Akyon"
},
"homepage": "https://github.com/fcakyon/claude-codex-settings#plugins",
"repository": "https://github.com/fcakyon/claude-codex-settings",
"license": "Apache-2.0"
}

View File

@@ -0,0 +1,36 @@
{
"description": "Code formatting and quality hooks for Ultralytics development",
"hooks": {
"PostToolUse": [
{
"matcher": "Edit|MultiEdit|Write",
"hooks": [
{
"type": "command",
"command": "file_path=$(jq -r '.tool_input.file_path // empty' 2>/dev/null); if [[ -n \"$file_path\" && -f \"$file_path\" ]]; then case \"$file_path\" in *.py|*.js|*.jsx|*.ts|*.tsx) if [[ \"$OSTYPE\" == \"darwin\"* ]]; then sed -i '' 's/^[[:space:]]*$//g' \"$file_path\" 2>/dev/null || true; else sed -i 's/^[[:space:]]*$//g' \"$file_path\" 2>/dev/null || true; fi ;; esac; fi"
},
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/hooks/scripts/format_python_docstrings.py"
},
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/hooks/scripts/python_code_quality.py"
},
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/hooks/scripts/prettier_formatting.py"
},
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/hooks/scripts/markdown_formatting.py"
},
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/hooks/scripts/bash_formatting.py"
}
]
}
]
}
}

View File

@@ -0,0 +1,58 @@
#!/usr/bin/env python3
"""
PostToolUse hook: Auto-format Bash/Shell scripts with prettier-plugin-sh
"""
import json
import shutil
import subprocess
import sys
from pathlib import Path
def check_prettier_version() -> bool:
"""Check if prettier is installed and warn if version differs from 3.6.2."""
if not shutil.which('npx'):
return False
try:
result = subprocess.run(['npx', 'prettier', '--version'],
capture_output=True, text=True, check=False, timeout=5)
if result.returncode == 0:
version = result.stdout.strip()
if '3.6.2' not in version:
print(f"⚠️ Prettier version mismatch: expected 3.6.2, found {version}")
return True
except Exception:
pass
return False
def main():
try:
data = json.load(sys.stdin)
file_path = data.get("tool_input", {}).get("file_path", "")
if not file_path.endswith(('.sh', '.bash')):
sys.exit(0)
sh_file = Path(file_path)
if not sh_file.exists() or any(p in sh_file.parts for p in ['.git', '.venv', 'venv', 'env', '.env', '__pycache__', '.mypy_cache', '.pytest_cache', '.tox', '.nox', '.eggs', 'eggs', '.idea', '.vscode', 'node_modules', 'site-packages', 'build', 'dist', '.claude']):
sys.exit(0)
# Check if prettier is available
if not check_prettier_version():
sys.exit(0)
# Try prettier with prettier-plugin-sh, handle any failure gracefully
try:
cmd = f'npx prettier --write --list-different --print-width 120 --plugin=$(npm root -g)/prettier-plugin-sh/lib/index.cjs "{sh_file}"'
subprocess.run(cmd, shell=True, capture_output=True, check=False, cwd=sh_file.parent, timeout=10)
except Exception:
pass # Silently handle any failure (missing plugin, timeout, etc.)
except Exception:
pass
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,547 @@
#!/usr/bin/env python3
"""Format Python docstrings in Google style without external dependencies."""
from __future__ import annotations
import ast
import json
import re
import sys
from pathlib import Path
URLS = {"https", "http", "ftp"}
SECTIONS = (
"Args",
"Attributes",
"Methods",
"Returns",
"Yields",
"Raises",
"Example",
"Examples",
"Notes",
"References",
)
SECTION_ALIASES = {
"Arguments": "Args",
"Usage": "Examples",
"Usage Example": "Examples",
"Usage Examples": "Examples",
"Example Usage": "Examples",
"Example": "Examples",
"Return": "Returns",
"Yield": "Yields",
"Raise": "Raises",
"Note": "Notes",
"Reference": "References",
}
LIST_RX = re.compile(r"""^(\s*)(?:[-*•]\s+|(?:\d+|[A-Za-z]+)[\.\)]\s+)""")
TABLE_RX = re.compile(r"^\s*\|.*\|\s*$")
TABLE_RULE_RX = re.compile(r"^\s*[:\-\|\s]{3,}$")
TREE_CHARS = ("", "", "", "")
# Antipatterns for non-Google docstring styles
RST_FIELD_RX = re.compile(r"^\s*:(param|type|return|rtype|raises)\b", re.M)
EPYDOC_RX = re.compile(r"^\s*@(?:param|type|return|rtype|raise)\b", re.M)
NUMPY_UNDERLINE_SECTION_RX = re.compile(r"^\s*(Parameters|Returns|Yields|Raises|Notes|Examples)\n[-]{3,}\s*$", re.M)
GOOGLE_SECTION_RX = re.compile(
r"^\s*(Args|Attributes|Methods|Returns|Yields|Raises|Example|Examples|Notes|References):\s*$", re.M
)
NON_GOOGLE = {"numpy", "rest", "epydoc"}
def wrap_words(words: list[str], width: int, indent: int, min_words_per_line: int = 1) -> list[str]:
"""Wrap words to width with indent; optionally avoid very short orphan lines."""
pad = " " * indent
if not words:
return []
lines: list[list[str]] = []
cur: list[str] = []
cur_len = indent
for w in words:
need = len(w) + (1 if cur else 0)
if cur and cur_len + need > width:
lines.append(cur)
cur, cur_len = [w], indent + len(w)
else:
cur.append(w)
cur_len += need
if cur:
lines.append(cur)
if min_words_per_line > 1:
i = 1
while i < len(lines):
if len(lines[i]) < min_words_per_line and len(lines[i - 1]) > 1:
donor = lines[i - 1][-1]
this_len = len(pad) + sum(len(x) for x in lines[i]) + (len(lines[i]) - 1)
if this_len + (1 if lines[i] else 0) + len(donor) <= width:
lines[i - 1].pop()
lines[i].insert(0, donor)
if i > 1 and len(lines[i - 1]) == 1:
i -= 1
continue
i += 1
return [pad + " ".join(line) for line in lines]
def wrap_para(text: str, width: int, indent: int, min_words_per_line: int = 1) -> list[str]:
"""Wrap a paragraph string; orphan control via min_words_per_line."""
if text := text.strip():
return wrap_words(text.split(), width, indent, min_words_per_line)
else:
return []
def wrap_hanging(head: str, desc: str, width: int, cont_indent: int) -> list[str]:
"""Wrap 'head + desc' with hanging indent; ensure first continuation has >=2 words."""
room = width - len(head)
words = desc.split()
if not words:
return [head.rstrip()]
take, used = [], 0
for w in words:
need = len(w) + (1 if take else 0)
if used + need <= room:
take.append(w)
used += need
else:
break
out: list[str] = []
if take:
out.append(head + " ".join(take))
rest = words[len(take) :]
else:
out.append(head.rstrip())
rest = words
out.extend(wrap_words(rest, width, cont_indent, min_words_per_line=2))
return out
def is_list_item(s: str) -> bool:
"""Return True if s looks like a bullet/numbered list item."""
return bool(LIST_RX.match(s.lstrip()))
def is_fence_line(s: str) -> bool:
"""Return True if s is a Markdown code-fence line."""
t = s.lstrip()
return t.startswith("```")
def is_table_like(s: str) -> bool:
"""Return True if s resembles a simple Markdown table or rule line."""
return bool(TABLE_RX.match(s)) or bool(TABLE_RULE_RX.match(s))
def is_tree_like(s: str) -> bool:
"""Return True if s contains common ASCII tree characters."""
return any(ch in s for ch in TREE_CHARS)
def is_indented_block_line(s: str) -> bool:
"""Return True if s looks like a deeply-indented preformatted block."""
return bool(s.startswith(" ")) or s.startswith("\t")
def header_name(line: str) -> str | None:
"""Return canonical section header or None."""
s = line.strip()
if not s.endswith(":") or len(s) <= 1:
return None
name = s[:-1].strip()
name = SECTION_ALIASES.get(name, name)
return name if name in SECTIONS else None
def add_header(lines: list[str], indent: int, title: str) -> None:
"""Append a section header with a blank line before it."""
while lines and lines[-1] == "":
lines.pop()
if lines:
lines.append("")
lines.append(" " * indent + f"{title}:")
def emit_paragraphs(
src: list[str], width: int, indent: int, list_indent: int | None = None, orphan_min: int = 1
) -> list[str]:
"""Wrap text while preserving lists, fenced code, tables, trees, and deeply-indented blocks."""
out: list[str] = []
buf: list[str] = []
in_fence = False
def flush():
"""Flush buffered paragraph with wrapping."""
nonlocal buf
if buf:
out.extend(wrap_para(" ".join(x.strip() for x in buf), width, indent, min_words_per_line=orphan_min))
buf = []
for raw in src:
s = raw.rstrip("\n")
stripped = s.strip()
if not stripped:
flush()
out.append("")
continue
if is_fence_line(s):
flush()
out.append(s.rstrip())
in_fence = not in_fence
continue
if in_fence or is_table_like(s) or is_tree_like(s) or is_indented_block_line(s):
flush()
out.append(s.rstrip())
continue
if is_list_item(s):
flush()
out.append((" " * list_indent + stripped) if list_indent is not None else s.rstrip())
continue
buf.append(s)
flush()
while out and out[-1] == "":
out.pop()
return out
def parse_sections(text: str) -> dict[str, list[str]]:
"""Parse Google-style docstring into sections."""
parts = {k: [] for k in ("summary", "description", *SECTIONS)}
cur = "summary"
for raw in text.splitlines():
line = raw.rstrip("\n")
if h := header_name(line):
cur = h
continue
if not line.strip():
if cur == "summary" and parts["summary"]:
cur = "description"
if parts[cur]:
parts[cur].append("")
continue
parts[cur].append(line)
return parts
def looks_like_param(s: str) -> bool:
"""Heuristic: True if line looks like a 'name: desc' param without being a list item."""
if is_list_item(s) or ":" not in s:
return False
head = s.split(":", 1)[0].strip()
return False if head in URLS else bool(head)
def iter_items(lines: list[str]) -> list[list[str]]:
"""Group lines into logical items separated by next param-like line."""
items, i, n = [], 0, len(lines)
while i < n:
while i < n and not lines[i].strip():
i += 1
if i >= n:
break
item = [lines[i]]
i += 1
while i < n:
st = lines[i].strip()
if st and looks_like_param(st):
break
item.append(lines[i])
i += 1
items.append(item)
return items
def format_structured_block(lines: list[str], width: int, base: int) -> list[str]:
"""Format Args/Returns/etc.; continuation at base+4, lists at base+8."""
out: list[str] = []
cont, lst = base + 4, base + 8
for item in iter_items(lines):
first = item[0].strip()
name, desc = ([*first.split(":", 1), ""])[:2]
name, desc = name.strip(), desc.strip()
had_colon = ":" in first
if not name or (" " in name and "(" not in name and ")" not in name):
out.extend(emit_paragraphs(item, width, cont, lst, orphan_min=2))
continue
# Join continuation lines that aren't new paragraphs into desc
parts = [desc] if desc else []
tail, i = [], 1
while i < len(item):
line = item[i].strip()
if not line or is_list_item(item[i]) or is_fence_line(item[i]) or is_table_like(item[i]):
tail = item[i:]
break
parts.append(line)
i += 1
else:
tail = []
desc = " ".join(parts)
head = " " * cont + (f"{name}: " if (desc or had_colon) else name)
out.extend(wrap_hanging(head, desc, width, cont + 4))
if tail:
if body := emit_paragraphs(tail, width, cont + 4, lst, orphan_min=2):
out.extend(body)
return out
def detect_opener(original_literal: str) -> tuple[str, str, bool]:
"""Return (prefix, quotes, inline_hint) from the original string token safely."""
s = original_literal.lstrip()
i = 0
while i < len(s) and s[i] in "rRuUbBfF":
i += 1
quotes = '"""'
if i + 3 <= len(s) and s[i : i + 3] in ('"""', "'''"):
quotes = s[i : i + 3]
keep = "".join(ch for ch in s[:i] if ch in "rRuU")
j = i + len(quotes)
inline_hint = j < len(s) and s[j : j + 1] not in {"", "\n", "\r"}
return keep, quotes, inline_hint
def format_google(text: str, indent: int, width: int, quotes: str, prefix: str, start_newline: bool) -> str:
"""Format multi-line Google-style docstring with start_newline controlling summary placement."""
p = parse_sections(text)
opener = prefix + quotes
out: list[str] = []
if p["summary"]:
summary_text = " ".join(x.strip() for x in p["summary"]).strip()
if summary_text and summary_text[-1] not in ".!?":
summary_text += "."
if start_newline:
out.append(opener)
out.extend(emit_paragraphs([summary_text], width, indent, list_indent=indent, orphan_min=1))
else:
eff_width = max(1, width - indent)
out.extend(wrap_hanging(opener, summary_text, eff_width, indent))
else:
out.append(opener)
if any(x.strip() for x in p["description"]):
out.append("")
out.extend(emit_paragraphs(p["description"], width, indent, list_indent=indent, orphan_min=1))
has_content = bool(p["summary"]) or any(x.strip() for x in p["description"])
for sec in ("Args", "Attributes", "Methods", "Returns", "Yields", "Raises"):
if any(x.strip() for x in p[sec]):
if has_content:
add_header(out, indent, sec)
else:
out.append(" " * indent + f"{sec}:")
has_content = True
out.extend(format_structured_block(p[sec], width, indent))
for sec in ("Examples", "Notes", "References"):
if any(x.strip() for x in p[sec]):
add_header(out, indent, sec)
out.extend(x.rstrip() for x in p[sec])
while out and out[-1] == "":
out.pop()
out.append(" " * indent + quotes)
return "\n".join(out)
def likely_docstring_style(text: str) -> str:
"""Return 'google' | 'numpy' | 'rest' | 'epydoc' | 'unknown' for docstring text."""
t = "\n".join(line.rstrip() for line in text.strip().splitlines())
if RST_FIELD_RX.search(t):
return "rest"
if EPYDOC_RX.search(t):
return "epydoc"
if NUMPY_UNDERLINE_SECTION_RX.search(t):
return "numpy"
return "google" if GOOGLE_SECTION_RX.search(t) else "unknown"
def format_docstring(
content: str, indent: int, width: int, quotes: str, prefix: str, start_newline: bool = False
) -> str:
"""Single-line if short/sectionless/no-lists; else Google-style; preserve quotes/prefix."""
if not content or not content.strip():
return f"{prefix}{quotes}{quotes}"
style = likely_docstring_style(content)
if style in NON_GOOGLE:
body = "\n".join(line.rstrip() for line in content.rstrip("\n").splitlines())
return f"{prefix}{quotes}{body}{quotes}"
text = content.strip()
has_section = any(f"{s}:" in text for s in SECTIONS)
has_list = any(is_list_item(line) for line in text.splitlines())
single_ok = (
("\n" not in text)
and not has_section
and not has_list
and (indent + len(prefix) + len(quotes) * 2 + len(text) <= width)
)
if single_ok:
words = text.split()
if words and not words[0].startswith(("http://", "https://")) and not words[0][0].isupper():
words[0] = words[0][0].upper() + words[0][1:]
out = " ".join(words)
if out and out[-1] not in ".!?":
out += "."
return f"{prefix}{quotes}{out}{quotes}"
return format_google(text, indent, width, quotes, prefix, start_newline)
class Visitor(ast.NodeVisitor):
"""Collect docstring replacements for classes and functions."""
def __init__(self, src: list[str], width: int = 120, start_newline: bool = False):
"""Init with source lines, target width, and start_newline flag."""
self.src, self.width, self.repl, self.start_newline = src, width, [], start_newline
def visit_Module(self, node):
"""Skip module docstring; visit children."""
self.generic_visit(node)
def visit_ClassDef(self, node):
"""Visit class definition and handle its docstring."""
self._handle(node)
self.generic_visit(node)
def visit_FunctionDef(self, node):
"""Visit function definition and handle its docstring."""
self._handle(node)
self.generic_visit(node)
def visit_AsyncFunctionDef(self, node):
"""Visit async function definition and handle its docstring."""
self._handle(node)
self.generic_visit(node)
def _handle(self, node):
"""If first stmt is a string expr, schedule replacement."""
try:
doc = ast.get_docstring(node, clean=False)
if not doc or not node.body or not isinstance(node.body[0], ast.Expr):
return
s = node.body[0].value
if not (isinstance(s, ast.Constant) and isinstance(s.value, str)):
return
if likely_docstring_style(doc) in NON_GOOGLE:
return
sl, el = node.body[0].lineno - 1, node.body[0].end_lineno - 1
sc, ec = node.body[0].col_offset, node.body[0].end_col_offset
if sl < 0 or el >= len(self.src):
return
original = (
self.src[sl][sc:ec]
if sl == el
else "\n".join([self.src[sl][sc:], *self.src[sl + 1 : el], self.src[el][:ec]])
)
prefix, quotes, _ = detect_opener(original)
formatted = format_docstring(doc, sc, self.width, quotes, prefix, self.start_newline)
if formatted.strip() != original.strip():
self.repl.append((sl, el, sc, ec, formatted))
except Exception:
return
def format_python_file(text: str, width: int = 120, start_newline: bool = False) -> str:
"""Return source with reformatted docstrings; on failure, return original."""
s = text
if not s.strip():
return s
if ('"""' not in s and "'''" not in s) or ("def " not in s and "class " not in s and "async def " not in s):
return s
try:
tree = ast.parse(s)
except SyntaxError:
return s
src = s.splitlines()
v = Visitor(src, width, start_newline=start_newline)
try:
v.visit(tree)
except Exception:
return s
if not v.repl:
return s
for sl, el, sc, ec, rep in reversed(v.repl):
try:
if sl == el:
src[sl] = src[sl][:sc] + rep + src[sl][ec:]
else:
nl = rep.splitlines()
nl[0] = src[sl][:sc] + nl[0]
nl[-1] += src[el][ec:]
src[sl : el + 1] = nl
except Exception:
continue
return "\n".join(src)
def preserve_trailing_newlines(original: str, formatted: str) -> str:
"""Preserve the original trailing newline count."""
o = len(original) - len(original.rstrip("\n"))
f = len(formatted) - len(formatted.rstrip("\n"))
return formatted if o == f else formatted.rstrip("\n") + ("\n" * o)
def read_python_path() -> Path | None:
"""Read the Python path from stdin payload.
Returns:
(Path | None): Python file path when present and valid.
"""
try:
data = json.load(sys.stdin)
except Exception:
return None
file_path = data.get("tool_input", {}).get("file_path", "")
path = Path(file_path) if file_path else None
if not path or path.suffix != ".py" or not path.exists():
return None
if any(
p in path.parts
for p in [
".git",
".venv",
"venv",
"env",
".env",
"__pycache__",
".mypy_cache",
".pytest_cache",
".tox",
".nox",
".eggs",
"eggs",
".idea",
".vscode",
"node_modules",
"site-packages",
"build",
"dist",
".claude",
]
):
return None
return path
def main() -> None:
"""Format Python docstrings in files."""
python_file = read_python_path()
if python_file:
try:
content = python_file.read_text()
formatted = preserve_trailing_newlines(content, format_python_file(content))
if formatted != content:
python_file.write_text(formatted)
print(f"Formatted: {python_file}")
except Exception as e:
output = {
"hookSpecificOutput": {
"hookEventName": "PostToolUse",
"additionalContext": f"Docstring formatting failed for {python_file.name}: {e}",
}
}
print(json.dumps(output))
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,311 @@
#!/usr/bin/env python3
"""
PostToolUse hook: Format Markdown files and embedded code blocks.
Inspired by https://github.com/ultralytics/actions/blob/main/actions/update_markdown_code_blocks.py
"""
from __future__ import annotations
import hashlib
import json
import re
import shutil
import subprocess
import sys
from pathlib import Path
from tempfile import TemporaryDirectory
PYTHON_BLOCK_PATTERN = r"^( *)```(?:python|py|\{[ ]*\.py[ ]*\.annotate[ ]*\})\n(.*?)\n\1```"
BASH_BLOCK_PATTERN = r"^( *)```(?:bash|sh|shell)\n(.*?)\n\1```"
LANGUAGE_TAGS = {"python": ["python", "py", "{ .py .annotate }"], "bash": ["bash", "sh", "shell"]}
def check_prettier_version() -> bool:
"""Check if prettier is installed and warn if version differs from 3.6.2."""
if not shutil.which("npx"):
return False
try:
result = subprocess.run(["npx", "prettier", "--version"],
capture_output=True, text=True, check=False, timeout=5)
if result.returncode == 0:
version = result.stdout.strip()
if "3.6.2" not in version:
print(f"⚠️ Prettier version mismatch: expected 3.6.2, found {version}")
return True
except Exception:
pass
return False
def extract_code_blocks(markdown_content: str) -> dict[str, list[tuple[str, str]]]:
"""Extract code blocks from markdown content.
Args:
markdown_content (str): Markdown text to inspect.
Returns:
(dict): Mapping of language names to lists of (indentation, block) pairs.
"""
python_blocks = re.compile(PYTHON_BLOCK_PATTERN, re.DOTALL | re.MULTILINE).findall(markdown_content)
bash_blocks = re.compile(BASH_BLOCK_PATTERN, re.DOTALL | re.MULTILINE).findall(markdown_content)
return {"python": python_blocks, "bash": bash_blocks}
def remove_indentation(code_block: str, num_spaces: int) -> str:
"""Remove indentation from a block of code.
Args:
code_block (str): Code snippet to adjust.
num_spaces (int): Leading space count to strip.
Returns:
(str): Code with indentation removed.
"""
lines = code_block.split("\n")
stripped_lines = [line[num_spaces:] if len(line) >= num_spaces else line for line in lines]
return "\n".join(stripped_lines)
def add_indentation(code_block: str, num_spaces: int) -> str:
"""Add indentation back to non-empty lines in a code block.
Args:
code_block (str): Code snippet to indent.
num_spaces (int): Space count to prefix.
Returns:
(str): Code with indentation restored.
"""
indent = " " * num_spaces
lines = code_block.split("\n")
return "\n".join([indent + line if line.strip() else line for line in lines])
def format_code_with_ruff(temp_dir: Path) -> None:
"""Format Python files in a temporary directory with Ruff and docstring formatter.
Args:
temp_dir (Path): Directory containing extracted Python blocks.
"""
try:
subprocess.run(["ruff", "format", "--line-length=120", str(temp_dir)], check=True)
print("Completed ruff format ✅")
except Exception as exc:
print(f"ERROR running ruff format ❌ {exc}")
try:
subprocess.run(
[
"ruff",
"check",
"--fix",
"--extend-select=F,I,D,UP,RUF",
"--target-version=py39",
"--ignore=D100,D101,D103,D104,D203,D205,D212,D213,D401,D406,D407,D413,F821,F841,RUF001,RUF002,RUF012",
str(temp_dir),
],
check=True,
)
print("Completed ruff check ✅")
except Exception as exc:
print(f"ERROR running ruff check ❌ {exc}")
# Format docstrings in extracted Python blocks (matches actions pipeline)
try:
from format_python_docstrings import format_python_file
for py_file in Path(temp_dir).glob("*.py"):
content = py_file.read_text()
formatted = format_python_file(content)
if formatted != content:
py_file.write_text(formatted)
print("Completed docstring formatting ✅")
except Exception as exc:
print(f"ERROR running docstring formatter ❌ {exc}")
def format_bash_with_prettier(temp_dir: Path) -> None:
"""Format Bash files in a temporary directory with prettier-plugin-sh.
Args:
temp_dir (Path): Directory containing extracted Bash blocks.
"""
try:
result = subprocess.run(
"npx prettier --write --print-width 120 --plugin=$(npm root -g)/prettier-plugin-sh/lib/index.cjs ./**/*.sh",
shell=True,
capture_output=True,
text=True,
cwd=temp_dir,
)
if result.returncode != 0:
print(f"ERROR running prettier-plugin-sh ❌ {result.stderr}")
else:
print("Completed bash formatting ✅")
except Exception as exc:
print(f"ERROR running prettier-plugin-sh ❌ {exc}")
def generate_temp_filename(file_path: Path, index: int, code_type: str) -> str:
"""Generate a deterministic filename for a temporary code block.
Args:
file_path (Path): Source markdown path.
index (int): Block index for uniqueness.
code_type (str): Language identifier.
Returns:
(str): Safe filename for the temporary code file.
"""
stem = file_path.stem
code_letter = code_type[0]
path_part = str(file_path.parent).replace("/", "_").replace("\\", "_").replace(" ", "-")
hash_val = hashlib.md5(f"{file_path}_{index}".encode(), usedforsecurity=False).hexdigest()[:6]
ext = ".py" if code_type == "python" else ".sh"
filename = f"{stem}_{path_part}_{code_letter}{index}_{hash_val}{ext}"
return re.sub(r"[^\w\-.]", "_", filename)
def process_markdown_file(
file_path: Path,
temp_dir: Path,
process_python: bool = True,
process_bash: bool = True,
) -> tuple[str, list[tuple[int, str, Path, str]]]:
"""Extract code blocks from a markdown file and store them as temporary files.
Args:
file_path (Path): Markdown path to process.
temp_dir (Path): Directory to store temporary files.
process_python (bool, optional): Enable Python block extraction.
process_bash (bool, optional): Enable Bash block extraction.
Returns:
markdown_content (str): Original markdown content.
temp_files (list): Extracted block metadata.
"""
try:
markdown_content = file_path.read_text()
except Exception as exc:
print(f"Error reading file {file_path}: {exc}")
return "", []
code_blocks_by_type = extract_code_blocks(markdown_content)
temp_files: list[tuple[int, str, Path, str]] = []
code_types: list[tuple[str, int]] = []
if process_python:
code_types.append(("python", 0))
if process_bash:
code_types.append(("bash", 1000))
for code_type, offset in code_types:
for i, (indentation, code_block) in enumerate(code_blocks_by_type[code_type]):
num_spaces = len(indentation)
code_without_indentation = remove_indentation(code_block, num_spaces)
temp_file_path = temp_dir / generate_temp_filename(file_path, i + offset, code_type)
try:
temp_file_path.write_text(code_without_indentation)
except Exception as exc:
print(f"Error writing temp file {temp_file_path}: {exc}")
continue
temp_files.append((num_spaces, code_block, temp_file_path, code_type))
return markdown_content, temp_files
def update_markdown_file(file_path: Path, markdown_content: str, temp_files: list[tuple[int, str, Path, str]]) -> None:
"""Replace markdown code blocks with formatted versions.
Args:
file_path (Path): Markdown file to update.
markdown_content (str): Original content.
temp_files (list): Metadata for formatted code blocks.
"""
for num_spaces, original_code_block, temp_file_path, code_type in temp_files:
try:
formatted_code = temp_file_path.read_text().rstrip("\n")
except Exception as exc:
print(f"Error reading temp file {temp_file_path}: {exc}")
continue
formatted_code_with_indentation = add_indentation(formatted_code, num_spaces)
for lang in LANGUAGE_TAGS[code_type]:
markdown_content = markdown_content.replace(
f"{' ' * num_spaces}```{lang}\n{original_code_block}\n{' ' * num_spaces}```",
f"{' ' * num_spaces}```{lang}\n{formatted_code_with_indentation}\n{' ' * num_spaces}```",
)
try:
file_path.write_text(markdown_content)
except Exception as exc:
print(f"Error writing file {file_path}: {exc}")
def run_prettier(markdown_file: Path) -> None:
"""Format a markdown file with Prettier when available.
Args:
markdown_file (Path): Markdown file to format.
"""
if not check_prettier_version():
return
is_docs = "docs" in markdown_file.parts and "reference" not in markdown_file.parts
command = ["npx", "prettier", "--write", "--list-different", "--print-width", "120", str(markdown_file)]
if is_docs:
command = ["npx", "prettier", "--tab-width", "4", "--print-width", "120", "--write", "--list-different", str(markdown_file)]
subprocess.run(command, capture_output=True, check=False, cwd=markdown_file.parent)
def format_markdown_file(markdown_file: Path) -> None:
"""Format markdown-embedded code and run Prettier on the file.
Args:
markdown_file (Path): Markdown file to process.
"""
with TemporaryDirectory() as tmp_dir_name:
temp_dir = Path(tmp_dir_name)
markdown_content, temp_files = process_markdown_file(markdown_file, temp_dir)
if not temp_files:
run_prettier(markdown_file)
return
has_python = any(code_type == "python" for *_, code_type in temp_files)
has_bash = any(code_type == "bash" for *_, code_type in temp_files)
if has_python:
format_code_with_ruff(temp_dir)
if has_bash:
format_bash_with_prettier(temp_dir)
update_markdown_file(markdown_file, markdown_content, temp_files)
run_prettier(markdown_file)
def read_markdown_path() -> Path | None:
"""Read the markdown path from stdin payload.
Returns:
markdown_path (Path | None): Markdown path when present and valid.
"""
try:
data = json.load(sys.stdin)
except Exception:
return None
file_path = data.get("tool_input", {}).get("file_path", "")
path = Path(file_path) if file_path else None
if not path or path.suffix.lower() != ".md" or not path.exists():
return None
if any(p in path.parts for p in ['.git', '.venv', 'venv', 'env', '.env', '__pycache__', '.mypy_cache', '.pytest_cache', '.tox', '.nox', '.eggs', 'eggs', '.idea', '.vscode', 'node_modules', 'site-packages', 'build', 'dist', '.claude']):
return None
return path
def main() -> None:
"""Run markdown formatting hook."""
markdown_file = read_markdown_path()
if markdown_file:
format_markdown_file(markdown_file)
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,66 @@
#!/usr/bin/env python3
"""
PostToolUse hook: Auto-format JS/TS/CSS/JSON/YAML/HTML/Vue/Svelte files with prettier
"""
import json
import re
import shutil
import subprocess
import sys
from pathlib import Path
# File extensions that prettier handles
PRETTIER_EXTENSIONS = {'.js', '.jsx', '.ts', '.tsx', '.css', '.less', '.scss',
'.json', '.yml', '.yaml', '.html', '.vue', '.svelte'}
LOCK_FILE_PATTERN = re.compile(r'.*lock\.(json|yaml|yml)$|.*\.lock$')
def check_prettier_version() -> bool:
"""Check if prettier is installed and warn if version differs from 3.6.2."""
if not shutil.which('npx'):
return False
try:
result = subprocess.run(['npx', 'prettier', '--version'],
capture_output=True, text=True, check=False, timeout=5)
if result.returncode == 0:
version = result.stdout.strip()
if '3.6.2' not in version:
print(f"⚠️ Prettier version mismatch: expected 3.6.2, found {version}")
return True
except Exception:
pass
return False
def main():
try:
data = json.load(sys.stdin)
file_path = data.get("tool_input", {}).get("file_path", "")
if not file_path:
sys.exit(0)
py_file = Path(file_path)
if not py_file.exists() or py_file.suffix not in PRETTIER_EXTENSIONS:
sys.exit(0)
# Skip virtual env, cache, .claude directories, lock files, model.json, and minified assets
if any(p in py_file.parts for p in ['.git', '.venv', 'venv', 'env', '.env', '__pycache__', '.mypy_cache', '.pytest_cache', '.tox', '.nox', '.eggs', 'eggs', '.idea', '.vscode', 'node_modules', 'site-packages', 'build', 'dist', '.claude']) or LOCK_FILE_PATTERN.match(py_file.name) or py_file.name == 'model.json' or py_file.name.endswith(('.min.js', '.min.css')):
sys.exit(0)
# Check if prettier is available
if not check_prettier_version():
sys.exit(0)
# Run prettier
subprocess.run([
'npx', 'prettier', '--write', '--list-different', '--print-width', '120', str(py_file)
], capture_output=True, check=False, cwd=py_file.parent)
except Exception:
pass
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,67 @@
#!/usr/bin/env python3
"""PostToolUse hook: Auto-format Python files with ruff, provide feedback on errors."""
import json
import shutil
import subprocess
import sys
from pathlib import Path
EXCLUDED_DIRS = {'.git', '.venv', 'venv', 'env', '.env', '__pycache__', '.mypy_cache', '.pytest_cache',
'.tox', '.nox', '.eggs', 'eggs', '.idea', '.vscode', 'node_modules', 'site-packages',
'build', 'dist', '.claude'}
def main():
try:
data = json.load(sys.stdin)
except Exception:
sys.exit(0)
file_path = data.get("tool_input", {}).get("file_path", "")
if not file_path.endswith('.py'):
sys.exit(0)
py_file = Path(file_path)
if not py_file.exists() or any(p in py_file.parts for p in EXCLUDED_DIRS):
sys.exit(0)
if not shutil.which('ruff'):
sys.exit(0)
work_dir = py_file.parent
issues = []
# Run ruff check with fixes
check_result = subprocess.run([
'ruff', 'check', '--fix',
'--extend-select', 'F,I,D,UP,RUF,FA',
'--target-version', 'py39',
'--ignore', 'D100,D104,D203,D205,D212,D213,D401,D406,D407,D413,RUF001,RUF002,RUF012',
str(py_file)
], capture_output=True, text=True, cwd=work_dir)
if check_result.returncode != 0:
error_output = check_result.stdout.strip() or check_result.stderr.strip()
issues.append(f'Ruff check found unfixable errors in {py_file.name}:\n{error_output}')
# Run ruff format regardless of check result
format_result = subprocess.run([
'ruff', 'format', '--line-length', '120', str(py_file)
], capture_output=True, text=True, cwd=work_dir)
if format_result.returncode != 0:
error_output = format_result.stderr.strip()
issues.append(f'Ruff format failed for {py_file.name}:\n{error_output}')
# Output single JSON with all collected feedback
if issues:
output = {"hookSpecificOutput": {"hookEventName": "PostToolUse",
"additionalContext": "\n\n".join(issues)}}
print(json.dumps(output))
sys.exit(0)
if __name__ == '__main__':
main()