Source code for sphinx_immaterial.apidoc.cpp.api_parser
"""Converts a preprocessed C++ source file into a JSON API description.
This program expects as input preprocessed C++ source code obtained using the
`-E` (preprocessed), `-C` (preserve comments), and `-dD` (preserve macro
definitions) GCC/Clang options. It also accepts a list of compiler options to
be passed to Clang, although these are mostly irrelevant due to the prior
preprocessing.
It uses a combination of the libclang Python bindings and the C++ parser in the
Sphinx Python package to parse declarations and definitions and produce a JSON
representation of the API.
The JSON representation describes the API as a set of "entities", each with a
unique id, and relations between them.
Libclang can handle arbitrary C++ syntax, but provides very restricted access to
the resultant AST. To workaround those limitations, in some cases extracted
declarations or portions of declarations are converted back to source
representation and re-parsed using the C++ parser provided by the Sphinx C++
domain. The C++ parser in Sphinx is extremely limited and as it does not rely
on a symbol table does not always correctly handle template arguments, but
provides full access to its AST.
This program performs a number of transformations on the declarations:
- Exported entities are filtered using various criteria
- `std::enable_if_t` uses are converted to C++20 requires clauses.
- Internal return types are elided (replaced with auto).
"""
import argparse
import dataclasses
import functools
import json
import os
import pathlib
import re
import time
import typing
from typing import (
cast,
Dict,
Any,
List,
Tuple,
Optional,
Sequence,
Union,
Pattern,
Literal,
Callable,
TypedDict,
Iterator,
)
from textwrap import dedent
import ctypes
import clang.cindex
from clang.cindex import (
Cursor,
CursorKind,
Token,
TokenKind,
TranslationUnit,
SourceLocation,
SourceRange,
)
import docutils.nodes
import pydantic.dataclasses
import sphinx.domains.cpp
import sphinx.util.logging
from typing_extensions import NotRequired
from . import ast_fixes # noqa: F401
logger = sphinx.util.logging.getLogger(__name__)
_UNMATCHABLE_REGEXP = re.compile("a^")
def _combine_regexp_list(items: Sequence[Union[str, Pattern[str]]]) -> re.Pattern:
if not items:
return _UNMATCHABLE_REGEXP
def get_parenthesized_source(x: Union[str, Pattern[str]]):
if isinstance(x, re.Pattern):
x = x.pattern
return f"(?:{x})"
return re.compile("|".join(get_parenthesized_source(x) for x in items))
def _make_replacement_pattern(
strings: List[str], prefix: str, suffix: str
) -> re.Pattern:
if not strings:
return _UNMATCHABLE_REGEXP
return re.compile(
"|".join(rf"(?:{prefix}{re.escape(before)}{suffix})" for before in strings)
)
TEMPLATE_PARAMETER_ENABLE_IF_TYPE_PATTERN = re.compile(
r"\s*(?:typename|class)\s*=\s*std\s*::\s*enable_if_t\s*<(.*)>\s*"
)
TEMPLATE_PARAMETER_ENABLE_IF_NON_TYPE_PATTERN = re.compile(
r"\s*std\s*::\s*enable_if_t\s*<(.*)>\s*\*\s*=\s*(nullptr|0)\s*"
)
SPECIAL_GROUP_COMMAND_PATTERN = re.compile(
r"^(?:\\|@)(ingroup|relates|membergroup|id)\s+(.*[^\s])\s*$", re.MULTILINE
)
[docs]
@pydantic.dataclasses.dataclass
class Config:
"""Specifies a C++ API parsing configuration.
Based on this configuration, a description of the API is generated.
"""
input_path: str = "__input.cpp"
"""Path to the input file to parse.
This may either be a path to an existing file, or `.input_content` may
specify its content, in which case the filesystem is not accessed.
If `.input_content` is specified and merely contains :cpp:`#include`
directives, then the actual path does not matter and may be left as the
default value.
"""
input_content: Optional[bytes] = None
"""Specifies the content of `.input_path`.
If unspecified, the content is read from filesystem.
"""
compiler_flags: List[str] = dataclasses.field(default_factory=list)
"""List of compiler flags to pass to Clang."""
verbose: bool = False
"""Parse in verbose mode."""
include_directory_map: Dict[str, str] = dataclasses.field(default_factory=dict)
"""Maps actual include directories to a displayed directory name.
The keys should be prefixes of paths specified in error messages/source
locations identified by clang.
The values should be the corresponding prefix to use in the documented
:cpp:`#include` paths.
"""
allow_paths: List[Pattern] = dataclasses.field(
default_factory=lambda: [re.compile("")]
)
"""List of regular expressions matching *allowed* paths.
Only entities defined in files that match `.allow_paths`, and don't match
`.disallow_paths`, are documented. By default all entities are documented,
but this default is not normally usable, because it will include entities
defined in the standard library and third-party libraries.
.. important::
When building on Windows, all path separators are normalized to :python:`"/"`.
Therefore, in the specified regular expressions, always use :python:`"/"` to
match a path separator.
"""
disallow_paths: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *disallowed* paths.
Entities defined in files matching any of these patterns are not documented.
.. important::
When building on Windows, all path separators are normalized to :python:`"/"`.
Therefore, in the specified regular expressions, always use :python:`"/"` to
match a path separator.
"""
disallow_namespaces: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *disallowed* namespaces.
Entities defined in namespaces matching any of the specified patterns are
not documented.
"""
allow_symbols: List[Pattern] = dataclasses.field(
default_factory=lambda: [re.compile("")]
)
"""List of regular expressions matching *allowed* symbols.
Only symbols matching one of the `.allow_symbols` patterns, and not matching
`.disallow_symbols`, are documented. By default, all symbols are allowed.
"""
disallow_symbols: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *disallowed* symbols.
Symbols matching any of these patterns are undocumented.
"""
allow_macros: List[Pattern] = dataclasses.field(
default_factory=lambda: [re.compile("")]
)
"""List of regular expressions matching *allowed* macros.
Only macros names matching `.allow_macros`, and not matching
`.disallow_macros`, are documented.
"""
disallow_macros: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *disallowed* macro names.
Macros matching any of these patterns are undocumented.
"""
ignore_diagnostics: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching diagnostics to ignore.
Diagnostics matching any of these patterns are ignored.
"""
template_parameter_enable_if_patterns: List[Pattern] = dataclasses.field(
default_factory=lambda: [
TEMPLATE_PARAMETER_ENABLE_IF_TYPE_PATTERN,
TEMPLATE_PARAMETER_ENABLE_IF_NON_TYPE_PATTERN,
]
)
type_replacements: Dict[str, str] = dataclasses.field(default_factory=dict)
"""Remaps type names."""
hide_types: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *hidden* types.
Matching return types are replaced with :cpp:`auto`, and matching
initializers are elided.
"""
ignore_template_parameters: List[Pattern] = dataclasses.field(default_factory=list)
"""List of regular expressions matching *ignored* template parameters.
Template parameters with a declaration matching any of these patterns are
excluded from the generated documentation.
"""
hide_initializers: List[Pattern] = dataclasses.field(
default_factory=lambda: [re.compile(r"^=\s*(?:(true|false)\s*$|\[)")]
)
"""List of regular expressions matching initializers to elide.
Any matching initializer expression is elided from the generated
documentation.
"""
# Derived from `allow_paths`.
allow_path_pattern: Pattern = dataclasses.field(init=False)
# Derived from `disallow_paths`.
disallow_path_pattern: Pattern = dataclasses.field(init=False)
# Derived from `allow_symbols`.
allow_symbols_pattern: Pattern = dataclasses.field(init=False)
# Derived from `disallow_symbols`.
disallow_symbols_pattern: Pattern = dataclasses.field(init=False)
# Derived from `allow_macros`.
allow_macros_pattern: Pattern = dataclasses.field(init=False)
# Derived from `disallow_macros`.
disallow_macros_pattern: Pattern = dataclasses.field(init=False)
# Derived from `ignore_diagnostics`.
ignore_diagnostics_pattern: Pattern = dataclasses.field(init=False)
# Derived from `hide_types`.
hide_types_pattern: Pattern = dataclasses.field(init=False)
# Derived from `type_replacements`.
type_replacements_pattern: Pattern = dataclasses.field(init=False)
# Derived from `ignore_template_parameters`.
ignore_template_parameters_pattern: Pattern = dataclasses.field(init=False)
# Derived from `hide_initializers`.
hide_initializers_pattern: Pattern = dataclasses.field(init=False)
include_directory_map_pattern: Pattern = dataclasses.field(init=False)
disallow_namespaces_pattern: Pattern = dataclasses.field(init=False)
def __post_init__(self):
self.allow_path_pattern = _combine_regexp_list(self.allow_paths) # type: ignore[misc]
self.disallow_path_pattern = _combine_regexp_list(self.disallow_paths) # type: ignore[misc]
self.allow_path_pattern = _combine_regexp_list(self.allow_paths) # type: ignore[misc]
self.disallow_namespaces_pattern = _combine_regexp_list(
self.disallow_namespaces
)
self.allow_symbols_pattern = _combine_regexp_list(self.allow_symbols) # type: ignore[misc]
self.disallow_symbols_pattern = _combine_regexp_list(self.disallow_symbols) # type: ignore[misc]
self.allow_macros_pattern = _combine_regexp_list(self.allow_macros) # type: ignore[misc]
self.disallow_macros_pattern = _combine_regexp_list(self.disallow_macros) # type: ignore[misc]
self.ignore_diagnostics_pattern = _combine_regexp_list(self.ignore_diagnostics) # type: ignore[misc]
self.hide_types_pattern = _combine_regexp_list(self.hide_types) # type: ignore[misc]
self.type_replacements_pattern = _make_replacement_pattern( # type: ignore[misc]
list(self.type_replacements.keys()), prefix=r"\b", suffix=r"\b"
)
self.ignore_template_parameters_pattern = _combine_regexp_list( # type: ignore[misc]
self.ignore_template_parameters
)
self.hide_initializers_pattern = _combine_regexp_list(self.hide_initializers) # type: ignore[misc]
if os.name == "nt":
self.normalized_include_directory_map = { # type: ignore[misc]
key.replace("\\", "/"): value
for key, value in self.include_directory_map.items()
}
else:
self.normalized_include_directory_map = self.include_directory_map # type: ignore[misc]
self.include_directory_map_pattern = _make_replacement_pattern( # type: ignore[misc]
list(self.normalized_include_directory_map.keys()), prefix="^", suffix=""
)
self.cached_mapped_include_directories = {} # type: ignore[misc]
normalized_include_directory_map: Dict[str, str] = dataclasses.field(init=False)
cached_mapped_include_directories: Dict[str, str] = dataclasses.field(init=False)
def map_include_path(self, path: str) -> str:
mapped = self.cached_mapped_include_directories.get(path)
if mapped is not None:
return mapped
if os.name == "nt":
path = path.replace("\\", "/")
if path.startswith("./"):
path = path[2:]
new_mapped = self.include_directory_map_pattern.sub(
lambda m: self.normalized_include_directory_map[m.group(0)], path
)
self.cached_mapped_include_directories[path] = new_mapped
return new_mapped
EntityId = str
EntityKind = Literal[
"class",
"conversion_function",
"function",
"method",
"constructor",
"var",
"alias",
"enum",
]
FunctionEntityKind = Literal[
"conversion_function", "function", "method", "constructor", "destructor"
]
ClassKeyword = Literal["class", "struct"]
class JsonLocation(TypedDict):
file: str
line: int
col: int
class JsonDocComment(TypedDict):
text: str
location: JsonLocation
TemplateParameterKind = Literal["type", "template", "non_type"]
class TemplateParameter(TypedDict):
declaration: str
name: str
kind: TemplateParameterKind
pack: bool
class CppApiEntityBase(TypedDict, total=False):
id: EntityId
parent: NotRequired[EntityId]
scope: NotRequired[str]
doc: NotRequired[Optional[JsonDocComment]]
document_with: NotRequired[EntityId]
siblings: NotRequired[List[EntityId]]
name: str
template_parameters: NotRequired[Optional[List[TemplateParameter]]]
location: JsonLocation
special_id: NotRequired[Optional[str]]
page_name: str
requires: Optional[List[str]]
specializes: Union[None, EntityId, Literal[True]]
related_members: Dict[str, List[EntityId]]
related_nonmembers: Dict[str, List[EntityId]]
special_membergroup: str
special_ingroup: str
special_relates: str
document_prefix: str
nonitpick: List[str]
class FunctionEntity(CppApiEntityBase):
kind: FunctionEntityKind
arity: int
name_substitute: str
friend: bool
declaration: str
class BaseClass(TypedDict):
type: str
access: str
class ClassEntity(CppApiEntityBase):
kind: Literal["class"]
keyword: ClassKeyword
prefix: List[str]
bases: List[BaseClass]
class VarEntity(CppApiEntityBase):
kind: Literal["var"]
declaration: str
name_substitute: str
initializer: Optional[str]
class TypeAliasEntity(CppApiEntityBase):
kind: Literal["alias"]
underlying_type: Optional[str]
class MacroEntity(CppApiEntityBase):
kind: Literal["macro"]
parameters: Optional[List[str]]
class EnumeratorEntity(TypedDict):
kind: Literal["enumerator"]
id: EntityId
name: str
decl: str
doc: Optional[JsonDocComment]
location: JsonLocation
class EnumEntity(CppApiEntityBase):
kind: Literal["enum"]
keyword: Optional[ClassKeyword]
enumerators: List[EnumeratorEntity]
CppApiEntity = Union[
ClassEntity, FunctionEntity, VarEntity, TypeAliasEntity, MacroEntity, EnumEntity
]
def json_location_to_string(location: Optional[JsonLocation]) -> Optional[str]:
if location is None:
return None
return "%s:%s:%s" % (location["file"], location["line"], location["col"])
def get_entity_id(cursor: Cursor) -> EntityId:
# USR workarounds from:
# https://github.com/foonathan/cppast/blob/e558e2d58f519e3a83af770d460672b1d4ba2886/src/libclang/parse_functions.cpp#L13
usr = cursor.get_usr()
if cursor.kind in (CursorKind.FUNCTION_TEMPLATE, CursorKind.CONVERSION_FUNCTION):
# Combine return type with USR to prevent collisions
return f"{usr} {cursor.result_type.spelling}"
if cursor.kind == CursorKind.CLASS_TEMPLATE_PARTIAL_SPECIALIZATION:
# libclang issue: templ<T()> vs templ<T() &>
# but identical USR
# same workaround: combine display name with usr
# (and hope this prevents all collisions...)
return f"{usr} {cursor.displayname}"
return usr
def _substitute_internal_type_names(config: Config, decl: str) -> str:
return config.type_replacements_pattern.sub(
lambda m: config.type_replacements[m.group(0)], decl
)
def get_previous_line_location(tu, location: SourceLocation):
file = location.file
line = location.line
return SourceLocation.from_position(tu, file, line - 1, 1)
def get_presumed_location(location: SourceLocation) -> typing.Tuple[str, int, int]:
file, line, col = clang.cindex._CXString(), ctypes.c_uint(), ctypes.c_uint()
clang.cindex.conf.lib.clang_getPresumedLocation(
location, ctypes.byref(file), ctypes.byref(line), ctypes.byref(col)
)
return (clang.cindex._CXString.from_result(file), int(line.value), int(col.value))
_clang_getFileContents = clang.cindex.conf.lib.clang_getFileContents
_clang_getFileContents.restype = ctypes.c_void_p
_PyMemoryView_FromMemory = ctypes.pythonapi.PyMemoryView_FromMemory
_PyMemoryView_FromMemory.argtypes = (ctypes.c_char_p, ctypes.c_ssize_t, ctypes.c_int)
_PyMemoryView_FromMemory.restype = ctypes.py_object
def _get_file_contents(tu, f):
size = ctypes.c_size_t()
ptr = _clang_getFileContents(tu, f, ctypes.byref(size))
buf = _PyMemoryView_FromMemory(ctypes.cast(ptr, ctypes.c_char_p), size.value, 0x100)
return buf
def _get_template_cursor_kind(cursor: Cursor) -> CursorKind:
return CursorKind.from_id(clang.cindex.conf.lib.clang_getTemplateCursorKind(cursor))
def _get_specialized_cursor_template(cursor: Cursor) -> typing.Optional[Cursor]:
return clang.cindex.conf.lib.clang_getSpecializedCursorTemplate(cursor)
def _get_full_nested_name(cursor: typing.Optional[Cursor]) -> str:
if cursor is None:
return ""
ancestors = []
while True:
if cursor.kind == CursorKind.TRANSLATION_UNIT:
break
if cursor.kind == CursorKind.NAMESPACE:
name = cursor.spelling
else:
name = cursor.displayname
ancestors.append(name + "::")
cursor = cursor.semantic_parent
ancestors.reverse()
return "".join(ancestors)
CLASS_KINDS = (
CursorKind.STRUCT_DECL,
CursorKind.CLASS_DECL,
CursorKind.CLASS_TEMPLATE,
CursorKind.CLASS_TEMPLATE_PARTIAL_SPECIALIZATION,
)
def _get_all_decls(
config: Config, cursor: Cursor, allow_file
) -> Iterator[tuple[Cursor, SourceLocation]]:
NAMESPACE = CursorKind.NAMESPACE
doc_comment_start_bound = cursor.location
for child in cursor.get_children():
location = child.location
if location.file is None:
continue
prev_doc_comment_start_bound = doc_comment_start_bound
doc_comment_start_bound = child.extent.end
kind = child.kind
if kind == NAMESPACE:
if (
not allow_file or allow_file(get_presumed_location(location)[0])
) and not config.disallow_namespaces_pattern.match(child.spelling):
yield from _get_all_decls(config, child, None)
continue
if kind not in ALLOWED_KINDS:
continue
if allow_file and not allow_file(get_presumed_location(location)[0]):
continue
if child.kind == CursorKind.MACRO_DEFINITION:
yield (child, prev_doc_comment_start_bound)
continue
yield (child, prev_doc_comment_start_bound)
if kind in CLASS_KINDS:
yield from _get_all_decls(config, child, None)
# Matches the start of a doc comment.
#
# This is used to test if an individual comment token is a doc comment.
_DOC_COMMENT_START = re.compile(
r"""
(?:
//
(?:/|!)
)
|
(?:
/\*
(?:!|\*)
)
""",
re.VERBOSE,
)
# Matches one or more doc comments with a "<" introducer to indicate that the
# doc comment applies to the entity before it, rather than the entity after it.
#
# This is used by `_get_raw_comments_after`.
_DOC_COMMENT_AFTER = re.compile(
rb"""
(
\s* # Skip leading whitespace
(?:
(
// # Comment introducer
(?:/|!)< # Doc comment indicator
[^\r\n]* # Comment text
\r? # Optionally ignored CR
$ # End of comment line
)
|
(
/\* # Comment introducer
(?:!|\*)< # Doc comment indicator
(?:.|\n)*? # Comment text
\*/ # Comment terminator
)
)
)+
""",
re.MULTILINE | re.VERBOSE,
)
def _get_raw_comments(
cursor: Cursor, doc_comment_start_bound: SourceLocation
) -> Optional[tuple[str, SourceLocation]]:
# libclang exposes `cursor.raw_comment` but in some cases it appears to be
# `None` even if there is in fact a comment. Instead, extract the comments
# by searching for comment tokens directly.
translation_unit = cursor.translation_unit
if cursor.kind == CursorKind.MACRO_DEFINITION:
# The extent for macro definitions skips the initial "#define". As a
# workaround, set the end location to the beginning of the line.
orig_location = cursor.location
end_location = SourceLocation.from_position(
translation_unit, orig_location.file, orig_location.line, 1
)
else:
for token in cursor.get_tokens():
end_location = token.location
break
else:
end_location = cursor.location
if (
doc_comment_start_bound.file is None
or end_location.file is None
or doc_comment_start_bound.file.name != end_location.file.name # type: ignore[attr-defined]
):
doc_comment_start_bound = SourceLocation.from_offset(
translation_unit, end_location.file, 0
)
tokens = list(
translation_unit.get_tokens(
extent=SourceRange.from_locations(doc_comment_start_bound, end_location)
)
)
tokens.reverse()
COMMENT = TokenKind.COMMENT
comment_tokens: list[Token] = []
for token in tokens:
token_location = token.extent.end
if token_location.file.name != end_location.file.name: # type: ignore[attr-defined]
break
if token_location.line < end_location.line - 1:
break
if token_location.offset >= end_location.offset:
continue
if token.kind != COMMENT:
break
end_location = token_location
comment_tokens.append(token)
if not comment_tokens:
return None
comment_tokens.reverse()
# Convert comment tokens back into a string, preserving indentation and line
# breaks.
comment_text_parts = []
prev_line = None
prev_indent = 0
first_doc_comment_token_i = -1
doc_comment_end_part_i = 0
for token_i, token in enumerate(comment_tokens):
spelling = token.spelling
is_doc_comment = _DOC_COMMENT_START.match(spelling) is not None
if first_doc_comment_token_i == -1:
if not is_doc_comment:
continue
first_doc_comment_token_i = token_i
token_location = token.location
line = token_location.line
if prev_line is not None and prev_line != line:
comment_text_parts.append("\n")
prev_line = line
prev_indent = 0
token_end_location = token.extent.end
column = token_location.column
extra_indent = column - prev_indent - 1
if extra_indent > 0:
comment_text_parts.append(" " * extra_indent)
comment_text_parts.append(spelling)
if is_doc_comment:
doc_comment_end_part_i = len(comment_text_parts)
prev_line = token_end_location.line
prev_indent = token_end_location.column
if not comment_text_parts:
return None
return (
"".join(comment_text_parts[:doc_comment_end_part_i]),
comment_tokens[first_doc_comment_token_i].location,
)
def _get_raw_comments_after(
tu, location: SourceLocation
) -> Optional[tuple[str, SourceLocation]]:
buf = memoryview(_get_file_contents(tu, location.file))
m = _DOC_COMMENT_AFTER.match(buf, location.offset + 1)
if m is None:
return None
return (" " * (location.column - 1) + m.group(0).decode("utf-8") + "\n", location)
# Matches a single multi-line comment, a single-line non-doc comment, or a
# sequence of single-line same-style doc comments.
_COMMENT_PATTERN = re.compile(
r"""
( # "//" comment (capture group 1)
[ \t]* # Skip leading whitespace on first line
// # Comment introducer
((?:/|!)<?)? # Optional doc comment indicator (capture group 2)
[^\n]* # Comment text
\n # End of first line.
(?: # Zero or more lines with the same doc comment indicator
[ \t]* # Skip leading whitspace
//\2 # Comment introducer and doc comment indicator.
[^\n]* # Comment text
\n # End of comment line
)*
)
|
( # "/*" comment (capture group 3)
[ \t]* # Skip leading whitespace
/\* # Comment introducer
((?:\*|!)<?) # Optional doc comment indicator (capture group 4)
(?:.|\n)*? # Comment text
\*/ # Comment terminator
)
""",
re.VERBOSE,
)
def _convert_raw_comment_into_doc_comment(raw_comment: str) -> str:
# Eliminate CR characters
raw_comment = raw_comment.replace("\r", "") + "\n"
pos = 0
parts: list[str] = []
while (m := _COMMENT_PATTERN.match(raw_comment, pos)) is not None:
pos = m.end(0)
if not m.group(2) and not m.group(4):
# Non-doc comment, replace with empty lines to preserve line number mapping
parts.append("\n" * m.group(0).count("\n"))
continue
if m.group(1):
# // comment
without_comment_prefix = re.sub(
r"^[ \t]*//" + re.escape(m.group(2)), "", m.group(0), flags=re.MULTILINE
)
else:
# /* comment
without_comment_prefix = (
raw_comment[m.start(0) : m.start(4) - 2]
+ " " * (2 + len(m.group(4)))
+ raw_comment[m.end(4) : m.end(0) - 2]
)
# Check if every line is prefixed with an asterisk at the same
# column as the initial "/*".
orig_text = m.group(0)
if re.fullmatch(r"([ \t]*)/\*[^\n]*(\n\1 \*[^\n]*)*(\s*\*/)?", orig_text):
without_comment_prefix = re.sub(
r"^([ \t]*)\*", r"\1 ", without_comment_prefix, flags=re.MULTILINE
)
parts.append(dedent(without_comment_prefix))
assert not raw_comment[pos:].strip(), "Unexpected syntax in raw comment"
return "".join(parts).rstrip()
_CURSOR_KINDS_THAT_ALLOW_DOC_COMMENTS_AFTER = frozenset(
[
CursorKind.VAR_DECL,
CursorKind.FIELD_DECL,
# May be variable template.
CursorKind.UNEXPOSED_DECL,
CursorKind.TYPE_ALIAS_DECL,
CursorKind.TYPEDEF_DECL,
CursorKind.TYPE_ALIAS_TEMPLATE_DECL,
CursorKind.ENUM_CONSTANT_DECL,
]
)
def _get_doc_comment(
config: Config, cursor: Cursor, doc_comment_start_bound: SourceLocation
) -> Optional[JsonDocComment]:
raw_comment = _get_raw_comments(cursor, doc_comment_start_bound)
if (
raw_comment is None
and cursor.kind in _CURSOR_KINDS_THAT_ALLOW_DOC_COMMENTS_AFTER
):
raw_comment = _get_raw_comments_after(
cursor.translation_unit, cursor.extent.end
)
if raw_comment is None:
return None
raw_comment_text, comment_location = raw_comment
comment_text = _convert_raw_comment_into_doc_comment(raw_comment_text)
return {
"text": comment_text,
"location": _get_location_json(config, comment_location),
}
class Extractor:
def __init__(self, config: Config):
self.config = config
input_path = config.input_path
input_content = config.input_content
if input_content is None:
input_content = pathlib.Path(input_path).read_bytes()
input_content = re.sub(
b"#pragma clang module", b"//#pragma clang module", input_content
)
self.input_source = input_content
self.index = clang.cindex.Index.create()
start_time = time.time()
self.tu = self.index.parse(
input_path,
unsaved_files=[(input_path, input_content)],
args=tuple(config.compiler_flags) + ("-ferror-limit=0",),
options=( # TranslationUnit.PARSE_SKIP_FUNCTION_BODIES +
TranslationUnit.PARSE_DETAILED_PROCESSING_RECORD
),
)
end_time = time.time()
if config.verbose:
logger.info("Parsed C++ input in %.5f seconds", end_time - start_time)
for diag in self.tu.diagnostics:
if config.ignore_diagnostics_pattern.search(diag.spelling):
if config.verbose:
logger.info(
diag.spelling,
location=_get_location_string(config, diag.location),
)
continue
logger.error(
diag.spelling, location=_get_location_string(config, diag.location)
)
def _allow_file(path: str) -> bool:
path = config.map_include_path(path)
if not config.allow_path_pattern.search(path):
return False
if config.disallow_path_pattern.search(path):
return False
return True
self.decls = list(
_get_all_decls(
config, self.tu.cursor, functools.lru_cache(maxsize=None)(_allow_file)
)
)
EXCLUDED_COMPILER_FLAGS = frozenset(
[
"-Xclang=-disable-noundef-analysis",
]
)
def _transform_type_alias_decl(config: Config, decl: Cursor):
underlying_type: Optional[str] = _substitute_internal_type_names(
config, decl.underlying_typedef_type.spelling
)
assert underlying_type is not None
if config.hide_types_pattern.search(underlying_type):
underlying_type = None
return {
"kind": "alias",
"name": decl.spelling,
"underlying_type": underlying_type,
}
def get_extent_spelling(translation_unit: TranslationUnit, extent: SourceRange) -> str:
"""Returns the C++ source representation for the specified extent.
Comments are excluded and for simplicity all tokens are separated by
whitespace. This results in excessive whitespace, but that does not matter
because this is intended to be parsed by the Sphinx cpp domain anyway.
"""
def get_spellings():
prev_token = None
COMMENT = TokenKind.COMMENT
for token in translation_unit.get_tokens(extent=extent):
if prev_token is not None:
yield prev_token.spelling
prev_token = None
if token.kind == COMMENT:
continue
prev_token = token
# We need to handle the last token specially, because clang sometimes parses
# ">>" as a single token but the extent may cover only the first of the two
# angle brackets.
if prev_token is not None:
spelling = prev_token.spelling
token_end = cast(SourceLocation, prev_token.extent.end)
offset_diff = token_end.offset - cast(SourceLocation, extent.end).offset
if offset_diff != 0:
yield spelling[:-offset_diff]
else:
yield spelling
return " ".join(get_spellings())
def get_related_comments(decl: Cursor):
# casts below are workaround for: https://github.com/tgockel/types-clang/pull/2
start = cast(SourceLocation, decl.extent.start)
end = cast(SourceLocation, decl.extent.end)
# Move forward one line to avoid skipping any comments on the last line
tu = decl.translation_unit
end = SourceLocation.from_position(tu, end.file, end.line + 1, 1)
COMMENT = TokenKind.COMMENT
for token in tu.get_tokens(
extent=SourceRange.from_locations(cast(int, start), cast(int, end))
):
if token.kind != COMMENT:
continue
yield token
NONITPICK_PATTERN = re.compile(r"//\s*NONITPICK:\s*(.*[^\s])\s*")
def get_nonitpick_directives(decl: Cursor) -> List[str]:
directives = []
for comment in get_related_comments(decl):
text = comment.spelling
m = NONITPICK_PATTERN.match(text)
if m is None:
continue
directives.append(m.group(1))
return directives
TEMPLATE_CURSOR_KINDS = frozenset(
[
CursorKind.FUNCTION_TEMPLATE,
CursorKind.CLASS_TEMPLATE,
CursorKind.CLASS_TEMPLATE_PARTIAL_SPECIALIZATION,
CursorKind.TYPE_ALIAS_TEMPLATE_DECL,
]
)
TEMPLATE_PARAMETER_KIND_TO_JSON_KIND = {
CursorKind.TEMPLATE_TYPE_PARAMETER: "type",
CursorKind.TEMPLATE_NON_TYPE_PARAMETER: "non_type",
CursorKind.TEMPLATE_TEMPLATE_PARAMETER: "template",
}
def _clang_template_parameter_to_json(config: Config, decl: Cursor):
param_decl_str = get_extent_spelling(decl.translation_unit, decl.extent)
param = _parse_template_parameter(param_decl_str)
spelling = decl.spelling
if param is None:
return {
"declaration": param_decl_str,
"name": spelling,
"kind": TEMPLATE_PARAMETER_KIND_TO_JSON_KIND[decl.kind],
# Heuristic to determine if it is a pack.
"pack": "..." in param_decl_str,
}
return _sphinx_ast_template_parameter_to_json(config, param)
def _get_template_parameters(config: Config, decl: Cursor):
if decl.kind not in TEMPLATE_CURSOR_KINDS:
return None
result = []
for child in decl.get_children():
if child.kind not in (
CursorKind.TEMPLATE_TYPE_PARAMETER,
CursorKind.TEMPLATE_NON_TYPE_PARAMETER,
CursorKind.TEMPLATE_TEMPLATE_PARAMETER,
):
continue
result.append(_clang_template_parameter_to_json(config, child))
return result
def _get_non_template_kind(cursor: Cursor):
kind = cursor.kind
if kind not in TEMPLATE_CURSOR_KINDS:
return kind
return _get_template_cursor_kind(cursor)
def _transform_type_alias_template_decl(
config: Config, decl: Cursor
) -> TypeAliasEntity:
underlying_type: Optional[str]
for child in decl.get_children():
if child.kind == CursorKind.TYPE_ALIAS_DECL:
underlying_type = _substitute_internal_type_names(
config, child.underlying_typedef_type.spelling
)
break
else:
raise ValueError("Could not determine underlying type")
requires = []
if re.search(r"^\s*std\s*::\s*enable_if_t\s*<", underlying_type) is not None:
presumed_file, presumed_line, _ = get_presumed_location(decl.location)
parser = sphinx.domains.cpp.DefinitionParser(
underlying_type,
location=(presumed_file, presumed_line),
config=cast(sphinx.config.Config, SphinxConfig()),
)
ast = parser._parse_type(False)
parser.skip_ws()
parser.assert_end()
assert isinstance(ast, sphinx.domains.cpp.ASTType)
requires_expr = _extract_requires_from_enable_if_t_type(config, ast)
if requires_expr is not None:
requires.append(requires_expr)
underlying_type = str(ast)
if config.hide_types_pattern.search(underlying_type) is not None:
underlying_type = None
return {
"kind": "alias",
"name": decl.spelling,
"underlying_type": underlying_type,
"requires": requires,
}
def _get_class_keyword(kind: CursorKind) -> ClassKeyword:
return "class" if kind == CursorKind.CLASS_DECL else "struct"
def _get_bases(config: Config, decl: Cursor):
for child in decl.get_children():
if child.kind != CursorKind.CXX_BASE_SPECIFIER:
continue
type_spelling = _substitute_internal_type_names(config, child.type.spelling)
if config.hide_types_pattern.search(type_spelling) is not None:
continue
yield {"type": type_spelling, "access": child.access_specifier.name.lower()}
def _transform_class_decl(config: Config, decl: Cursor) -> ClassEntity:
obj: ClassEntity = {
"kind": "class",
"keyword": _get_class_keyword(decl.kind),
"name": decl.displayname,
"prefix": _parse_declaration_prefix(decl, is_class=True),
"bases": list(_get_bases(config, decl)),
}
specializes = _get_specialized_cursor_template(decl)
if specializes:
obj["specializes"] = get_entity_id(specializes)
return obj
def _transform_class_template_decl(config: Config, decl: Cursor) -> ClassEntity:
return {
"kind": "class",
"keyword": _get_class_keyword(_get_template_cursor_kind(decl)),
"name": decl.spelling,
"prefix": _parse_declaration_prefix(decl, is_class=True),
"bases": list(_get_bases(config, decl)),
}
def _transform_class_template_partial_specialization_decl(
config: Config, decl: Cursor
) -> ClassEntity:
return {
"kind": "class",
"keyword": _get_class_keyword(_get_template_cursor_kind(decl)),
"name": decl.displayname,
"specializes": get_entity_id(
cast(Cursor, _get_specialized_cursor_template(decl))
),
"prefix": _parse_declaration_prefix(decl, is_class=True),
"bases": list(_get_bases(config, decl)),
}
def _get_function_parameters(decl: Cursor):
if decl.kind == CursorKind.FUNCTION_DECL:
yield from decl.get_arguments()
return
for child in decl.get_children():
if child.kind != CursorKind.PARM_DECL:
continue
yield child
FUNCTION_CURSOR_KIND_TO_JSON_KIND = {
CursorKind.FUNCTION_DECL: "function",
CursorKind.CXX_METHOD: "method",
CursorKind.CONSTRUCTOR: "constructor",
CursorKind.DESTRUCTOR: "destructor",
CursorKind.CONVERSION_FUNCTION: "conversion_function",
}
def _parse_declaration_prefix(decl: Cursor, is_class: bool) -> typing.List[str]:
decl_extent = decl.extent
start_location = decl_extent.start
end_location = None
prefix_parts = []
for child in decl.get_children():
# Skip template introduction
if child.kind in (
CursorKind.TEMPLATE_TYPE_PARAMETER,
CursorKind.TEMPLATE_NON_TYPE_PARAMETER,
CursorKind.TEMPLATE_TEMPLATE_PARAMETER,
):
start_location = child.extent.end
continue
if child.kind.is_attribute():
attr_spelling = get_extent_spelling(decl.translation_unit, child.extent)
prefix_parts.append(f"[[{attr_spelling}]]")
continue
end_location = child.extent.start
break
if not is_class:
for token in decl.translation_unit.get_tokens(
extent=SourceRange.from_locations(
start_location, end_location or decl_extent.end
)
):
# skip `inline` since that is not an important part of the API
if token.spelling in ("explicit", "constexpr"):
prefix_parts.append(token.spelling)
return prefix_parts
def _get_declaration_spelling(decl: Cursor) -> str:
decl_extent = decl.extent
start_location = decl_extent.start
end_location = None
for child in decl.get_children():
if child.kind.is_statement():
end_location = child.extent.start
break
else:
end_location = decl_extent.end
return get_extent_spelling(
decl.translation_unit,
extent=SourceRange.from_locations(start_location, end_location),
)
def _transform_function_decl(config: Config, decl: Cursor):
name = decl.spelling
if name.startswith("<deduction guide for "):
# Exclude deduction guides for now
return None
non_template_kind = decl.kind
if decl.kind == CursorKind.FUNCTION_TEMPLATE:
non_template_kind = _get_template_cursor_kind(decl)
specializes = _get_specialized_cursor_template(decl)
if non_template_kind == CursorKind.CONSTRUCTOR:
# TODO: handle = default, = delete
first_bracket = name.find("<")
if first_bracket != -1:
name = name[:first_bracket]
prefix = _parse_declaration_prefix(decl, is_class=False)
if decl.storage_class == clang.cindex.StorageClass.STATIC:
prefix.insert(0, "static")
source_code = _get_declaration_spelling(decl)
name_substitute = _pick_name_substitute(source_code)
decl_string = (
"".join(x + " " for x in prefix)
+ name_substitute
+ "("
+ ", ".join(
get_extent_spelling(decl.translation_unit, arg.extent)
for arg in _get_function_parameters(decl)
)
+ ")"
)
requires_expr = None
else:
(
decl_string,
bare_name,
template_args,
name_substitute,
requires_expr,
) = _parse_function(config, decl)
name = bare_name
if specializes and template_args is not None:
name += template_args
arity = sum(x.kind == CursorKind.PARM_DECL for x in decl.get_children())
obj = {
"kind": FUNCTION_CURSOR_KIND_TO_JSON_KIND[non_template_kind],
"name": name,
"arity": arity,
"declaration": decl_string,
"name_substitute": name_substitute,
"requires": [requires_expr] if requires_expr else None,
}
if specializes:
obj["specializes"] = get_entity_id(specializes)
return obj
def _transform_enum_decl(config: Config, decl: Cursor) -> EnumEntity:
keyword = None
tokens = list(decl.get_tokens())
assert len(tokens) >= 2
assert tokens[0].spelling == "enum"
token1_spelling = tokens[1].spelling
if token1_spelling in ("class", "struct"):
keyword = cast(ClassKeyword, token1_spelling)
enumerators: List[EnumeratorEntity] = []
prev_decl_location = decl.location
for child in decl.get_children():
if child.kind == CursorKind.ENUM_CONSTANT_DECL:
enumerators.append(
{
"kind": "enumerator",
"id": get_entity_id(child),
"name": child.spelling,
"decl": get_extent_spelling(decl.translation_unit, child.extent),
"doc": _get_doc_comment(config, child, prev_decl_location),
"location": _get_location_json(config, child.location),
}
)
prev_decl_location = child.extent.end
return {
"kind": "enum",
"keyword": keyword,
"name": decl.spelling,
"enumerators": enumerators,
}
def _pick_name_substitute(code: str) -> str:
i = 0
while True:
substitute = f"__x{i}"
if substitute not in code:
return substitute
i += 1
def _transform_var_decl(config: Config, decl: Cursor) -> VarEntity:
exprs = [x for x in decl.get_children() if x.kind.is_expression()]
presumed_filename, presumed_line, _ = get_presumed_location(decl.location)
if len(exprs) > 1:
raise ValueError(
"%s:%d: Expected VAR decl to have at most one expression as a child: %r, but has: %d"
% (presumed_filename, presumed_line, decl.spelling, len(exprs))
)
prefix = _parse_declaration_prefix(decl, is_class=False)
type_spelling = decl.type.spelling
if "(lambda at " in type_spelling:
type_spelling = "auto"
name_substitute = _pick_name_substitute(type_spelling)
initializer = None
if len(exprs) == 1:
initializer = "= " + get_extent_spelling(
decl.translation_unit, exprs[0].extent
).rstrip(";")
if _is_internal_initializer(config, initializer):
initializer = None
declaration = " ".join(prefix) + " " + type_spelling + " " + name_substitute
return {
"kind": "var",
"name": decl.spelling,
"declaration": declaration,
"name_substitute": name_substitute,
"initializer": initializer,
}
class SphinxConfig:
cpp_id_attributes: Any = []
cpp_paren_attributes: Any = []
def _parse_name(name: str, template_prefix: str) -> sphinx.domains.cpp.ASTNestedName:
parser = sphinx.domains.cpp.DefinitionParser(
f"{template_prefix} int {name}",
location=("", 0),
config=cast(sphinx.config.Config, SphinxConfig()),
)
ast = parser.parse_declaration("member", "member")
parser.skip_ws()
parser.assert_end(allowSemicolon=True)
return ast.name
def _substitute_name(
top_ast: sphinx.domains.cpp.ASTDeclaration,
ast: sphinx.domains.cpp.ASTType,
source_code: str,
) -> str:
name_substitute = _pick_name_substitute(source_code)
template_args = ast.name.names[-1].templateArgs
name_substitute_with_args = name_substitute
if template_args is not None:
name_substitute_with_args += str(template_args)
template_prefix = ""
if (
top_ast.templatePrefix is not None
and top_ast.templatePrefix.templates is not None
):
template_prefix = str(top_ast.templatePrefix.templates[-1])
ast.name = _parse_name(name_substitute_with_args, template_prefix=template_prefix)
return name_substitute_with_args
def _maybe_wrap_requires_expr_in_parentheses(expr: str) -> str:
parser = sphinx.domains.cpp.DefinitionParser(
"requires " + expr,
location=("", 0),
config=cast(sphinx.config.Config, SphinxConfig()),
)
try:
parser._parse_requires_clause()
parser.skip_ws()
parser.assert_end()
return expr
except Exception:
return f"({expr})"
def _extract_requires_from_enable_if_t_type(
config: Config, ast: sphinx.domains.cpp.ASTType
) -> typing.Optional[str]:
if not isinstance(
ast.declSpecs.trailingTypeSpec, sphinx.domains.cpp.ASTTrailingTypeSpecName
):
return None
decl_specs = ast.declSpecs
trailing_type_spec = decl_specs.trailingTypeSpec
if not str(trailing_type_spec).startswith("std::enable_if_t<"):
return None
template_args = trailing_type_spec.name.names[1].templateArgs.args # type: ignore[attr-defined]
requires_expr = str(template_args[0])
if len(template_args) == 2:
result_type = str(template_args[1])
else:
result_type = "void"
parser = sphinx.domains.cpp.DefinitionParser(
result_type, location=("", 0), config=cast(sphinx.config.Config, SphinxConfig())
)
new_ast = parser._parse_type(False)
parser.skip_ws()
parser.assert_end()
new_decl_specs = new_ast.declSpecs
def copy_qualifiers(
orig_d: sphinx.domains.cpp.ASTDeclarator,
new_d: sphinx.domains.cpp.ASTDeclarator,
):
if isinstance(new_d, sphinx.domains.cpp.ASTDeclaratorRef):
return sphinx.domains.cpp.ASTDeclaratorRef(
next=copy_qualifiers(orig_d, new_d.next), attrs=new_d.attrs
)
if isinstance(new_d, sphinx.domains.cpp.ASTDeclaratorPtr):
return sphinx.domains.cpp.ASTDeclaratorPtr(
next=copy_qualifiers(orig_d, new_d.next),
volatile=new_d.volatile,
const=new_d.const,
attrs=new_d.attrs,
)
return orig_d
ast.decl = copy_qualifiers(ast.decl, new_ast.decl)
decl_specs.trailingTypeSpec = new_decl_specs.trailingTypeSpec
decl_specs.leftSpecs.const = (
decl_specs.leftSpecs.const or new_decl_specs.leftSpecs.const
)
decl_specs.leftSpecs.volatile = (
decl_specs.leftSpecs.volatile or new_decl_specs.leftSpecs.volatile
)
decl_specs.rightSpecs.const = (
decl_specs.rightSpecs.const or new_decl_specs.rightSpecs.const
)
decl_specs.rightSpecs.volatile = (
decl_specs.rightSpecs.volatile or new_decl_specs.rightSpecs.volatile
)
return _substitute_internal_type_names(config, requires_expr)
_FUNCTION_NAME_REPLACEMENTS = {
"operator[ ]": "operator[]",
"operator( )": "operator()",
}
def _parse_function(config: Config, decl: Cursor):
presumed_file, presumed_line, _ = get_presumed_location(decl.location)
source_code = _get_declaration_spelling(decl)
parser = sphinx.domains.cpp.DefinitionParser(
source_code,
location=(presumed_file, presumed_line),
config=cast(sphinx.config.Config, SphinxConfig()),
)
ast = parser.parse_declaration("function", "function")
parser.skip_ws()
parser.assert_end(allowSemicolon=True)
assert isinstance(ast.declaration, sphinx.domains.cpp.ASTType)
requires_expr = _extract_requires_from_enable_if_t_type(config, ast.declaration)
last_name_element = ast.declaration.name.names[-1]
bare_name = str(last_name_element.identOrOp)
bare_name = _FUNCTION_NAME_REPLACEMENTS.get(bare_name, bare_name)
template_args = last_name_element.templateArgs
template_args_str = str(template_args) if template_args is not None else None
name_substitute = _substitute_name(ast, ast.declaration, source_code)
# Exclude `inline` specifier since it isn't really informative in API
# documentation.
ast.declaration.declSpecs.leftSpecs.inline = False
decl_string = _substitute_internal_type_names(config, str(ast.declaration))
return decl_string, bare_name, template_args_str, name_substitute, requires_expr
def _is_internal_initializer(config: Config, initializer: str) -> bool:
return (
config.hide_initializers_pattern.search(initializer) is not None
or config.hide_types_pattern.search(initializer) is not None
)
def _sphinx_ast_template_parameter_to_json(
config: Config, param: sphinx.domains.cpp.ASTTemplateParam
) -> TemplateParameter:
if isinstance(param, sphinx.domains.cpp.ASTTemplateParamType):
kind = "type"
elif isinstance(param, sphinx.domains.cpp.ASTTemplateParamTemplateType):
kind = "template"
else:
kind = "non_type"
identifier = param.get_identifier()
return {
"declaration": _substitute_internal_type_names(config, str(param)),
"name": str(identifier) if identifier else "",
"kind": cast(TemplateParameterKind, kind),
"pack": param.isPack, # type: ignore[attr-defined]
}
def _transform_unexposed_decl(config: Config, decl: Cursor) -> Optional[VarEntity]:
# libclang unfortunately does not support variable templates; they are only
# exposed as an unexposed decl.
source_code = get_extent_spelling(decl.translation_unit, decl.extent)
# Note: Since `source_code` is reconstructed from the tokens, we don't need to
# worry about inconsistency in spacing.
if not source_code.startswith("template <"):
return None
# Assume that it is a variable template
# Attempt to parse it via sphinx's c++ domain parser
presumed_file, presumed_line, _ = get_presumed_location(decl.location)
try:
parser = sphinx.domains.cpp.DefinitionParser(
source_code,
location=(presumed_file, presumed_line),
config=cast(sphinx.config.Config, SphinxConfig()),
)
ast = parser.parse_declaration("member", "member")
parser.skip_ws()
parser.assert_end(allowSemicolon=True)
declaration = cast(
Union[
sphinx.domains.cpp.ASTTypeWithInit,
sphinx.domains.cpp.ASTTemplateParamConstrainedTypeWithInit,
],
ast.declaration,
)
template_args = declaration.type.name.names[-1].templateArgs
name = str(declaration.type.name.names[-1])
name_substitute = _substitute_name(ast, declaration.type, source_code)
decl_string = _substitute_internal_type_names(config, str(declaration.type))
decl_string = re.sub("(^| )inline ", " ", decl_string)
initializer: Optional[str] = _substitute_internal_type_names(
config, str(declaration.init).strip().rstrip(";").strip()
)
assert initializer is not None
if _is_internal_initializer(config, initializer):
initializer = None
template_params = []
templates = cast(
sphinx.domains.cpp.ASTTemplateDeclarationPrefix, ast.templatePrefix
).templates
assert templates is not None
for templ_param in templates[-1].params:
template_params.append(
_sphinx_ast_template_parameter_to_json(
config, cast(sphinx.domains.cpp.ASTTemplateParam, templ_param)
)
)
obj: VarEntity = {
"kind": "var",
"name": name,
"template_parameters": template_params,
"declaration": decl_string,
"name_substitute": name_substitute,
"initializer": initializer,
}
if template_args is not None:
obj["specializes"] = True
return obj
except Exception as e:
raise ValueError("Failed to parse unexposed") from e
return None
def _parse_macro_parameters(decl: Cursor) -> typing.Optional[typing.List[str]]:
# Check if the macro is a function-like macro
# `cast` below is workaround for: https://github.com/tgockel/types-clang/pull/2
token_iterator = cast(typing.Iterator[Token], decl.get_tokens())
first_token = next(token_iterator)
assert first_token.spelling == decl.spelling
def_start_offset = cast(SourceLocation, first_token.extent.end).offset
try:
next_token = next(token_iterator)
except StopIteration:
return None
if next_token.spelling != "(":
return None
if next_token.location.offset != def_start_offset:
# There is a space between the macro name and the first "(", which means
# this is not a function-like macro.
return None
parameters = []
for token in token_iterator:
if token.kind == TokenKind.COMMENT:
continue
spelling = token.spelling
if spelling == ")":
break
if spelling == ",":
continue
parameters.append(spelling)
else:
presumed_file, presumed_line, _ = get_presumed_location(decl.location)
raise ValueError(
"%s:%d: Failed to parse macro %s"
% (presumed_file, presumed_line, decl.spelling)
)
return parameters
def _transform_macro(config: Config, decl: Cursor) -> Optional[MacroEntity]:
name = decl.spelling
if config.disallow_macros_pattern.search(name) is not None:
return None
return {
"kind": "macro",
"name": name,
"parameters": _parse_macro_parameters(decl),
}
TRANSFORMERS: Dict[CursorKind, Callable[[Config, Cursor], Optional[CppApiEntity]]] = {
CursorKind.TYPE_ALIAS_DECL: _transform_type_alias_decl,
CursorKind.TYPEDEF_DECL: _transform_type_alias_decl,
CursorKind.TYPE_ALIAS_TEMPLATE_DECL: _transform_type_alias_template_decl,
CursorKind.STRUCT_DECL: _transform_class_decl,
CursorKind.CLASS_DECL: _transform_class_decl,
CursorKind.CLASS_TEMPLATE: _transform_class_template_decl,
CursorKind.CLASS_TEMPLATE_PARTIAL_SPECIALIZATION: _transform_class_template_partial_specialization_decl,
CursorKind.FUNCTION_DECL: _transform_function_decl,
CursorKind.CXX_METHOD: _transform_function_decl,
CursorKind.CONVERSION_FUNCTION: _transform_function_decl,
CursorKind.CONSTRUCTOR: _transform_function_decl,
CursorKind.DESTRUCTOR: _transform_function_decl,
CursorKind.FUNCTION_TEMPLATE: _transform_function_decl,
CursorKind.ENUM_DECL: _transform_enum_decl,
CursorKind.VAR_DECL: _transform_var_decl,
CursorKind.FIELD_DECL: _transform_var_decl,
CursorKind.UNEXPOSED_DECL: _transform_unexposed_decl,
CursorKind.MACRO_DEFINITION: _transform_macro,
}
ALLOWED_KINDS = frozenset(list(TRANSFORMERS.keys()) + [CursorKind.FRIEND_DECL])
def _parse_args(output_required: bool):
ap = argparse.ArgumentParser()
ap.add_argument("--config", type=str, required=True)
ap.add_argument("--output", type=str, required=output_required)
return ap.parse_args()
def _merge_decl_json(existing_json, new_json):
if existing_json["doc"] and new_json["doc"]:
raise ValueError("Duplicate doc strings: %r and %r" % (existing_json, new_json))
existing_json["doc"] = existing_json["doc"] or new_json["doc"]
template_parameters = existing_json.get("template_parameters")
if template_parameters:
new_template_parameters = new_json.get("template_parameters")
for i, old_param in enumerate(template_parameters):
new_param = new_template_parameters[i]
if new_param.startswith(old_param):
template_parameters[i] = new_param
elif not old_param.startswith(new_param):
raise ValueError(
"Conflicting template parameter %d: %r and %r"
% (i, existing_json, new_json)
)
def _get_location_json(config: Config, location: SourceLocation) -> JsonLocation:
filename, line, col = get_presumed_location(location)
filename = config.map_include_path(filename)
return {"file": filename, "line": line, "col": col}
def _get_location_string(config: Config, location: SourceLocation) -> str:
filename, line, col = get_presumed_location(location)
filename = config.map_include_path(filename)
return f"{filename}:{line}:{col}"
def _is_immediately_after(decl: Cursor, prev_decl: Cursor) -> bool:
# casts below are workaround for: https://github.com/tgockel/types-clang/pull/2
prev_end = cast(SourceLocation, prev_decl.extent.end)
cur_start = cast(SourceLocation, decl.extent.start)
cur_file, cur_line, _ = get_presumed_location(cur_start)
prev_file, prev_line, _ = get_presumed_location(prev_end)
return cur_file == prev_file and cur_line == prev_line + 1
_NORMALIZED_KIND = {
"constructor": "function",
"conversion_function": "function",
"method": "function",
"function": "function",
"class": "class",
"alias": "alias",
"enum": "enum",
"macro": "macro",
"var": "var",
}
def _kinds_are_compatible(a: str, b: str) -> bool:
return _NORMALIZED_KIND[a] == _NORMALIZED_KIND[b]
class JsonApiGenerator:
def __init__(self, extractor):
self.extractor = extractor
self.config = extractor.config
self.seen_decls = {}
self.output_json = []
self._prev_decl = None
self._document_with_parent = {}
self._seen_unexposed_entities: set[tuple[str, int, str]] = set()
def _resolve_document_with(self, entity_id: EntityId) -> EntityId:
while True:
document_with_parent = self._document_with_parent.get(entity_id)
if document_with_parent is None:
break
entity_id = document_with_parent
return entity_id
def _transform_cursor_to_json(
self,
decl: Cursor,
parent: Optional[Cursor],
doc_comment_start_bound: SourceLocation,
):
doc = _get_doc_comment(self.config, decl, doc_comment_start_bound)
document_with = None
location = _get_location_json(self.config, decl.location)
if not doc:
if self._prev_decl is not None and _is_immediately_after(
decl, self._prev_decl[0]
):
document_with = self._resolve_document_with(self._prev_decl[1]["id"])
else:
# Exclude undocumented entities
return None
else:
if (
self._prev_decl is not None
and self._prev_decl[1]["location"] == location
):
# Same line as previous declaration, presumably due to macro expansion
# generating multiple declarations.
#
# Document as a sibling of the previous declaration.
document_with = self._resolve_document_with(self._prev_decl[1]["id"])
transformer = TRANSFORMERS.get(decl.kind)
if transformer is None:
return None
json_repr = transformer(self.config, decl)
if json_repr is None:
return None
if parent is None or parent.kind in (
CursorKind.NAMESPACE,
CursorKind.TRANSLATION_UNIT,
):
json_repr["scope"] = _get_full_nested_name(parent)
else:
json_repr["parent"] = get_entity_id(parent)
if decl.kind != CursorKind.UNEXPOSED_DECL:
template_parameters = _get_template_parameters(self.config, decl)
if json_repr.get("specializes") and template_parameters is None:
template_parameters = []
json_repr["template_parameters"] = template_parameters
# Exclude duplicate UNEXPOSED_DECL entities.
#
# Some versions of libclang seem to also generate an UNEXPOSED_DECL for
# instantations of variable templates. These occur at the same source
# location as the original declaration, and are assumed to always occur
# after the original declaration.
if decl.kind == CursorKind.UNEXPOSED_DECL:
duplicate_key = (
decl.location.file.name, # type: ignore[attr-defined]
decl.location.offset,
json.dumps(json_repr),
)
if duplicate_key in self._seen_unexposed_entities:
return None
self._seen_unexposed_entities.add(duplicate_key)
entity_id = get_entity_id(decl)
if document_with:
prev_json = cast(Any, self._prev_decl)[1]
if (
prev_json is None
or not _kinds_are_compatible(prev_json["kind"], json_repr["kind"])
or prev_json.get("parent") != json_repr.get("parent")
or prev_json.get("scope") != json_repr.get("scope")
):
if not doc:
# Undocumented and can't document with previous decl
return None
document_with = None
if document_with is not None:
doc = None
self._document_with_parent[entity_id] = document_with
json_repr["document_with"] = document_with
json_repr["location"] = location
nonitpick = get_nonitpick_directives(decl)
if nonitpick:
json_repr["nonitpick"] = nonitpick
json_repr["doc"] = doc
json_repr["id"] = entity_id
return json_repr
def add(self, decl: Cursor, doc_comment_start_bound: SourceLocation):
is_friend = False
if decl.kind == CursorKind.FRIEND_DECL:
# Check if this is a hidden friend function.
children = list(decl.get_children())
if len(children) != 1:
return
decl = children[0]
if not decl.kind.is_declaration():
return
is_friend = True
parent = decl.lexical_parent
else:
parent = decl.semantic_parent
json_repr = self._transform_cursor_to_json(
decl, parent, doc_comment_start_bound
)
if json_repr is None:
self._prev_decl = None
return
json_repr["friend"] = is_friend
parent_id = json_repr.get("parent")
if parent_id is not None and parent_id not in self.seen_decls:
# Parent is undocumented, skip.
return
self._prev_decl = (decl, json_repr)
entity_id = json_repr["id"]
existing_json_repr = self.seen_decls.get(entity_id)
if existing_json_repr is not None:
_merge_decl_json(existing_json_repr, json_repr)
return
self.seen_decls[entity_id] = json_repr
def _parse_template_parameter(
decl: str,
) -> Optional[sphinx.domains.cpp.ASTTemplateParam]:
# Note: We must include an extra trailing ">" because
# `_parse_template_parameter` fails if the parameter is not followed by "," or
# ">".
parser = sphinx.domains.cpp.DefinitionParser(
decl + ">", location=("", 0), config=cast(sphinx.config.Config, SphinxConfig())
)
parser.allowFallbackExpressionParsing = False
try:
param = parser._parse_template_parameter()
assert parser.skip_string(">")
parser.assert_end()
return param
except sphinx.domains.cpp.DefinitionError:
return None
def _extract_sfinae_replacement(template_parameter: str) -> Optional[Tuple[str, str]]:
param = _parse_template_parameter(template_parameter)
if param is None:
return None
name = str(param.get_identifier())
if not name.lower().startswith("sfinae"):
return None
if isinstance(param, sphinx.domains.cpp.ASTTemplateParamType):
default_type = param.data.default
if default_type is None:
return None
return (name, str(default_type))
if isinstance(param, sphinx.domains.cpp.ASTTemplateParamNonType):
default_value: Optional[sphinx.domains.cpp.ASTBase] = param.param.init
if default_value is None:
return None
if isinstance(default_value, sphinx.domains.cpp.ASTInitializer):
default_value = default_value.value
return (name, str(default_value))
return None
CONDITIONALLY_EXPLICIT_PATTERN = re.compile(r"\(ExplicitRequires\((.+)\)\)")
def _match_template_parameter_enable_if_pattern(
config: Config, decl: str
) -> Optional[str]:
for pattern in config.template_parameter_enable_if_patterns:
m = pattern.fullmatch(decl)
if m is not None:
return m.group(1)
return None
def _transform_template_parameters(config: Config, template_parameters: List[Any]):
"""Transforms template parameters to C++20 form."""
requires = []
new_template_parameters = []
replacements: Dict[str, str] = {}
for template_parameter in template_parameters:
decl = template_parameter["declaration"]
requires_expr = _match_template_parameter_enable_if_pattern(config, decl)
if requires_expr is not None:
requires.append(requires_expr)
continue
if config.ignore_template_parameters_pattern.fullmatch(decl):
continue
# If the template parameter is of the form `YYY SfinaeXXX = Condition`, then
# we want to exclude it from the template parameter list and instead return
# the substitution `{"SfinaeXXX": "Condition"}`. To avoid parsing in cases
# that can't possibly match, first look to see if the name starts with
# `"sfinae"`.
if re.match(r"sfinae", template_parameter["name"], re.IGNORECASE) is not None:
# Possibly match, parse to verify.
replacement = _extract_sfinae_replacement(decl)
if replacement is not None:
replacements[replacement[0]] = replacement[1]
continue
new_template_parameters.append(template_parameter)
return (
new_template_parameters,
requires,
replacements,
)
def _strip_return_type(
declaration: str, template_prefix: str, location: Tuple[str, int]
) -> str:
parser = sphinx.domains.cpp.DefinitionParser(
template_prefix + declaration,
location=location,
config=cast(sphinx.config.Config, SphinxConfig()),
)
ast = parser.parse_declaration("function", "function")
parser.skip_ws()
parser.assert_end()
assert isinstance(ast.declaration, sphinx.domains.cpp.ASTType)
ast.declaration.declSpecs.trailingTypeSpec = (
sphinx.domains.cpp.ASTTrailingTypeSpecFundamental(["auto"], ["auto"])
)
return str(ast.declaration)
_OPERATOR_PAGE_NAMES = {
("operator+", 1): "operator-unary_plus",
("operator-", 1): "operator-negate",
("operator*", 1): "operator-dereference",
("operator~", 1): "operator-complement",
("operator!", 1): "operator-logical_not",
("operator++", 1): "operator-pre_inc",
("operator++", 2): "operator-post_inc",
("operator--", 1): "operator-pre_dec",
("operator--", 2): "operator-post_dec",
("operator<<", 2): "operator-shift_left",
("operator>>", 2): "operator-shift_right",
("operator+", 2): "operator-plus",
("operator-", 2): "operator-minus",
("operator*", 2): "operator-multiplies",
("operator/", 2): "operator-divides",
("operator%", 2): "operator-modulus",
("operator<", 2): "operator-less",
("operator<=", 2): "operator-less_equal",
("operator>=", 2): "operator-greater_equal",
("operator==", 2): "operator-equal_to",
("operator!=", 2): "operator-not_equal_to",
("operator=", 2): "operator-assign",
("operator<<=", 2): "operator-shift_left_assign",
("operator>>=", 2): "operator-shift_right_assign",
("operator*=", 2): "operator-multiplies_assign",
("operator/=", 2): "operator-divides_assign",
("operator%=", 2): "operator-modulus_assign",
("operator+=", 2): "operator-plus_assign",
("operator-=", 2): "operator-minus_assign",
("operator&=", 2): "operator-bitwise_and_assign",
("operator|=", 2): "operator-bitwise_or_assign",
("operator^=", 2): "operator-bitwise_xor_assign",
("operator&&", 2): "operator-logical_and",
("operator||", 2): "operator-logical_or",
("operator|", 2): "operator-bitwise_or",
("operator&", 2): "operator-bitwise_and",
("operator^", 2): "operator-bitwise_xor",
("operator,", 2): "operator-comma",
("operator->", 1): "operator-arrow",
("operator&", 1): "operator-address_of",
"operator()": "operator-call",
"operator[]": "operator-subscript",
}
DEFAULT_MEMBER_GROUP_FOR_MEMBER_ENTITY_TYPE = {
"constructor": "Constructors",
"destructor": "Constructors",
"class": "Types",
"method": "Methods",
"enum": "Types",
"alias": "Types",
"conversion_function": "Conversion operators",
"var": "Data members",
"function": "Friend functions",
}
DEFAULT_MEMBER_GROUP_FOR_NON_MEMBER_ENTITY_TYPE = {
"alias": "Related Types",
"enum": "Related Types",
"class": "Related Types",
"function": "Related Functions",
"var": "Related Constants",
"macro": "Related Macros",
}
def _get_default_member_group(entity: CppApiEntity) -> str:
if entity.get("parent"):
return DEFAULT_MEMBER_GROUP_FOR_MEMBER_ENTITY_TYPE[entity["kind"]]
return DEFAULT_MEMBER_GROUP_FOR_NON_MEMBER_ENTITY_TYPE[entity["kind"]]
def _normalize_doc_text(text: str) -> str:
text = re.sub(r"^((?:\\|@)(?:brief|details)\s+)", "", text, flags=re.MULTILINE)
text = re.sub(
r"^(?:\\|@)(t?param)(\[(?:in|out|in,\sout)\])?\s+([a-zA-Z_][^ ]*)",
":\\1 \\3\\2:",
text,
flags=re.MULTILINE,
)
text = re.sub(
r"^(?:\\|@)(error)\s+`([^`]+)`", ":\\1 \\2:", text, flags=re.MULTILINE
)
text = re.sub(
r"^(?:\\|@)(returns?|pre|post|[ds]?checks|invariant|requires)(?: |\n )",
":\\1: ",
text,
flags=re.MULTILINE,
)
text = re.sub(r"^(?:\\|@)(retval)\s+(\S+)", ":\\1 \\2:", text, flags=re.MULTILINE)
text = SPECIAL_GROUP_COMMAND_PATTERN.sub("", text)
return text
FUNCTION_ENTITY_KINDS = frozenset(
["function", "method", "constructor", "destructor", "conversion_function"]
)
def _is_function(entity: CppApiEntity) -> bool:
return entity["kind"] in FUNCTION_ENTITY_KINDS
def _get_path_component_from_special_id(special_id: str) -> str:
special_id = re.sub("[^a-zA-Z0-9_]+", "-", special_id)
return special_id.strip("-")
def _apply_identifier_replacements(
requires_term: str, replacements: Dict[str, str]
) -> str:
for orig_identifier, replacement in replacements.items():
requires_term = re.sub(
r"\b" + re.escape(orig_identifier) + r"\b", replacement, requires_term
)
return requires_term
def _make_explicit_conditional(decl: str, explicit: str) -> str:
new_str = re.sub(r"\bexplicit\b", f"explicit({explicit})", decl, 1)
if new_str == decl:
raise ValueError(
"Failed to insert explicit condition %r into: %s"
% (
explicit,
decl,
)
)
return new_str
def _is_uniform_binary_expr(
expr: sphinx.domains.cpp.ASTBase, allowed_ops: Tuple[str, ...]
) -> bool:
if not isinstance(expr, sphinx.domains.cpp.ASTBinOpExpr):
return False
return all(op in allowed_ops for op in expr.ops)
def _is_logical_and_expr(expr: sphinx.domains.cpp.ASTBase) -> bool:
return _is_uniform_binary_expr(expr, ("&&", "and"))
def _is_primary_expr(expr: sphinx.domains.cpp.ASTBase) -> bool:
return isinstance(
expr,
(
sphinx.domains.cpp.ASTLiteral,
sphinx.domains.cpp.ASTIdExpression,
sphinx.domains.cpp.ASTThisLiteral,
),
)
def _normalize_requires_terms(terms: List[str]) -> List[str]:
if not terms:
return terms
expr = " && ".join(f"({term})" for term in terms)
parser = sphinx.domains.cpp.DefinitionParser(
"requires " + expr,
location=("", 0),
config=cast(sphinx.config.Config, SphinxConfig()),
)
# If we allow fallback parsing, we end up with incorrect parsing and a
# spurious warning.
parser.allowFallbackExpressionParsing = False
ast = parser._parse_requires_clause()
parser.skip_ws()
parser.assert_end()
assert ast is not None
new_terms = []
def process(
expr: Union[sphinx.domains.cpp.ASTType, sphinx.domains.cpp.ASTExpression],
):
while True:
if isinstance(expr, sphinx.domains.cpp.ASTParenExpr):
expr = expr.expr
continue
if isinstance(expr, sphinx.domains.cpp.ASTBinOpExpr) and not expr.ops:
expr = expr.exprs[0]
continue
if (
isinstance(expr, sphinx.domains.cpp.ASTPostfixExpr)
and not expr.postFixes
):
expr = expr.prefix
continue
break
if _is_logical_and_expr(expr):
for sub_expr in cast(sphinx.domains.cpp.ASTBinOpExpr, expr).exprs:
process(sub_expr)
return
if _is_primary_expr(expr):
new_terms.append(str(expr))
return
new_terms.append(f"({expr})")
process(ast.expr)
return new_terms
def _format_template_arguments(entity: CppApiEntity) -> str:
if entity.get("specializes"):
# Template arguments already included in `entity["name"]`.
return ""
template_parameters = entity.get("template_parameters")
if not template_parameters:
return ""
strs = []
for param in template_parameters:
arg = param["name"]
if not arg:
continue
if param["pack"]:
arg += "..."
strs.append(arg)
args_str = ", ".join(strs)
return f"<{args_str}>"
def _get_entity_base_page_name_component(entity: CppApiEntity) -> str:
base_name = entity["name"]
if (entity["kind"] == "class" or entity["kind"] == "var") and entity.get(
"specializes"
):
# Strip any template arguments
base_name = re.sub("([^<]*).*", r"\1", base_name)
elif entity["kind"] == "conversion_function":
base_name = "operator-cast"
elif entity["kind"] in ("function", "method") and re.match(
r"operator\b", base_name
):
arity = cast(FunctionEntity, entity)["arity"]
if entity["kind"] == "method":
arity += 1
op_page_name = _OPERATOR_PAGE_NAMES.get((base_name, arity))
if op_page_name is None:
op_page_name = _OPERATOR_PAGE_NAMES[base_name]
base_name = op_page_name
return base_name
def _get_entity_page_name_component(entity: CppApiEntity) -> str:
page_name = _get_entity_base_page_name_component(entity)
special_id = entity.get("special_id")
if special_id is not None:
page_name += f"-{_get_path_component_from_special_id(special_id)}"
return page_name
def _ensure_unique_page_names(
entities_with_page_names: List[EntityId],
entities: Dict[EntityId, CppApiEntity],
warning,
) -> None:
names: Dict[
Tuple[Optional[str], Optional[str], str, Optional[str]], List[EntityId]
] = {}
for entity_id in entities_with_page_names:
entity = entities[entity_id]
parent_id = entity.get("parent")
special_id = entity.get("special_id")
scope = entity.get("scope")
base_name = _get_entity_base_page_name_component(entity)
key = (parent_id, scope, base_name, special_id)
names.setdefault(key, []).append(entity_id)
for (parent_id, scope, base_name, special_id), entity_ids in names.items():
if len(entity_ids) == 1:
continue
page_entities = [entities[entity_id] for entity_id in entity_ids]
warning(
"Disambiguating %d overloads of %s using numerical ids. Definitions at %s",
len(entity_ids),
base_name,
", ".join(
"%s:%d" % (entity["location"]["file"], entity["location"]["line"])
for entity in page_entities
),
)
for i, entity in enumerate(page_entities):
entity["special_id"] = str(i + 1)
class JsonDiagnostic(TypedDict):
message: str
location: Optional[JsonLocation]
class JsonNitpickExclusion(TypedDict):
file: str
line: int
target: str
class JsonApiData(TypedDict):
errors: List[JsonDiagnostic]
warnings: List[JsonDiagnostic]
nonitpick: List[JsonNitpickExclusion]
groups: Dict[str, List[EntityId]]
entities: Dict[str, CppApiEntity]
def organize_entities(
config: Config, entities: Dict[EntityId, CppApiEntity]
) -> JsonApiData:
errors: List[JsonDiagnostic] = []
warnings: List[JsonDiagnostic] = []
def error(msg: str, *args, location: Optional[JsonLocation] = None):
errors.append({"message": msg % args, "location": location})
def warning(msg: str, *args, location: Optional[JsonLocation] = None):
warnings.append({"message": msg % args, "location": location})
def _handle_document_with(entity: CppApiEntity) -> bool:
document_with = entity.get("document_with")
if document_with is None:
return False
sibling_entity: Optional[CppApiEntity] = entities.get(document_with)
if sibling_entity is None:
return False
sibling_entity.setdefault("siblings", []).append(entity["id"])
return True
def _normalize_entity_requires(entity: CppApiEntity):
template_parameters = entity.get("template_parameters")
if template_parameters:
(
template_parameters,
requires,
replacements,
) = _transform_template_parameters(config, template_parameters)
if entity.get("specializes") is None and not template_parameters:
entity["template_parameters"] = None
else:
entity["template_parameters"] = template_parameters
else:
requires = []
replacements = None
explicit = None
existing_requires = entity.get("requires")
if existing_requires:
requires = existing_requires + requires
if _is_function(entity):
func_entity = cast(FunctionEntity, entity)
declaration = func_entity["declaration"]
if replacements:
declaration = _apply_identifier_replacements(declaration, replacements)
if (
func_entity["kind"] != "constructor"
and config.hide_types_pattern.search(
declaration[: declaration.index(func_entity["name_substitute"])]
)
is not None
):
declaration = _strip_return_type(
declaration,
"template <> " if template_parameters is not None else "",
location=(entity["location"]["file"], entity["location"]["line"]),
)
func_entity["declaration"] = declaration
else:
if replacements:
for key in cast(
Tuple[Literal["declaration", "underlying_type"], ...],
("declaration", "underlying_type"),
):
x = cast(Optional[str], entity.get(key, None))
if x is not None:
entity[key] = _apply_identifier_replacements(x, replacements) # type: ignore[typeddict-item]
if replacements:
requires = [
_apply_identifier_replacements(x, replacements) for x in requires
]
requires = _normalize_requires_terms(requires)
new_requires = []
explicit = None
for term in requires:
m = CONDITIONALLY_EXPLICIT_PATTERN.fullmatch(term)
if m is not None:
if explicit is not None:
raise ValueError(
"cannot have more than one conditionally-explicit term"
)
explicit = m.group(1)
else:
new_requires.append(term)
requires = new_requires
if explicit:
if entity["kind"] != "constructor":
raise ValueError(
"conditionally-explicit terms only valid on constructors"
)
entity["declaration"] = _make_explicit_conditional(
entity["declaration"], explicit
)
requires = [x for x in requires if config.hide_types_pattern.search(x) is None]
entity["requires"] = requires
def get_entity_page_name(entity: CppApiEntity) -> str:
components = []
cur_entity = entity
while True:
components.append(_get_entity_page_name_component(cur_entity))
parent_id = cur_entity.get("parent")
if parent_id is None:
break
parent_entity = entities.get(parent_id)
assert parent_entity is not None
cur_entity = parent_entity
components.reverse()
page_name = (cur_entity["scope"] + "::".join(components)).replace("::", ".")
return page_name
def _parse_entity_doc(entity: CppApiEntity):
doc = entity["doc"]
if doc is None:
if _handle_document_with(entity):
return True
return False
doc_text = doc["text"]
for m in SPECIAL_GROUP_COMMAND_PATTERN.finditer(doc_text):
entity[cast(Literal["special_id"], "special_" + m.group(1))] = m.group( # noqa: F821
2
).strip()
return True
def get_entity_scope(entity: CppApiEntity) -> str:
components = []
cur_entity = entity
while True:
parent_id = cur_entity.get("parent")
if parent_id is None:
break
parent_entity = entities.get(parent_id)
assert parent_entity is not None
cur_entity = parent_entity
name_with_args = cur_entity["name"]
if not cur_entity.get("specializes"):
name_with_args += _format_template_arguments(cur_entity)
components.append(name_with_args)
components.reverse()
if components:
components.append("")
return cur_entity.get("scope", "") + "::".join(components)
def get_entity_object_name(entity: CppApiEntity) -> str:
name = get_entity_scope(entity) + entity["name"]
special_id = entity.get("special_id")
if special_id:
name += f"[{special_id}]"
return name
unspecialized_names: Dict[
Tuple[Optional[EntityId], Optional[str], str], EntityId
] = {}
names: Dict[str, EntityId] = {}
def resolve_entity_name(
scope: str, relative_entity_name: str
) -> Optional[EntityId]:
if relative_entity_name.startswith("::"):
resolved = relative_entity_name[2:]
entity_id = names.get(resolved)
if entity_id is None:
return None
return entity_id
truncate_idx = len(scope)
while True:
full_name = scope[:truncate_idx] + relative_entity_name
entity_id = names.get(full_name)
if entity_id is not None:
return entity_id
if truncate_idx == 0:
return None
truncate_idx = scope.rfind("::", 0, truncate_idx - 2)
if truncate_idx == -1:
truncate_idx = 0
else:
truncate_idx = truncate_idx + 2
must_resolve_specializes: List[CppApiEntity] = []
all_nonitpick: List[JsonNitpickExclusion] = []
def _handle_nitpick(entity: CppApiEntity, targets: List[str]) -> None:
document_with = entity.get("document_with")
if document_with:
entity = entities[document_with]
location: JsonLocation = entity["location"]
filename: str = location["file"]
line: int = location["line"]
for target in targets:
all_nonitpick.append({"file": filename, "line": line, "target": target})
entities_with_page_names: List[EntityId] = []
for entity in entities.values():
specializes = entity.get("specializes")
if (
entity["kind"] == "var"
and entity.get("template_parameters") is not None
and specializes is None
):
key = (entity.get("parent"), entity.get("scope"), entity["name"])
entity_id = entity["id"]
if unspecialized_names.setdefault(key, entity_id) != entity_id:
other_entity_id = unspecialized_names[key]
other_entity = entities[other_entity_id]
raise ValueError(
"Duplicate unspecialized entity name: %r %r %r"
% (
key,
entity,
other_entity,
)
)
if specializes is True:
must_resolve_specializes.append(entity)
if not _parse_entity_doc(entity):
continue
_normalize_entity_requires(entity)
nonitpick = entity.get("nonitpick")
if nonitpick:
_handle_nitpick(entity, nonitpick)
if not entity["doc"]:
continue
entities_with_page_names.append(entity["id"])
for entity in must_resolve_specializes:
name = cast(str, entity["name"])
name = name[: name.index("<")]
unspecialized_key = (entity.get("parent"), entity.get("scope"), name)
unspecialized_id = unspecialized_names.get(unspecialized_key)
if unspecialized_id is not None:
entity["specializes"] = unspecialized_id
_ensure_unique_page_names(entities_with_page_names, entities, warning)
for entity_id in entities_with_page_names:
entity = entities[entity_id]
names[get_entity_object_name(entity)] = entity_id
entity["page_name"] = get_entity_page_name(entity)
doc = entity["doc"]
assert doc is not None
doc["text"] = _normalize_doc_text(doc["text"])
groups: Dict[str, List[EntityId]] = {}
for entity in entities.values():
entity_id = entity["id"]
doc = entity["doc"]
if doc is None:
continue
ingroup = entity.get("special_ingroup")
relates_name = entity.get("special_relates")
member_group = entity.get("special_membergroup")
if (ingroup is not None) and (relates_name is not None):
error(
"Cannot specify both \\ingroup and \\relates for %r",
entity,
location=doc["location"],
)
continue
if ingroup is not None:
ingroup = docutils.nodes.make_id(ingroup)
groups.setdefault(ingroup, []).append(entity_id)
if member_group is not None:
error(
"Cannot specify both \\ingroup and \\membergroup for %r",
entity,
location=doc["location"],
)
continue
parent_id = entity.get("parent")
if relates_name is not None:
scope = get_entity_scope(entity)
relates_id = resolve_entity_name(scope, relates_name)
if relates_id is None:
error(
"Cannot resolve entity reference %r from scope %r",
relates_name,
scope,
location=doc["location"],
)
continue
parent_id = None
else:
if parent_id is None:
warning(
"No group or relates specified for entity %s%s",
entity.get("scope"),
entity["name"],
location=doc["location"],
)
continue
relates_id = parent_id
if member_group is None:
member_group = _get_default_member_group(entity)
assert relates_id is not None
entities[relates_id].setdefault(
cast(
Literal["related_members", "related_nonmembers"],
"related_members" if parent_id is not None else "related_nonmembers",
),
cast(Dict[str, List[EntityId]], {}),
).setdefault(member_group, []).append(entity_id)
return {
"entities": entities,
"groups": groups,
"errors": errors,
"warnings": warnings,
"nonitpick": all_nonitpick,
}
def _get_output_json(extractor: Extractor) -> JsonApiData:
generator = JsonApiGenerator(extractor)
if extractor.config.verbose:
logger.info("Found %d C++ declarations", len(extractor.decls))
for decl, doc_comment_start_bound in extractor.decls:
generator.add(decl, doc_comment_start_bound)
return organize_entities(extractor.config, generator.seen_decls)
def generate_output(config: Config) -> JsonApiData:
extractor = Extractor(config)
return _get_output_json(extractor)
def _load_config(config_path: str) -> Config:
config_content = pathlib.Path(config_path).read_text(encoding="utf-8")
context: dict = {}
exec(config_content, context)
config = context["config"]
assert isinstance(config, Config)
return config
def main():
args = _parse_args(output_required=True)
config = _load_config(args.config)
output_json = generate_output(config)
if args.output is not None:
with open(args.output, "w", encoding="utf-8") as f:
json.dump(output_json, f)
if __name__ == "__main__":
main()
Last update:
Nov 16, 2024