mirror of
https://github.com/thegeeklab/ansible-doctor.git
synced 2024-06-17 15:50:54 +02:00
Compare commits
125 Commits
Author | SHA1 | Date | |
---|---|---|---|
8e22e87a31 | |||
|
ed113e37ea | ||
|
c33738c8a1 | ||
|
d5e4e6ba33 | ||
08e2178333 | |||
|
42892daa74 | ||
|
ab6c988869 | ||
|
eb2cf69468 | ||
|
abee343056 | ||
|
29320b6b96 | ||
8e042c739e | |||
9b20c11660 | |||
|
5760ee0832 | ||
|
db94c07396 | ||
|
73bbd746d3 | ||
ab0372bef5 | |||
3df7e465db | |||
89c6a11be4 | |||
|
4051d2915d | ||
|
fe4e4e5f7a | ||
|
172e4f4380 | ||
|
fada900568 | ||
|
013f760c8a | ||
|
00adc389a2 | ||
|
84fdc06315 | ||
|
db68e80372 | ||
|
af702628eb | ||
|
81d4e97af6 | ||
|
a33f3c53bb | ||
|
ccc2d249f8 | ||
|
f7ff5fd624 | ||
|
acee6e1285 | ||
|
2375ad118d | ||
|
a2f02527d9 | ||
|
0bf59ac34f | ||
|
94ec1a632b | ||
|
075e1f91ca | ||
4cf63bf2fe | |||
|
e3f797d5e3 | ||
|
3a0c5ae35f | ||
|
9f7f943c93 | ||
|
b38c4aa2b8 | ||
|
ed167d1443 | ||
|
da6fd26c6d | ||
|
522c21f8fc | ||
|
28e39055e3 | ||
|
894965286b | ||
|
c2e0f787ce | ||
|
d524537fd3 | ||
|
a559b654ca | ||
|
aa78adf912 | ||
6e88c18375 | |||
7b9ba09f1d | |||
|
08883952c1 | ||
|
e1ef4937dd | ||
|
571222f6f5 | ||
|
6d50525021 | ||
|
3ade4698e7 | ||
|
ee81c8ee73 | ||
|
03df5bd79b | ||
|
1e32a8f87a | ||
|
dab9bc8691 | ||
e2eaa81c4f | |||
|
732f588aa9 | ||
6619351fbd | |||
|
8f6f444931 | ||
|
49b861cabc | ||
|
59b497d745 | ||
b2f4fd2bd8 | |||
|
864af95606 | ||
|
758c87ee80 | ||
|
9b0edda70a | ||
|
1d32f7548a | ||
|
704cdb9d7c | ||
|
c7f3fe57a0 | ||
|
1270d7cb7d | ||
|
461eeb2d74 | ||
|
0817646004 | ||
|
5242fd882a | ||
|
052c668d92 | ||
1e8a8beef7 | |||
|
c124460c11 | ||
|
38bd53f7bc | ||
|
dbf9c979ac | ||
|
fe12548387 | ||
ae14704b74 | |||
|
2232a12bc8 | ||
78d4c5f44b | |||
c3068a573f | |||
|
505f9b58cc | ||
|
593df92d32 | ||
|
568b91654d | ||
|
4afabc4284 | ||
|
10ff283ec2 | ||
|
1d0ff92bf3 | ||
|
7245a4149c | ||
|
df155dcf8a | ||
2ce29b2bff | |||
1dc53d1970 | |||
2f1f42318b | |||
|
69d682df79 | ||
223b1d8814 | |||
|
70539cc9a2 | ||
|
982e2db2df | ||
|
787c09a741 | ||
|
95c2a6aeaf | ||
|
855f48894a | ||
2270051d0d | |||
|
c0f100b70e | ||
|
41ed10270d | ||
|
eef09f4a42 | ||
|
0460f09627 | ||
6a7ae3011d | |||
0db500b0a8 | |||
dfa10dd209 | |||
|
82398da75d | ||
|
0257b874e9 | ||
|
40b96ae285 | ||
0f65f50e06 | |||
|
9476810896 | ||
|
2eb9aad213 | ||
cacc92f831 | |||
9536cd400d | |||
593a90ff10 | |||
|
a1e3e669d4 |
1
.lycheeignore
Normal file
1
.lycheeignore
Normal file
|
@ -0,0 +1 @@
|
|||
https://hub.docker.com/r/thegeeklab/*
|
|
@ -5,3 +5,6 @@ MD041: False
|
|||
MD024: False
|
||||
MD004:
|
||||
style: dash
|
||||
MD033:
|
||||
allowed_elements:
|
||||
- "br"
|
||||
|
|
|
@ -2,7 +2,9 @@
|
|||
- id: ansible-doctor
|
||||
name: ansible-doctor
|
||||
description: Create annotation based documentation for your Ansible roles.
|
||||
entry: ansible-doctor -f
|
||||
entry: ansible-doctor -f -qqq
|
||||
language: python
|
||||
pass_filenames: False
|
||||
always_run: True
|
||||
additional_dependencies:
|
||||
- .[ansible-core]
|
||||
|
|
|
@ -6,15 +6,14 @@ when:
|
|||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
steps:
|
||||
build:
|
||||
- name: build
|
||||
image: docker.io/library/python:3.12
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry build
|
||||
|
||||
dryrun:
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:1
|
||||
- name: dryrun
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:4
|
||||
settings:
|
||||
containerfile: Containerfile.multiarch
|
||||
dry_run: true
|
||||
|
@ -26,9 +25,9 @@ steps:
|
|||
when:
|
||||
- event: [pull_request]
|
||||
|
||||
publish-dockerhub:
|
||||
- name: publish-dockerhub
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:4
|
||||
group: container
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:1
|
||||
settings:
|
||||
auto_tag: true
|
||||
containerfile: Containerfile.multiarch
|
||||
|
@ -47,9 +46,9 @@ steps:
|
|||
branch:
|
||||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
publish-quay:
|
||||
- name: publish-quay
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:4
|
||||
group: container
|
||||
image: quay.io/thegeeklab/wp-docker-buildx:1
|
||||
settings:
|
||||
auto_tag: true
|
||||
containerfile: Containerfile.multiarch
|
||||
|
|
|
@ -6,27 +6,25 @@ when:
|
|||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
steps:
|
||||
build:
|
||||
- name: build
|
||||
image: docker.io/library/python:3.12
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry build
|
||||
|
||||
checksum:
|
||||
- name: checksum
|
||||
image: quay.io/thegeeklab/alpine-tools
|
||||
commands:
|
||||
- cd dist/ && sha256sum * > ../sha256sum.txt
|
||||
|
||||
changelog:
|
||||
- name: changelog
|
||||
image: quay.io/thegeeklab/git-sv
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- git sv current-version
|
||||
- git sv release-notes -t ${CI_COMMIT_TAG:-next} -o CHANGELOG.md
|
||||
- cat CHANGELOG.md
|
||||
|
||||
publish-github:
|
||||
- name: publish-github
|
||||
image: docker.io/plugins/github-release
|
||||
settings:
|
||||
api_key:
|
||||
|
@ -40,15 +38,14 @@ steps:
|
|||
when:
|
||||
- event: [tag]
|
||||
|
||||
publish-pypi:
|
||||
- name: publish-pypi
|
||||
image: docker.io/library/python:3.12
|
||||
secrets:
|
||||
- source: pypi_password
|
||||
target: POETRY_HTTP_BASIC_PYPI_PASSWORD
|
||||
- source: pypi_username
|
||||
target: POETRY_HTTP_BASIC_PYPI_USERNAME
|
||||
- source: pypi_password
|
||||
target: POETRY_HTTP_BASIC_PYPI_PASSWORD
|
||||
- source: pypi_username
|
||||
target: POETRY_HTTP_BASIC_PYPI_USERNAME
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry publish -n
|
||||
when:
|
||||
|
|
|
@ -6,51 +6,44 @@ when:
|
|||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
steps:
|
||||
assets:
|
||||
- name: assets
|
||||
image: quay.io/thegeeklab/alpine-tools
|
||||
commands:
|
||||
- make doc
|
||||
|
||||
markdownlint:
|
||||
- name: markdownlint
|
||||
image: quay.io/thegeeklab/markdownlint-cli
|
||||
group: test
|
||||
commands:
|
||||
- markdownlint 'README.md' 'CONTRIBUTING.md'
|
||||
|
||||
spellcheck:
|
||||
- name: spellcheck
|
||||
image: quay.io/thegeeklab/alpine-tools
|
||||
group: test
|
||||
commands:
|
||||
- spellchecker --files '_docs/**/*.md' 'README.md' 'CONTRIBUTING.md' -d .dictionary -p spell indefinite-article syntax-urls
|
||||
- spellchecker --files 'docs/**/*.md' 'README.md' 'CONTRIBUTING.md' -d .dictionary -p spell indefinite-article syntax-urls
|
||||
environment:
|
||||
FORCE_COLOR: "true"
|
||||
NPM_CONFIG_LOGLEVEL: "error"
|
||||
|
||||
link-validation:
|
||||
- name: link-validation
|
||||
image: docker.io/lycheeverse/lychee
|
||||
group: test
|
||||
commands:
|
||||
- lychee --no-progress --format detailed docs/content README.md
|
||||
|
||||
testbuild:
|
||||
image: quay.io/thegeeklab/hugo:0.115.2
|
||||
commands:
|
||||
- hugo --panicOnWarning -s docs/ -b http://localhost:8000/
|
||||
|
||||
build:
|
||||
image: quay.io/thegeeklab/hugo:0.115.2
|
||||
- name: build
|
||||
image: quay.io/thegeeklab/hugo:0.127.0
|
||||
commands:
|
||||
- hugo --panicOnWarning -s docs/
|
||||
|
||||
beautify:
|
||||
- name: beautify
|
||||
image: quay.io/thegeeklab/alpine-tools
|
||||
commands:
|
||||
- html-beautify -r -f 'docs/public/**/*.html'
|
||||
environment:
|
||||
FORCE_COLOR: "true"
|
||||
NPM_CONFIG_LOGLEVEL: error
|
||||
|
||||
publish:
|
||||
- name: publish
|
||||
image: quay.io/thegeeklab/wp-s3-action
|
||||
settings:
|
||||
access_key:
|
||||
|
@ -69,15 +62,15 @@ steps:
|
|||
- event: [push, manual]
|
||||
branch:
|
||||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
status: [success]
|
||||
status: [success, failure]
|
||||
|
||||
pushrm-dockerhub:
|
||||
- name: pushrm-dockerhub
|
||||
image: docker.io/chko/docker-pushrm:1
|
||||
secrets:
|
||||
- source: docker_password
|
||||
target: DOCKER_PASS
|
||||
target: DOCKER_PASS
|
||||
- source: docker_username
|
||||
target: DOCKER_USER
|
||||
target: DOCKER_USER
|
||||
environment:
|
||||
PUSHRM_FILE: README.md
|
||||
PUSHRM_SHORT: Annotation based documentation for your Ansible roles
|
||||
|
@ -88,7 +81,7 @@ steps:
|
|||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
status: [success]
|
||||
|
||||
pushrm-quay:
|
||||
- name: pushrm-quay
|
||||
image: docker.io/chko/docker-pushrm:1
|
||||
secrets:
|
||||
- source: quay_token
|
||||
|
|
|
@ -6,22 +6,29 @@ when:
|
|||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
steps:
|
||||
check-format:
|
||||
- name: check-format
|
||||
image: docker.io/library/python:3.12
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry install
|
||||
- poetry run yapf -dr ./${CI_REPO_NAME//-/}
|
||||
- poetry install -E ansible-core
|
||||
- poetry run ruff format --check --diff ./${CI_REPO_NAME//-/}
|
||||
environment:
|
||||
PY_COLORS: "1"
|
||||
|
||||
check-coding:
|
||||
- name: check-coding
|
||||
image: docker.io/library/python:3.12
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry install
|
||||
- poetry run ruff ./${CI_REPO_NAME//-/}
|
||||
- poetry install -E ansible-core
|
||||
- poetry run ruff check ./${CI_REPO_NAME//-/}
|
||||
environment:
|
||||
PY_COLORS: "1"
|
||||
|
||||
- name: check-jinja
|
||||
image: docker.io/library/python:3.12
|
||||
commands:
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry install -E ansible-core
|
||||
- poetry run j2lint ansibledoctor/templates/ -i jinja-statements-indentation jinja-statements-delimiter
|
||||
environment:
|
||||
PY_COLORS: "1"
|
||||
|
|
|
@ -8,7 +8,7 @@ when:
|
|||
runs_on: [success, failure]
|
||||
|
||||
steps:
|
||||
matrix:
|
||||
- name: matrix
|
||||
image: quay.io/thegeeklab/wp-matrix
|
||||
settings:
|
||||
homeserver:
|
||||
|
|
|
@ -5,22 +5,30 @@ when:
|
|||
branch:
|
||||
- ${CI_REPO_DEFAULT_BRANCH}
|
||||
|
||||
matrix:
|
||||
PYTHON_VERSION:
|
||||
- "3.8"
|
||||
- "3.9"
|
||||
- "3.10"
|
||||
- "3.11"
|
||||
- "3.12"
|
||||
|
||||
steps:
|
||||
pytest:
|
||||
image: docker.io/library/python:${PYTHON_VERSION}
|
||||
variables:
|
||||
- &pytest_base
|
||||
group: pytest
|
||||
commands:
|
||||
- git fetch --depth=2147483647
|
||||
- pip install poetry poetry-dynamic-versioning -qq
|
||||
- poetry install
|
||||
- poetry install -E ansible-core
|
||||
- poetry version
|
||||
- poetry run ${CI_REPO_NAME} --help
|
||||
environment:
|
||||
PY_COLORS: "1"
|
||||
|
||||
steps:
|
||||
- name: python-312
|
||||
image: docker.io/library/python:3.12
|
||||
<<: *pytest_base
|
||||
|
||||
- name: python-311
|
||||
image: docker.io/library/python:3.11
|
||||
<<: *pytest_base
|
||||
|
||||
- name: python-310
|
||||
image: docker.io/library/python:3.10
|
||||
<<: *pytest_base
|
||||
|
||||
- name: python-39
|
||||
image: docker.io/library/python:3.9
|
||||
<<: *pytest_base
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
FROM python:3.12-alpine@sha256:a5d1738d6abbdff3e81c10b7f86923ebcb340ca536e21e8c5ee7d938d263dba1
|
||||
FROM python:3.12-alpine@sha256:a982997504b8ec596f553d78f4de4b961bbdf5254e0177f6e99bb34f4ef16f95
|
||||
|
||||
LABEL maintainer="Robert Kaussow <mail@thegeeklab.de>"
|
||||
LABEL org.opencontainers.image.authors="Robert Kaussow <mail@thegeeklab.de>"
|
||||
|
@ -12,9 +12,9 @@ ENV TZ=UTC
|
|||
|
||||
ADD dist/ansible_doctor-*.whl /
|
||||
|
||||
RUN apk --update add --virtual .build-deps build-base libffi-dev openssl-dev && \
|
||||
RUN apk --update add --virtual .build-deps build-base libffi-dev openssl-dev git && \
|
||||
pip install --upgrade --no-cache-dir pip && \
|
||||
pip install --no-cache-dir $(find / -name "ansible_doctor-*.whl") && \
|
||||
pip install --no-cache-dir $(find / -name "ansible_doctor-*.whl")[ansible-core] && \
|
||||
rm -f ansible_doctor-*.whl && \
|
||||
rm -rf /var/cache/apk/* && \
|
||||
rm -rf /root/.cache/
|
||||
|
|
2
Makefile
2
Makefile
|
@ -1,5 +1,5 @@
|
|||
# renovate: datasource=github-releases depName=thegeeklab/hugo-geekdoc
|
||||
THEME_VERSION := v0.44.0
|
||||
THEME_VERSION := v0.46.0
|
||||
THEME := hugo-geekdoc
|
||||
BASEDIR := docs
|
||||
THEMEDIR := $(BASEDIR)/themes
|
||||
|
|
|
@ -1,3 +1,10 @@
|
|||
"""Provide version information."""
|
||||
|
||||
__version__ = "0.0.0"
|
||||
|
||||
import sys
|
||||
|
||||
try:
|
||||
import ansible # noqa
|
||||
except ImportError:
|
||||
sys.exit("ERROR: Python requirements are missing: 'ansible-core' not found.")
|
||||
|
|
|
@ -6,9 +6,10 @@ import re
|
|||
from collections import defaultdict
|
||||
|
||||
import anyconfig
|
||||
import structlog
|
||||
|
||||
from ansibledoctor.config import SingleConfig
|
||||
from ansibledoctor.utils import SingleLog, _split_string
|
||||
from ansibledoctor.utils import _split_string, sysexit_with_message
|
||||
|
||||
|
||||
class AnnotationItem:
|
||||
|
@ -37,8 +38,7 @@ class Annotation:
|
|||
self._all_items = defaultdict(dict)
|
||||
self._file_handler = None
|
||||
self.config = SingleConfig()
|
||||
self.log = SingleLog()
|
||||
self.logger = self.log.logger
|
||||
self.log = structlog.get_logger()
|
||||
self._files_registry = files_registry
|
||||
|
||||
self._all_annotations = self.config.get_annotations_definition()
|
||||
|
@ -67,7 +67,7 @@ class Annotation:
|
|||
num, line, self._annotation_definition["name"], rfile
|
||||
)
|
||||
if item:
|
||||
self.logger.info(str(item))
|
||||
self.log.info(f"Found {item!s}")
|
||||
self._populate_item(
|
||||
item.get_obj().items(), self._annotation_definition["name"]
|
||||
)
|
||||
|
@ -85,7 +85,7 @@ class Annotation:
|
|||
try:
|
||||
anyconfig.merge(self._all_items[key], value, ac_merge=anyconfig.MS_DICTS)
|
||||
except ValueError as e:
|
||||
self.log.sysexit_with_message(f"Unable to merge annotation values:\n{e}")
|
||||
sysexit_with_message("Failed to merge annotation values", error=e)
|
||||
|
||||
def _get_annotation_data(self, num, line, name, rfile):
|
||||
"""
|
||||
|
@ -171,15 +171,15 @@ class Annotation:
|
|||
|
||||
if parts[2].startswith("$"):
|
||||
source = "".join([x.strip() for x in multiline])
|
||||
multiline = self._str_to_json(key, source, rfile, num, line)
|
||||
multiline = self._str_to_json(key, source, rfile, num)
|
||||
|
||||
item.data[key][parts[1]] = multiline
|
||||
return item
|
||||
|
||||
def _str_to_json(self, key, string, rfile, num, line):
|
||||
def _str_to_json(self, key, string, rfile, num):
|
||||
try:
|
||||
return {key: json.loads(string)}
|
||||
except ValueError:
|
||||
self.log.sysexit_with_message(
|
||||
f"Json value error: Can't parse json in {rfile}:{num!s}:\n{line.strip()}"
|
||||
sysexit_with_message(
|
||||
f"ValueError: Failed to parse json in {rfile}:{num!s}", file=rfile
|
||||
)
|
||||
|
|
|
@ -4,25 +4,32 @@
|
|||
import argparse
|
||||
import os
|
||||
|
||||
import structlog
|
||||
|
||||
import ansibledoctor.exception
|
||||
from ansibledoctor import __version__
|
||||
from ansibledoctor.config import SingleConfig
|
||||
from ansibledoctor.doc_generator import Generator
|
||||
from ansibledoctor.doc_parser import Parser
|
||||
from ansibledoctor.utils import SingleLog
|
||||
from ansibledoctor.utils import sysexit_with_message
|
||||
|
||||
|
||||
class AnsibleDoctor:
|
||||
"""Create main object."""
|
||||
|
||||
def __init__(self):
|
||||
self.log = SingleLog()
|
||||
self.logger = self.log.logger
|
||||
self.args = self._cli_args()
|
||||
self.config = self._get_config()
|
||||
self._execute()
|
||||
log = structlog.get_logger()
|
||||
|
||||
def _cli_args(self):
|
||||
def __init__(self):
|
||||
try:
|
||||
self.config = SingleConfig()
|
||||
self.config.load(args=self._parse_args())
|
||||
self._execute()
|
||||
except ansibledoctor.exception.DoctorError as e:
|
||||
sysexit_with_message(e)
|
||||
except KeyboardInterrupt:
|
||||
sysexit_with_message("Aborted...")
|
||||
|
||||
def _parse_args(self):
|
||||
"""
|
||||
Use argparse for parsing CLI arguments.
|
||||
|
||||
|
@ -33,90 +40,102 @@ class AnsibleDoctor:
|
|||
description="Generate documentation from annotated Ansible roles using templates"
|
||||
)
|
||||
parser.add_argument(
|
||||
"base_dir", nargs="?", help="base directory (default: current working directory)"
|
||||
"base_dir",
|
||||
nargs="?",
|
||||
default=self.config.config.base_dir,
|
||||
help="base directory (default: current working directory)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-c", "--config", dest="config_file", help="path to configuration file"
|
||||
"-c",
|
||||
"--config",
|
||||
dest="config_file",
|
||||
help="path to configuration file",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-o", "--output", dest="output_dir", action="store", help="output directory"
|
||||
"-o",
|
||||
"--output",
|
||||
dest="renderer__dest",
|
||||
action="store",
|
||||
default=self.config.config.renderer.dest,
|
||||
help="output directory",
|
||||
metavar="OUTPUT_DIR",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-r",
|
||||
"--recursive",
|
||||
dest="recursive",
|
||||
action="store_true",
|
||||
default=None,
|
||||
help="run recursively over the base directory subfolders"
|
||||
default=self.config.config.recursive,
|
||||
help="run recursively over the base directory subfolders",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-f",
|
||||
"--force",
|
||||
dest="force_overwrite",
|
||||
dest="renderer.force_overwrite",
|
||||
action="store_true",
|
||||
default=None,
|
||||
help="force overwrite output file"
|
||||
default=self.config.config.renderer.force_overwrite,
|
||||
help="force overwrite output file",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-d",
|
||||
"--dry-run",
|
||||
dest="dry_run",
|
||||
action="store_true",
|
||||
default=None,
|
||||
help="dry run without writing"
|
||||
default=self.config.config.dry_run,
|
||||
help="dry run without writing",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-n",
|
||||
"--no-role-detection",
|
||||
dest="role_detection",
|
||||
action="store_false",
|
||||
default=None,
|
||||
help="disable automatic role detection"
|
||||
default=self.config.config.role.autodetect,
|
||||
help="disable automatic role detection",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-v", dest="logging.level", action="append_const", const=-1, help="increase log level"
|
||||
"-v",
|
||||
dest="logging.level",
|
||||
action="append_const",
|
||||
const=-1,
|
||||
help="increase log level",
|
||||
)
|
||||
parser.add_argument(
|
||||
"-q", dest="logging.level", action="append_const", const=1, help="decrease log level"
|
||||
"-q",
|
||||
dest="logging.level",
|
||||
action="append_const",
|
||||
const=1,
|
||||
help="decrease log level",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--version",
|
||||
action="version",
|
||||
version=f"%(prog)s {__version__}",
|
||||
)
|
||||
parser.add_argument("--version", action="version", version=f"%(prog)s {__version__}")
|
||||
|
||||
return parser.parse_args().__dict__
|
||||
|
||||
def _get_config(self):
|
||||
try:
|
||||
config = SingleConfig(args=self.args)
|
||||
except ansibledoctor.exception.ConfigError as e:
|
||||
self.log.sysexit_with_message(e)
|
||||
|
||||
return config
|
||||
|
||||
def _execute(self):
|
||||
cwd = self.config.base_dir
|
||||
cwd = os.path.abspath(self.config.config.base_dir)
|
||||
walkdirs = [cwd]
|
||||
|
||||
if self.config.recursive:
|
||||
if self.config.config.recursive:
|
||||
walkdirs = [f.path for f in os.scandir(cwd) if f.is_dir()]
|
||||
|
||||
for item in walkdirs:
|
||||
os.chdir(item)
|
||||
self.config.load(root_path=os.getcwd())
|
||||
|
||||
self.config.set_config(base_dir=os.getcwd())
|
||||
try:
|
||||
self.log.set_level(self.config.config["logging"]["level"])
|
||||
except ValueError as e:
|
||||
self.log.sysexit_with_message(f"Can not set log level.\n{e!s}")
|
||||
self.logger.info(f"Using config file: {self.config.config_file}")
|
||||
self.log.debug("Switch working directory", path=item)
|
||||
self.log.info("Lookup config file", path=self.config.config_files)
|
||||
|
||||
self.logger.debug(f"Using working dir: {item}")
|
||||
|
||||
if self.config.config["role_detection"]:
|
||||
if self.config.is_role:
|
||||
self.logger.info(f"Ansible role detected: {self.config.config['role_name']}")
|
||||
if self.config.config.role.autodetect:
|
||||
if self.config.is_role():
|
||||
structlog.contextvars.bind_contextvars(role=self.config.config.role_name)
|
||||
self.log.info("Ansible role detected")
|
||||
else:
|
||||
self.log.sysexit_with_message("No Ansible role detected")
|
||||
sysexit_with_message("No Ansible role detected")
|
||||
else:
|
||||
self.logger.info("Ansible role detection disabled")
|
||||
self.log.info("Ansible role detection disabled")
|
||||
|
||||
doc_parser = Parser()
|
||||
doc_generator = Generator(doc_parser)
|
||||
|
|
|
@ -1,337 +1,226 @@
|
|||
#!/usr/bin/env python3
|
||||
"""Global settings definition."""
|
||||
|
||||
import logging
|
||||
import os
|
||||
import re
|
||||
from io import StringIO
|
||||
|
||||
import anyconfig
|
||||
import environs
|
||||
import jsonschema.exceptions
|
||||
import ruamel.yaml
|
||||
import colorama
|
||||
import structlog
|
||||
from appdirs import AppDirs
|
||||
from jsonschema._utils import format_as_index
|
||||
from dynaconf import Dynaconf, ValidationError, Validator
|
||||
|
||||
import ansibledoctor.exception
|
||||
from ansibledoctor.utils import Singleton
|
||||
|
||||
config_dir = AppDirs("ansible-doctor").user_config_dir
|
||||
default_config_file = os.path.join(config_dir, "config.yml")
|
||||
default_envs_prefix = "ANSIBLE_DOCTOR_"
|
||||
|
||||
|
||||
class Config:
|
||||
"""
|
||||
Create an object with all necessary settings.
|
||||
|
||||
Settings are loade from multiple locations in defined order (last wins):
|
||||
- default settings defined by `self._get_defaults()`
|
||||
- yaml config file, defaults to OS specific user config dir (https://pypi.org/project/appdirs/)
|
||||
- provides cli parameters
|
||||
"""
|
||||
|
||||
SETTINGS = {
|
||||
"config_file": {
|
||||
"default": default_config_file,
|
||||
"env": "CONFIG_FILE",
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"base_dir": {
|
||||
"default": os.getcwd(),
|
||||
"refresh": os.getcwd,
|
||||
"env": "BASE_DIR",
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"role_name": {
|
||||
"default": "",
|
||||
"env": "ROLE_NAME",
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"dry_run": {
|
||||
"default": False,
|
||||
"env": "DRY_RUN",
|
||||
"file": True,
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
"logging.level": {
|
||||
"default": "WARNING",
|
||||
"env": "LOG_LEVEL",
|
||||
"file": True,
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"logging.json": {
|
||||
"default": False,
|
||||
"env": "LOG_JSON",
|
||||
"file": True,
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
"output_dir": {
|
||||
"default": os.getcwd(),
|
||||
"refresh": os.getcwd,
|
||||
"env": "OUTPUT_DIR",
|
||||
"file": True,
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"recursive": {
|
||||
"default": False,
|
||||
"env": "RECURSIVE",
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
"template_dir": {
|
||||
"default": os.path.join(os.path.dirname(os.path.realpath(__file__)), "templates"),
|
||||
"env": "TEMPLATE_DIR",
|
||||
"file": True,
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"template": {
|
||||
"default": "readme",
|
||||
"env": "TEMPLATE",
|
||||
"file": True,
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"template_autotrim": {
|
||||
"default": True,
|
||||
"env": "TEMPLATE_AUTOTRIM",
|
||||
"file": True,
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
"force_overwrite": {
|
||||
"default": False,
|
||||
"env": "FORCE_OVERWRITE",
|
||||
"file": True,
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
"custom_header": {
|
||||
"default": "",
|
||||
"env": "CUSTOM_HEADER",
|
||||
"file": True,
|
||||
"type": environs.Env().str
|
||||
},
|
||||
"exclude_files": {
|
||||
"default": [],
|
||||
"env": "EXCLUDE_FILES",
|
||||
"file": True,
|
||||
"type": environs.Env().list
|
||||
},
|
||||
"exclude_tags": {
|
||||
"default": [],
|
||||
"env": "EXCLUDE_TAGS",
|
||||
"file": True,
|
||||
"type": environs.Env().list
|
||||
},
|
||||
"role_detection": {
|
||||
"default": True,
|
||||
"env": "ROLE_DETECTION",
|
||||
"file": True,
|
||||
"type": environs.Env().bool
|
||||
},
|
||||
}
|
||||
"""Create configuration object."""
|
||||
|
||||
ANNOTATIONS = {
|
||||
"meta": {
|
||||
"name": "meta",
|
||||
"automatic": True,
|
||||
"subtypes": ["value"],
|
||||
"allow_multiple": False
|
||||
"allow_multiple": False,
|
||||
},
|
||||
"todo": {
|
||||
"name": "todo",
|
||||
"automatic": True,
|
||||
"subtypes": ["value"],
|
||||
"allow_multiple": True
|
||||
"allow_multiple": True,
|
||||
},
|
||||
"var": {
|
||||
"name": "var",
|
||||
"automatic": True,
|
||||
"subtypes": ["value", "example", "description", "type", "deprecated"],
|
||||
"allow_multiple": False
|
||||
"allow_multiple": False,
|
||||
},
|
||||
"example": {
|
||||
"name": "example",
|
||||
"automatic": True,
|
||||
"subtypes": [],
|
||||
"allow_multiple": False
|
||||
"allow_multiple": False,
|
||||
},
|
||||
"tag": {
|
||||
"name": "tag",
|
||||
"automatic": True,
|
||||
"subtypes": ["value", "description"],
|
||||
"allow_multiple": False
|
||||
"allow_multiple": False,
|
||||
},
|
||||
}
|
||||
|
||||
def __init__(self, args=None):
|
||||
"""
|
||||
Initialize a new settings class.
|
||||
def __init__(self):
|
||||
self.config_files = [
|
||||
os.path.join(AppDirs("ansible-doctor").user_config_dir, "config.yml"),
|
||||
".ansibledoctor",
|
||||
".ansibledoctor.yml",
|
||||
".ansibledoctor.yaml",
|
||||
]
|
||||
self.config_merge = True
|
||||
self.args = {}
|
||||
self.load()
|
||||
|
||||
:param args: An optional dict of options, arguments and commands from the CLI.
|
||||
:param config_file: An optional path to a yaml config file.
|
||||
:returns: None
|
||||
def load(self, root_path=None, args=None):
|
||||
tmpl_src = os.path.join(os.path.dirname(os.path.realpath(__file__)), "templates")
|
||||
tmpl_provider = ["local", "git"]
|
||||
|
||||
"""
|
||||
if args is None:
|
||||
self._args = {}
|
||||
else:
|
||||
self._args = args
|
||||
self._schema = None
|
||||
self.config = None
|
||||
self.is_role = False
|
||||
self.set_config()
|
||||
if args:
|
||||
if args.get("config_file"):
|
||||
self.config_merge = False
|
||||
self.config_files = [os.path.abspath(args.get("config_file"))]
|
||||
args.pop("config_file")
|
||||
|
||||
def _get_args(self, args):
|
||||
cleaned = dict(filter(lambda item: item[1] is not None, args.items()))
|
||||
self.args = args
|
||||
|
||||
normalized = {}
|
||||
for key, value in cleaned.items():
|
||||
normalized = self._add_dict_branch(normalized, key.split("."), value)
|
||||
self.config = Dynaconf(
|
||||
envvar_prefix="ANSIBLE_DOCTOR",
|
||||
merge_enabled=self.config_merge,
|
||||
core_loaders=["YAML"],
|
||||
root_path=root_path,
|
||||
settings_files=self.config_files,
|
||||
fresh_vars=["base_dir", "output_dir"],
|
||||
validators=[
|
||||
Validator(
|
||||
"base_dir",
|
||||
default=os.getcwd(),
|
||||
apply_default_on_none=True,
|
||||
is_type_of=str,
|
||||
),
|
||||
Validator(
|
||||
"dry_run",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"recursive",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"exclude_files",
|
||||
default=[],
|
||||
is_type_of=list,
|
||||
),
|
||||
Validator(
|
||||
"exclude_tags",
|
||||
default=[],
|
||||
is_type_of=list,
|
||||
),
|
||||
Validator(
|
||||
"role.name",
|
||||
is_type_of=str,
|
||||
),
|
||||
Validator(
|
||||
"role.autodetect",
|
||||
default=True,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"logging.level",
|
||||
default="WARNING",
|
||||
is_in=[
|
||||
"DEBUG",
|
||||
"INFO",
|
||||
"WARNING",
|
||||
"ERROR",
|
||||
"CRITICAL",
|
||||
"debug",
|
||||
"info",
|
||||
"warning",
|
||||
"error",
|
||||
"critical",
|
||||
],
|
||||
),
|
||||
Validator(
|
||||
"logging.json",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"recursive",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"template.src",
|
||||
default=f"local>{tmpl_src}",
|
||||
is_type_of=str,
|
||||
condition=lambda x: re.match(r"^(local|git)\s*>\s*", x),
|
||||
messages={
|
||||
"condition": f"Template provider must be one of {tmpl_provider}.",
|
||||
},
|
||||
),
|
||||
Validator(
|
||||
"template.name",
|
||||
default="readme",
|
||||
is_type_of=str,
|
||||
),
|
||||
Validator(
|
||||
"template.options.tabulate_variables",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"renderer.autotrim",
|
||||
default=True,
|
||||
is_type_of=bool,
|
||||
),
|
||||
Validator(
|
||||
"renderer.include_header",
|
||||
default="",
|
||||
is_type_of=str,
|
||||
),
|
||||
Validator(
|
||||
"renderer.dest",
|
||||
default=os.path.relpath(os.getcwd()),
|
||||
is_type_of=str,
|
||||
),
|
||||
Validator(
|
||||
"renderer.force_overwrite",
|
||||
default=False,
|
||||
is_type_of=bool,
|
||||
),
|
||||
],
|
||||
)
|
||||
|
||||
self.validate()
|
||||
|
||||
# Override correct log level from argparse
|
||||
levels = ["DEBUG", "INFO", "WARNING", "ERROR", "CRITICAL"]
|
||||
log_level = levels.index(self.SETTINGS["logging.level"]["default"])
|
||||
if normalized.get("logging"):
|
||||
for adjustment in normalized["logging"]["level"]:
|
||||
log_level = min(len(levels) - 1, max(log_level + adjustment, 0))
|
||||
normalized["logging"]["level"] = levels[log_level]
|
||||
log_level = levels.index(self.config.logging.level.upper())
|
||||
if self.args.get("logging.level") and isinstance(self.args["logging.level"], list):
|
||||
for lvl in self.args["logging.level"]:
|
||||
log_level = min(len(levels) - 1, max(log_level + lvl, 0))
|
||||
|
||||
return normalized
|
||||
self.args["logging__level"] = levels[log_level]
|
||||
|
||||
def _get_defaults(self):
|
||||
normalized = {}
|
||||
for key, item in self.SETTINGS.items():
|
||||
if item.get("refresh"):
|
||||
item["default"] = item["refresh"]()
|
||||
normalized = self._add_dict_branch(normalized, key.split("."), item["default"])
|
||||
if root_path:
|
||||
self.args["base_dir"] = root_path
|
||||
|
||||
self.schema = anyconfig.gen_schema(normalized)
|
||||
return normalized
|
||||
self.config.update(self.args)
|
||||
self.validate()
|
||||
|
||||
def _get_envs(self):
|
||||
normalized = {}
|
||||
for key, item in self.SETTINGS.items():
|
||||
if item.get("env"):
|
||||
envname = f"{default_envs_prefix}{item['env']}"
|
||||
try:
|
||||
value = item["type"](envname)
|
||||
normalized = self._add_dict_branch(normalized, key.split("."), value)
|
||||
except environs.EnvError as e:
|
||||
if f'"{envname}" not set' in str(e):
|
||||
pass
|
||||
else:
|
||||
raise ansibledoctor.exception.ConfigError(
|
||||
"Unable to read environment variable", str(e)
|
||||
) from e
|
||||
self._init_logger()
|
||||
|
||||
return normalized
|
||||
|
||||
def set_config(self, base_dir=None):
|
||||
args = self._get_args(self._args)
|
||||
envs = self._get_envs()
|
||||
defaults = self._get_defaults()
|
||||
|
||||
self.recursive = defaults.get("recursive")
|
||||
if envs.get("recursive"):
|
||||
self.recursive = envs.get("recursive")
|
||||
if args.get("recursive"):
|
||||
self.recursive = args.get("recursive")
|
||||
if "recursive" in defaults:
|
||||
defaults.pop("recursive")
|
||||
|
||||
self.config_file = defaults.get("config_file")
|
||||
if envs.get("config_file"):
|
||||
self.config_file = self._normalize_path(envs.get("config_file"))
|
||||
if args.get("config_file"):
|
||||
self.config_file = self._normalize_path(args.get("config_file"))
|
||||
if "config_file" in defaults:
|
||||
defaults.pop("config_file")
|
||||
|
||||
self.base_dir = defaults.get("base_dir")
|
||||
if envs.get("base_dir"):
|
||||
self.base_dir = self._normalize_path(envs.get("base_dir"))
|
||||
if args.get("base_dir"):
|
||||
self.base_dir = self._normalize_path(args.get("base_dir"))
|
||||
if base_dir:
|
||||
self.base_dir = base_dir
|
||||
if "base_dir" in defaults:
|
||||
defaults.pop("base_dir")
|
||||
|
||||
self.is_role = os.path.isdir(os.path.join(self.base_dir, "tasks"))
|
||||
|
||||
# compute role_name default
|
||||
defaults["role_name"] = os.path.basename(self.base_dir)
|
||||
|
||||
source_files = []
|
||||
source_files.append((self.config_file, False))
|
||||
source_files.append((os.path.join(os.getcwd(), ".ansibledoctor"), True))
|
||||
source_files.append((os.path.join(os.getcwd(), ".ansibledoctor.yml"), True))
|
||||
source_files.append((os.path.join(os.getcwd(), ".ansibledoctor.yaml"), True))
|
||||
|
||||
for (config, first_found) in source_files:
|
||||
if config and os.path.exists(config):
|
||||
with open(config, encoding="utf8") as stream:
|
||||
s = stream.read()
|
||||
try:
|
||||
file_dict = ruamel.yaml.safe_load(s)
|
||||
except (
|
||||
ruamel.yaml.composer.ComposerError, ruamel.yaml.scanner.ScannerError
|
||||
) as e:
|
||||
message = f"{e.context} {e.problem}"
|
||||
raise ansibledoctor.exception.ConfigError(
|
||||
f"Unable to read config file: {config}", message
|
||||
) from e
|
||||
|
||||
if self._validate(file_dict):
|
||||
anyconfig.merge(defaults, file_dict, ac_merge=anyconfig.MS_DICTS)
|
||||
defaults["logging"]["level"] = defaults["logging"]["level"].upper()
|
||||
|
||||
self.config_file = config
|
||||
if first_found:
|
||||
break
|
||||
|
||||
if self._validate(envs):
|
||||
anyconfig.merge(defaults, envs, ac_merge=anyconfig.MS_DICTS)
|
||||
|
||||
if self._validate(args):
|
||||
anyconfig.merge(defaults, args, ac_merge=anyconfig.MS_DICTS)
|
||||
|
||||
fix_files = ["output_dir", "template_dir", "custom_header"]
|
||||
for filename in fix_files:
|
||||
if defaults[filename] and defaults[filename] != "":
|
||||
defaults[filename] = self._normalize_path(defaults[filename])
|
||||
|
||||
defaults["logging"]["level"] = defaults["logging"]["level"].upper()
|
||||
|
||||
self.config = defaults
|
||||
|
||||
def _normalize_path(self, path):
|
||||
if not os.path.isabs(path):
|
||||
base = os.path.join(os.getcwd(), path)
|
||||
return os.path.abspath(os.path.expanduser(os.path.expandvars(base)))
|
||||
|
||||
return path
|
||||
|
||||
def _validate(self, config):
|
||||
def validate(self):
|
||||
try:
|
||||
anyconfig.validate(config, self.schema, ac_schema_safe=False)
|
||||
except jsonschema.exceptions.ValidationError as e:
|
||||
schema_error = "Failed validating '{validator}' in schema{schema}\n{message}".format(
|
||||
validator=e.validator,
|
||||
schema=format_as_index(list(e.relative_schema_path)[:-1]),
|
||||
message=e.message
|
||||
)
|
||||
raise ansibledoctor.exception.ConfigError("Configuration error", schema_error) from e
|
||||
self.config.validators.validate_all()
|
||||
except ValidationError as e:
|
||||
raise ansibledoctor.exception.ConfigError("Configuration error", e.message) from e
|
||||
|
||||
return True
|
||||
|
||||
def _add_dict_branch(self, tree, vector, value):
|
||||
key = vector[0]
|
||||
tree[key] = value \
|
||||
if len(vector) == 1 \
|
||||
else self._add_dict_branch(tree[key] if key in tree else {}, vector[1:], value)
|
||||
return tree
|
||||
def is_role(self):
|
||||
self.config.role_name = self.config.get(
|
||||
"role_name", os.path.basename(self.config.base_dir)
|
||||
)
|
||||
return os.path.isdir(os.path.join(self.config.base_dir, "tasks"))
|
||||
|
||||
def get_annotations_definition(self, automatic=True):
|
||||
annotations = {}
|
||||
if automatic:
|
||||
for k, item in self.ANNOTATIONS.items():
|
||||
if "automatic" in item and item["automatic"]:
|
||||
if item.get("automatic"):
|
||||
annotations[k] = item
|
||||
return annotations
|
||||
|
||||
|
@ -339,19 +228,84 @@ class Config:
|
|||
annotations = []
|
||||
if automatic:
|
||||
for k, item in self.ANNOTATIONS.items():
|
||||
if "automatic" in item and item["automatic"]:
|
||||
if item.get("automatic"):
|
||||
annotations.append(k)
|
||||
return annotations
|
||||
|
||||
def get_template(self):
|
||||
"""
|
||||
Get the base dir for the template to use.
|
||||
def _init_logger(self):
|
||||
styles = structlog.dev.ConsoleRenderer.get_default_level_styles()
|
||||
styles["debug"] = colorama.Fore.BLUE
|
||||
|
||||
:return: str abs path
|
||||
"""
|
||||
template_dir = self.config.get("template_dir")
|
||||
template = self.config.get("template")
|
||||
return os.path.realpath(os.path.join(template_dir, template))
|
||||
processors = [
|
||||
structlog.contextvars.merge_contextvars,
|
||||
structlog.processors.add_log_level,
|
||||
structlog.processors.StackInfoRenderer(),
|
||||
structlog.dev.set_exc_info,
|
||||
structlog.processors.TimeStamper(fmt="%Y-%m-%d %H:%M:%S", utc=False),
|
||||
]
|
||||
|
||||
if self.config.logging.json:
|
||||
processors.append(ErrorStringifier())
|
||||
processors.append(structlog.processors.JSONRenderer())
|
||||
else:
|
||||
processors.append(MultilineConsoleRenderer(level_styles=styles))
|
||||
|
||||
try:
|
||||
structlog.configure(
|
||||
processors=processors,
|
||||
wrapper_class=structlog.make_filtering_bound_logger(
|
||||
logging.getLevelName(self.config.get("logging.level")),
|
||||
),
|
||||
)
|
||||
structlog.contextvars.unbind_contextvars()
|
||||
except KeyError as e:
|
||||
raise ansibledoctor.exception.ConfigError(f"Can not set log level: {e!s}") from e
|
||||
|
||||
|
||||
class ErrorStringifier:
|
||||
"""A processor that converts exceptions to a string representation."""
|
||||
|
||||
def __call__(self, _, __, event_dict):
|
||||
if "error" not in event_dict:
|
||||
return event_dict
|
||||
|
||||
err = event_dict.get("error")
|
||||
|
||||
if isinstance(err, Exception):
|
||||
event_dict["error"] = f"{err.__class__.__name__}: {err}"
|
||||
|
||||
return event_dict
|
||||
|
||||
|
||||
class MultilineConsoleRenderer(structlog.dev.ConsoleRenderer):
|
||||
"""A processor for printing multiline strings."""
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
super().__init__(*args, **kwargs)
|
||||
|
||||
def __call__(self, _, __, event_dict):
|
||||
err = None
|
||||
|
||||
if "error" in event_dict:
|
||||
err = event_dict.pop("error")
|
||||
|
||||
event_dict = super().__call__(_, __, event_dict)
|
||||
|
||||
if not err:
|
||||
return event_dict
|
||||
|
||||
sio = StringIO()
|
||||
sio.write(event_dict)
|
||||
|
||||
if isinstance(err, Exception):
|
||||
sio.write(
|
||||
f"\n{colorama.Fore.RED}{err.__class__.__name__}:"
|
||||
f"{colorama.Style.RESET_ALL} {str(err).strip()}"
|
||||
)
|
||||
else:
|
||||
sio.write(f"\n{err.strip()}")
|
||||
|
||||
return sio.getvalue()
|
||||
|
||||
|
||||
class SingleConfig(Config, metaclass=Singleton):
|
||||
|
|
|
@ -1,75 +1,52 @@
|
|||
#!/usr/bin/env python3
|
||||
"""Prepare output and write compiled jinja2 templates."""
|
||||
|
||||
import glob
|
||||
import ntpath
|
||||
import os
|
||||
import re
|
||||
from functools import reduce
|
||||
|
||||
import jinja2.exceptions
|
||||
import ruamel.yaml
|
||||
import structlog
|
||||
from jinja2 import Environment, FileSystemLoader
|
||||
from jinja2.filters import pass_eval_context
|
||||
|
||||
import ansibledoctor.exception
|
||||
from ansibledoctor.config import SingleConfig
|
||||
from ansibledoctor.utils import FileUtils, SingleLog
|
||||
from ansibledoctor.template import Template
|
||||
from ansibledoctor.utils import FileUtils, sysexit_with_message
|
||||
|
||||
|
||||
class Generator:
|
||||
"""Generate documentation from jinja2 templates."""
|
||||
|
||||
def __init__(self, doc_parser):
|
||||
self.template_files = []
|
||||
self.extension = "j2"
|
||||
self._parser = None
|
||||
self.log = structlog.get_logger()
|
||||
self.config = SingleConfig()
|
||||
self.log = SingleLog()
|
||||
self.logger = self.log.logger
|
||||
self.template = Template(
|
||||
self.config.config.get("template.name"),
|
||||
self.config.config.get("template.src"),
|
||||
)
|
||||
self._parser = doc_parser
|
||||
self._scan_template()
|
||||
|
||||
def _scan_template(self):
|
||||
"""
|
||||
Search for Jinja2 (.j2) files to apply to the destination.
|
||||
|
||||
:return: None
|
||||
"""
|
||||
template_dir = self.config.get_template()
|
||||
if os.path.isdir(template_dir):
|
||||
self.logger.info(f"Using template dir: {template_dir}")
|
||||
else:
|
||||
self.log.sysexit_with_message(f"Can not open template dir {template_dir}")
|
||||
|
||||
for file in glob.iglob(template_dir + "/**/*." + self.extension, recursive=True):
|
||||
relative_file = file[len(template_dir) + 1:]
|
||||
if ntpath.basename(file)[:1] != "_":
|
||||
self.logger.debug(f"Found template file: {relative_file}")
|
||||
self.template_files.append(relative_file)
|
||||
else:
|
||||
self.logger.debug(f"Ignoring template file: {relative_file}")
|
||||
|
||||
def _create_dir(self, directory):
|
||||
if not self.config.config["dry_run"] and not os.path.isdir(directory):
|
||||
try:
|
||||
os.makedirs(directory, exist_ok=True)
|
||||
self.logger.info(f"Creating dir: {directory}")
|
||||
self.log.info(f"Creating dir: {directory}")
|
||||
except FileExistsError as e:
|
||||
self.log.sysexit_with_message(str(e))
|
||||
sysexit_with_message(e)
|
||||
|
||||
def _write_doc(self):
|
||||
files_to_overwite = []
|
||||
|
||||
for file in self.template_files:
|
||||
for tf in self.template.files:
|
||||
doc_file = os.path.join(
|
||||
self.config.config.get("output_dir"),
|
||||
os.path.splitext(file)[0]
|
||||
self.config.config.get("renderer.dest"), os.path.splitext(tf)[0]
|
||||
)
|
||||
if os.path.isfile(doc_file):
|
||||
files_to_overwite.append(doc_file)
|
||||
|
||||
header_file = self.config.config.get("custom_header")
|
||||
header_file = self.config.config.get("renderer.include_header")
|
||||
role_data = self._parser.get_data()
|
||||
header_content = ""
|
||||
if bool(header_file):
|
||||
|
@ -78,70 +55,69 @@ class Generator:
|
|||
with open(header_file) as a:
|
||||
header_content = a.read()
|
||||
except FileNotFoundError as e:
|
||||
self.log.sysexit_with_message(f"Can not open custom header file\n{e!s}")
|
||||
sysexit_with_message("Can not open custom header file", path=header_file, error=e)
|
||||
|
||||
if (
|
||||
len(files_to_overwite) > 0 and self.config.config.get("force_overwrite") is False
|
||||
len(files_to_overwite) > 0
|
||||
and self.config.config.get("renderer.force_overwrite") is False
|
||||
and not self.config.config["dry_run"]
|
||||
):
|
||||
files_to_overwite_string = "\n".join(files_to_overwite)
|
||||
self.logger.warning(f"This files will be overwritten:\n{files_to_overwite_string}")
|
||||
prompt = f"These files will be overwritten:\n{files_to_overwite_string}".replace(
|
||||
"\n", "\n... "
|
||||
)
|
||||
|
||||
try:
|
||||
if not FileUtils.query_yes_no("Do you want to continue?"):
|
||||
self.log.sysexit_with_message("Aborted...")
|
||||
except ansibledoctor.exception.InputError as e:
|
||||
self.logger.debug(str(e))
|
||||
self.log.sysexit_with_message("Aborted...")
|
||||
if not FileUtils.query_yes_no(f"{prompt}\nDo you want to continue?"):
|
||||
sysexit_with_message("Aborted...")
|
||||
except KeyboardInterrupt:
|
||||
sysexit_with_message("Aborted...")
|
||||
|
||||
for file in self.template_files:
|
||||
for tf in self.template.files:
|
||||
doc_file = os.path.join(
|
||||
self.config.config.get("output_dir"),
|
||||
os.path.splitext(file)[0]
|
||||
self.config.config.get("renderer.dest"), os.path.splitext(tf)[0]
|
||||
)
|
||||
source_file = self.config.get_template() + "/" + file
|
||||
template = os.path.join(self.template.path, tf)
|
||||
|
||||
self.logger.debug(f"Writing doc output to: {doc_file} from: {source_file}")
|
||||
self.log.debug("Writing renderer output", path=doc_file, src=os.path.dirname(template))
|
||||
|
||||
# make sure the directory exists
|
||||
self._create_dir(os.path.dirname(doc_file))
|
||||
|
||||
if os.path.exists(source_file) and os.path.isfile(source_file):
|
||||
with open(source_file) as template:
|
||||
if os.path.exists(template) and os.path.isfile(template):
|
||||
with open(template) as template:
|
||||
data = template.read()
|
||||
if data is not None:
|
||||
try:
|
||||
jenv = Environment( # nosec
|
||||
loader=FileSystemLoader(self.config.get_template()),
|
||||
loader=FileSystemLoader(self.template.path),
|
||||
lstrip_blocks=True,
|
||||
trim_blocks=True,
|
||||
autoescape=jinja2.select_autoescape()
|
||||
autoescape=jinja2.select_autoescape(),
|
||||
)
|
||||
jenv.filters["to_nice_yaml"] = self._to_nice_yaml
|
||||
jenv.filters["deep_get"] = self._deep_get
|
||||
jenv.filters["safe_join"] = self._safe_join
|
||||
# keep the old name of the function to not break custom templates.
|
||||
jenv.filters["save_join"] = self._safe_join
|
||||
data = jenv.from_string(data).render(role_data, role=role_data)
|
||||
template_options = self.config.config.get("template.options")
|
||||
data = jenv.from_string(data).render(
|
||||
role_data, role=role_data, options=template_options
|
||||
)
|
||||
if not self.config.config["dry_run"]:
|
||||
with open(doc_file, "wb") as outfile:
|
||||
outfile.write(header_content.encode("utf-8"))
|
||||
outfile.write(data.encode("utf-8"))
|
||||
self.logger.info(f"Writing to: {doc_file}")
|
||||
else:
|
||||
self.logger.info(f"Writing to: {doc_file}")
|
||||
except (
|
||||
jinja2.exceptions.UndefinedError,
|
||||
jinja2.exceptions.TemplateSyntaxError,
|
||||
jinja2.exceptions.TemplateRuntimeError
|
||||
jinja2.exceptions.TemplateRuntimeError,
|
||||
) as e:
|
||||
self.log.sysexit_with_message(
|
||||
f"Jinja2 templating error while loading file: '{file}'\n{e!s}"
|
||||
sysexit_with_message(
|
||||
"Jinja2 template error while loading file", path=tf, error=e
|
||||
)
|
||||
except UnicodeEncodeError as e:
|
||||
self.log.sysexit_with_message(
|
||||
f"Unable to print special characters\n{e!s}"
|
||||
)
|
||||
sysexit_with_message("Failed to print special characters", error=e)
|
||||
|
||||
def _to_nice_yaml(self, a, indent=4, **kw):
|
||||
"""Make verbose, human readable yaml."""
|
||||
|
@ -154,8 +130,9 @@ class Generator:
|
|||
def _deep_get(self, _, dictionary, keys):
|
||||
default = None
|
||||
return reduce(
|
||||
lambda d, key: d.get(key, default)
|
||||
if isinstance(d, dict) else default, keys.split("."), dictionary
|
||||
lambda d, key: d.get(key, default) if isinstance(d, dict) else default,
|
||||
keys.split("."),
|
||||
dictionary,
|
||||
)
|
||||
|
||||
@pass_eval_context
|
||||
|
@ -165,12 +142,11 @@ class Generator:
|
|||
|
||||
normalized = jinja2.filters.do_join(eval_ctx, value, d, attribute=None)
|
||||
|
||||
if self.config.config["template_autotrim"]:
|
||||
if self.config.config.renderer.autotrim:
|
||||
for s in [r" +(\n|\t| )", r"(\n|\t) +"]:
|
||||
normalized = re.sub(s, "\\1", normalized)
|
||||
|
||||
return jinja2.filters.do_mark_safe(normalized)
|
||||
|
||||
def render(self):
|
||||
self.logger.info(f"Using output dir: {self.config.config.get('output_dir')}")
|
||||
self._write_doc()
|
||||
|
|
|
@ -3,17 +3,17 @@
|
|||
|
||||
import fnmatch
|
||||
from collections import defaultdict
|
||||
from contextlib import suppress
|
||||
|
||||
import anyconfig
|
||||
import ruamel.yaml
|
||||
from nested_lookup import nested_lookup
|
||||
import structlog
|
||||
|
||||
from ansibledoctor.annotation import Annotation
|
||||
from ansibledoctor.config import SingleConfig
|
||||
from ansibledoctor.contstants import YAML_EXTENSIONS
|
||||
from ansibledoctor.exception import YAMLError
|
||||
from ansibledoctor.file_registry import Registry
|
||||
from ansibledoctor.utils import SingleLog, UnsafeTag, flatten
|
||||
from ansibledoctor.utils import flatten, sysexit_with_message
|
||||
from ansibledoctor.utils.yamlhelper import parse_yaml, parse_yaml_ansible
|
||||
|
||||
|
||||
class Parser:
|
||||
|
@ -23,110 +23,73 @@ class Parser:
|
|||
self._annotation_objs = {}
|
||||
self._data = defaultdict(dict)
|
||||
self.config = SingleConfig()
|
||||
self.log = SingleLog()
|
||||
self.logger = SingleLog().logger
|
||||
self.log = structlog.get_logger()
|
||||
self._files_registry = Registry()
|
||||
self._parse_meta_file()
|
||||
self._parse_var_files()
|
||||
self._parse_task_tags()
|
||||
self._populate_doc_data()
|
||||
|
||||
def _yaml_remove_comments(self, d):
|
||||
if isinstance(d, dict):
|
||||
for k, v in d.items():
|
||||
self._yaml_remove_comments(k)
|
||||
self._yaml_remove_comments(v)
|
||||
elif isinstance(d, list):
|
||||
for elem in d:
|
||||
self._yaml_remove_comments(elem)
|
||||
|
||||
with suppress(AttributeError):
|
||||
attr = "comment" if isinstance(
|
||||
d, ruamel.yaml.scalarstring.ScalarString
|
||||
) else ruamel.yaml.comments.Comment.attrib
|
||||
delattr(d, attr)
|
||||
|
||||
def _parse_var_files(self):
|
||||
for rfile in self._files_registry.get_files():
|
||||
if any(fnmatch.fnmatch(rfile, "*/defaults/*." + ext) for ext in YAML_EXTENSIONS):
|
||||
with open(rfile, encoding="utf8") as yaml_file:
|
||||
with open(rfile, encoding="utf8") as yamlfile:
|
||||
try:
|
||||
ruamel.yaml.add_constructor(
|
||||
UnsafeTag.yaml_tag,
|
||||
UnsafeTag.yaml_constructor,
|
||||
constructor=ruamel.yaml.SafeConstructor
|
||||
)
|
||||
raw = parse_yaml(yamlfile)
|
||||
except YAMLError as e:
|
||||
sysexit_with_message("Failed to read yaml file", path=rfile, error=e)
|
||||
|
||||
raw = ruamel.yaml.YAML(typ="rt").load(yaml_file)
|
||||
self._yaml_remove_comments(raw)
|
||||
data = defaultdict(dict, raw or {})
|
||||
|
||||
data = defaultdict(dict, raw or {})
|
||||
for key, value in data.items():
|
||||
self._data["var"][key] = {"value": {key: value}}
|
||||
except (
|
||||
ruamel.yaml.composer.ComposerError,
|
||||
ruamel.yaml.scanner.ScannerError,
|
||||
ruamel.yaml.constructor.ConstructorError,
|
||||
ruamel.yaml.constructor.DuplicateKeyError,
|
||||
) as e:
|
||||
message = f"{e.context} {e.problem}"
|
||||
self.log.sysexit_with_message(
|
||||
f"Unable to read yaml file {rfile}\n{message}"
|
||||
)
|
||||
for key, value in data.items():
|
||||
self._data["var"][key] = {"value": {key: value}}
|
||||
|
||||
def _parse_meta_file(self):
|
||||
self._data["meta"]["name"] = {"value": self.config.config["role_name"]}
|
||||
|
||||
for rfile in self._files_registry.get_files():
|
||||
if any("meta/main." + ext in rfile for ext in YAML_EXTENSIONS):
|
||||
with open(rfile, encoding="utf8") as yaml_file:
|
||||
with open(rfile, encoding="utf8") as yamlfile:
|
||||
try:
|
||||
raw = ruamel.yaml.YAML(typ="rt").load(yaml_file)
|
||||
self._yaml_remove_comments(raw)
|
||||
raw = parse_yaml(yamlfile)
|
||||
except YAMLError as e:
|
||||
sysexit_with_message("Failed to read yaml file", path=rfile, error=e)
|
||||
|
||||
data = defaultdict(dict, raw)
|
||||
if data.get("galaxy_info"):
|
||||
for key, value in data.get("galaxy_info").items():
|
||||
self._data["meta"][key] = {"value": value}
|
||||
data = defaultdict(dict, raw)
|
||||
if data.get("galaxy_info"):
|
||||
for key, value in data.get("galaxy_info").items():
|
||||
self._data["meta"][key] = {"value": value}
|
||||
|
||||
if data.get("dependencies") is not None:
|
||||
self._data["meta"]["dependencies"] = {
|
||||
"value": data.get("dependencies")
|
||||
}
|
||||
except (
|
||||
ruamel.yaml.composer.ComposerError, ruamel.yaml.scanner.ScannerError
|
||||
) as e:
|
||||
message = f"{e.context} {e.problem}"
|
||||
self.log.sysexit_with_message(
|
||||
f"Unable to read yaml file {rfile}\n{message}"
|
||||
)
|
||||
if data.get("dependencies") is not None:
|
||||
self._data["meta"]["dependencies"] = {"value": data.get("dependencies")}
|
||||
|
||||
def _parse_task_tags(self):
|
||||
for rfile in self._files_registry.get_files():
|
||||
if any(fnmatch.fnmatch(rfile, "*/tasks/*." + ext) for ext in YAML_EXTENSIONS):
|
||||
with open(rfile, encoding="utf8") as yaml_file:
|
||||
with open(rfile, encoding="utf8") as yamlfile:
|
||||
try:
|
||||
raw = ruamel.yaml.YAML(typ="rt").load(yaml_file)
|
||||
self._yaml_remove_comments(raw)
|
||||
raw = parse_yaml_ansible(yamlfile)
|
||||
except YAMLError as e:
|
||||
sysexit_with_message("Failed to read yaml file", path=rfile, error=e)
|
||||
|
||||
tags = list(set(flatten(nested_lookup("tags", raw))))
|
||||
for tag in [
|
||||
x for x in tags if x not in self.config.config["exclude_tags"]
|
||||
]:
|
||||
self._data["tag"][tag] = {"value": tag}
|
||||
except (
|
||||
ruamel.yaml.composer.ComposerError, ruamel.yaml.scanner.ScannerError
|
||||
) as e:
|
||||
message = f"{e.context} {e.problem}"
|
||||
self.log.sysexit_with_message(
|
||||
f"Unable to read yaml file {rfile}\n{message}"
|
||||
)
|
||||
tags = []
|
||||
for task in raw:
|
||||
task_tags = task.get("tags")
|
||||
if isinstance(task_tags, str):
|
||||
task_tags = [task_tags]
|
||||
|
||||
for tag in task_tags:
|
||||
if tag not in self.config.config["exclude_tags"]:
|
||||
tags.append(tag)
|
||||
|
||||
for tag in flatten(tags):
|
||||
self._data["tag"][tag] = {"value": tag}
|
||||
|
||||
def _populate_doc_data(self):
|
||||
"""Generate the documentation data object."""
|
||||
tags = defaultdict(dict)
|
||||
for annotation in self.config.get_annotations_names(automatic=True):
|
||||
self.logger.info(f"Finding annotations for: @{annotation}")
|
||||
self.log.info(f"Lookup annotation @{annotation}")
|
||||
self._annotation_objs[annotation] = Annotation(
|
||||
name=annotation, files_registry=self._files_registry
|
||||
)
|
||||
|
@ -135,7 +98,7 @@ class Parser:
|
|||
try:
|
||||
anyconfig.merge(self._data, tags, ac_merge=anyconfig.MS_DICTS)
|
||||
except ValueError as e:
|
||||
self.log.sysexit_with_message(f"Unable to merge annotation values:\n{e}")
|
||||
sysexit_with_message("Failed to merge annotation values", error=e)
|
||||
|
||||
def get_data(self):
|
||||
return self._data
|
||||
|
|
|
@ -10,13 +10,17 @@ class DoctorError(Exception):
|
|||
self.original_exception = original_exception
|
||||
|
||||
|
||||
class YAMLError(DoctorError):
|
||||
"""Errors while reading a yaml file."""
|
||||
|
||||
pass
|
||||
|
||||
|
||||
class ConfigError(DoctorError):
|
||||
"""Errors related to config file handling."""
|
||||
|
||||
pass
|
||||
|
||||
|
||||
class InputError(DoctorError):
|
||||
"""Errors related to config file handling."""
|
||||
|
||||
pass
|
||||
class TemplateError(DoctorError):
|
||||
"""Errors related to template file handling."""
|
||||
|
|
|
@ -5,10 +5,10 @@ import glob
|
|||
import os
|
||||
|
||||
import pathspec
|
||||
import structlog
|
||||
|
||||
from ansibledoctor.config import SingleConfig
|
||||
from ansibledoctor.contstants import YAML_EXTENSIONS
|
||||
from ansibledoctor.utils import SingleLog
|
||||
|
||||
|
||||
class Registry:
|
||||
|
@ -21,7 +21,7 @@ class Registry:
|
|||
def __init__(self):
|
||||
self._doc = []
|
||||
self.config = SingleConfig()
|
||||
self.log = SingleLog().logger
|
||||
self.log = structlog.get_logger()
|
||||
self._scan_for_yamls()
|
||||
|
||||
def get_files(self):
|
||||
|
@ -35,20 +35,17 @@ class Registry:
|
|||
:return: None
|
||||
"""
|
||||
extensions = YAML_EXTENSIONS
|
||||
base_dir = self.config.base_dir
|
||||
role_name = os.path.basename(base_dir)
|
||||
base_dir = self.config.config.base_dir
|
||||
excludes = self.config.config.get("exclude_files")
|
||||
excludespec = pathspec.PathSpec.from_lines("gitwildmatch", excludes)
|
||||
|
||||
self.log.debug(f"Scan for files: {base_dir}")
|
||||
self.log.debug("Lookup role files", path=base_dir)
|
||||
|
||||
for extension in extensions:
|
||||
pattern = os.path.join(base_dir, "**/*." + extension)
|
||||
for filename in glob.iglob(pattern, recursive=True):
|
||||
if not excludespec.match_file(filename):
|
||||
self.log.debug(
|
||||
f"Adding file to '{role_name}': {os.path.relpath(filename, base_dir)}"
|
||||
)
|
||||
self.log.debug("Found role file", path=os.path.relpath(filename, base_dir))
|
||||
self._doc.append(filename)
|
||||
else:
|
||||
self.log.debug(f"Excluding file: {os.path.relpath(filename, base_dir)}")
|
||||
self.log.debug("Skippped role file", path=os.path.relpath(filename, base_dir))
|
||||
|
|
113
ansibledoctor/template.py
Normal file
113
ansibledoctor/template.py
Normal file
|
@ -0,0 +1,113 @@
|
|||
"""Module for handling templates."""
|
||||
|
||||
import atexit
|
||||
import glob
|
||||
import ntpath
|
||||
import os
|
||||
import shutil
|
||||
import tempfile
|
||||
|
||||
import structlog
|
||||
from git import GitCommandError, Repo
|
||||
|
||||
import ansibledoctor.exception
|
||||
from ansibledoctor.utils import sysexit_with_message
|
||||
|
||||
|
||||
class Template:
|
||||
"""
|
||||
Represents a template that can be used to generate content.
|
||||
|
||||
Templates can be sourced from a local file or a Git repository. The `Template` class handles
|
||||
the initialization and setup of a template, including cloning a Git repository if necessary.
|
||||
|
||||
Args:
|
||||
----
|
||||
name (str): The name of the template.
|
||||
src (str): The source of the template, in the format `<provider>><path>`.
|
||||
Supported providers are `local` and `git`.
|
||||
|
||||
Raises:
|
||||
------
|
||||
ansibledoctor.exception.TemplateError
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, name, src):
|
||||
self.log = structlog.get_logger()
|
||||
self.name = name
|
||||
self.src = src
|
||||
|
||||
try:
|
||||
provider, path = self.src.split(">", 1)
|
||||
except ValueError as e:
|
||||
raise ansibledoctor.exception.TemplateError(
|
||||
"Error reading template src", str(e)
|
||||
) from e
|
||||
|
||||
self.provider = provider.strip().lower()
|
||||
self.path = path.strip()
|
||||
|
||||
if self.provider == "local":
|
||||
self.path = os.path.realpath(os.path.join(self.path, self.name))
|
||||
elif self.provider == "git":
|
||||
repo_url, branch_or_tag = (
|
||||
self.path.split("#", 1) if "#" in self.path else (self.path, None)
|
||||
)
|
||||
temp_dir = self._clone_repo(repo_url, branch_or_tag)
|
||||
self.path = os.path.join(temp_dir, self.name)
|
||||
else:
|
||||
raise ansibledoctor.exception.TemplateError(
|
||||
f"Unsupported template provider: {provider}"
|
||||
)
|
||||
|
||||
self.files = self._scan_files()
|
||||
|
||||
def _clone_repo(self, repo_url, branch_or_tag=None):
|
||||
temp_dir = tempfile.mkdtemp(prefix="ansibledoctor-")
|
||||
atexit.register(self._cleanup_temp_dir, temp_dir)
|
||||
|
||||
try:
|
||||
self.log.debug("Cloning template repo", src=repo_url)
|
||||
repo = Repo.clone_from(repo_url, temp_dir)
|
||||
if branch_or_tag:
|
||||
self.log.debug(f"Checking out branch or tag: {branch_or_tag}")
|
||||
try:
|
||||
repo.git.checkout(branch_or_tag)
|
||||
except GitCommandError as e:
|
||||
raise ansibledoctor.exception.TemplateError(
|
||||
f"Error checking out branch or tag: {branch_or_tag}: {e}"
|
||||
) from e
|
||||
|
||||
return temp_dir
|
||||
except GitCommandError as e:
|
||||
msg = e.stderr.strip("'").strip()
|
||||
msg = msg.removeprefix("stderr: ")
|
||||
|
||||
raise ansibledoctor.exception.TemplateError(
|
||||
f"Error cloning Git repository: {msg}"
|
||||
) from e
|
||||
|
||||
def _scan_files(self):
|
||||
"""Search for Jinja2 (.j2) files to apply to the destination."""
|
||||
template_files = []
|
||||
|
||||
if os.path.isdir(self.path):
|
||||
self.log.info("Lookup template files", src=self.src)
|
||||
else:
|
||||
sysexit_with_message("Can not open template directory", path=self.path)
|
||||
|
||||
for file in glob.iglob(self.path + "/**/*.j2", recursive=True):
|
||||
relative_file = file[len(self.path) + 1 :]
|
||||
if ntpath.basename(file)[:1] != "_":
|
||||
self.log.debug("Found template file", path=relative_file)
|
||||
template_files.append(relative_file)
|
||||
else:
|
||||
self.log.debug("Skipped template file", path=relative_file)
|
||||
|
||||
return template_files
|
||||
|
||||
@staticmethod
|
||||
def _cleanup_temp_dir(temp_dir):
|
||||
if temp_dir and os.path.exists(temp_dir):
|
||||
shutil.rmtree(temp_dir)
|
|
@ -2,10 +2,12 @@
|
|||
{% set var = role.var | default({}) %}
|
||||
{% if var %}
|
||||
- [Default Variables](#default-variables)
|
||||
{% if not options.tabulate_vars %}
|
||||
{% for key, item in var | dictsort %}
|
||||
- [{{ key }}](#{{ key }})
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
{% endif %}
|
||||
{% if tag %}
|
||||
- [Discovered Tags](#discovered-tags)
|
||||
{% endif %}
|
||||
|
|
49
ansibledoctor/templates/hugo-book/_vars_tabulated.j2
Normal file
49
ansibledoctor/templates/hugo-book/_vars_tabulated.j2
Normal file
|
@ -0,0 +1,49 @@
|
|||
{% set var = role.var | default({}) %}
|
||||
{% if var %}
|
||||
## Default Variables
|
||||
|
||||
{% set columns = ["variable", "default", "description", "type", "deprecated", "example"] %}
|
||||
{% set found_columns = ["variable", "default"] + var.values() | map("list") | sum(start=["key"]) | unique | list %}
|
||||
{% for c in columns %}
|
||||
{% if c in found_columns %}
|
||||
|{{ c | capitalize -}}
|
||||
{% endif %}
|
||||
{% endfor %}
|
||||
|
|
||||
{% for c in columns %}
|
||||
{% if c in found_columns %}
|
||||
|{{ "-" * (c | length) -}}
|
||||
{% endif %}
|
||||
{% endfor %}
|
||||
|
|
||||
{% for key, item in var | dictsort %}
|
||||
|{{ key -}}
|
||||
|{{ (item.value | default({}))[key] | default -}}
|
||||
{% if "description" in found_columns %}
|
||||
|{{ item.description | default([]) | safe_join("<br />") | replace("\n", "<br />") | replace("|", "\|") -}}
|
||||
{% endif %}
|
||||
{% if "type" in found_columns %}
|
||||
|{{ item.type | default([]) | join("<br />") -}}
|
||||
{% endif %}
|
||||
{% if "deprecated" in found_columns %}
|
||||
|
|
||||
{%- if "deprecated" in found_columns %}
|
||||
{% if item.deprecated is defined %}
|
||||
{% set deprecated = [item.deprecated] if item.deprecated is string else item.deprecated %}
|
||||
{% set deprecated_string = deprecated | map("replace", "\n", "<br />") | safe_join("<br />") %}
|
||||
{% if deprecated_string -%}
|
||||
{{ deprecated_string }}
|
||||
{%- else -%}
|
||||
True
|
||||
{%- endif %}
|
||||
{%- else -%}
|
||||
False
|
||||
{%- endif %}
|
||||
{% endif %}
|
||||
{% endif %}
|
||||
{% if "example" in found_columns %}
|
||||
|{{ item.example | default([]) | safe_join("<br />") | replace("\n", "<br />") | replace("|", "\|") -}}
|
||||
{% endif %}
|
||||
|
|
||||
{% endfor %}
|
||||
{% endif %}
|
|
@ -23,7 +23,11 @@ summary: {{ meta.summary.value | safe_join(" ") }}
|
|||
{% include '_requirements.j2' %}
|
||||
|
||||
{# Vars #}
|
||||
{% if options.tabulate_vars %}
|
||||
{% include '_vars_tabulated.j2' %}
|
||||
{% else %}
|
||||
{% include '_vars.j2' %}
|
||||
{% endif %}
|
||||
|
||||
{# Tag #}
|
||||
{% include '_tag.j2' %}
|
||||
|
|
|
@ -15,7 +15,11 @@
|
|||
{% include '_requirements.j2' %}
|
||||
|
||||
{# Vars #}
|
||||
{% if options.tabulate_vars %}
|
||||
{% include '_vars_tabulated.j2' %}
|
||||
{% else %}
|
||||
{% include '_vars.j2' %}
|
||||
{% endif %}
|
||||
|
||||
{# Tag #}
|
||||
{% include '_tag.j2' %}
|
||||
|
|
|
@ -4,10 +4,12 @@
|
|||
{% set var = role.var | default({}) %}
|
||||
{% if var %}
|
||||
- [Default Variables](#default-variables)
|
||||
{% if not options.tabulate_vars %}
|
||||
{% for key, item in var | dictsort %}
|
||||
- [{{ key }}](#{{ key }})
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
{% endif %}
|
||||
{% if tag %}
|
||||
- [Discovered Tags](#discovered-tags)
|
||||
{% endif %}
|
||||
|
|
49
ansibledoctor/templates/readme/_vars_tabulated.j2
Normal file
49
ansibledoctor/templates/readme/_vars_tabulated.j2
Normal file
|
@ -0,0 +1,49 @@
|
|||
{% set var = role.var | default({}) %}
|
||||
{% if var %}
|
||||
## Default Variables
|
||||
|
||||
{% set columns = ["variable", "default", "description", "type", "deprecated", "example"] %}
|
||||
{% set found_columns = ["variable", "default"] + var.values() | map("list") | sum(start=["key"]) | unique | list %}
|
||||
{% for c in columns %}
|
||||
{% if c in found_columns %}
|
||||
|{{ c | capitalize -}}
|
||||
{% endif %}
|
||||
{% endfor %}
|
||||
|
|
||||
{% for c in columns %}
|
||||
{% if c in found_columns %}
|
||||
|{{ "-" * (c | length) -}}
|
||||
{% endif %}
|
||||
{% endfor %}
|
||||
|
|
||||
{% for key, item in var | dictsort %}
|
||||
|{{ key -}}
|
||||
|{{ (item.value | default({}))[key] | default -}}
|
||||
{% if "description" in found_columns %}
|
||||
|{{ item.description | default([]) | safe_join("<br />") | replace("\n", "<br />") | replace("|", "\|") -}}
|
||||
{% endif %}
|
||||
{% if "type" in found_columns %}
|
||||
|{{ item.type | default([]) | join("<br />") -}}
|
||||
{% endif %}
|
||||
{% if "deprecated" in found_columns %}
|
||||
|
|
||||
{%- if "deprecated" in found_columns %}
|
||||
{% if item.deprecated is defined %}
|
||||
{% set deprecated = [item.deprecated] if item.deprecated is string else item.deprecated %}
|
||||
{% set deprecated_string = deprecated | map("replace", "\n", "<br />") | safe_join("<br />") %}
|
||||
{% if deprecated_string -%}
|
||||
{{ deprecated_string }}
|
||||
{%- else -%}
|
||||
True
|
||||
{%- endif %}
|
||||
{%- else -%}
|
||||
False
|
||||
{%- endif %}
|
||||
{% endif %}
|
||||
{% endif %}
|
||||
{% if "example" in found_columns %}
|
||||
|{{ item.example | default([]) | safe_join("<br />") | replace("\n", "<br />") | replace("|", "\|") -}}
|
||||
{% endif %}
|
||||
|
|
||||
{% endfor %}
|
||||
{% endif %}
|
|
@ -1,339 +0,0 @@
|
|||
#!/usr/bin/env python3
|
||||
"""Global utility methods and classes."""
|
||||
|
||||
import logging
|
||||
import os
|
||||
import sys
|
||||
from collections.abc import Iterable
|
||||
|
||||
import colorama
|
||||
from pythonjsonlogger import jsonlogger
|
||||
|
||||
import ansibledoctor.exception
|
||||
|
||||
CONSOLE_FORMAT = "{}{}[%(levelname)s]{} %(message)s"
|
||||
JSON_FORMAT = "%(asctime)s %(levelname)s %(message)s"
|
||||
|
||||
|
||||
def strtobool(value):
|
||||
"""Convert a string representation of truth to true or false."""
|
||||
|
||||
_map = {
|
||||
"y": True,
|
||||
"yes": True,
|
||||
"t": True,
|
||||
"true": True,
|
||||
"on": True,
|
||||
"1": True,
|
||||
"n": False,
|
||||
"no": False,
|
||||
"f": False,
|
||||
"false": False,
|
||||
"off": False,
|
||||
"0": False
|
||||
}
|
||||
|
||||
try:
|
||||
return _map[str(value).lower()]
|
||||
except KeyError as err:
|
||||
raise ValueError(f'"{value}" is not a valid bool value') from err
|
||||
|
||||
|
||||
def to_bool(string):
|
||||
return bool(strtobool(str(string)))
|
||||
|
||||
|
||||
def flatten(items):
|
||||
for x in items:
|
||||
if isinstance(x, Iterable) and not isinstance(x, (str, bytes)):
|
||||
yield from flatten(x)
|
||||
else:
|
||||
yield x
|
||||
|
||||
|
||||
def _should_do_markup():
|
||||
py_colors = os.environ.get("PY_COLORS", None)
|
||||
if py_colors is not None:
|
||||
return to_bool(py_colors)
|
||||
|
||||
return sys.stdout.isatty() and os.environ.get("TERM") != "dumb"
|
||||
|
||||
|
||||
def _split_string(string, delimiter, escape, maxsplit=None):
|
||||
result = []
|
||||
current_element = []
|
||||
iterator = iter(string)
|
||||
count_split = 0
|
||||
skip_split = False
|
||||
|
||||
for character in iterator:
|
||||
if maxsplit and count_split >= maxsplit:
|
||||
skip_split = True
|
||||
|
||||
if character == escape and not skip_split:
|
||||
try:
|
||||
next_character = next(iterator)
|
||||
if next_character != delimiter and next_character != escape:
|
||||
# Do not copy the escape character if it is intended to escape either the
|
||||
# delimiter or the escape character itself. Copy the escape character
|
||||
# if it is not used to escape either of these characters.
|
||||
current_element.append(escape)
|
||||
current_element.append(next_character)
|
||||
count_split += 1
|
||||
except StopIteration:
|
||||
current_element.append(escape)
|
||||
elif character == delimiter and not skip_split:
|
||||
result.append("".join(current_element))
|
||||
current_element = []
|
||||
count_split += 1
|
||||
else:
|
||||
current_element.append(character)
|
||||
|
||||
result.append("".join(current_element))
|
||||
return result
|
||||
|
||||
|
||||
colorama.init(autoreset=True, strip=not _should_do_markup())
|
||||
|
||||
|
||||
class Singleton(type):
|
||||
"""Meta singleton class."""
|
||||
|
||||
_instances = {}
|
||||
|
||||
def __call__(cls, *args, **kwargs):
|
||||
if cls not in cls._instances:
|
||||
cls._instances[cls] = super().__call__(*args, **kwargs)
|
||||
return cls._instances[cls]
|
||||
|
||||
|
||||
class LogFilter:
|
||||
"""Exclude log messages above the logged level."""
|
||||
|
||||
def __init__(self, level):
|
||||
"""
|
||||
Initialize a new custom log filter.
|
||||
|
||||
:param level: Log level limit
|
||||
:returns: None
|
||||
|
||||
"""
|
||||
self.__level = level
|
||||
|
||||
def filter(self, logRecord): # noqa
|
||||
# https://docs.python.org/3/library/logging.html#logrecord-attributes
|
||||
return logRecord.levelno <= self.__level
|
||||
|
||||
|
||||
class MultilineFormatter(logging.Formatter):
|
||||
"""Reset color after newline characters."""
|
||||
|
||||
def format(self, record): # noqa
|
||||
record.msg = record.msg.replace("\n", f"\n{colorama.Style.RESET_ALL}... ")
|
||||
return logging.Formatter.format(self, record)
|
||||
|
||||
|
||||
class MultilineJsonFormatter(jsonlogger.JsonFormatter):
|
||||
"""Remove newline characters."""
|
||||
|
||||
def format(self, record): # noqa
|
||||
record.msg = record.msg.replace("\n", " ")
|
||||
return jsonlogger.JsonFormatter.format(self, record)
|
||||
|
||||
|
||||
class Log:
|
||||
"""Handle logging."""
|
||||
|
||||
def __init__(self, level=logging.WARNING, name="ansibledoctor", json=False):
|
||||
self.logger = logging.getLogger(name)
|
||||
self.logger.setLevel(level)
|
||||
self.logger.addHandler(self._get_error_handler(json=json))
|
||||
self.logger.addHandler(self._get_warning_handler(json=json))
|
||||
self.logger.addHandler(self._get_info_handler(json=json))
|
||||
self.logger.addHandler(self._get_critical_handler(json=json))
|
||||
self.logger.addHandler(self._get_debug_handler(json=json))
|
||||
self.logger.propagate = False
|
||||
|
||||
def _get_error_handler(self, json=False):
|
||||
handler = logging.StreamHandler(sys.stderr)
|
||||
handler.setLevel(logging.ERROR)
|
||||
handler.addFilter(LogFilter(logging.ERROR))
|
||||
handler.setFormatter(
|
||||
MultilineFormatter(
|
||||
self.error(
|
||||
CONSOLE_FORMAT.format(
|
||||
colorama.Fore.RED, colorama.Style.BRIGHT, colorama.Style.RESET_ALL
|
||||
)
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
if json:
|
||||
handler.setFormatter(MultilineJsonFormatter(JSON_FORMAT))
|
||||
|
||||
return handler
|
||||
|
||||
def _get_warning_handler(self, json=False):
|
||||
handler = logging.StreamHandler(sys.stdout)
|
||||
handler.setLevel(logging.WARNING)
|
||||
handler.addFilter(LogFilter(logging.WARNING))
|
||||
handler.setFormatter(
|
||||
MultilineFormatter(
|
||||
self.warning(
|
||||
CONSOLE_FORMAT.format(
|
||||
colorama.Fore.YELLOW, colorama.Style.BRIGHT, colorama.Style.RESET_ALL
|
||||
)
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
if json:
|
||||
handler.setFormatter(MultilineJsonFormatter(JSON_FORMAT))
|
||||
|
||||
return handler
|
||||
|
||||
def _get_info_handler(self, json=False):
|
||||
handler = logging.StreamHandler(sys.stdout)
|
||||
handler.setLevel(logging.INFO)
|
||||
handler.addFilter(LogFilter(logging.INFO))
|
||||
handler.setFormatter(
|
||||
MultilineFormatter(
|
||||
self.info(
|
||||
CONSOLE_FORMAT.format(
|
||||
colorama.Fore.CYAN, colorama.Style.BRIGHT, colorama.Style.RESET_ALL
|
||||
)
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
if json:
|
||||
handler.setFormatter(MultilineJsonFormatter(JSON_FORMAT))
|
||||
|
||||
return handler
|
||||
|
||||
def _get_critical_handler(self, json=False):
|
||||
handler = logging.StreamHandler(sys.stderr)
|
||||
handler.setLevel(logging.CRITICAL)
|
||||
handler.addFilter(LogFilter(logging.CRITICAL))
|
||||
handler.setFormatter(
|
||||
MultilineFormatter(
|
||||
self.critical(
|
||||
CONSOLE_FORMAT.format(
|
||||
colorama.Fore.RED, colorama.Style.BRIGHT, colorama.Style.RESET_ALL
|
||||
)
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
if json:
|
||||
handler.setFormatter(MultilineJsonFormatter(JSON_FORMAT))
|
||||
|
||||
return handler
|
||||
|
||||
def _get_debug_handler(self, json=False):
|
||||
handler = logging.StreamHandler(sys.stderr)
|
||||
handler.setLevel(logging.DEBUG)
|
||||
handler.addFilter(LogFilter(logging.DEBUG))
|
||||
handler.setFormatter(
|
||||
MultilineFormatter(
|
||||
self.critical(
|
||||
CONSOLE_FORMAT.format(
|
||||
colorama.Fore.BLUE, colorama.Style.BRIGHT, colorama.Style.RESET_ALL
|
||||
)
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
if json:
|
||||
handler.setFormatter(MultilineJsonFormatter(JSON_FORMAT))
|
||||
|
||||
return handler
|
||||
|
||||
def set_level(self, s):
|
||||
self.logger.setLevel(s)
|
||||
|
||||
def debug(self, msg):
|
||||
"""Format info messages and return string."""
|
||||
return msg
|
||||
|
||||
def critical(self, msg):
|
||||
"""Format critical messages and return string."""
|
||||
return msg
|
||||
|
||||
def error(self, msg):
|
||||
"""Format error messages and return string."""
|
||||
return msg
|
||||
|
||||
def warning(self, msg):
|
||||
"""Format warning messages and return string."""
|
||||
return msg
|
||||
|
||||
def info(self, msg):
|
||||
"""Format info messages and return string."""
|
||||
return msg
|
||||
|
||||
def _color_text(self, color, msg):
|
||||
"""
|
||||
Colorize strings.
|
||||
|
||||
:param color: colorama color settings
|
||||
:param msg: string to colorize
|
||||
:returns: string
|
||||
|
||||
"""
|
||||
return f"{color}{msg}{colorama.Style.RESET_ALL}"
|
||||
|
||||
def sysexit(self, code=1):
|
||||
sys.exit(code)
|
||||
|
||||
def sysexit_with_message(self, msg, code=1):
|
||||
self.logger.critical(str(msg))
|
||||
self.sysexit(code)
|
||||
|
||||
|
||||
class SingleLog(Log, metaclass=Singleton):
|
||||
"""Singleton logging class."""
|
||||
|
||||
pass
|
||||
|
||||
|
||||
class UnsafeTag:
|
||||
"""Handle custom yaml unsafe tag."""
|
||||
|
||||
yaml_tag = "!unsafe"
|
||||
|
||||
def __init__(self, value):
|
||||
self.unsafe = value
|
||||
|
||||
@staticmethod
|
||||
def yaml_constructor(loader, node):
|
||||
return loader.construct_scalar(node)
|
||||
|
||||
|
||||
class FileUtils:
|
||||
"""Mics static methods for file handling."""
|
||||
|
||||
@staticmethod
|
||||
def create_path(path):
|
||||
os.makedirs(path, exist_ok=True)
|
||||
|
||||
@staticmethod
|
||||
def query_yes_no(question, default=True):
|
||||
"""
|
||||
Ask a yes/no question via input() and return their answer.
|
||||
|
||||
"question" is a string that is presented to the user.
|
||||
"default" is the presumed answer if the user just hits <Enter>.
|
||||
It must be "yes" (the default), "no" or None (meaning
|
||||
an answer is required of the user).
|
||||
|
||||
The "answer" return value is one of "yes" or "no".
|
||||
"""
|
||||
prompt = "[Y/n]" if default else "[N/y]"
|
||||
|
||||
try:
|
||||
# input method is safe in python3
|
||||
choice = input(f"{question} {prompt} ") or default # nosec
|
||||
return to_bool(choice)
|
||||
except (KeyboardInterrupt, ValueError) as e:
|
||||
raise ansibledoctor.exception.InputError("Error while reading input", e) from e
|
130
ansibledoctor/utils/__init__.py
Normal file
130
ansibledoctor/utils/__init__.py
Normal file
|
@ -0,0 +1,130 @@
|
|||
#!/usr/bin/env python3
|
||||
"""Global utility methods and classes."""
|
||||
|
||||
import os
|
||||
import sys
|
||||
from collections.abc import Iterable
|
||||
|
||||
import structlog
|
||||
|
||||
|
||||
def strtobool(value):
|
||||
"""Convert a string representation of truth to true or false."""
|
||||
|
||||
_map = {
|
||||
"y": True,
|
||||
"yes": True,
|
||||
"t": True,
|
||||
"true": True,
|
||||
"on": True,
|
||||
"1": True,
|
||||
"n": False,
|
||||
"no": False,
|
||||
"f": False,
|
||||
"false": False,
|
||||
"off": False,
|
||||
"0": False,
|
||||
}
|
||||
|
||||
try:
|
||||
return _map[str(value).lower()]
|
||||
except KeyError as err:
|
||||
raise ValueError(f'"{value}" is not a valid bool value') from err
|
||||
|
||||
|
||||
def to_bool(string):
|
||||
return bool(strtobool(str(string)))
|
||||
|
||||
|
||||
def flatten(items):
|
||||
for x in items:
|
||||
if isinstance(x, Iterable) and not isinstance(x, (str, bytes)):
|
||||
yield from flatten(x)
|
||||
else:
|
||||
yield x
|
||||
|
||||
|
||||
def _split_string(string, delimiter, escape, maxsplit=None):
|
||||
result = []
|
||||
current_element = []
|
||||
iterator = iter(string)
|
||||
count_split = 0
|
||||
skip_split = False
|
||||
|
||||
for character in iterator:
|
||||
if maxsplit and count_split >= maxsplit:
|
||||
skip_split = True
|
||||
|
||||
if character == escape and not skip_split:
|
||||
try:
|
||||
next_character = next(iterator)
|
||||
if next_character != delimiter and next_character != escape:
|
||||
# Do not copy the escape character if it is intended to escape either the
|
||||
# delimiter or the escape character itself. Copy the escape character
|
||||
# if it is not used to escape either of these characters.
|
||||
current_element.append(escape)
|
||||
current_element.append(next_character)
|
||||
count_split += 1
|
||||
except StopIteration:
|
||||
current_element.append(escape)
|
||||
elif character == delimiter and not skip_split:
|
||||
result.append("".join(current_element))
|
||||
current_element = []
|
||||
count_split += 1
|
||||
else:
|
||||
current_element.append(character)
|
||||
|
||||
result.append("".join(current_element))
|
||||
return result
|
||||
|
||||
|
||||
def sysexit(code=1):
|
||||
sys.exit(code)
|
||||
|
||||
|
||||
def sysexit_with_message(msg, code=1, **kwargs):
|
||||
structlog.get_logger().critical(str(msg).strip(), **kwargs)
|
||||
sysexit(code)
|
||||
|
||||
|
||||
class Singleton(type):
|
||||
"""Meta singleton class."""
|
||||
|
||||
_instances = {}
|
||||
|
||||
def __call__(cls, *args, **kwargs):
|
||||
if cls not in cls._instances:
|
||||
cls._instances[cls] = super().__call__(*args, **kwargs)
|
||||
return cls._instances[cls]
|
||||
|
||||
|
||||
class FileUtils:
|
||||
"""Mics static methods for file handling."""
|
||||
|
||||
@staticmethod
|
||||
def create_path(path):
|
||||
os.makedirs(path, exist_ok=True)
|
||||
|
||||
@staticmethod
|
||||
def query_yes_no(question, default=True):
|
||||
"""
|
||||
Ask a yes/no question via input() and return their answer.
|
||||
|
||||
"question" is a string that is presented to the user.
|
||||
"default" is the presumed answer if the user just hits <Enter>.
|
||||
It must be "yes" (the default), "no" or None (meaning
|
||||
an answer is required of the user).
|
||||
|
||||
The "answer" return value is one of "yes" or "no".
|
||||
"""
|
||||
prompt = "[Y/n]" if default else "[N/y]"
|
||||
|
||||
while True:
|
||||
try:
|
||||
# input method is safe in python3
|
||||
choice = input(f"{question} {prompt} ") or default # nosec
|
||||
return to_bool(choice)
|
||||
except ValueError:
|
||||
print("Invalid input. Please enter 'y' or 'n'.") # noqa: T201
|
||||
except KeyboardInterrupt as e:
|
||||
raise e
|
78
ansibledoctor/utils/yamlhelper.py
Normal file
78
ansibledoctor/utils/yamlhelper.py
Normal file
|
@ -0,0 +1,78 @@
|
|||
"""Utils for YAML file operations."""
|
||||
|
||||
from collections import defaultdict
|
||||
from contextlib import suppress
|
||||
|
||||
import ruamel.yaml
|
||||
import yaml
|
||||
from ansible.parsing.yaml.loader import AnsibleLoader
|
||||
|
||||
import ansibledoctor.exception
|
||||
|
||||
|
||||
class UnsafeTag:
|
||||
"""Handle custom yaml unsafe tag."""
|
||||
|
||||
yaml_tag = "!unsafe"
|
||||
|
||||
def __init__(self, value):
|
||||
self.unsafe = value
|
||||
|
||||
@staticmethod
|
||||
def yaml_constructor(loader, node):
|
||||
return loader.construct_scalar(node)
|
||||
|
||||
|
||||
def parse_yaml_ansible(yamlfile):
|
||||
try:
|
||||
loader = AnsibleLoader(yamlfile)
|
||||
data = loader.get_single_data() or []
|
||||
except (
|
||||
yaml.parser.ParserError,
|
||||
yaml.scanner.ScannerError,
|
||||
yaml.constructor.ConstructorError,
|
||||
yaml.composer.ComposerError,
|
||||
) as e:
|
||||
raise ansibledoctor.exception.YAMLError(e) from e
|
||||
|
||||
return data
|
||||
|
||||
|
||||
def parse_yaml(yamlfile):
|
||||
try:
|
||||
ruamel.yaml.add_constructor(
|
||||
UnsafeTag.yaml_tag,
|
||||
UnsafeTag.yaml_constructor,
|
||||
constructor=ruamel.yaml.SafeConstructor,
|
||||
)
|
||||
|
||||
data = ruamel.yaml.YAML(typ="rt").load(yamlfile)
|
||||
_yaml_remove_comments(data)
|
||||
data = defaultdict(dict, data or {})
|
||||
except (
|
||||
ruamel.yaml.parser.ParserError,
|
||||
ruamel.yaml.scanner.ScannerError,
|
||||
ruamel.yaml.constructor.ConstructorError,
|
||||
ruamel.yaml.composer.ComposerError,
|
||||
) as e:
|
||||
raise ansibledoctor.exception.YAMLError(e) from e
|
||||
|
||||
return data
|
||||
|
||||
|
||||
def _yaml_remove_comments(d):
|
||||
if isinstance(d, dict):
|
||||
for k, v in d.items():
|
||||
_yaml_remove_comments(k)
|
||||
_yaml_remove_comments(v)
|
||||
elif isinstance(d, list):
|
||||
for elem in d:
|
||||
_yaml_remove_comments(elem)
|
||||
|
||||
with suppress(AttributeError):
|
||||
attr = (
|
||||
"comment"
|
||||
if isinstance(d, ruamel.yaml.scalarstring.ScalarString)
|
||||
else ruamel.yaml.comments.Comment.attrib
|
||||
)
|
||||
delattr(d, attr)
|
|
@ -4,11 +4,12 @@ title: Using pip
|
|||
|
||||
```Shell
|
||||
# From PyPI as unprivileged user
|
||||
$ pip install ansible-doctor --user
|
||||
$ pip install ansible-doctor[ansible-core] --user
|
||||
|
||||
# .. or as root
|
||||
$ sudo pip install ansible-doctor
|
||||
$ sudo pip install ansible-doctor[ansible-core]
|
||||
|
||||
# From Wheel file
|
||||
$ pip install https://github.com/thegeeklab/ansible-doctor/releases/download/v0.1.1/ansible_doctor-0.1.1-py2.py3-none-any.whl
|
||||
# Please check first whether a newer version is available.
|
||||
$ pip install https://github.com/thegeeklab/ansible-doctor/releases/download/v3.1.4/ansible_doctor-3.1.4-py2.py3-none-any.whl[ansible-core]
|
||||
```
|
||||
|
|
|
@ -20,36 +20,17 @@ Configuration options can be set in different places, which are processed in the
|
|||
---
|
||||
# Default is the current working directory.
|
||||
base_dir:
|
||||
# Default is the basename of 'role_name'.
|
||||
role_name:
|
||||
# Auto-detect if the given directory is a role, can be disabled
|
||||
# to parse loose files instead.
|
||||
role_detection: True
|
||||
# Don't write anything to file system
|
||||
|
||||
role:
|
||||
# Default is the basename of 'role_name'.
|
||||
name:
|
||||
# Auto-detect if the given directory is a role, can be disabled
|
||||
# to parse loose files instead.
|
||||
autodetect: True
|
||||
|
||||
# Don't write anything to file system.
|
||||
dry_run: False
|
||||
|
||||
logging:
|
||||
# Possible options debug | info | warning | error | critical
|
||||
level: "warning"
|
||||
# Json logging can be enabled if a parsable output is required
|
||||
json: False
|
||||
|
||||
# Path to write rendered template file. Default is the current working directory.
|
||||
output_dir:
|
||||
# Default is built-in templates directory.
|
||||
template_dir:
|
||||
template: readme
|
||||
# By default, double spaces, spaces before and after line breaks or tab characters, etc.
|
||||
# are automatically removed before the template is rendered. As a result, indenting
|
||||
# with spaces does not work. If you want to use spaces to indent text, you must disable
|
||||
# this option.
|
||||
template_autotrim: True
|
||||
|
||||
# Don't ask to overwrite if output file exists.
|
||||
force_overwrite: False
|
||||
# Load custom header from given file and append template output to it before write.
|
||||
custom_header: ""
|
||||
|
||||
exclude_files: []
|
||||
# Examples
|
||||
# exclude_files:
|
||||
|
@ -59,6 +40,62 @@ exclude_files: []
|
|||
# Exclude tags from automatic detection. Configured tags are only skipped
|
||||
# if the tag is not used in an annotation.
|
||||
exclude_tags: []
|
||||
|
||||
logging:
|
||||
# Possible options: debug|info|warning| error|critical
|
||||
level: "warning"
|
||||
# JSON logging can be enabled if a parsable output is required.
|
||||
json: False
|
||||
|
||||
template:
|
||||
# Name of the template to be used. In most cases, this is the name of a directory that is attached to the
|
||||
# the `src` path or Git repo (see example below).
|
||||
name: readme
|
||||
|
||||
# Template provider source. Currently supported providers are `local|git`.
|
||||
# The `local` provider loads templates from the local file system. This provider
|
||||
# is used by default and uses the built-in templates.
|
||||
#
|
||||
# Examples:
|
||||
# template:
|
||||
# name: readme
|
||||
# src: local>/tmp/custom_templates/
|
||||
#
|
||||
# The `git` provider allows templates to be loaded from a git repository. At the moment
|
||||
# the functions of this provider are limited and only public repositories are supported.
|
||||
#
|
||||
# Examples:
|
||||
# template:
|
||||
# src: git>https://github.com/thegeeklab/ansible-doctor
|
||||
# name: ansibledoctor/templates/readme
|
||||
#
|
||||
# template:
|
||||
# src: git>git@github.com:thegeeklab/ansible-doctor.git
|
||||
# name: ansibledoctor/templates/readme
|
||||
#
|
||||
# template:
|
||||
# src: git>git@github.com:thegeeklab/ansible-doctor.git#branch-or-tag
|
||||
# name: ansibledoctor/templates/readme
|
||||
src:
|
||||
|
||||
options:
|
||||
# Configures whether to tabulate variables in the output. When set to `True`,
|
||||
# variables will be displayed in a tabular format intsead of plain marktdown sections.
|
||||
# NOTE: This option does not support rendering multiline code blocks.
|
||||
tabulate_vars: False
|
||||
|
||||
renderer:
|
||||
# By default, double spaces, spaces before and after line breaks or tab characters, etc.
|
||||
# are automatically removed before the template is rendered. As a result, indenting
|
||||
# with spaces does not work. If you want to use spaces to indent text, you must disable
|
||||
# this option.
|
||||
autotrim: True
|
||||
# Load custom header from given file and append template output to it before write.
|
||||
include_header: ""
|
||||
# Path to write rendered template file. Default is the current working directory.
|
||||
dest:
|
||||
# Don't ask to overwrite if output file exists.
|
||||
force_overwrite: False
|
||||
```
|
||||
|
||||
## CLI
|
||||
|
@ -91,22 +128,25 @@ options:
|
|||
## Environment Variables
|
||||
|
||||
```Shell
|
||||
ANSIBLE_DOCTOR_CONFIG_FILE=
|
||||
ANSIBLE_DOCTOR_ROLE_DETECTION=true
|
||||
ANSIBLE_DOCTOR_BASE_DIR=
|
||||
ANSIBLE_DOCTOR_RECURSIVE=false
|
||||
ANSIBLE_DOCTOR_ROLE_NAME=
|
||||
ANSIBLE_DOCTOR_DRY_RUN=false
|
||||
ANSIBLE_DOCTOR_LOG_LEVEL=warning
|
||||
ANSIBLE_DOCTOR_LOG_JSON=false
|
||||
ANSIBLE_DOCTOR_OUTPUT_DIR=
|
||||
ANSIBLE_DOCTOR_TEMPLATE_DIR=
|
||||
ANSIBLE_DOCTOR_TEMPLATE=readme
|
||||
ANSIBLE_DOCTOR_TEMPLATE_AUTOTRIM=true
|
||||
ANSIBLE_DOCTOR_FORCE_OVERWRITE=false
|
||||
ANSIBLE_DOCTOR_CUSTOM_HEADER=
|
||||
ANSIBLE_DOCTOR_DRY_RUN=False
|
||||
ANSIBLE_DOCTOR_EXCLUDE_FILES=
|
||||
ANSIBLE_DOCTOR_EXCLUDE_FILES=molecule/,files/**/*.py
|
||||
ANSIBLE_DOCTOR_EXCLUDE_TAGS=
|
||||
|
||||
ANSIBLE_DOCTOR_ROLE__NAME=
|
||||
ANSIBLE_DOCTOR_ROLE__AUTODETECT=True
|
||||
|
||||
ANSIBLE_DOCTOR_LOGGING__LEVEL="warning"
|
||||
ANSIBLE_DOCTOR_LOGGING__JSON=False
|
||||
|
||||
ANSIBLE_DOCTOR_TEMPLATE__NAME=readme
|
||||
ANSIBLE_DOCTOR_TEMPLATE__SRC=
|
||||
ANSIBLE_DOCTOR_TEMPLATE__OPTIONS__TABULATE_VARS=False
|
||||
|
||||
ANSIBLE_DOCTOR_RENDERER__AUTOTRIM=True
|
||||
ANSIBLE_DOCTOR_RENDERER__INCLUDE_HEADER=
|
||||
ANSIBLE_DOCTOR_RENDERER__DEST=
|
||||
ANSIBLE_DOCTOR_RENDERER__FORCE_OVERWRITE=False
|
||||
```
|
||||
|
||||
## Pre-Commit setup
|
||||
|
@ -119,8 +159,8 @@ To use _ansible-doctor_ with the [pre-commit](https://pre-commit.com/) framework
|
|||
|
||||
{{< highlight yaml "linenos=table" >}}
|
||||
- repo: https://github.com/thegeeklab/ansible-doctor
|
||||
# change ref to the latest release from https://github.com/thegeeklab/ansible-doctor/releases
|
||||
rev: v1.4.8
|
||||
# update version with `pre-commit autoupdate`
|
||||
rev: v4.0.4
|
||||
hooks:
|
||||
- id: ansible-doctor
|
||||
{{< /highlight >}}
|
||||
|
|
|
@ -1,5 +1,10 @@
|
|||
---
|
||||
custom_header: HEADER.md
|
||||
logging:
|
||||
level: debug
|
||||
template: readme
|
||||
|
||||
template:
|
||||
src: git>https://github.com/thegeeklab/ansible-doctor
|
||||
name: ansibledoctor/templates/readme
|
||||
|
||||
renderer:
|
||||
include_header: HEADER.md
|
||||
|
|
|
@ -204,7 +204,6 @@ demo_role_unset: some_value
|
|||
|
||||
## Dependencies
|
||||
|
||||
- role1
|
||||
- role2
|
||||
|
||||
## License
|
||||
|
|
|
@ -18,6 +18,5 @@ galaxy_info:
|
|||
- documentation
|
||||
|
||||
dependencies:
|
||||
- role1
|
||||
- role: role2
|
||||
- name: namespace.role3
|
||||
|
|
|
@ -11,6 +11,8 @@
|
|||
- name: Demo task with a tag list
|
||||
debug:
|
||||
msg: "Demo message"
|
||||
tags:
|
||||
- module-tag
|
||||
tags:
|
||||
- role-tag1
|
||||
- role-tag2
|
||||
|
|
|
@ -1,5 +1,9 @@
|
|||
---
|
||||
custom_header: HEADER.md
|
||||
logging:
|
||||
level: debug
|
||||
template: readme
|
||||
|
||||
template:
|
||||
name: readme
|
||||
|
||||
renderer:
|
||||
include_header: HEADER.md
|
||||
|
|
1019
poetry.lock
generated
1019
poetry.lock
generated
File diff suppressed because it is too large
Load Diff
138
pyproject.toml
138
pyproject.toml
|
@ -10,7 +10,6 @@ classifiers = [
|
|||
"Natural Language :: English",
|
||||
"Operating System :: POSIX",
|
||||
"Programming Language :: Python :: 3",
|
||||
"Programming Language :: Python :: 3.8",
|
||||
"Programming Language :: Python :: 3.9",
|
||||
"Programming Language :: Python :: 3.10",
|
||||
"Programming Language :: Python :: 3.11",
|
||||
|
@ -22,42 +21,42 @@ classifiers = [
|
|||
description = "Generate documentation from annotated Ansible roles using templates."
|
||||
documentation = "https://ansible-doctor.geekdocs.de/"
|
||||
homepage = "https://ansible-doctor.geekdocs.de/"
|
||||
include = [
|
||||
"LICENSE",
|
||||
]
|
||||
include = ["LICENSE"]
|
||||
keywords = ["ansible", "role", "documentation"]
|
||||
license = "GPL-3.0-only"
|
||||
name = "ansible-doctor"
|
||||
packages = [
|
||||
{include = "ansibledoctor"},
|
||||
]
|
||||
packages = [{ include = "ansibledoctor" }]
|
||||
readme = "README.md"
|
||||
repository = "https://github.com/thegeeklab/ansible-doctor/"
|
||||
version = "0.0.0"
|
||||
|
||||
[tool.poetry.dependencies]
|
||||
Jinja2 = "3.1.2"
|
||||
anyconfig = "0.13.0"
|
||||
Jinja2 = "3.1.4"
|
||||
anyconfig = "0.14.0"
|
||||
appdirs = "1.4.4"
|
||||
colorama = "0.4.6"
|
||||
environs = "9.5.0"
|
||||
jsonschema = "4.19.2"
|
||||
nested-lookup = "0.2.25"
|
||||
pathspec = "0.11.2"
|
||||
python = "^3.8.0"
|
||||
python-json-logger = "2.0.7"
|
||||
"ruamel.yaml" = "0.18.5"
|
||||
pathspec = "0.12.1"
|
||||
python = "^3.9.0"
|
||||
"ruamel.yaml" = "0.18.6"
|
||||
dynaconf = "3.2.5"
|
||||
gitpython = "3.1.43"
|
||||
ansible-core = { version = "2.14.17", optional = true }
|
||||
structlog = "24.2.0"
|
||||
|
||||
[tool.poetry.extras]
|
||||
ansible-core = ["ansible-core"]
|
||||
|
||||
[tool.poetry.scripts]
|
||||
ansible-doctor = "ansibledoctor.cli:main"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
ruff = "0.1.3"
|
||||
pytest = "7.4.3"
|
||||
pytest-mock = "3.12.0"
|
||||
pytest-cov = "4.1.0"
|
||||
ruff = "0.4.9"
|
||||
pytest = "8.2.2"
|
||||
pytest-mock = "3.14.0"
|
||||
pytest-cov = "5.0.0"
|
||||
toml = "0.10.2"
|
||||
yapf = "0.40.2"
|
||||
j2lint = "1.1.0"
|
||||
|
||||
|
||||
[tool.poetry-dynamic-versioning]
|
||||
enable = true
|
||||
|
@ -81,17 +80,22 @@ requires = ["poetry-core>=1.0.0", "poetry-dynamic-versioning"]
|
|||
|
||||
[tool.ruff]
|
||||
exclude = [
|
||||
".git",
|
||||
"__pycache__",
|
||||
"build",
|
||||
"dist",
|
||||
"test",
|
||||
"*.pyc",
|
||||
"*.egg-info",
|
||||
".cache",
|
||||
".eggs",
|
||||
"env*",
|
||||
".git",
|
||||
"__pycache__",
|
||||
"build",
|
||||
"dist",
|
||||
"test",
|
||||
"*.pyc",
|
||||
"*.egg-info",
|
||||
".cache",
|
||||
".eggs",
|
||||
"env*",
|
||||
]
|
||||
|
||||
line-length = 99
|
||||
indent-width = 4
|
||||
|
||||
[tool.ruff.lint]
|
||||
# Explanation of errors
|
||||
#
|
||||
# D102: Missing docstring in public method
|
||||
|
@ -102,47 +106,39 @@ exclude = [
|
|||
# D203: One blank line required before class docstring
|
||||
# D212: Multi-line docstring summary should start at the first line
|
||||
ignore = [
|
||||
"D102",
|
||||
"D103",
|
||||
"D105",
|
||||
"D107",
|
||||
"D202",
|
||||
"D203",
|
||||
"D212",
|
||||
"UP038",
|
||||
"RUF012",
|
||||
"D102",
|
||||
"D103",
|
||||
"D105",
|
||||
"D107",
|
||||
"D202",
|
||||
"D203",
|
||||
"D212",
|
||||
"UP038",
|
||||
"RUF012",
|
||||
]
|
||||
line-length = 99
|
||||
select = [
|
||||
"D",
|
||||
"E",
|
||||
"F",
|
||||
"Q",
|
||||
"W",
|
||||
"I",
|
||||
"S",
|
||||
"BLE",
|
||||
"N",
|
||||
"UP",
|
||||
"B",
|
||||
"A",
|
||||
"C4",
|
||||
"T20",
|
||||
"SIM",
|
||||
"RET",
|
||||
"ARG",
|
||||
"ERA",
|
||||
"RUF",
|
||||
"D",
|
||||
"E",
|
||||
"F",
|
||||
"Q",
|
||||
"W",
|
||||
"I",
|
||||
"S",
|
||||
"BLE",
|
||||
"N",
|
||||
"UP",
|
||||
"B",
|
||||
"A",
|
||||
"C4",
|
||||
"T20",
|
||||
"SIM",
|
||||
"RET",
|
||||
"ARG",
|
||||
"ERA",
|
||||
"RUF",
|
||||
]
|
||||
|
||||
[tool.ruff.flake8-quotes]
|
||||
inline-quotes = "double"
|
||||
|
||||
[tool.yapf]
|
||||
based_on_style = "google"
|
||||
column_limit = 99
|
||||
dedent_closing_brackets = true
|
||||
coalesce_brackets = true
|
||||
split_before_logical_operator = true
|
||||
indent_dictionary_value = true
|
||||
allow_split_before_dict_value = false
|
||||
[tool.ruff.format]
|
||||
quote-style = "double"
|
||||
indent-style = "space"
|
||||
line-ending = "lf"
|
||||
|
|
|
@ -1,4 +1,17 @@
|
|||
{
|
||||
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
|
||||
"extends": ["github>thegeeklab/renovate-presets"]
|
||||
"extends": ["github>thegeeklab/renovate-presets"],
|
||||
"packageRules": [
|
||||
{
|
||||
"description": "Ansible base dependencies",
|
||||
"matchPackageNames": ["ansible-core"],
|
||||
"separateMinorPatch": true
|
||||
},
|
||||
{
|
||||
"matchManagers": ["woodpecker"],
|
||||
"matchFileNames": [".woodpecker/test.yml"],
|
||||
"matchPackageNames": ["docker.io/library/python"],
|
||||
"enabled": false
|
||||
}
|
||||
]
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue
Block a user