Compare commits

..

1 Commits

Author SHA1 Message Date
d0fbe9a838 Update tox template
All checks were successful
gl/check check status: success (d0fbe9a83828b2ced0ae751df5edcd7895738439)
2023-01-06 14:31:14 +01:00
8 changed files with 53 additions and 214 deletions

File diff suppressed because it is too large Load Diff

View File

@ -209,22 +209,3 @@ class Services(object):
doc_struct[srv["service_category"]].append(srv_res)
return dict(categories=doc_struct)
def get_service_with_docs_by_service_type(self, service_type):
"""Retrieve service and service docs by service_type
:param str service_type: Filter by service_type
"""
res = dict()
res['service'] = {}
docs = []
services = self._service_data
for doc in services['documents']:
if doc['service_type'] == service_type:
docs.append(doc)
res['documents'] = docs
for service in services['services']:
if service['service_type'] == service_type:
res['service'] = service
break
return res

View File

@ -104,9 +104,6 @@ html_title = "{{ title }}"
# so a file named "default.css" will overwrite the builtin "default.css".
html_static_path = ['_static']
# Do not include sources into the rendered results
html_copy_source = False
# -- Options for PDF output --------------------------------------------------
latex_documents = [
{%- if pdf_name is defined %}

View File

@ -21,21 +21,19 @@ deps =
-r{toxinidir}/requirements.txt
commands = {posargs}
# This env is invoked in the periodic pipeline and is therefore responsible to
# build all relevant docs at once.
[testenv:docs]
deps = -r{toxinidir}/doc/requirements.txt
allowlist_externals =
mkdir
cp
sh
find
commands =
{%- for doc in docs %}
{[testenv:{{ doc.type }}]commands}
{[testenv:json-{{ doc.type }}]commands}
{%- endfor %}
[testenv:pdf-docs]
[testenv:docs-pdf]
deps = -r{toxinidir}/doc/requirements.txt
allowlist_externals =
rm
@ -59,8 +57,10 @@ deps = -r{toxinidir}/doc/requirements.txt
allowlist_externals =
cp
mkdir
find
commands =
sphinx-build -W --keep-going -b html {{ loc }}/source doc/build/html/{{ doc.type }}
find doc/build/html -type d -name '_sources' -exec rm -rf {:} ;
{%- if doc.type == 'api-ref' %}
mkdir -p api-ref/build/html
cp -av doc/build/html/api-ref api-ref/build/html
@ -78,12 +78,11 @@ deps = -r{toxinidir}/doc/requirements.txt
allowlist_externals =
cp
mkdir
sh
find
commands =
sphinx-build -W --keep-going -b json {{ loc }}/source doc/build/json/{{ doc.type }}
# Drop data useless for the search - wrap it also with sh/xargs due to bugs
# in tox
sh -c "find doc/build/json -type d -and '(' -name '_images' -or -name '_static' -or -name '_sources' ')' -print0 | xargs -0 rm -rf"
# Drop data useless for the search
find doc/build/json -type d -and ( -name '_images' -or -name '_static' -or -name '_sources' ) -exec rm -rf {:} ;
{%- if doc.type == 'api-ref' %}
mkdir -p api-ref/build/json
cp -av doc/build/json/api-ref api-ref/build/json

View File

@ -1,6 +1,2 @@
GitPython
ruamel.yaml
requests
jinja2
dirsync
cookiecutter
ruamel

View File

@ -46,7 +46,7 @@ def process_services(args, services):
has_wiki=False,
default_delete_branch_after_merge=True,
allow_merge_commit=False,
allow_squash_merge=True,
allow_squash_merge=False,
allow_rebase_merge=False,
default_merge_style="squash",
branch_protections=[],

View File

@ -18,9 +18,9 @@ import argparse
import logging
import os
import pathlib
import shutil
import subprocess
import warnings
from dirsync import sync
from git import Repo
from git import SymbolicReference
@ -111,21 +111,13 @@ def process_repositories(args, service):
new_branch.set_tracking_branch(remote_ref)
new_branch.checkout()
source_path = pathlib.Path(copy_from, doc["rst_location"])
target_path = pathlib.Path(copy_to, doc["rst_location"])
sync(
source_path,
target_path,
'sync',
purge=True,
create=True,
content=True,
ignore=['conf.py']
shutil.copytree(
pathlib.Path(copy_from, doc["rst_location"]),
pathlib.Path(copy_to, doc["rst_location"]),
ignore=lambda a, b: ["conf.py"],
dirs_exist_ok=True,
)
repo_to.index.add([doc["rst_location"]])
for obj in repo_to.index.diff(None).iter_change_type('D'):
repo_to.index.remove([obj.b_path])
if len(repo_to.index.diff("HEAD")) == 0:
# Nothing to commit
logging.debug("No changes.")

View File

@ -130,7 +130,6 @@ def process_repositories(args, service):
zuul_templates = None
zuul_jobs = dict()
zuul_new_jobs = list()
zuul_vars = dict()
zuul_config_updated = False
for item in zuul_config:
if "project" in item.keys():
@ -138,11 +137,10 @@ def process_repositories(args, service):
zuul_templates = project.setdefault("templates", [])
if not zuul_templates:
zuul_templates = []
zuul_vars = project.setdefault("vars", {})
elif "job" in item.keys():
job = item["job"]
zuul_jobs[job["name"]] = job
logging.debug(f"Existing jobs {zuul_jobs}")
print(f"Existing jobs {zuul_jobs}")
if "helpcenter-base-jobs" not in zuul_templates:
zuul_templates.append("helpcenter-base-jobs")
zuul_config_updated = True
@ -150,7 +148,6 @@ def process_repositories(args, service):
job_suffix = (
"-hc-int-jobs" if args.environment == "internal" else "-hc-jobs"
)
sphinx_pdf_files = zuul_vars.setdefault('sphinx_pdf_files', [])
for doc in data.docs_by_service_type(service["service_type"]):
logging.debug(f"Analyzing document {doc}")
if not doc.get("type"):
@ -159,12 +156,6 @@ def process_repositories(args, service):
doc_type = "dev-guide"
else:
doc_type = doc["type"]
# Collect all PDF files into sphinx_pdf_files var
pdf_name = doc.get('pdf_name')
if pdf_name and f"{pdf_name}.pdf" not in sphinx_pdf_files:
sphinx_pdf_files.append(f"{pdf_name}.pdf")
zuul_config_updated = True
template_name = f"{doc_type}{job_suffix}"
if doc_type in ["api-ref", "umn", "dev-guide"]:
if template_name not in zuul_templates:
@ -199,7 +190,6 @@ def process_repositories(args, service):
if "project" in item.keys():
project = item["project"]
project["templates"] = zuul_templates
project["vars"] = zuul_vars
# Ensure new jobs are in check
if len(zuul_new_jobs) > 0:
project.setdefault(