show keyword documentation in ansible-doc (#72476)
* show keyword documentation in ansible-doc Co-authored-by: Felix Fontein <felix@fontein.de>
This commit is contained in:
parent
96ad5b799e
commit
8eab113cb1
8 changed files with 197 additions and 92 deletions
|
@ -11,6 +11,7 @@ include examples/hosts
|
||||||
include examples/ansible.cfg
|
include examples/ansible.cfg
|
||||||
include examples/scripts/ConfigureRemotingForAnsible.ps1
|
include examples/scripts/ConfigureRemotingForAnsible.ps1
|
||||||
include examples/scripts/upgrade_to_ps3.ps1
|
include examples/scripts/upgrade_to_ps3.ps1
|
||||||
|
include lib/ansible/keyword_desc.yml
|
||||||
recursive-include lib/ansible/executor/powershell *.ps1
|
recursive-include lib/ansible/executor/powershell *.ps1
|
||||||
recursive-include lib/ansible/module_utils/csharp *.cs
|
recursive-include lib/ansible/module_utils/csharp *.cs
|
||||||
recursive-include lib/ansible/module_utils/powershell *.psm1
|
recursive-include lib/ansible/module_utils/powershell *.psm1
|
||||||
|
|
2
changelogs/fragments/add_keywords_to_ansible_doc.yml
Normal file
2
changelogs/fragments/add_keywords_to_ansible_doc.yml
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
minor_changes:
|
||||||
|
- ansible-doc has new option to show keyword documentation.
|
|
@ -110,7 +110,7 @@ cli:
|
||||||
$(GENERATE_CLI) --template-file=../templates/cli_rst.j2 --output-dir=rst/cli/ --output-format rst ../../lib/ansible/cli/*.py
|
$(GENERATE_CLI) --template-file=../templates/cli_rst.j2 --output-dir=rst/cli/ --output-format rst ../../lib/ansible/cli/*.py
|
||||||
|
|
||||||
keywords: ../templates/playbooks_keywords.rst.j2
|
keywords: ../templates/playbooks_keywords.rst.j2
|
||||||
$(KEYWORD_DUMPER) --template-dir=../templates --output-dir=rst/reference_appendices/ ./keyword_desc.yml
|
$(KEYWORD_DUMPER) --template-dir=../templates --output-dir=rst/reference_appendices/ ../../lib/ansible/keyword_desc.yml
|
||||||
|
|
||||||
config: ../templates/config.rst.j2
|
config: ../templates/config.rst.j2
|
||||||
$(CONFIG_DUMPER) --template-file=../templates/config.rst.j2 --output-dir=rst/reference_appendices/ ../../lib/ansible/config/base.yml
|
$(CONFIG_DUMPER) --template-file=../templates/config.rst.j2 --output-dir=rst/reference_appendices/ ../../lib/ansible/config/base.yml
|
||||||
|
|
|
@ -7,6 +7,7 @@ __metaclass__ = type
|
||||||
|
|
||||||
import datetime
|
import datetime
|
||||||
import json
|
import json
|
||||||
|
import pkgutil
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import textwrap
|
import textwrap
|
||||||
|
@ -24,8 +25,10 @@ from ansible.errors import AnsibleError, AnsibleOptionsError
|
||||||
from ansible.module_utils._text import to_native, to_text
|
from ansible.module_utils._text import to_native, to_text
|
||||||
from ansible.module_utils.common._collections_compat import Container, Sequence
|
from ansible.module_utils.common._collections_compat import Container, Sequence
|
||||||
from ansible.module_utils.common.json import AnsibleJSONEncoder
|
from ansible.module_utils.common.json import AnsibleJSONEncoder
|
||||||
|
from ansible.module_utils.compat import importlib
|
||||||
from ansible.module_utils.six import string_types
|
from ansible.module_utils.six import string_types
|
||||||
from ansible.parsing.plugin_docs import read_docstub
|
from ansible.parsing.plugin_docs import read_docstub
|
||||||
|
from ansible.parsing.utils.yaml import from_yaml
|
||||||
from ansible.parsing.yaml.dumper import AnsibleDumper
|
from ansible.parsing.yaml.dumper import AnsibleDumper
|
||||||
from ansible.plugins.loader import action_loader, fragment_loader
|
from ansible.plugins.loader import action_loader, fragment_loader
|
||||||
from ansible.utils.collection_loader import AnsibleCollectionConfig
|
from ansible.utils.collection_loader import AnsibleCollectionConfig
|
||||||
|
@ -41,6 +44,11 @@ from ansible.utils.plugin_docs import (
|
||||||
display = Display()
|
display = Display()
|
||||||
|
|
||||||
|
|
||||||
|
TARGET_OPTIONS = C.DOCUMENTABLE_PLUGINS + ('keyword',)
|
||||||
|
PB_OBJECTS = ['Play', 'Role', 'Block', 'Task']
|
||||||
|
PB_LOADED = {}
|
||||||
|
|
||||||
|
|
||||||
def jdump(text):
|
def jdump(text):
|
||||||
try:
|
try:
|
||||||
display.display(json.dumps(text, cls=AnsibleJSONEncoder, sort_keys=True, indent=4))
|
display.display(json.dumps(text, cls=AnsibleJSONEncoder, sort_keys=True, indent=4))
|
||||||
|
@ -83,6 +91,12 @@ class DocCLI(CLI):
|
||||||
_CONST = re.compile(r"\bC\(([^)]+)\)")
|
_CONST = re.compile(r"\bC\(([^)]+)\)")
|
||||||
_RULER = re.compile(r"\bHORIZONTALLINE\b")
|
_RULER = re.compile(r"\bHORIZONTALLINE\b")
|
||||||
|
|
||||||
|
# rst specific
|
||||||
|
_REFTAG = re.compile(r":ref:")
|
||||||
|
_TERM = re.compile(r":term:")
|
||||||
|
_NOTES = re.compile(r".. note:")
|
||||||
|
_SEEALSO = re.compile(r"^\s*.. seealso:.*$", re.MULTILINE)
|
||||||
|
|
||||||
def __init__(self, args):
|
def __init__(self, args):
|
||||||
|
|
||||||
super(DocCLI, self).__init__(args)
|
super(DocCLI, self).__init__(args)
|
||||||
|
@ -100,6 +114,11 @@ class DocCLI(CLI):
|
||||||
t = cls._CONST.sub("`" + r"\1" + "'", t) # C(word) => `word'
|
t = cls._CONST.sub("`" + r"\1" + "'", t) # C(word) => `word'
|
||||||
t = cls._RULER.sub("\n{0}\n".format("-" * 13), t) # HORIZONTALLINE => -------
|
t = cls._RULER.sub("\n{0}\n".format("-" * 13), t) # HORIZONTALLINE => -------
|
||||||
|
|
||||||
|
t = cls._REFTAG.sub(r"", t) # remove rst :ref:
|
||||||
|
t = cls._TERM.sub(r"", t) # remove rst :term:
|
||||||
|
t = cls._NOTES.sub(r" Note:", t) # nicer note
|
||||||
|
t = cls._SEEALSO.sub(r"", t) # remove seealso
|
||||||
|
|
||||||
return t
|
return t
|
||||||
|
|
||||||
def init_parser(self):
|
def init_parser(self):
|
||||||
|
@ -114,10 +133,11 @@ class DocCLI(CLI):
|
||||||
opt_help.add_basedir_options(self.parser)
|
opt_help.add_basedir_options(self.parser)
|
||||||
|
|
||||||
self.parser.add_argument('args', nargs='*', help='Plugin', metavar='plugin')
|
self.parser.add_argument('args', nargs='*', help='Plugin', metavar='plugin')
|
||||||
|
|
||||||
self.parser.add_argument("-t", "--type", action="store", default='module', dest='type',
|
self.parser.add_argument("-t", "--type", action="store", default='module', dest='type',
|
||||||
help='Choose which plugin type (defaults to "module"). '
|
help='Choose which plugin type (defaults to "module"). '
|
||||||
'Available plugin types are : {0}'.format(C.DOCUMENTABLE_PLUGINS),
|
'Available plugin types are : {0}'.format(TARGET_OPTIONS),
|
||||||
choices=C.DOCUMENTABLE_PLUGINS)
|
choices=TARGET_OPTIONS)
|
||||||
self.parser.add_argument("-j", "--json", action="store_true", default=False, dest='json_format',
|
self.parser.add_argument("-j", "--json", action="store_true", default=False, dest='json_format',
|
||||||
help='Change output into json format.')
|
help='Change output into json format.')
|
||||||
|
|
||||||
|
@ -172,36 +192,69 @@ class DocCLI(CLI):
|
||||||
# display results
|
# display results
|
||||||
DocCLI.pager("\n".join(text))
|
DocCLI.pager("\n".join(text))
|
||||||
|
|
||||||
def run(self):
|
@staticmethod
|
||||||
|
def _list_keywords():
|
||||||
|
return from_yaml(pkgutil.get_data('ansible', 'keyword_desc.yml'))
|
||||||
|
|
||||||
super(DocCLI, self).run()
|
@staticmethod
|
||||||
|
def _get_keywords_docs(keys):
|
||||||
|
|
||||||
plugin_type = context.CLIARGS['type']
|
data = {}
|
||||||
do_json = context.CLIARGS['json_format']
|
descs = DocCLI._list_keywords()
|
||||||
|
for keyword in keys:
|
||||||
|
if keyword.startswith('with_'):
|
||||||
|
keyword = 'loop'
|
||||||
|
try:
|
||||||
|
# if no desc, typeerror raised ends this block
|
||||||
|
kdata = {'description': descs[keyword]}
|
||||||
|
|
||||||
if plugin_type in C.DOCUMENTABLE_PLUGINS:
|
# get playbook objects for keyword and use first to get keyword attributes
|
||||||
loader = getattr(plugin_loader, '%s_loader' % plugin_type)
|
kdata['applies_to'] = []
|
||||||
|
for pobj in PB_OBJECTS:
|
||||||
|
if pobj not in PB_LOADED:
|
||||||
|
obj_class = 'ansible.playbook.%s' % pobj.lower()
|
||||||
|
loaded_class = importlib.import_module(obj_class)
|
||||||
|
PB_LOADED[pobj] = getattr(loaded_class, pobj, None)
|
||||||
|
|
||||||
|
if keyword in PB_LOADED[pobj]._valid_attrs:
|
||||||
|
kdata['applies_to'].append(pobj)
|
||||||
|
|
||||||
|
# we should only need these once
|
||||||
|
if 'type' not in kdata:
|
||||||
|
|
||||||
|
fa = getattr(PB_LOADED[pobj], '_%s' % keyword)
|
||||||
|
if getattr(fa, 'private'):
|
||||||
|
kdata = {}
|
||||||
|
raise KeyError
|
||||||
|
|
||||||
|
kdata['type'] = getattr(fa, 'isa', 'string')
|
||||||
|
|
||||||
|
if keyword.endswith('when'):
|
||||||
|
kdata['template'] = 'implicit'
|
||||||
|
elif getattr(fa, 'static'):
|
||||||
|
kdata['template'] = 'static'
|
||||||
else:
|
else:
|
||||||
raise AnsibleOptionsError("Unknown or undocumentable plugin type: %s" % plugin_type)
|
kdata['template'] = 'explicit'
|
||||||
|
|
||||||
# add to plugin paths from command line
|
# those that require no processing
|
||||||
basedir = context.CLIARGS['basedir']
|
for visible in ('alias', 'priority'):
|
||||||
if basedir:
|
kdata[visible] = getattr(fa, visible)
|
||||||
AnsibleCollectionConfig.playbook_paths = basedir
|
|
||||||
loader.add_directory(basedir, with_subdir=True)
|
|
||||||
|
|
||||||
if context.CLIARGS['module_path']:
|
# remove None keys
|
||||||
for path in context.CLIARGS['module_path']:
|
for k in list(kdata.keys()):
|
||||||
if path:
|
if kdata[k] is None:
|
||||||
loader.add_directory(path)
|
del kdata[k]
|
||||||
|
|
||||||
# save only top level paths for errors
|
data[keyword] = kdata
|
||||||
search_paths = DocCLI.print_paths(loader)
|
|
||||||
loader._paths = None # reset so we can use subdirs below
|
|
||||||
|
|
||||||
# list plugins names or filepath for type, both options share most code
|
except KeyError as e:
|
||||||
if context.CLIARGS['list_files'] or context.CLIARGS['list_dir']:
|
display.warning("Skipping Invalid keyword '%s' specified: %s" % (keyword, to_native(e)))
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
def _list_plugins(self, plugin_type, loader):
|
||||||
|
|
||||||
|
results = {}
|
||||||
coll_filter = None
|
coll_filter = None
|
||||||
if len(context.CLIARGS['args']) == 1:
|
if len(context.CLIARGS['args']) == 1:
|
||||||
coll_filter = context.CLIARGS['args'][0]
|
coll_filter = context.CLIARGS['args'][0]
|
||||||
|
@ -209,8 +262,7 @@ class DocCLI(CLI):
|
||||||
if coll_filter in ('', None):
|
if coll_filter in ('', None):
|
||||||
paths = loader._get_paths_with_context()
|
paths = loader._get_paths_with_context()
|
||||||
for path_context in paths:
|
for path_context in paths:
|
||||||
self.plugin_list.update(
|
self.plugin_list.update(DocCLI.find_plugins(path_context.path, path_context.internal, plugin_type))
|
||||||
DocCLI.find_plugins(path_context.path, path_context.internal, plugin_type))
|
|
||||||
|
|
||||||
add_collection_plugins(self.plugin_list, plugin_type, coll_filter=coll_filter)
|
add_collection_plugins(self.plugin_list, plugin_type, coll_filter=coll_filter)
|
||||||
|
|
||||||
|
@ -219,24 +271,20 @@ class DocCLI(CLI):
|
||||||
results = self._get_plugin_list_descriptions(loader)
|
results = self._get_plugin_list_descriptions(loader)
|
||||||
elif context.CLIARGS['list_files']:
|
elif context.CLIARGS['list_files']:
|
||||||
results = self._get_plugin_list_filenames(loader)
|
results = self._get_plugin_list_filenames(loader)
|
||||||
|
|
||||||
if do_json:
|
|
||||||
jdump(results)
|
|
||||||
elif self.plugin_list:
|
|
||||||
self.display_plugin_list(results)
|
|
||||||
else:
|
|
||||||
display.warning("No plugins found.")
|
|
||||||
# dump plugin desc/data as JSON
|
# dump plugin desc/data as JSON
|
||||||
elif context.CLIARGS['dump']:
|
elif context.CLIARGS['dump']:
|
||||||
plugin_data = {}
|
|
||||||
plugin_names = DocCLI.get_all_plugins_of_type(plugin_type)
|
plugin_names = DocCLI.get_all_plugins_of_type(plugin_type)
|
||||||
for plugin_name in plugin_names:
|
for plugin_name in plugin_names:
|
||||||
plugin_info = DocCLI.get_plugin_metadata(plugin_type, plugin_name)
|
plugin_info = DocCLI.get_plugin_metadata(plugin_type, plugin_name)
|
||||||
if plugin_info is not None:
|
if plugin_info is not None:
|
||||||
plugin_data[plugin_name] = plugin_info
|
results[plugin_name] = plugin_info
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
|
def _get_plugins_docs(self, plugin_type, loader):
|
||||||
|
|
||||||
|
search_paths = DocCLI.print_paths(loader)
|
||||||
|
|
||||||
jdump(plugin_data)
|
|
||||||
else:
|
|
||||||
# display specific plugin docs
|
# display specific plugin docs
|
||||||
if len(context.CLIARGS['args']) == 0:
|
if len(context.CLIARGS['args']) == 0:
|
||||||
raise AnsibleOptionsError("Incorrect options passed")
|
raise AnsibleOptionsError("Incorrect options passed")
|
||||||
|
@ -261,13 +309,57 @@ class DocCLI(CLI):
|
||||||
|
|
||||||
plugin_docs[plugin] = DocCLI._combine_plugin_doc(plugin, plugin_type, doc, plainexamples, returndocs, metadata)
|
plugin_docs[plugin] = DocCLI._combine_plugin_doc(plugin, plugin_type, doc, plainexamples, returndocs, metadata)
|
||||||
|
|
||||||
if do_json:
|
return plugin_docs
|
||||||
jdump(plugin_docs)
|
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
|
||||||
|
super(DocCLI, self).run()
|
||||||
|
|
||||||
|
plugin_type = context.CLIARGS['type']
|
||||||
|
do_json = context.CLIARGS['json_format']
|
||||||
|
listing = context.CLIARGS['list_files'] or context.CLIARGS['list_dir'] or context.CLIARGS['dump']
|
||||||
|
docs = {}
|
||||||
|
|
||||||
|
if plugin_type not in TARGET_OPTIONS:
|
||||||
|
raise AnsibleOptionsError("Unknown or undocumentable plugin type: %s" % plugin_type)
|
||||||
|
elif plugin_type == 'keyword':
|
||||||
|
|
||||||
|
if listing:
|
||||||
|
docs = DocCLI._list_keywords()
|
||||||
|
else:
|
||||||
|
docs = DocCLI._get_keywords_docs(context.CLIARGS['args'])
|
||||||
|
else:
|
||||||
|
loader = getattr(plugin_loader, '%s_loader' % plugin_type)
|
||||||
|
|
||||||
|
# add to plugin paths from command line
|
||||||
|
basedir = context.CLIARGS['basedir']
|
||||||
|
if basedir:
|
||||||
|
AnsibleCollectionConfig.playbook_paths = basedir
|
||||||
|
loader.add_directory(basedir, with_subdir=True)
|
||||||
|
|
||||||
|
if context.CLIARGS['module_path']:
|
||||||
|
for path in context.CLIARGS['module_path']:
|
||||||
|
if path:
|
||||||
|
loader.add_directory(path)
|
||||||
|
|
||||||
|
# save only top level paths for errors
|
||||||
|
loader._paths = None # reset so we can use subdirs below
|
||||||
|
|
||||||
|
if listing:
|
||||||
|
docs = self._list_plugins(plugin_type, loader)
|
||||||
|
else:
|
||||||
|
docs = self._get_plugins_docs(plugin_type, loader)
|
||||||
|
|
||||||
|
if do_json:
|
||||||
|
jdump(docs)
|
||||||
|
else:
|
||||||
|
text = []
|
||||||
|
if plugin_type in C.DOCUMENTABLE_PLUGINS:
|
||||||
|
if listing and docs:
|
||||||
|
self.display_plugin_list(docs)
|
||||||
else:
|
else:
|
||||||
# Some changes to how plain text docs are formatted
|
# Some changes to how plain text docs are formatted
|
||||||
text = []
|
for plugin, doc_data in docs.items():
|
||||||
for plugin, doc_data in plugin_docs.items():
|
|
||||||
textret = DocCLI.format_plugin_doc(plugin, plugin_type,
|
textret = DocCLI.format_plugin_doc(plugin, plugin_type,
|
||||||
doc_data['doc'], doc_data['examples'],
|
doc_data['doc'], doc_data['examples'],
|
||||||
doc_data['return'], doc_data['metadata'])
|
doc_data['return'], doc_data['metadata'])
|
||||||
|
@ -275,6 +367,8 @@ class DocCLI(CLI):
|
||||||
text.append(textret)
|
text.append(textret)
|
||||||
else:
|
else:
|
||||||
display.warning("No valid documentation was retrieved from '%s'" % plugin)
|
display.warning("No valid documentation was retrieved from '%s'" % plugin)
|
||||||
|
elif docs:
|
||||||
|
text = DocCLI._dump_yaml(docs, '')
|
||||||
|
|
||||||
if text:
|
if text:
|
||||||
DocCLI.pager(''.join(text))
|
DocCLI.pager(''.join(text))
|
||||||
|
|
|
@ -947,6 +947,7 @@ def _get_collection_name_from_path(path):
|
||||||
|
|
||||||
original_path_prefix = os.path.join('/', *path_parts[0:ac_pos + 3])
|
original_path_prefix = os.path.join('/', *path_parts[0:ac_pos + 3])
|
||||||
|
|
||||||
|
imported_pkg_path = to_native(os.path.abspath(to_bytes(imported_pkg_path)))
|
||||||
if original_path_prefix != imported_pkg_path:
|
if original_path_prefix != imported_pkg_path:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
|
@ -3,6 +3,12 @@
|
||||||
set -eux
|
set -eux
|
||||||
ansible-playbook test.yml -i inventory "$@"
|
ansible-playbook test.yml -i inventory "$@"
|
||||||
|
|
||||||
|
# test keyword docs
|
||||||
|
ansible-doc -t keyword -l | grep 'vars_prompt: list of variables to prompt for.'
|
||||||
|
ansible-doc -t keyword vars_prompt | grep 'description: list of variables to prompt for.'
|
||||||
|
ansible-doc -t keyword asldkfjaslidfhals 2>&1 | grep 'Skipping Invalid keyword'
|
||||||
|
|
||||||
|
# collections testing
|
||||||
(
|
(
|
||||||
unset ANSIBLE_PLAYBOOK_DIR
|
unset ANSIBLE_PLAYBOOK_DIR
|
||||||
cd "$(dirname "$0")"
|
cd "$(dirname "$0")"
|
||||||
|
|
|
@ -31,6 +31,7 @@ hacking/build_library/build_ansible/command_plugins/update_intersphinx.py compil
|
||||||
hacking/build_library/build_ansible/commands.py compile-2.6!skip # release and docs process only, 3.6+ required
|
hacking/build_library/build_ansible/commands.py compile-2.6!skip # release and docs process only, 3.6+ required
|
||||||
hacking/build_library/build_ansible/commands.py compile-2.7!skip # release and docs process only, 3.6+ required
|
hacking/build_library/build_ansible/commands.py compile-2.7!skip # release and docs process only, 3.6+ required
|
||||||
hacking/build_library/build_ansible/commands.py compile-3.5!skip # release and docs process only, 3.6+ required
|
hacking/build_library/build_ansible/commands.py compile-3.5!skip # release and docs process only, 3.6+ required
|
||||||
|
lib/ansible/keyword_desc.yml no-unwanted-files
|
||||||
lib/ansible/cli/console.py pylint:blacklisted-name
|
lib/ansible/cli/console.py pylint:blacklisted-name
|
||||||
lib/ansible/cli/scripts/ansible_cli_stub.py shebang
|
lib/ansible/cli/scripts/ansible_cli_stub.py shebang
|
||||||
lib/ansible/cli/scripts/ansible_connection_cli_stub.py shebang
|
lib/ansible/cli/scripts/ansible_connection_cli_stub.py shebang
|
||||||
|
|
Loading…
Reference in a new issue