2016-11-15 00:31:02 +01:00
|
|
|
#!/usr/bin/env python
|
2017-03-14 17:07:22 +01:00
|
|
|
# (c) 2016-2017, Toshio Kuratomi <tkuratomi@ansible.com>
|
|
|
|
#
|
|
|
|
# This file is part of Ansible
|
|
|
|
#
|
|
|
|
# Ansible is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU General Public License as published by
|
|
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
#
|
|
|
|
# Ansible is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
|
|
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
# Make coding more python3-ish
|
|
|
|
from __future__ import (absolute_import, division, print_function)
|
|
|
|
__metaclass__ = type
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
import ast
|
|
|
|
import csv
|
|
|
|
import os
|
|
|
|
import sys
|
2016-12-06 22:17:49 +01:00
|
|
|
from collections import defaultdict
|
2016-11-15 00:31:02 +01:00
|
|
|
from distutils.version import StrictVersion
|
2016-12-06 15:27:10 +01:00
|
|
|
from pprint import pformat, pprint
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
import yaml
|
|
|
|
|
|
|
|
from ansible.module_utils._text import to_text
|
|
|
|
from ansible.plugins import module_loader
|
|
|
|
|
|
|
|
|
|
|
|
# There's a few files that are not new-style modules. Have to blacklist them
|
|
|
|
NONMODULE_PY_FILES = frozenset(('async_wrapper.py',))
|
|
|
|
NONMODULE_MODULE_NAMES = frozenset(os.path.splitext(p)[0] for p in NONMODULE_PY_FILES)
|
|
|
|
|
2016-12-06 22:17:49 +01:00
|
|
|
# Default metadata
|
2017-05-16 19:52:07 +02:00
|
|
|
DEFAULT_METADATA = {'metadata_version': '1.0', 'status': ['preview'], 'supported_by': 'community'}
|
2016-12-06 22:17:49 +01:00
|
|
|
|
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
class ParseError(Exception):
|
|
|
|
"""Thrown when parsing a file fails"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class MissingModuleError(Exception):
|
|
|
|
"""Thrown when unable to find a plugin"""
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def usage():
|
|
|
|
print("""Usage:
|
2017-03-14 17:07:22 +01:00
|
|
|
metadata-tool.py report [--version X]
|
|
|
|
metadata-tool.py add [--version X] [--overwrite] CSVFILE
|
|
|
|
metadata-tool.py add-default [--version X] [--overwrite]
|
|
|
|
medatada-tool.py upgrade [--version X]""")
|
2016-11-15 00:31:02 +01:00
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
|
|
|
|
def parse_args(arg_string):
|
|
|
|
if len(arg_string) < 1:
|
|
|
|
usage()
|
|
|
|
|
|
|
|
action = arg_string[0]
|
|
|
|
|
|
|
|
version = None
|
|
|
|
if '--version' in arg_string:
|
|
|
|
version_location = arg_string.index('--version')
|
|
|
|
arg_string.pop(version_location)
|
|
|
|
version = arg_string.pop(version_location)
|
|
|
|
|
|
|
|
overwrite = False
|
|
|
|
if '--overwrite' in arg_string:
|
|
|
|
overwrite = True
|
|
|
|
arg_string.remove('--overwrite')
|
|
|
|
|
|
|
|
csvfile = None
|
|
|
|
if len(arg_string) == 2:
|
|
|
|
csvfile = arg_string[1]
|
|
|
|
elif len(arg_string) > 2:
|
|
|
|
usage()
|
|
|
|
|
|
|
|
return action, {'version': version, 'overwrite': overwrite, 'csvfile': csvfile}
|
|
|
|
|
2017-05-16 19:52:07 +02:00
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
def seek_end_of_dict(module_data, start_line, start_col, next_node_line, next_node_col):
|
|
|
|
"""Look for the end of a dict in a set of lines
|
|
|
|
|
|
|
|
We know the starting position of the dict and we know the start of the
|
|
|
|
next code node but in between there may be multiple newlines and comments.
|
|
|
|
There may also be multiple python statements on the same line (separated
|
|
|
|
by semicolons)
|
|
|
|
|
|
|
|
Examples::
|
|
|
|
ANSIBLE_METADATA = {[..]}
|
|
|
|
DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = {[..]} # Optional comments with confusing junk => {}
|
|
|
|
# Optional comments {}
|
|
|
|
DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = {
|
|
|
|
[..]
|
|
|
|
}
|
|
|
|
# Optional comments {}
|
|
|
|
DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = {[..]} ; DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = {}EOF
|
|
|
|
"""
|
2017-01-28 09:12:11 +01:00
|
|
|
if next_node_line is None:
|
2016-11-15 00:31:02 +01:00
|
|
|
# The dict is the last statement in the file
|
|
|
|
snippet = module_data.splitlines()[start_line:]
|
|
|
|
next_node_col = 0
|
|
|
|
# Include the last line in the file
|
|
|
|
last_line_offset = 0
|
|
|
|
else:
|
|
|
|
# It's somewhere in the middle so we need to separate it from the rest
|
|
|
|
snippet = module_data.splitlines()[start_line:next_node_line]
|
|
|
|
# Do not include the last line because that's where the next node
|
|
|
|
# starts
|
|
|
|
last_line_offset = 1
|
|
|
|
|
|
|
|
if next_node_col == 0:
|
|
|
|
# This handles all variants where there are only comments and blank
|
|
|
|
# lines between the dict and the next code node
|
|
|
|
|
|
|
|
# Step backwards through all the lines in the snippet
|
|
|
|
for line_idx, line in tuple(reversed(tuple(enumerate(snippet))))[last_line_offset:]:
|
|
|
|
end_col = None
|
|
|
|
# Step backwards through all the characters in the line
|
|
|
|
for col_idx, char in reversed(tuple(enumerate(c for c in line))):
|
|
|
|
if char == '}' and end_col is None:
|
|
|
|
# Potentially found the end of the dict
|
|
|
|
end_col = col_idx
|
|
|
|
|
|
|
|
elif char == '#' and end_col is not None:
|
|
|
|
# The previous '}' was part of a comment. Keep trying
|
|
|
|
end_col = None
|
|
|
|
|
|
|
|
if end_col is not None:
|
|
|
|
# Found the end!
|
|
|
|
end_line = start_line + line_idx
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
# Harder cases involving multiple statements on one line
|
|
|
|
# Good Ansible Module style doesn't do this so we're just going to
|
|
|
|
# treat this as an error for now:
|
|
|
|
raise ParseError('Multiple statements per line confuses the module metadata parser.')
|
|
|
|
|
|
|
|
return end_line, end_col
|
|
|
|
|
|
|
|
|
|
|
|
def seek_end_of_string(module_data, start_line, start_col, next_node_line, next_node_col):
|
|
|
|
"""
|
|
|
|
This is much trickier than finding the end of a dict. A dict has only one
|
|
|
|
ending character, "}". Strings have four potential ending characters. We
|
|
|
|
have to parse the beginning of the string to determine what the ending
|
|
|
|
character will be.
|
|
|
|
|
|
|
|
Examples:
|
|
|
|
ANSIBLE_METADATA = '''[..]''' # Optional comment with confusing chars '''
|
|
|
|
# Optional comment with confusing chars '''
|
|
|
|
DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = '''
|
|
|
|
[..]
|
|
|
|
'''
|
|
|
|
DOCUMENTATIONS = [..]
|
|
|
|
|
|
|
|
ANSIBLE_METADATA = '''[..]''' ; DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
SHORT_NAME = ANSIBLE_METADATA = '''[..]''' ; DOCUMENTATION = [..]
|
|
|
|
|
|
|
|
String marker variants:
|
|
|
|
* '[..]'
|
|
|
|
* "[..]"
|
|
|
|
* '''[..]'''
|
|
|
|
* \"\"\"[..]\"\"\"
|
|
|
|
|
|
|
|
Each of these come in u, r, and b variants:
|
|
|
|
* '[..]'
|
|
|
|
* u'[..]'
|
|
|
|
* b'[..]'
|
|
|
|
* r'[..]'
|
|
|
|
* ur'[..]'
|
|
|
|
* ru'[..]'
|
|
|
|
* br'[..]'
|
|
|
|
* b'[..]'
|
|
|
|
* rb'[..]'
|
|
|
|
"""
|
|
|
|
raise NotImplementedError('Finding end of string not yet implemented')
|
|
|
|
|
|
|
|
|
|
|
|
def extract_metadata(module_data):
|
|
|
|
"""Extract the metadata from a module
|
|
|
|
|
|
|
|
:arg module_data: Byte string containing a module's code
|
|
|
|
:returns: a tuple of metadata (a dict), line the metadata starts on,
|
|
|
|
column the metadata starts on, line the metadata ends on, column the
|
|
|
|
metadata ends on, and the names the metadata is assigned to. One of
|
|
|
|
the names the metadata is assigned to will be ANSIBLE_METADATA If no
|
|
|
|
metadata is found, the tuple will be (None, -1, -1, -1, -1, None)
|
|
|
|
"""
|
|
|
|
metadata = None
|
|
|
|
start_line = -1
|
|
|
|
start_col = -1
|
|
|
|
end_line = -1
|
|
|
|
end_col = -1
|
|
|
|
targets = None
|
|
|
|
mod_ast_tree = ast.parse(module_data)
|
|
|
|
for root_idx, child in enumerate(mod_ast_tree.body):
|
|
|
|
if isinstance(child, ast.Assign):
|
|
|
|
for target in child.targets:
|
|
|
|
if target.id == 'ANSIBLE_METADATA':
|
|
|
|
if isinstance(child.value, ast.Dict):
|
|
|
|
metadata = ast.literal_eval(child.value)
|
|
|
|
|
|
|
|
try:
|
|
|
|
# Determine where the next node starts
|
2017-05-16 19:52:07 +02:00
|
|
|
next_node = mod_ast_tree.body[root_idx + 1]
|
2016-11-15 00:31:02 +01:00
|
|
|
next_lineno = next_node.lineno
|
|
|
|
next_col_offset = next_node.col_offset
|
|
|
|
except IndexError:
|
|
|
|
# Metadata is defined in the last node of the file
|
|
|
|
next_lineno = None
|
|
|
|
next_col_offset = None
|
|
|
|
|
|
|
|
# Determine where the current metadata ends
|
|
|
|
end_line, end_col = seek_end_of_dict(module_data,
|
2017-05-16 19:52:07 +02:00
|
|
|
child.lineno - 1,
|
|
|
|
child.col_offset,
|
|
|
|
next_lineno,
|
|
|
|
next_col_offset)
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
elif isinstance(child.value, ast.Str):
|
|
|
|
metadata = yaml.safe_load(child.value.s)
|
|
|
|
end_line = seek_end_of_string(module_data)
|
|
|
|
elif isinstance(child.value, ast.Bytes):
|
|
|
|
metadata = yaml.safe_load(to_text(child.value.s, errors='surrogate_or_strict'))
|
|
|
|
end_line = seek_end_of_string(module_data)
|
|
|
|
else:
|
|
|
|
# Example:
|
|
|
|
# ANSIBLE_METADATA = 'junk'
|
|
|
|
# ANSIBLE_METADATA = { [..the real metadata..] }
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Do these after the if-else so we don't pollute them in
|
|
|
|
# case this was a false positive
|
|
|
|
start_line = child.lineno - 1
|
|
|
|
start_col = child.col_offset
|
|
|
|
targets = [t.id for t in child.targets]
|
|
|
|
break
|
|
|
|
|
|
|
|
if metadata is not None:
|
|
|
|
# Once we've found the metadata we're done
|
|
|
|
break
|
|
|
|
|
|
|
|
return metadata, start_line, start_col, end_line, end_col, targets
|
|
|
|
|
|
|
|
|
|
|
|
def find_documentation(module_data):
|
|
|
|
"""Find the DOCUMENTATION metadata for a module file"""
|
|
|
|
start_line = -1
|
|
|
|
mod_ast_tree = ast.parse(module_data)
|
|
|
|
for child in mod_ast_tree.body:
|
|
|
|
if isinstance(child, ast.Assign):
|
|
|
|
for target in child.targets:
|
|
|
|
if target.id == 'DOCUMENTATION':
|
|
|
|
start_line = child.lineno - 1
|
|
|
|
break
|
|
|
|
|
|
|
|
return start_line
|
|
|
|
|
|
|
|
|
|
|
|
def remove_metadata(module_data, start_line, start_col, end_line, end_col):
|
|
|
|
"""Remove a section of a module file"""
|
|
|
|
lines = module_data.split('\n')
|
|
|
|
new_lines = lines[:start_line]
|
|
|
|
if start_col != 0:
|
|
|
|
new_lines.append(lines[start_line][:start_col])
|
|
|
|
|
|
|
|
next_line = lines[end_line]
|
|
|
|
if len(next_line) - 1 != end_col:
|
|
|
|
new_lines.append(next_line[end_col:])
|
|
|
|
|
|
|
|
if len(lines) > end_line:
|
|
|
|
new_lines.extend(lines[end_line + 1:])
|
|
|
|
return '\n'.join(new_lines)
|
|
|
|
|
|
|
|
|
|
|
|
def insert_metadata(module_data, new_metadata, insertion_line, targets=('ANSIBLE_METADATA',)):
|
|
|
|
"""Insert a new set of metadata at a specified line"""
|
2016-12-06 15:27:10 +01:00
|
|
|
assignments = ' = '.join(targets)
|
|
|
|
pretty_metadata = pformat(new_metadata, width=1).split('\n')
|
|
|
|
|
|
|
|
new_lines = []
|
|
|
|
new_lines.append('{} = {}'.format(assignments, pretty_metadata[0]))
|
|
|
|
|
|
|
|
if len(pretty_metadata) > 1:
|
|
|
|
for line in pretty_metadata[1:]:
|
|
|
|
new_lines.append('{}{}'.format(' ' * (len(assignments) - 1 + len(' = {')), line))
|
|
|
|
|
|
|
|
old_lines = module_data.split('\n')
|
|
|
|
lines = old_lines[:insertion_line] + new_lines + [''] + old_lines[insertion_line:]
|
2016-11-15 00:31:02 +01:00
|
|
|
return '\n'.join(lines)
|
|
|
|
|
|
|
|
|
|
|
|
def parse_assigned_metadata_initial(csvfile):
|
|
|
|
"""
|
|
|
|
Fields:
|
|
|
|
:0: Module name
|
|
|
|
:1: Core (x if so)
|
|
|
|
:2: Extras (x if so)
|
|
|
|
:3: Category
|
|
|
|
:4: Supported/SLA
|
2017-03-14 17:07:22 +01:00
|
|
|
:5: Curated
|
2016-11-15 00:31:02 +01:00
|
|
|
:6: Stable
|
|
|
|
:7: Deprecated
|
|
|
|
:8: Notes
|
|
|
|
:9: Team Notes
|
2016-12-05 15:44:37 +01:00
|
|
|
:10: Notes 2
|
|
|
|
:11: final supported_by field
|
2016-11-15 00:31:02 +01:00
|
|
|
"""
|
|
|
|
with open(csvfile, 'rb') as f:
|
|
|
|
for record in csv.reader(f):
|
|
|
|
module = record[0]
|
|
|
|
|
2016-12-06 15:27:10 +01:00
|
|
|
if record[12] == 'core':
|
2016-11-15 00:31:02 +01:00
|
|
|
supported_by = 'core'
|
2016-12-06 15:27:10 +01:00
|
|
|
elif record[12] == 'curated':
|
2017-03-14 17:07:22 +01:00
|
|
|
supported_by = 'curated'
|
2016-12-06 15:27:10 +01:00
|
|
|
elif record[12] == 'community':
|
2016-12-05 15:44:37 +01:00
|
|
|
supported_by = 'community'
|
|
|
|
else:
|
2016-12-06 15:27:10 +01:00
|
|
|
print('Module %s has no supported_by field. Using community' % record[0])
|
|
|
|
supported_by = 'community'
|
2016-12-06 22:17:49 +01:00
|
|
|
supported_by = DEFAULT_METADATA['supported_by']
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
status = []
|
|
|
|
if record[6]:
|
|
|
|
status.append('stableinterface')
|
|
|
|
if record[7]:
|
|
|
|
status.append('deprecated')
|
|
|
|
if not status:
|
2016-12-06 22:17:49 +01:00
|
|
|
status.extend(DEFAULT_METADATA['status'])
|
2016-11-15 00:31:02 +01:00
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
yield (module, {'version': DEFAULT_METADATA['metadata_version'], 'supported_by': supported_by, 'status': status})
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
|
|
|
|
def parse_assigned_metadata(csvfile):
|
|
|
|
"""
|
|
|
|
Fields:
|
|
|
|
:0: Module name
|
|
|
|
:1: supported_by string. One of the valid support fields
|
2017-03-14 17:07:22 +01:00
|
|
|
core, community, curated
|
2016-11-15 00:31:02 +01:00
|
|
|
:2: stableinterface
|
|
|
|
:3: preview
|
|
|
|
:4: deprecated
|
|
|
|
:5: removed
|
|
|
|
|
|
|
|
https://github.com/ansible/proposals/issues/30
|
|
|
|
"""
|
|
|
|
with open(csvfile, 'rb') as f:
|
|
|
|
for record in csv.reader(f):
|
|
|
|
module = record[0]
|
|
|
|
supported_by = record[1]
|
|
|
|
|
|
|
|
status = []
|
|
|
|
if record[2]:
|
|
|
|
status.append('stableinterface')
|
|
|
|
if record[4]:
|
|
|
|
status.append('deprecated')
|
|
|
|
if record[5]:
|
|
|
|
status.append('removed')
|
|
|
|
if not status or record[3]:
|
|
|
|
status.append('preview')
|
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
yield (module, {'metadata_version': '1.0', 'supported_by': supported_by, 'status': status})
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
|
|
|
|
def write_metadata(filename, new_metadata, version=None, overwrite=False):
|
|
|
|
with open(filename, 'rb') as f:
|
|
|
|
module_data = f.read()
|
|
|
|
|
|
|
|
try:
|
|
|
|
current_metadata, start_line, start_col, end_line, end_col, targets = extract_metadata(module_data)
|
|
|
|
except SyntaxError:
|
|
|
|
if filename.endswith('.py'):
|
|
|
|
raise
|
|
|
|
# Probably non-python modules. These should all have python
|
|
|
|
# documentation files where we can place the data
|
|
|
|
raise ParseError('Could not add metadata to {}'.format(filename))
|
|
|
|
|
|
|
|
if current_metadata is None:
|
|
|
|
# No curent metadata so we can just add it
|
|
|
|
start_line = find_documentation(module_data)
|
|
|
|
if start_line < 0:
|
|
|
|
if os.path.basename(filename) in NONMODULE_PY_FILES:
|
|
|
|
# These aren't new-style modules
|
|
|
|
return
|
|
|
|
|
|
|
|
raise Exception('Module file {} had no ANSIBLE_METADATA or DOCUMENTATION'.format(filename))
|
|
|
|
|
|
|
|
module_data = insert_metadata(module_data, new_metadata, start_line, targets=('ANSIBLE_METADATA',))
|
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
elif overwrite or (version is not None and ('metadata_version' not in current_metadata or
|
|
|
|
StrictVersion(current_metadata['metadata_version']) < StrictVersion(version))):
|
2016-11-15 00:31:02 +01:00
|
|
|
# Current metadata that we do not want. Remove the current
|
|
|
|
# metadata and put the new version in its place
|
|
|
|
module_data = remove_metadata(module_data, start_line, start_col, end_line, end_col)
|
|
|
|
module_data = insert_metadata(module_data, new_metadata, start_line, targets=targets)
|
|
|
|
|
|
|
|
else:
|
|
|
|
# Current metadata and we don't want to overwrite it
|
|
|
|
return
|
|
|
|
|
|
|
|
# Save the new version of the module
|
|
|
|
with open(filename, 'wb') as f:
|
|
|
|
f.write(module_data)
|
|
|
|
|
|
|
|
|
2016-12-06 22:17:49 +01:00
|
|
|
def return_metadata(plugins):
|
2017-03-14 17:07:22 +01:00
|
|
|
"""Get the metadata for all modules
|
2016-12-06 22:17:49 +01:00
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
Handle duplicate module names
|
|
|
|
|
|
|
|
:arg plugins: List of plugins to look for
|
|
|
|
:returns: Mapping of plugin name to metadata dictionary
|
|
|
|
"""
|
2016-12-06 22:17:49 +01:00
|
|
|
metadata = {}
|
|
|
|
for name, filename in plugins:
|
|
|
|
# There may be several files for a module (if it is written in another
|
|
|
|
# language, for instance) but only one of them (the .py file) should
|
|
|
|
# contain the metadata.
|
|
|
|
if name not in metadata or metadata[name] is not None:
|
|
|
|
with open(filename, 'rb') as f:
|
|
|
|
module_data = f.read()
|
|
|
|
metadata[name] = extract_metadata(module_data)[0]
|
|
|
|
return metadata
|
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
def metadata_summary(plugins, version=None):
|
|
|
|
"""Compile information about the metadata status for a list of modules
|
|
|
|
|
|
|
|
:arg plugins: List of plugins to look for. Each entry in the list is
|
|
|
|
a tuple of (module name, full path to module)
|
|
|
|
:kwarg version: If given, make sure the modules have this version of
|
2016-12-06 22:17:49 +01:00
|
|
|
metadata or higher.
|
2016-11-15 00:31:02 +01:00
|
|
|
:returns: A tuple consisting of a list of modules with no metadata at the
|
|
|
|
required version and a list of files that have metadata at the
|
|
|
|
required version.
|
|
|
|
"""
|
|
|
|
no_metadata = {}
|
|
|
|
has_metadata = {}
|
2016-12-06 22:17:49 +01:00
|
|
|
supported_by = defaultdict(set)
|
|
|
|
status = defaultdict(set)
|
2017-03-14 17:07:22 +01:00
|
|
|
requested_version = StrictVersion(version)
|
2016-12-06 22:17:49 +01:00
|
|
|
|
|
|
|
all_mods_metadata = return_metadata(plugins)
|
2016-11-15 00:31:02 +01:00
|
|
|
for name, filename in plugins:
|
2016-12-06 22:17:49 +01:00
|
|
|
# Does the module have metadata?
|
2016-11-15 00:31:02 +01:00
|
|
|
if name not in no_metadata and name not in has_metadata:
|
2016-12-06 22:17:49 +01:00
|
|
|
metadata = all_mods_metadata[name]
|
2016-11-15 00:31:02 +01:00
|
|
|
if metadata is None:
|
|
|
|
no_metadata[name] = filename
|
2017-03-14 17:07:22 +01:00
|
|
|
elif version is not None and ('metadata_version' not in metadata or StrictVersion(metadata['metadata_version']) < requested_version):
|
2016-11-15 00:31:02 +01:00
|
|
|
no_metadata[name] = filename
|
|
|
|
else:
|
|
|
|
has_metadata[name] = filename
|
|
|
|
|
2016-12-06 22:17:49 +01:00
|
|
|
# What categories does the plugin belong in?
|
|
|
|
if all_mods_metadata[name] is None:
|
|
|
|
# No metadata for this module. Use the default metadata
|
|
|
|
supported_by[DEFAULT_METADATA['supported_by']].add(filename)
|
|
|
|
status[DEFAULT_METADATA['status'][0]].add(filename)
|
|
|
|
else:
|
|
|
|
supported_by[all_mods_metadata[name]['supported_by']].add(filename)
|
|
|
|
for one_status in all_mods_metadata[name]['status']:
|
|
|
|
status[one_status].add(filename)
|
|
|
|
|
|
|
|
return list(no_metadata.values()), list(has_metadata.values()), supported_by, status
|
2016-11-15 00:31:02 +01:00
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
# Filters to convert between metadata versions
|
2017-05-16 19:52:07 +02:00
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
|
|
|
|
def convert_metadata_pre_1_0_to_1_0(metadata):
|
|
|
|
"""
|
|
|
|
Convert pre-1.0 to 1.0 metadata format
|
|
|
|
|
|
|
|
:arg metadata: The old metadata
|
|
|
|
:returns: The new metadata
|
|
|
|
|
|
|
|
Changes from pre-1.0 to 1.0:
|
|
|
|
* ``version`` field renamed to ``metadata_version``
|
|
|
|
* ``supported_by`` field value ``unmaintained`` has been removed (change to
|
|
|
|
``community`` and let an external list track whether a module is unmaintained)
|
|
|
|
* ``supported_by`` field value ``committer`` has been renamed to ``curated``
|
|
|
|
"""
|
|
|
|
new_metadata = {'metadata_version': '1.0',
|
|
|
|
'supported_by': metadata['supported_by'],
|
|
|
|
'status': metadata['status']
|
|
|
|
}
|
|
|
|
if new_metadata['supported_by'] == 'unmaintained':
|
|
|
|
new_metadata['supported_by'] = 'community'
|
|
|
|
elif new_metadata['supported_by'] == 'committer':
|
|
|
|
new_metadata['supported_by'] = 'curated'
|
|
|
|
|
|
|
|
return new_metadata
|
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
# Subcommands
|
2017-05-16 19:52:07 +02:00
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
def add_from_csv(csv_file, version=None, overwrite=False):
|
|
|
|
"""Implement the subcommand to add metadata from a csv file
|
|
|
|
"""
|
|
|
|
# Add metadata for everything from the CSV file
|
|
|
|
diagnostic_messages = []
|
|
|
|
for module_name, new_metadata in parse_assigned_metadata_initial(csv_file):
|
|
|
|
filename = module_loader.find_plugin(module_name, mod_type='.py')
|
|
|
|
if filename is None:
|
|
|
|
diagnostic_messages.append('Unable to find the module file for {}'.format(module_name))
|
|
|
|
continue
|
|
|
|
|
|
|
|
try:
|
|
|
|
write_metadata(filename, new_metadata, version, overwrite)
|
|
|
|
except ParseError as e:
|
|
|
|
diagnostic_messages.append(e.args[0])
|
|
|
|
continue
|
|
|
|
|
|
|
|
if diagnostic_messages:
|
|
|
|
pprint(diagnostic_messages)
|
|
|
|
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
|
|
|
def add_default(version=None, overwrite=False):
|
|
|
|
"""Implement the subcommand to add default metadata to modules
|
|
|
|
|
|
|
|
Add the default metadata to any plugin which lacks it.
|
|
|
|
:kwarg version: If given, the metadata must be at least this version.
|
|
|
|
Otherwise, treat the module as not having existing metadata.
|
|
|
|
:kwarg overwrite: If True, overwrite any existing metadata. Otherwise,
|
|
|
|
do not modify files which have metadata at an appropriate version
|
|
|
|
"""
|
|
|
|
# List of all plugins
|
|
|
|
plugins = module_loader.all(path_only=True)
|
|
|
|
plugins = ((os.path.splitext((os.path.basename(p)))[0], p) for p in plugins)
|
|
|
|
plugins = (p for p in plugins if p[0] not in NONMODULE_MODULE_NAMES)
|
|
|
|
|
|
|
|
# Iterate through each plugin
|
|
|
|
processed = set()
|
|
|
|
diagnostic_messages = []
|
|
|
|
for name, filename in (info for info in plugins if info[0] not in processed):
|
|
|
|
try:
|
2016-12-06 22:17:49 +01:00
|
|
|
write_metadata(filename, DEFAULT_METADATA, version, overwrite)
|
2016-11-15 00:31:02 +01:00
|
|
|
except ParseError as e:
|
|
|
|
diagnostic_messages.append(e.args[0])
|
|
|
|
continue
|
|
|
|
processed.add(name)
|
|
|
|
|
|
|
|
if diagnostic_messages:
|
|
|
|
pprint(diagnostic_messages)
|
|
|
|
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
2017-03-14 17:07:22 +01:00
|
|
|
def upgrade_metadata(version=None):
|
|
|
|
"""Implement the subcommand to upgrade the default metadata in modules.
|
|
|
|
|
|
|
|
:kwarg version: If given, the version of the metadata to upgrade to. If
|
|
|
|
not given, upgrade to the latest format version.
|
|
|
|
"""
|
|
|
|
if version is None:
|
|
|
|
# Number larger than any of the defined metadata formats.
|
|
|
|
version = 9999999
|
|
|
|
requested_version = StrictVersion(version)
|
|
|
|
|
|
|
|
# List all plugins
|
|
|
|
plugins = module_loader.all(path_only=True)
|
|
|
|
plugins = ((os.path.splitext((os.path.basename(p)))[0], p) for p in plugins)
|
|
|
|
plugins = (p for p in plugins if p[0] not in NONMODULE_MODULE_NAMES)
|
|
|
|
|
|
|
|
processed = set()
|
|
|
|
diagnostic_messages = []
|
|
|
|
for name, filename in (info for info in plugins if info[0] not in processed):
|
|
|
|
# For each plugin, read the existing metadata
|
|
|
|
with open(filename, 'rb') as f:
|
|
|
|
module_data = f.read()
|
|
|
|
metadata = extract_metadata(module_data)[0]
|
|
|
|
|
|
|
|
# If the metadata isn't the requested version, convert it to the new
|
|
|
|
# version
|
|
|
|
if 'metadata_version' not in metadata or metadata['metadata_version'] != version:
|
|
|
|
#
|
|
|
|
# With each iteration of metadata, add a new conditional to
|
|
|
|
# upgrade from the previous version
|
|
|
|
#
|
|
|
|
|
|
|
|
if 'metadata_version' not in metadata:
|
|
|
|
# First version, pre-1.0 final metadata
|
|
|
|
metadata = convert_metadata_pre_1_0_to_1_0(metadata)
|
|
|
|
|
|
|
|
if metadata['metadata_version'] == '1.0' and StrictVersion('1.0') < requested_version:
|
|
|
|
# 1.0 version => XXX. We don't yet have anything beyond 1.0
|
|
|
|
# so there's nothing here
|
|
|
|
pass
|
|
|
|
|
|
|
|
# Replace the existing metadata with the new format
|
|
|
|
try:
|
|
|
|
write_metadata(filename, metadata, version, overwrite=True)
|
|
|
|
except ParseError as e:
|
|
|
|
diagnostic_messages.append(e.args[0])
|
|
|
|
continue
|
|
|
|
|
|
|
|
processed.add(name)
|
|
|
|
|
|
|
|
if diagnostic_messages:
|
|
|
|
pprint(diagnostic_messages)
|
|
|
|
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
def report(version=None):
|
|
|
|
"""Implement the report subcommand
|
|
|
|
|
|
|
|
Print out all the modules that have metadata and all the ones that do not.
|
|
|
|
|
|
|
|
:kwarg version: If given, the metadata must be at least this version.
|
|
|
|
Otherwise return it as not having metadata
|
|
|
|
"""
|
|
|
|
# List of all plugins
|
|
|
|
plugins = module_loader.all(path_only=True)
|
|
|
|
plugins = ((os.path.splitext((os.path.basename(p)))[0], p) for p in plugins)
|
2017-03-14 17:07:22 +01:00
|
|
|
plugins = (p for p in plugins if p[0] not in NONMODULE_MODULE_NAMES)
|
2016-12-06 22:17:49 +01:00
|
|
|
plugins = list(plugins)
|
2016-11-15 00:31:02 +01:00
|
|
|
|
2016-12-06 22:17:49 +01:00
|
|
|
no_metadata, has_metadata, support, status = metadata_summary(plugins, version=version)
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
print('== Has metadata ==')
|
|
|
|
pprint(sorted(has_metadata))
|
2016-12-06 22:17:49 +01:00
|
|
|
print('')
|
|
|
|
|
2016-11-15 00:31:02 +01:00
|
|
|
print('== Has no metadata ==')
|
|
|
|
pprint(sorted(no_metadata))
|
|
|
|
print('')
|
2016-12-06 22:17:49 +01:00
|
|
|
|
|
|
|
print('== Supported by core ==')
|
|
|
|
pprint(sorted(support['core']))
|
2017-03-14 17:07:22 +01:00
|
|
|
print('== Supported by value curated ==')
|
|
|
|
pprint(sorted(support['curated']))
|
2016-12-06 22:17:49 +01:00
|
|
|
print('== Supported by community ==')
|
|
|
|
pprint(sorted(support['community']))
|
|
|
|
print('')
|
|
|
|
|
|
|
|
print('== Status: stableinterface ==')
|
|
|
|
pprint(sorted(status['stableinterface']))
|
|
|
|
print('== Status: preview ==')
|
|
|
|
pprint(sorted(status['preview']))
|
|
|
|
print('== Status: deprecated ==')
|
|
|
|
pprint(sorted(status['deprecated']))
|
|
|
|
print('== Status: removed ==')
|
|
|
|
pprint(sorted(status['removed']))
|
|
|
|
print('')
|
|
|
|
|
|
|
|
print('== Summary ==')
|
|
|
|
print('No Metadata: {0} Has Metadata: {1}'.format(len(no_metadata), len(has_metadata)))
|
2017-03-14 17:07:22 +01:00
|
|
|
print('Supported by core: {0} Supported by community: {1} Supported by value curated: {2}'.format(len(support['core']),
|
2017-05-16 19:52:07 +02:00
|
|
|
len(support['community']), len(support['curated'])))
|
2017-02-07 16:49:55 +01:00
|
|
|
print('Status StableInterface: {0} Status Preview: {1} Status Deprecated: {2} Status Removed: {3}'.format(len(status['stableinterface']),
|
2017-05-16 19:52:07 +02:00
|
|
|
len(status['preview']), len(status['deprecated']), len(status['removed'])))
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
return 0
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
action, args = parse_args(sys.argv[1:])
|
|
|
|
|
|
|
|
if action == 'report':
|
|
|
|
rc = report(version=args['version'])
|
|
|
|
elif action == 'add':
|
|
|
|
rc = add_from_csv(args['csvfile'], version=args['version'], overwrite=args['overwrite'])
|
|
|
|
elif action == 'add-default':
|
|
|
|
rc = add_default(version=args['version'], overwrite=args['overwrite'])
|
2017-03-14 17:07:22 +01:00
|
|
|
elif action == 'upgrade':
|
|
|
|
rc = upgrade_metadata(version=args['version'])
|
2016-11-15 00:31:02 +01:00
|
|
|
|
|
|
|
sys.exit(rc)
|