2016-11-15 00:31:02 +01:00
#!/usr/bin/env python
import ast
import csv
import os
import sys
2016-12-06 22:17:49 +01:00
from collections import defaultdict
2016-11-15 00:31:02 +01:00
from distutils . version import StrictVersion
2016-12-06 15:27:10 +01:00
from pprint import pformat , pprint
2016-11-15 00:31:02 +01:00
import yaml
from ansible . module_utils . _text import to_text
from ansible . plugins import module_loader
# There's a few files that are not new-style modules. Have to blacklist them
NONMODULE_PY_FILES = frozenset ( ( ' async_wrapper.py ' , ) )
NONMODULE_MODULE_NAMES = frozenset ( os . path . splitext ( p ) [ 0 ] for p in NONMODULE_PY_FILES )
2016-12-06 22:17:49 +01:00
# Default metadata
DEFAULT_METADATA = { ' version ' : ' 1.0 ' , ' status ' : [ ' preview ' ] , ' supported_by ' : ' community ' }
2016-11-15 00:31:02 +01:00
class ParseError ( Exception ) :
""" Thrown when parsing a file fails """
pass
class MissingModuleError ( Exception ) :
""" Thrown when unable to find a plugin """
pass
def usage ( ) :
print ( """ Usage:
metadata - tester . py report [ - - version X ]
metadata - tester . py add [ - - version X ] [ - - overwrite ] CSVFILE
metadata - tester . py add - default [ - - version X ] [ - - overwrite ] """ )
sys . exit ( 1 )
def parse_args ( arg_string ) :
if len ( arg_string ) < 1 :
usage ( )
action = arg_string [ 0 ]
version = None
if ' --version ' in arg_string :
version_location = arg_string . index ( ' --version ' )
arg_string . pop ( version_location )
version = arg_string . pop ( version_location )
overwrite = False
if ' --overwrite ' in arg_string :
overwrite = True
arg_string . remove ( ' --overwrite ' )
csvfile = None
if len ( arg_string ) == 2 :
csvfile = arg_string [ 1 ]
elif len ( arg_string ) > 2 :
usage ( )
return action , { ' version ' : version , ' overwrite ' : overwrite , ' csvfile ' : csvfile }
def seek_end_of_dict ( module_data , start_line , start_col , next_node_line , next_node_col ) :
""" Look for the end of a dict in a set of lines
We know the starting position of the dict and we know the start of the
next code node but in between there may be multiple newlines and comments .
There may also be multiple python statements on the same line ( separated
by semicolons )
Examples : :
ANSIBLE_METADATA = { [ . . ] }
DOCUMENTATION = [ . . ]
ANSIBLE_METADATA = { [ . . ] } # Optional comments with confusing junk => {}
# Optional comments {}
DOCUMENTATION = [ . . ]
ANSIBLE_METADATA = {
[ . . ]
}
# Optional comments {}
DOCUMENTATION = [ . . ]
ANSIBLE_METADATA = { [ . . ] } ; DOCUMENTATION = [ . . ]
ANSIBLE_METADATA = { } EOF
"""
if next_node_line == None :
# The dict is the last statement in the file
snippet = module_data . splitlines ( ) [ start_line : ]
next_node_col = 0
# Include the last line in the file
last_line_offset = 0
else :
# It's somewhere in the middle so we need to separate it from the rest
snippet = module_data . splitlines ( ) [ start_line : next_node_line ]
# Do not include the last line because that's where the next node
# starts
last_line_offset = 1
if next_node_col == 0 :
# This handles all variants where there are only comments and blank
# lines between the dict and the next code node
# Step backwards through all the lines in the snippet
for line_idx , line in tuple ( reversed ( tuple ( enumerate ( snippet ) ) ) ) [ last_line_offset : ] :
end_col = None
# Step backwards through all the characters in the line
for col_idx , char in reversed ( tuple ( enumerate ( c for c in line ) ) ) :
if char == ' } ' and end_col is None :
# Potentially found the end of the dict
end_col = col_idx
elif char == ' # ' and end_col is not None :
# The previous '}' was part of a comment. Keep trying
end_col = None
if end_col is not None :
# Found the end!
end_line = start_line + line_idx
break
else :
# Harder cases involving multiple statements on one line
# Good Ansible Module style doesn't do this so we're just going to
# treat this as an error for now:
raise ParseError ( ' Multiple statements per line confuses the module metadata parser. ' )
return end_line , end_col
def seek_end_of_string ( module_data , start_line , start_col , next_node_line , next_node_col ) :
"""
This is much trickier than finding the end of a dict . A dict has only one
ending character , " } " . Strings have four potential ending characters . We
have to parse the beginning of the string to determine what the ending
character will be .
Examples :
ANSIBLE_METADATA = ''' [..] ''' # Optional comment with confusing chars '''
# Optional comment with confusing chars '''
DOCUMENTATION = [ . . ]
ANSIBLE_METADATA = '''
[ . . ]
'''
DOCUMENTATIONS = [ . . ]
ANSIBLE_METADATA = ''' [..] ''' ; DOCUMENTATION = [ . . ]
SHORT_NAME = ANSIBLE_METADATA = ''' [..] ''' ; DOCUMENTATION = [ . . ]
String marker variants :
* ' [..] '
* " [..] "
* ''' [..] '''
* \" \" \" [..] \" \" \"
Each of these come in u , r , and b variants :
* ' [..] '
* u ' [..] '
* b ' [..] '
* r ' [..] '
* ur ' [..] '
* ru ' [..] '
* br ' [..] '
* b ' [..] '
* rb ' [..] '
"""
raise NotImplementedError ( ' Finding end of string not yet implemented ' )
def extract_metadata ( module_data ) :
""" Extract the metadata from a module
: arg module_data : Byte string containing a module ' s code
: returns : a tuple of metadata ( a dict ) , line the metadata starts on ,
column the metadata starts on , line the metadata ends on , column the
metadata ends on , and the names the metadata is assigned to . One of
the names the metadata is assigned to will be ANSIBLE_METADATA If no
metadata is found , the tuple will be ( None , - 1 , - 1 , - 1 , - 1 , None )
"""
metadata = None
start_line = - 1
start_col = - 1
end_line = - 1
end_col = - 1
targets = None
mod_ast_tree = ast . parse ( module_data )
for root_idx , child in enumerate ( mod_ast_tree . body ) :
if isinstance ( child , ast . Assign ) :
for target in child . targets :
if target . id == ' ANSIBLE_METADATA ' :
if isinstance ( child . value , ast . Dict ) :
metadata = ast . literal_eval ( child . value )
try :
# Determine where the next node starts
next_node = mod_ast_tree . body [ root_idx + 1 ]
next_lineno = next_node . lineno
next_col_offset = next_node . col_offset
except IndexError :
# Metadata is defined in the last node of the file
next_lineno = None
next_col_offset = None
# Determine where the current metadata ends
end_line , end_col = seek_end_of_dict ( module_data ,
child . lineno - 1 , child . col_offset , next_lineno ,
next_col_offset )
elif isinstance ( child . value , ast . Str ) :
metadata = yaml . safe_load ( child . value . s )
end_line = seek_end_of_string ( module_data )
elif isinstance ( child . value , ast . Bytes ) :
metadata = yaml . safe_load ( to_text ( child . value . s , errors = ' surrogate_or_strict ' ) )
end_line = seek_end_of_string ( module_data )
else :
# Example:
# ANSIBLE_METADATA = 'junk'
# ANSIBLE_METADATA = { [..the real metadata..] }
continue
# Do these after the if-else so we don't pollute them in
# case this was a false positive
start_line = child . lineno - 1
start_col = child . col_offset
targets = [ t . id for t in child . targets ]
break
if metadata is not None :
# Once we've found the metadata we're done
break
return metadata , start_line , start_col , end_line , end_col , targets
def find_documentation ( module_data ) :
""" Find the DOCUMENTATION metadata for a module file """
start_line = - 1
mod_ast_tree = ast . parse ( module_data )
for child in mod_ast_tree . body :
if isinstance ( child , ast . Assign ) :
for target in child . targets :
if target . id == ' DOCUMENTATION ' :
start_line = child . lineno - 1
break
return start_line
def remove_metadata ( module_data , start_line , start_col , end_line , end_col ) :
""" Remove a section of a module file """
lines = module_data . split ( ' \n ' )
new_lines = lines [ : start_line ]
if start_col != 0 :
new_lines . append ( lines [ start_line ] [ : start_col ] )
next_line = lines [ end_line ]
if len ( next_line ) - 1 != end_col :
new_lines . append ( next_line [ end_col : ] )
if len ( lines ) > end_line :
new_lines . extend ( lines [ end_line + 1 : ] )
return ' \n ' . join ( new_lines )
def insert_metadata ( module_data , new_metadata , insertion_line , targets = ( ' ANSIBLE_METADATA ' , ) ) :
""" Insert a new set of metadata at a specified line """
2016-12-06 15:27:10 +01:00
assignments = ' = ' . join ( targets )
pretty_metadata = pformat ( new_metadata , width = 1 ) . split ( ' \n ' )
new_lines = [ ]
new_lines . append ( ' {} = {} ' . format ( assignments , pretty_metadata [ 0 ] ) )
if len ( pretty_metadata ) > 1 :
for line in pretty_metadata [ 1 : ] :
new_lines . append ( ' {} {} ' . format ( ' ' * ( len ( assignments ) - 1 + len ( ' = { ' ) ) , line ) )
old_lines = module_data . split ( ' \n ' )
lines = old_lines [ : insertion_line ] + new_lines + [ ' ' ] + old_lines [ insertion_line : ]
2016-11-15 00:31:02 +01:00
return ' \n ' . join ( lines )
def parse_assigned_metadata_initial ( csvfile ) :
"""
Fields :
: 0 : Module name
: 1 : Core ( x if so )
: 2 : Extras ( x if so )
: 3 : Category
: 4 : Supported / SLA
2016-12-05 13:26:08 +01:00
: 5 : Committer
2016-11-15 00:31:02 +01:00
: 6 : Stable
: 7 : Deprecated
: 8 : Notes
: 9 : Team Notes
2016-12-05 15:44:37 +01:00
: 10 : Notes 2
: 11 : final supported_by field
2016-11-15 00:31:02 +01:00
"""
with open ( csvfile , ' rb ' ) as f :
for record in csv . reader ( f ) :
module = record [ 0 ]
2016-12-06 15:27:10 +01:00
if record [ 12 ] == ' core ' :
2016-11-15 00:31:02 +01:00
supported_by = ' core '
2016-12-06 15:27:10 +01:00
elif record [ 12 ] == ' curated ' :
2016-12-05 13:26:08 +01:00
supported_by = ' committer '
2016-12-06 15:27:10 +01:00
elif record [ 12 ] == ' community ' :
2016-12-05 15:44:37 +01:00
supported_by = ' community '
else :
2016-12-06 15:27:10 +01:00
print ( ' Module %s has no supported_by field. Using community ' % record [ 0 ] )
supported_by = ' community '
2016-12-06 22:17:49 +01:00
supported_by = DEFAULT_METADATA [ ' supported_by ' ]
2016-11-15 00:31:02 +01:00
status = [ ]
if record [ 6 ] :
status . append ( ' stableinterface ' )
if record [ 7 ] :
status . append ( ' deprecated ' )
if not status :
2016-12-06 22:17:49 +01:00
status . extend ( DEFAULT_METADATA [ ' status ' ] )
2016-11-15 00:31:02 +01:00
2016-12-06 22:17:49 +01:00
yield ( module , { ' version ' : DEFAULT_METADATA [ ' version ' ] , ' supported_by ' : supported_by , ' status ' : status } )
2016-11-15 00:31:02 +01:00
def parse_assigned_metadata ( csvfile ) :
"""
Fields :
: 0 : Module name
: 1 : supported_by string . One of the valid support fields
2016-12-05 13:26:08 +01:00
core , community , unmaintained , committer
2016-11-15 00:31:02 +01:00
: 2 : stableinterface
: 3 : preview
: 4 : deprecated
: 5 : removed
: 6 : tested
https : / / github . com / ansible / proposals / issues / 30
"""
with open ( csvfile , ' rb ' ) as f :
for record in csv . reader ( f ) :
module = record [ 0 ]
supported_by = record [ 1 ]
status = [ ]
if record [ 2 ] :
status . append ( ' stableinterface ' )
if record [ 4 ] :
status . append ( ' deprecated ' )
if record [ 5 ] :
status . append ( ' removed ' )
if record [ 6 ] :
status . append ( ' tested ' )
if not status or record [ 3 ] :
status . append ( ' preview ' )
yield ( module , { ' version ' : ' 1.0 ' , ' supported_by ' : supported_by , ' status ' : status } )
def write_metadata ( filename , new_metadata , version = None , overwrite = False ) :
with open ( filename , ' rb ' ) as f :
module_data = f . read ( )
try :
current_metadata , start_line , start_col , end_line , end_col , targets = extract_metadata ( module_data )
except SyntaxError :
if filename . endswith ( ' .py ' ) :
raise
# Probably non-python modules. These should all have python
# documentation files where we can place the data
raise ParseError ( ' Could not add metadata to {} ' . format ( filename ) )
if current_metadata is None :
# No curent metadata so we can just add it
start_line = find_documentation ( module_data )
if start_line < 0 :
if os . path . basename ( filename ) in NONMODULE_PY_FILES :
# These aren't new-style modules
return
raise Exception ( ' Module file {} had no ANSIBLE_METADATA or DOCUMENTATION ' . format ( filename ) )
module_data = insert_metadata ( module_data , new_metadata , start_line , targets = ( ' ANSIBLE_METADATA ' , ) )
elif overwrite or ( version is not None and ( ' version ' not in current_metadata or StrictVersion ( current_metadata [ ' version ' ] ) < StrictVersion ( version ) ) ) :
# Current metadata that we do not want. Remove the current
# metadata and put the new version in its place
module_data = remove_metadata ( module_data , start_line , start_col , end_line , end_col )
module_data = insert_metadata ( module_data , new_metadata , start_line , targets = targets )
else :
# Current metadata and we don't want to overwrite it
return
# Save the new version of the module
with open ( filename , ' wb ' ) as f :
f . write ( module_data )
2016-12-06 22:17:49 +01:00
def return_metadata ( plugins ) :
metadata = { }
for name , filename in plugins :
# There may be several files for a module (if it is written in another
# language, for instance) but only one of them (the .py file) should
# contain the metadata.
if name not in metadata or metadata [ name ] is not None :
with open ( filename , ' rb ' ) as f :
module_data = f . read ( )
metadata [ name ] = extract_metadata ( module_data ) [ 0 ]
return metadata
2016-11-15 00:31:02 +01:00
def metadata_summary ( plugins , version = None ) :
""" Compile information about the metadata status for a list of modules
: arg plugins : List of plugins to look for . Each entry in the list is
a tuple of ( module name , full path to module )
: kwarg version : If given , make sure the modules have this version of
2016-12-06 22:17:49 +01:00
metadata or higher .
2016-11-15 00:31:02 +01:00
: returns : A tuple consisting of a list of modules with no metadata at the
required version and a list of files that have metadata at the
required version .
"""
no_metadata = { }
has_metadata = { }
2016-12-06 22:17:49 +01:00
supported_by = defaultdict ( set )
status = defaultdict ( set )
plugins = list ( plugins )
all_mods_metadata = return_metadata ( plugins )
2016-11-15 00:31:02 +01:00
for name , filename in plugins :
2016-12-06 22:17:49 +01:00
# Does the module have metadata?
2016-11-15 00:31:02 +01:00
if name not in no_metadata and name not in has_metadata :
2016-12-06 22:17:49 +01:00
metadata = all_mods_metadata [ name ]
2016-11-15 00:31:02 +01:00
if metadata is None :
no_metadata [ name ] = filename
elif version is not None and ( ' version ' not in metadata or StrictVersion ( metadata [ ' version ' ] ) < StrictVersion ( version ) ) :
no_metadata [ name ] = filename
else :
has_metadata [ name ] = filename
2016-12-06 22:17:49 +01:00
# What categories does the plugin belong in?
if all_mods_metadata [ name ] is None :
# No metadata for this module. Use the default metadata
supported_by [ DEFAULT_METADATA [ ' supported_by ' ] ] . add ( filename )
status [ DEFAULT_METADATA [ ' status ' ] [ 0 ] ] . add ( filename )
else :
supported_by [ all_mods_metadata [ name ] [ ' supported_by ' ] ] . add ( filename )
for one_status in all_mods_metadata [ name ] [ ' status ' ] :
status [ one_status ] . add ( filename )
return list ( no_metadata . values ( ) ) , list ( has_metadata . values ( ) ) , supported_by , status
2016-11-15 00:31:02 +01:00
#
# Subcommands
#
def add_from_csv ( csv_file , version = None , overwrite = False ) :
""" Implement the subcommand to add metadata from a csv file
"""
# Add metadata for everything from the CSV file
diagnostic_messages = [ ]
for module_name , new_metadata in parse_assigned_metadata_initial ( csv_file ) :
filename = module_loader . find_plugin ( module_name , mod_type = ' .py ' )
if filename is None :
diagnostic_messages . append ( ' Unable to find the module file for {} ' . format ( module_name ) )
continue
try :
write_metadata ( filename , new_metadata , version , overwrite )
except ParseError as e :
diagnostic_messages . append ( e . args [ 0 ] )
continue
if diagnostic_messages :
pprint ( diagnostic_messages )
return 0
def add_default ( version = None , overwrite = False ) :
""" Implement the subcommand to add default metadata to modules
Add the default metadata to any plugin which lacks it .
: kwarg version : If given , the metadata must be at least this version .
Otherwise , treat the module as not having existing metadata .
: kwarg overwrite : If True , overwrite any existing metadata . Otherwise ,
do not modify files which have metadata at an appropriate version
"""
# List of all plugins
plugins = module_loader . all ( path_only = True )
plugins = ( ( os . path . splitext ( ( os . path . basename ( p ) ) ) [ 0 ] , p ) for p in plugins )
plugins = ( p for p in plugins if p [ 0 ] not in NONMODULE_MODULE_NAMES )
# Iterate through each plugin
processed = set ( )
diagnostic_messages = [ ]
for name , filename in ( info for info in plugins if info [ 0 ] not in processed ) :
try :
2016-12-06 22:17:49 +01:00
write_metadata ( filename , DEFAULT_METADATA , version , overwrite )
2016-11-15 00:31:02 +01:00
except ParseError as e :
diagnostic_messages . append ( e . args [ 0 ] )
continue
processed . add ( name )
if diagnostic_messages :
pprint ( diagnostic_messages )
return 0
def report ( version = None ) :
""" Implement the report subcommand
Print out all the modules that have metadata and all the ones that do not .
: kwarg version : If given , the metadata must be at least this version .
Otherwise return it as not having metadata
"""
# List of all plugins
plugins = module_loader . all ( path_only = True )
2016-12-06 22:17:49 +01:00
plugins = list ( plugins )
2016-11-15 00:31:02 +01:00
plugins = ( ( os . path . splitext ( ( os . path . basename ( p ) ) ) [ 0 ] , p ) for p in plugins )
plugins = ( p for p in plugins if p [ 0 ] != NONMODULE_MODULE_NAMES )
2016-12-06 22:17:49 +01:00
plugins = list ( plugins )
2016-11-15 00:31:02 +01:00
2016-12-06 22:17:49 +01:00
no_metadata , has_metadata , support , status = metadata_summary ( plugins , version = version )
2016-11-15 00:31:02 +01:00
print ( ' == Has metadata == ' )
pprint ( sorted ( has_metadata ) )
2016-12-06 22:17:49 +01:00
print ( ' ' )
2016-11-15 00:31:02 +01:00
print ( ' == Has no metadata == ' )
pprint ( sorted ( no_metadata ) )
print ( ' ' )
2016-12-06 22:17:49 +01:00
print ( ' == Supported by core == ' )
pprint ( sorted ( support [ ' core ' ] ) )
print ( ' == Supported by committers == ' )
pprint ( sorted ( support [ ' committer ' ] ) )
print ( ' == Supported by community == ' )
pprint ( sorted ( support [ ' community ' ] ) )
print ( ' ' )
print ( ' == Status: stableinterface == ' )
pprint ( sorted ( status [ ' stableinterface ' ] ) )
print ( ' == Status: preview == ' )
pprint ( sorted ( status [ ' preview ' ] ) )
print ( ' == Status: deprecated == ' )
pprint ( sorted ( status [ ' deprecated ' ] ) )
print ( ' == Status: removed == ' )
pprint ( sorted ( status [ ' removed ' ] ) )
print ( ' ' )
print ( ' == Summary == ' )
print ( ' No Metadata: {0} Has Metadata: {1} ' . format ( len ( no_metadata ) , len ( has_metadata ) ) )
print ( ' Supported by core: {0} Supported by community: {1} Supported by committer: {2} ' . format ( len ( support [ ' core ' ] ) , len ( support [ ' community ' ] ) , len ( support [ ' committer ' ] ) ) )
print ( ' Status StableInterface: {0} Status Preview: {1} Status Deprecated: {2} Status Removed: {3} ' . format ( len ( status [ ' stableinterface ' ] ) , len ( status [ ' preview ' ] ) , len ( status [ ' deprecated ' ] ) , len ( status [ ' removed ' ] ) ) )
2016-11-15 00:31:02 +01:00
return 0
if __name__ == ' __main__ ' :
action , args = parse_args ( sys . argv [ 1 : ] )
### TODO: Implement upgrade metadata and upgrade metadata from csvfile
if action == ' report ' :
rc = report ( version = args [ ' version ' ] )
elif action == ' add ' :
rc = add_from_csv ( args [ ' csvfile ' ] , version = args [ ' version ' ] , overwrite = args [ ' overwrite ' ] )
elif action == ' add-default ' :
rc = add_default ( version = args [ ' version ' ] , overwrite = args [ ' overwrite ' ] )
sys . exit ( rc )