mirror of
https://github.com/thegeeklab/ansible-later.git
synced 2024-11-16 01:50:39 +00:00
1004 lines
45 KiB
Python
1004 lines
45 KiB
Python
|
# (c) 2013-2014, Michael DeHaan <michael.dehaan@gmail.com>
|
||
|
# (c) 2015 Toshio Kuratomi <tkuratomi@ansible.com>
|
||
|
#
|
||
|
# This file is part of Ansible
|
||
|
#
|
||
|
# Ansible is free software: you can redistribute it and/or modify
|
||
|
# it under the terms of the GNU General Public License as published by
|
||
|
# the Free Software Foundation, either version 3 of the License, or
|
||
|
# (at your option) any later version.
|
||
|
#
|
||
|
# Ansible is distributed in the hope that it will be useful,
|
||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
|
# GNU General Public License for more details.
|
||
|
#
|
||
|
# You should have received a copy of the GNU General Public License
|
||
|
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
|
||
|
|
||
|
# Make coding more python3-ish
|
||
|
from __future__ import (absolute_import, division, print_function)
|
||
|
__metaclass__ = type
|
||
|
|
||
|
import ast
|
||
|
import base64
|
||
|
import datetime
|
||
|
import imp
|
||
|
import json
|
||
|
import os
|
||
|
import shlex
|
||
|
import zipfile
|
||
|
import random
|
||
|
import re
|
||
|
from distutils.version import LooseVersion
|
||
|
from io import BytesIO
|
||
|
|
||
|
from ansible.release import __version__, __author__
|
||
|
from ansible import constants as C
|
||
|
from ansible.errors import AnsibleError
|
||
|
from ansible.module_utils._text import to_bytes, to_text, to_native
|
||
|
from ansible.plugins.loader import module_utils_loader, ps_module_utils_loader
|
||
|
from ansible.plugins.shell.powershell import async_watchdog, async_wrapper, become_wrapper, leaf_exec, exec_wrapper
|
||
|
# Must import strategy and use write_locks from there
|
||
|
# If we import write_locks directly then we end up binding a
|
||
|
# variable to the object and then it never gets updated.
|
||
|
from ansible.executor import action_write_locks
|
||
|
|
||
|
try:
|
||
|
from __main__ import display
|
||
|
except ImportError:
|
||
|
from ansible.utils.display import Display
|
||
|
display = Display()
|
||
|
|
||
|
|
||
|
REPLACER = b"#<<INCLUDE_ANSIBLE_MODULE_COMMON>>"
|
||
|
REPLACER_VERSION = b"\"<<ANSIBLE_VERSION>>\""
|
||
|
REPLACER_COMPLEX = b"\"<<INCLUDE_ANSIBLE_MODULE_COMPLEX_ARGS>>\""
|
||
|
REPLACER_WINDOWS = b"# POWERSHELL_COMMON"
|
||
|
REPLACER_JSONARGS = b"<<INCLUDE_ANSIBLE_MODULE_JSON_ARGS>>"
|
||
|
REPLACER_SELINUX = b"<<SELINUX_SPECIAL_FILESYSTEMS>>"
|
||
|
|
||
|
# We could end up writing out parameters with unicode characters so we need to
|
||
|
# specify an encoding for the python source file
|
||
|
ENCODING_STRING = u'# -*- coding: utf-8 -*-'
|
||
|
b_ENCODING_STRING = b'# -*- coding: utf-8 -*-'
|
||
|
|
||
|
# module_common is relative to module_utils, so fix the path
|
||
|
_MODULE_UTILS_PATH = os.path.join(os.path.dirname(__file__), '..', 'module_utils')
|
||
|
|
||
|
# ******************************************************************************
|
||
|
|
||
|
ANSIBALLZ_TEMPLATE = u'''%(shebang)s
|
||
|
%(coding)s
|
||
|
_ANSIBALLZ_WRAPPER = True # For test-module script to tell this is a ANSIBALLZ_WRAPPER
|
||
|
# This code is part of Ansible, but is an independent component.
|
||
|
# The code in this particular templatable string, and this templatable string
|
||
|
# only, is BSD licensed. Modules which end up using this snippet, which is
|
||
|
# dynamically combined together by Ansible still belong to the author of the
|
||
|
# module, and they may assign their own license to the complete work.
|
||
|
#
|
||
|
# Copyright (c), James Cammarata, 2016
|
||
|
# Copyright (c), Toshio Kuratomi, 2016
|
||
|
#
|
||
|
# Redistribution and use in source and binary forms, with or without modification,
|
||
|
# are permitted provided that the following conditions are met:
|
||
|
#
|
||
|
# * Redistributions of source code must retain the above copyright
|
||
|
# notice, this list of conditions and the following disclaimer.
|
||
|
# * Redistributions in binary form must reproduce the above copyright notice,
|
||
|
# this list of conditions and the following disclaimer in the documentation
|
||
|
# and/or other materials provided with the distribution.
|
||
|
#
|
||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
||
|
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||
|
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
||
|
# IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||
|
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||
|
# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
||
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
||
|
# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
|
||
|
# USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||
|
def _ansiballz_main():
|
||
|
import os
|
||
|
import os.path
|
||
|
import sys
|
||
|
import __main__
|
||
|
|
||
|
# For some distros and python versions we pick up this script in the temporary
|
||
|
# directory. This leads to problems when the ansible module masks a python
|
||
|
# library that another import needs. We have not figured out what about the
|
||
|
# specific distros and python versions causes this to behave differently.
|
||
|
#
|
||
|
# Tested distros:
|
||
|
# Fedora23 with python3.4 Works
|
||
|
# Ubuntu15.10 with python2.7 Works
|
||
|
# Ubuntu15.10 with python3.4 Fails without this
|
||
|
# Ubuntu16.04.1 with python3.5 Fails without this
|
||
|
# To test on another platform:
|
||
|
# * use the copy module (since this shadows the stdlib copy module)
|
||
|
# * Turn off pipelining
|
||
|
# * Make sure that the destination file does not exist
|
||
|
# * ansible ubuntu16-test -m copy -a 'src=/etc/motd dest=/var/tmp/m'
|
||
|
# This will traceback in shutil. Looking at the complete traceback will show
|
||
|
# that shutil is importing copy which finds the ansible module instead of the
|
||
|
# stdlib module
|
||
|
scriptdir = None
|
||
|
try:
|
||
|
scriptdir = os.path.dirname(os.path.realpath(__main__.__file__))
|
||
|
except (AttributeError, OSError):
|
||
|
# Some platforms don't set __file__ when reading from stdin
|
||
|
# OSX raises OSError if using abspath() in a directory we don't have
|
||
|
# permission to read (realpath calls abspath)
|
||
|
pass
|
||
|
if scriptdir is not None:
|
||
|
sys.path = [p for p in sys.path if p != scriptdir]
|
||
|
|
||
|
import base64
|
||
|
import imp
|
||
|
import shutil
|
||
|
import tempfile
|
||
|
import zipfile
|
||
|
|
||
|
if sys.version_info < (3,):
|
||
|
bytes = str
|
||
|
MOD_DESC = ('.py', 'U', imp.PY_SOURCE)
|
||
|
PY3 = False
|
||
|
else:
|
||
|
unicode = str
|
||
|
MOD_DESC = ('.py', 'r', imp.PY_SOURCE)
|
||
|
PY3 = True
|
||
|
|
||
|
ZIPDATA = """%(zipdata)s"""
|
||
|
|
||
|
# Note: temp_path isn't needed once we switch to zipimport
|
||
|
def invoke_module(modlib_path, temp_path, json_params):
|
||
|
# When installed via setuptools (including python setup.py install),
|
||
|
# ansible may be installed with an easy-install.pth file. That file
|
||
|
# may load the system-wide install of ansible rather than the one in
|
||
|
# the module. sitecustomize is the only way to override that setting.
|
||
|
z = zipfile.ZipFile(modlib_path, mode='a')
|
||
|
|
||
|
# py3: modlib_path will be text, py2: it's bytes. Need bytes at the end
|
||
|
sitecustomize = u'import sys\\nsys.path.insert(0,"%%s")\\n' %% modlib_path
|
||
|
sitecustomize = sitecustomize.encode('utf-8')
|
||
|
# Use a ZipInfo to work around zipfile limitation on hosts with
|
||
|
# clocks set to a pre-1980 year (for instance, Raspberry Pi)
|
||
|
zinfo = zipfile.ZipInfo()
|
||
|
zinfo.filename = 'sitecustomize.py'
|
||
|
zinfo.date_time = ( %(year)i, %(month)i, %(day)i, %(hour)i, %(minute)i, %(second)i)
|
||
|
z.writestr(zinfo, sitecustomize)
|
||
|
# Note: Remove the following section when we switch to zipimport
|
||
|
# Write the module to disk for imp.load_module
|
||
|
module = os.path.join(temp_path, '__main__.py')
|
||
|
with open(module, 'wb') as f:
|
||
|
f.write(z.read('__main__.py'))
|
||
|
f.close()
|
||
|
# End pre-zipimport section
|
||
|
z.close()
|
||
|
|
||
|
# Put the zipped up module_utils we got from the controller first in the python path so that we
|
||
|
# can monkeypatch the right basic
|
||
|
sys.path.insert(0, modlib_path)
|
||
|
|
||
|
# Monkeypatch the parameters into basic
|
||
|
from ansible.module_utils import basic
|
||
|
basic._ANSIBLE_ARGS = json_params
|
||
|
%(coverage)s
|
||
|
# Run the module! By importing it as '__main__', it thinks it is executing as a script
|
||
|
with open(module, 'rb') as mod:
|
||
|
imp.load_module('__main__', mod, module, MOD_DESC)
|
||
|
|
||
|
# Ansible modules must exit themselves
|
||
|
print('{"msg": "New-style module did not handle its own exit", "failed": true}')
|
||
|
sys.exit(1)
|
||
|
|
||
|
def debug(command, zipped_mod, json_params):
|
||
|
# The code here normally doesn't run. It's only used for debugging on the
|
||
|
# remote machine.
|
||
|
#
|
||
|
# The subcommands in this function make it easier to debug ansiballz
|
||
|
# modules. Here's the basic steps:
|
||
|
#
|
||
|
# Run ansible with the environment variable: ANSIBLE_KEEP_REMOTE_FILES=1 and -vvv
|
||
|
# to save the module file remotely::
|
||
|
# $ ANSIBLE_KEEP_REMOTE_FILES=1 ansible host1 -m ping -a 'data=october' -vvv
|
||
|
#
|
||
|
# Part of the verbose output will tell you where on the remote machine the
|
||
|
# module was written to::
|
||
|
# [...]
|
||
|
# <host1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
|
||
|
# PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o
|
||
|
# ControlPath=/home/badger/.ansible/cp/ansible-ssh-%%h-%%p-%%r -tt rhel7 '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8
|
||
|
# LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/badger/.ansible/tmp/ansible-tmp-1461173013.93-9076457629738/ping'"'"''
|
||
|
# [...]
|
||
|
#
|
||
|
# Login to the remote machine and run the module file via from the previous
|
||
|
# step with the explode subcommand to extract the module payload into
|
||
|
# source files::
|
||
|
# $ ssh host1
|
||
|
# $ /usr/bin/python /home/badger/.ansible/tmp/ansible-tmp-1461173013.93-9076457629738/ping explode
|
||
|
# Module expanded into:
|
||
|
# /home/badger/.ansible/tmp/ansible-tmp-1461173408.08-279692652635227/ansible
|
||
|
#
|
||
|
# You can now edit the source files to instrument the code or experiment with
|
||
|
# different parameter values. When you're ready to run the code you've modified
|
||
|
# (instead of the code from the actual zipped module), use the execute subcommand like this::
|
||
|
# $ /usr/bin/python /home/badger/.ansible/tmp/ansible-tmp-1461173013.93-9076457629738/ping execute
|
||
|
|
||
|
# Okay to use __file__ here because we're running from a kept file
|
||
|
basedir = os.path.join(os.path.abspath(os.path.dirname(__file__)), 'debug_dir')
|
||
|
args_path = os.path.join(basedir, 'args')
|
||
|
script_path = os.path.join(basedir, '__main__.py')
|
||
|
|
||
|
if command == 'excommunicate':
|
||
|
print('The excommunicate debug command is deprecated and will be removed in 2.11. Use execute instead.')
|
||
|
command = 'execute'
|
||
|
|
||
|
if command == 'explode':
|
||
|
# transform the ZIPDATA into an exploded directory of code and then
|
||
|
# print the path to the code. This is an easy way for people to look
|
||
|
# at the code on the remote machine for debugging it in that
|
||
|
# environment
|
||
|
z = zipfile.ZipFile(zipped_mod)
|
||
|
for filename in z.namelist():
|
||
|
if filename.startswith('/'):
|
||
|
raise Exception('Something wrong with this module zip file: should not contain absolute paths')
|
||
|
|
||
|
dest_filename = os.path.join(basedir, filename)
|
||
|
if dest_filename.endswith(os.path.sep) and not os.path.exists(dest_filename):
|
||
|
os.makedirs(dest_filename)
|
||
|
else:
|
||
|
directory = os.path.dirname(dest_filename)
|
||
|
if not os.path.exists(directory):
|
||
|
os.makedirs(directory)
|
||
|
f = open(dest_filename, 'wb')
|
||
|
f.write(z.read(filename))
|
||
|
f.close()
|
||
|
|
||
|
# write the args file
|
||
|
f = open(args_path, 'wb')
|
||
|
f.write(json_params)
|
||
|
f.close()
|
||
|
|
||
|
print('Module expanded into:')
|
||
|
print('%%s' %% basedir)
|
||
|
exitcode = 0
|
||
|
|
||
|
elif command == 'execute':
|
||
|
# Execute the exploded code instead of executing the module from the
|
||
|
# embedded ZIPDATA. This allows people to easily run their modified
|
||
|
# code on the remote machine to see how changes will affect it.
|
||
|
|
||
|
# Set pythonpath to the debug dir
|
||
|
sys.path.insert(0, basedir)
|
||
|
|
||
|
# read in the args file which the user may have modified
|
||
|
with open(args_path, 'rb') as f:
|
||
|
json_params = f.read()
|
||
|
|
||
|
# Monkeypatch the parameters into basic
|
||
|
from ansible.module_utils import basic
|
||
|
basic._ANSIBLE_ARGS = json_params
|
||
|
|
||
|
# Run the module! By importing it as '__main__', it thinks it is executing as a script
|
||
|
import imp
|
||
|
with open(script_path, 'r') as f:
|
||
|
importer = imp.load_module('__main__', f, script_path, ('.py', 'r', imp.PY_SOURCE))
|
||
|
|
||
|
# Ansible modules must exit themselves
|
||
|
print('{"msg": "New-style module did not handle its own exit", "failed": true}')
|
||
|
sys.exit(1)
|
||
|
|
||
|
else:
|
||
|
print('WARNING: Unknown debug command. Doing nothing.')
|
||
|
exitcode = 0
|
||
|
|
||
|
return exitcode
|
||
|
|
||
|
#
|
||
|
# See comments in the debug() method for information on debugging
|
||
|
#
|
||
|
|
||
|
ANSIBALLZ_PARAMS = %(params)s
|
||
|
if PY3:
|
||
|
ANSIBALLZ_PARAMS = ANSIBALLZ_PARAMS.encode('utf-8')
|
||
|
try:
|
||
|
# There's a race condition with the controller removing the
|
||
|
# remote_tmpdir and this module executing under async. So we cannot
|
||
|
# store this in remote_tmpdir (use system tempdir instead)
|
||
|
# Only need to use [ansible_module]_payload_ in the temp_path until we move to zipimport
|
||
|
# (this helps ansible-test produce coverage stats)
|
||
|
temp_path = tempfile.mkdtemp(prefix='ansible_%(ansible_module)s_payload_')
|
||
|
|
||
|
zipped_mod = os.path.join(temp_path, 'ansible_%(ansible_module)s_payload.zip')
|
||
|
with open(zipped_mod, 'wb') as modlib:
|
||
|
modlib.write(base64.b64decode(ZIPDATA))
|
||
|
|
||
|
if len(sys.argv) == 2:
|
||
|
exitcode = debug(sys.argv[1], zipped_mod, ANSIBALLZ_PARAMS)
|
||
|
else:
|
||
|
# Note: temp_path isn't needed once we switch to zipimport
|
||
|
invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)
|
||
|
finally:
|
||
|
try:
|
||
|
shutil.rmtree(temp_path)
|
||
|
except (NameError, OSError):
|
||
|
# tempdir creation probably failed
|
||
|
pass
|
||
|
sys.exit(exitcode)
|
||
|
|
||
|
if __name__ == '__main__':
|
||
|
_ansiballz_main()
|
||
|
'''
|
||
|
|
||
|
ANSIBALLZ_COVERAGE_TEMPLATE = '''
|
||
|
# Access to the working directory is required by coverage.
|
||
|
# Some platforms, such as macOS, may not allow querying the working directory when using become to drop privileges.
|
||
|
try:
|
||
|
os.getcwd()
|
||
|
except OSError:
|
||
|
os.chdir('/')
|
||
|
|
||
|
os.environ['COVERAGE_FILE'] = '%(coverage_output)s'
|
||
|
|
||
|
import atexit
|
||
|
import coverage
|
||
|
|
||
|
cov = coverage.Coverage(config_file='%(coverage_config)s')
|
||
|
|
||
|
def atexit_coverage():
|
||
|
cov.stop()
|
||
|
cov.save()
|
||
|
|
||
|
atexit.register(atexit_coverage)
|
||
|
|
||
|
cov.start()
|
||
|
'''
|
||
|
|
||
|
|
||
|
def _strip_comments(source):
|
||
|
# Strip comments and blank lines from the wrapper
|
||
|
buf = []
|
||
|
for line in source.splitlines():
|
||
|
l = line.strip()
|
||
|
if not l or l.startswith(u'#'):
|
||
|
continue
|
||
|
buf.append(line)
|
||
|
return u'\n'.join(buf)
|
||
|
|
||
|
|
||
|
if C.DEFAULT_KEEP_REMOTE_FILES:
|
||
|
# Keep comments when KEEP_REMOTE_FILES is set. That way users will see
|
||
|
# the comments with some nice usage instructions
|
||
|
ACTIVE_ANSIBALLZ_TEMPLATE = ANSIBALLZ_TEMPLATE
|
||
|
else:
|
||
|
# ANSIBALLZ_TEMPLATE stripped of comments for smaller over the wire size
|
||
|
ACTIVE_ANSIBALLZ_TEMPLATE = _strip_comments(ANSIBALLZ_TEMPLATE)
|
||
|
|
||
|
|
||
|
class ModuleDepFinder(ast.NodeVisitor):
|
||
|
# Caveats:
|
||
|
# This code currently does not handle:
|
||
|
# * relative imports from py2.6+ from . import urls
|
||
|
IMPORT_PREFIX_SIZE = len('ansible.module_utils.')
|
||
|
|
||
|
def __init__(self, *args, **kwargs):
|
||
|
"""
|
||
|
Walk the ast tree for the python module.
|
||
|
|
||
|
Save submodule[.submoduleN][.identifier] into self.submodules
|
||
|
|
||
|
self.submodules will end up with tuples like:
|
||
|
- ('basic',)
|
||
|
- ('urls', 'fetch_url')
|
||
|
- ('database', 'postgres')
|
||
|
- ('database', 'postgres', 'quote')
|
||
|
|
||
|
It's up to calling code to determine whether the final element of the
|
||
|
dotted strings are module names or something else (function, class, or
|
||
|
variable names)
|
||
|
"""
|
||
|
super(ModuleDepFinder, self).__init__(*args, **kwargs)
|
||
|
self.submodules = set()
|
||
|
|
||
|
def visit_Import(self, node):
|
||
|
# import ansible.module_utils.MODLIB[.MODLIBn] [as asname]
|
||
|
for alias in (a for a in node.names if a.name.startswith('ansible.module_utils.')):
|
||
|
py_mod = alias.name[self.IMPORT_PREFIX_SIZE:]
|
||
|
py_mod = tuple(py_mod.split('.'))
|
||
|
self.submodules.add(py_mod)
|
||
|
self.generic_visit(node)
|
||
|
|
||
|
def visit_ImportFrom(self, node):
|
||
|
# Specialcase: six is a special case because of its
|
||
|
# import logic
|
||
|
if node.names[0].name == '_six':
|
||
|
self.submodules.add(('_six',))
|
||
|
elif node.module.startswith('ansible.module_utils'):
|
||
|
where_from = node.module[self.IMPORT_PREFIX_SIZE:]
|
||
|
if where_from:
|
||
|
# from ansible.module_utils.MODULE1[.MODULEn] import IDENTIFIER [as asname]
|
||
|
# from ansible.module_utils.MODULE1[.MODULEn] import MODULEn+1 [as asname]
|
||
|
# from ansible.module_utils.MODULE1[.MODULEn] import MODULEn+1 [,IDENTIFIER] [as asname]
|
||
|
py_mod = tuple(where_from.split('.'))
|
||
|
for alias in node.names:
|
||
|
self.submodules.add(py_mod + (alias.name,))
|
||
|
else:
|
||
|
# from ansible.module_utils import MODLIB [,MODLIB2] [as asname]
|
||
|
for alias in node.names:
|
||
|
self.submodules.add((alias.name,))
|
||
|
self.generic_visit(node)
|
||
|
|
||
|
|
||
|
class PSModuleDepFinder():
|
||
|
|
||
|
def __init__(self):
|
||
|
self.modules = dict()
|
||
|
self.ps_version = None
|
||
|
self.os_version = None
|
||
|
self.become = False
|
||
|
|
||
|
self._re_module = re.compile(to_bytes(r'(?i)^#\s*requires\s+\-module(?:s?)\s*(Ansible\.ModuleUtils\..+)'))
|
||
|
self._re_ps_version = re.compile(to_bytes(r'(?i)^#requires\s+\-version\s+([0-9]+(\.[0-9]+){0,3})$'))
|
||
|
self._re_os_version = re.compile(to_bytes(r'(?i)^#ansiblerequires\s+\-osversion\s+([0-9]+(\.[0-9]+){0,3})$'))
|
||
|
self._re_become = re.compile(to_bytes(r'(?i)^#ansiblerequires\s+\-become$'))
|
||
|
|
||
|
def scan_module(self, module_data):
|
||
|
lines = module_data.split(b'\n')
|
||
|
module_utils = set()
|
||
|
|
||
|
for line in lines:
|
||
|
module_util_match = self._re_module.match(line)
|
||
|
if module_util_match:
|
||
|
# tolerate windows line endings by stripping any remaining newline chars
|
||
|
module_util_name = to_text(module_util_match.group(1).rstrip())
|
||
|
if module_util_name not in self.modules.keys():
|
||
|
module_utils.add(module_util_name)
|
||
|
|
||
|
ps_version_match = self._re_ps_version.match(line)
|
||
|
if ps_version_match:
|
||
|
self._parse_version_match(ps_version_match, "ps_version")
|
||
|
|
||
|
os_version_match = self._re_os_version.match(line)
|
||
|
if os_version_match:
|
||
|
self._parse_version_match(os_version_match, "os_version")
|
||
|
|
||
|
# once become is set, no need to keep on checking recursively
|
||
|
if not self.become:
|
||
|
become_match = self._re_become.match(line)
|
||
|
if become_match:
|
||
|
self.become = True
|
||
|
|
||
|
# recursively drill into each Requires to see if there are any more
|
||
|
# requirements
|
||
|
for m in set(module_utils):
|
||
|
m = to_text(m)
|
||
|
mu_path = ps_module_utils_loader.find_plugin(m, ".psm1")
|
||
|
if not mu_path:
|
||
|
raise AnsibleError('Could not find imported module support code for \'%s\'.' % m)
|
||
|
|
||
|
module_util_data = to_bytes(_slurp(mu_path))
|
||
|
self.modules[m] = module_util_data
|
||
|
self.scan_module(module_util_data)
|
||
|
|
||
|
def _parse_version_match(self, match, attribute):
|
||
|
new_version = to_text(match.group(1)).rstrip()
|
||
|
|
||
|
# PowerShell cannot cast a string of "1" to Version, it must have at
|
||
|
# least the major.minor for it to be valid so we append 0
|
||
|
if match.group(2) is None:
|
||
|
new_version = "%s.0" % new_version
|
||
|
|
||
|
existing_version = getattr(self, attribute, None)
|
||
|
if existing_version is None:
|
||
|
setattr(self, attribute, new_version)
|
||
|
else:
|
||
|
# determine which is the latest version and set that
|
||
|
if LooseVersion(new_version) > LooseVersion(existing_version):
|
||
|
setattr(self, attribute, new_version)
|
||
|
|
||
|
|
||
|
def _slurp(path):
|
||
|
if not os.path.exists(path):
|
||
|
raise AnsibleError("imported module support code does not exist at %s" % os.path.abspath(path))
|
||
|
fd = open(path, 'rb')
|
||
|
data = fd.read()
|
||
|
fd.close()
|
||
|
return data
|
||
|
|
||
|
|
||
|
def _get_shebang(interpreter, task_vars, templar, args=tuple()):
|
||
|
"""
|
||
|
Note not stellar API:
|
||
|
Returns None instead of always returning a shebang line. Doing it this
|
||
|
way allows the caller to decide to use the shebang it read from the
|
||
|
file rather than trust that we reformatted what they already have
|
||
|
correctly.
|
||
|
"""
|
||
|
interpreter_config = u'ansible_%s_interpreter' % os.path.basename(interpreter).strip()
|
||
|
|
||
|
if interpreter_config not in task_vars:
|
||
|
return (None, interpreter)
|
||
|
|
||
|
interpreter = templar.template(task_vars[interpreter_config].strip())
|
||
|
shebang = u'#!' + interpreter
|
||
|
|
||
|
if args:
|
||
|
shebang = shebang + u' ' + u' '.join(args)
|
||
|
|
||
|
return (shebang, interpreter)
|
||
|
|
||
|
|
||
|
def recursive_finder(name, data, py_module_names, py_module_cache, zf):
|
||
|
"""
|
||
|
Using ModuleDepFinder, make sure we have all of the module_utils files that
|
||
|
the module its module_utils files needs.
|
||
|
"""
|
||
|
# Parse the module and find the imports of ansible.module_utils
|
||
|
tree = ast.parse(data)
|
||
|
finder = ModuleDepFinder()
|
||
|
finder.visit(tree)
|
||
|
|
||
|
#
|
||
|
# Determine what imports that we've found are modules (vs class, function.
|
||
|
# variable names) for packages
|
||
|
#
|
||
|
|
||
|
normalized_modules = set()
|
||
|
# Loop through the imports that we've found to normalize them
|
||
|
# Exclude paths that match with paths we've already processed
|
||
|
# (Have to exclude them a second time once the paths are processed)
|
||
|
|
||
|
module_utils_paths = [p for p in module_utils_loader._get_paths(subdirs=False) if os.path.isdir(p)]
|
||
|
module_utils_paths.append(_MODULE_UTILS_PATH)
|
||
|
for py_module_name in finder.submodules.difference(py_module_names):
|
||
|
module_info = None
|
||
|
|
||
|
if py_module_name[0] == 'six':
|
||
|
# Special case the python six library because it messes up the
|
||
|
# import process in an incompatible way
|
||
|
module_info = imp.find_module('six', module_utils_paths)
|
||
|
py_module_name = ('six',)
|
||
|
idx = 0
|
||
|
elif py_module_name[0] == '_six':
|
||
|
# Special case the python six library because it messes up the
|
||
|
# import process in an incompatible way
|
||
|
module_info = imp.find_module('_six', [os.path.join(p, 'six') for p in module_utils_paths])
|
||
|
py_module_name = ('six', '_six')
|
||
|
idx = 0
|
||
|
else:
|
||
|
# Check whether either the last or the second to last identifier is
|
||
|
# a module name
|
||
|
for idx in (1, 2):
|
||
|
if len(py_module_name) < idx:
|
||
|
break
|
||
|
try:
|
||
|
module_info = imp.find_module(py_module_name[-idx],
|
||
|
[os.path.join(p, *py_module_name[:-idx]) for p in module_utils_paths])
|
||
|
break
|
||
|
except ImportError:
|
||
|
continue
|
||
|
|
||
|
# Could not find the module. Construct a helpful error message.
|
||
|
if module_info is None:
|
||
|
msg = ['Could not find imported module support code for %s. Looked for' % (name,)]
|
||
|
if idx == 2:
|
||
|
msg.append('either %s.py or %s.py' % (py_module_name[-1], py_module_name[-2]))
|
||
|
else:
|
||
|
msg.append(py_module_name[-1])
|
||
|
raise AnsibleError(' '.join(msg))
|
||
|
|
||
|
# Found a byte compiled file rather than source. We cannot send byte
|
||
|
# compiled over the wire as the python version might be different.
|
||
|
# imp.find_module seems to prefer to return source packages so we just
|
||
|
# error out if imp.find_module returns byte compiled files (This is
|
||
|
# fragile as it depends on undocumented imp.find_module behaviour)
|
||
|
if module_info[2][2] not in (imp.PY_SOURCE, imp.PKG_DIRECTORY):
|
||
|
msg = ['Could not find python source for imported module support code for %s. Looked for' % name]
|
||
|
if idx == 2:
|
||
|
msg.append('either %s.py or %s.py' % (py_module_name[-1], py_module_name[-2]))
|
||
|
else:
|
||
|
msg.append(py_module_name[-1])
|
||
|
raise AnsibleError(' '.join(msg))
|
||
|
|
||
|
if idx == 2:
|
||
|
# We've determined that the last portion was an identifier and
|
||
|
# thus, not part of the module name
|
||
|
py_module_name = py_module_name[:-1]
|
||
|
|
||
|
# If not already processed then we've got work to do
|
||
|
# If not in the cache, then read the file into the cache
|
||
|
# We already have a file handle for the module open so it makes
|
||
|
# sense to read it now
|
||
|
if py_module_name not in py_module_cache:
|
||
|
if module_info[2][2] == imp.PKG_DIRECTORY:
|
||
|
# Read the __init__.py instead of the module file as this is
|
||
|
# a python package
|
||
|
normalized_name = py_module_name + ('__init__',)
|
||
|
if normalized_name not in py_module_names:
|
||
|
normalized_path = os.path.join(os.path.join(module_info[1], '__init__.py'))
|
||
|
normalized_data = _slurp(normalized_path)
|
||
|
py_module_cache[normalized_name] = (normalized_data, normalized_path)
|
||
|
normalized_modules.add(normalized_name)
|
||
|
else:
|
||
|
normalized_name = py_module_name
|
||
|
if normalized_name not in py_module_names:
|
||
|
normalized_path = module_info[1]
|
||
|
normalized_data = module_info[0].read()
|
||
|
module_info[0].close()
|
||
|
py_module_cache[normalized_name] = (normalized_data, normalized_path)
|
||
|
normalized_modules.add(normalized_name)
|
||
|
|
||
|
# Make sure that all the packages that this module is a part of
|
||
|
# are also added
|
||
|
for i in range(1, len(py_module_name)):
|
||
|
py_pkg_name = py_module_name[:-i] + ('__init__',)
|
||
|
if py_pkg_name not in py_module_names:
|
||
|
pkg_dir_info = imp.find_module(py_pkg_name[-1],
|
||
|
[os.path.join(p, *py_pkg_name[:-1]) for p in module_utils_paths])
|
||
|
normalized_modules.add(py_pkg_name)
|
||
|
py_module_cache[py_pkg_name] = (_slurp(pkg_dir_info[1]), pkg_dir_info[1])
|
||
|
|
||
|
# FIXME: Currently the AnsiBallZ wrapper monkeypatches module args into a global
|
||
|
# variable in basic.py. If a module doesn't import basic.py, then the AnsiBallZ wrapper will
|
||
|
# traceback when it tries to monkypatch. So, for now, we have to unconditionally include
|
||
|
# basic.py.
|
||
|
#
|
||
|
# In the future we need to change the wrapper to monkeypatch the args into a global variable in
|
||
|
# their own, separate python module. That way we won't require basic.py. Modules which don't
|
||
|
# want basic.py can import that instead. AnsibleModule will need to change to import the vars
|
||
|
# from the separate python module and mirror the args into its global variable for backwards
|
||
|
# compatibility.
|
||
|
if ('basic',) not in py_module_names:
|
||
|
pkg_dir_info = imp.find_module('basic', module_utils_paths)
|
||
|
normalized_modules.add(('basic',))
|
||
|
py_module_cache[('basic',)] = (_slurp(pkg_dir_info[1]), pkg_dir_info[1])
|
||
|
# End of AnsiballZ hack
|
||
|
|
||
|
#
|
||
|
# iterate through all of the ansible.module_utils* imports that we haven't
|
||
|
# already checked for new imports
|
||
|
#
|
||
|
|
||
|
# set of modules that we haven't added to the zipfile
|
||
|
unprocessed_py_module_names = normalized_modules.difference(py_module_names)
|
||
|
|
||
|
for py_module_name in unprocessed_py_module_names:
|
||
|
py_module_path = os.path.join(*py_module_name)
|
||
|
py_module_file_name = '%s.py' % py_module_path
|
||
|
|
||
|
zf.writestr(os.path.join("ansible/module_utils",
|
||
|
py_module_file_name), py_module_cache[py_module_name][0])
|
||
|
display.vvvvv("Using module_utils file %s" % py_module_cache[py_module_name][1])
|
||
|
|
||
|
# Add the names of the files we're scheduling to examine in the loop to
|
||
|
# py_module_names so that we don't re-examine them in the next pass
|
||
|
# through recursive_finder()
|
||
|
py_module_names.update(unprocessed_py_module_names)
|
||
|
|
||
|
for py_module_file in unprocessed_py_module_names:
|
||
|
recursive_finder(py_module_file, py_module_cache[py_module_file][0], py_module_names, py_module_cache, zf)
|
||
|
# Save memory; the file won't have to be read again for this ansible module.
|
||
|
del py_module_cache[py_module_file]
|
||
|
|
||
|
|
||
|
def _is_binary(b_module_data):
|
||
|
textchars = bytearray(set([7, 8, 9, 10, 12, 13, 27]) | set(range(0x20, 0x100)) - set([0x7f]))
|
||
|
start = b_module_data[:1024]
|
||
|
return bool(start.translate(None, textchars))
|
||
|
|
||
|
|
||
|
def _find_module_utils(module_name, b_module_data, module_path, module_args, task_vars, templar, module_compression, async_timeout, become,
|
||
|
become_method, become_user, become_password, become_flags, environment):
|
||
|
"""
|
||
|
Given the source of the module, convert it to a Jinja2 template to insert
|
||
|
module code and return whether it's a new or old style module.
|
||
|
"""
|
||
|
module_substyle = module_style = 'old'
|
||
|
|
||
|
# module_style is something important to calling code (ActionBase). It
|
||
|
# determines how arguments are formatted (json vs k=v) and whether
|
||
|
# a separate arguments file needs to be sent over the wire.
|
||
|
# module_substyle is extra information that's useful internally. It tells
|
||
|
# us what we have to look to substitute in the module files and whether
|
||
|
# we're using module replacer or ansiballz to format the module itself.
|
||
|
if _is_binary(b_module_data):
|
||
|
module_substyle = module_style = 'binary'
|
||
|
elif REPLACER in b_module_data:
|
||
|
# Do REPLACER before from ansible.module_utils because we need make sure
|
||
|
# we substitute "from ansible.module_utils basic" for REPLACER
|
||
|
module_style = 'new'
|
||
|
module_substyle = 'python'
|
||
|
b_module_data = b_module_data.replace(REPLACER, b'from ansible.module_utils.basic import *')
|
||
|
elif b'from ansible.module_utils.' in b_module_data:
|
||
|
module_style = 'new'
|
||
|
module_substyle = 'python'
|
||
|
elif REPLACER_WINDOWS in b_module_data:
|
||
|
module_style = 'new'
|
||
|
module_substyle = 'powershell'
|
||
|
b_module_data = b_module_data.replace(REPLACER_WINDOWS, b'#Requires -Module Ansible.ModuleUtils.Legacy')
|
||
|
elif re.search(b'#Requires -Module', b_module_data, re.IGNORECASE) \
|
||
|
or re.search(b'#Requires -Version', b_module_data, re.IGNORECASE)\
|
||
|
or re.search(b'#AnsibleRequires -OSVersion', b_module_data, re.IGNORECASE):
|
||
|
module_style = 'new'
|
||
|
module_substyle = 'powershell'
|
||
|
elif REPLACER_JSONARGS in b_module_data:
|
||
|
module_style = 'new'
|
||
|
module_substyle = 'jsonargs'
|
||
|
elif b'WANT_JSON' in b_module_data:
|
||
|
module_substyle = module_style = 'non_native_want_json'
|
||
|
|
||
|
shebang = None
|
||
|
# Neither old-style, non_native_want_json nor binary modules should be modified
|
||
|
# except for the shebang line (Done by modify_module)
|
||
|
if module_style in ('old', 'non_native_want_json', 'binary'):
|
||
|
return b_module_data, module_style, shebang
|
||
|
|
||
|
output = BytesIO()
|
||
|
py_module_names = set()
|
||
|
|
||
|
if module_substyle == 'python':
|
||
|
params = dict(ANSIBLE_MODULE_ARGS=module_args,)
|
||
|
python_repred_params = repr(json.dumps(params))
|
||
|
|
||
|
try:
|
||
|
compression_method = getattr(zipfile, module_compression)
|
||
|
except AttributeError:
|
||
|
display.warning(u'Bad module compression string specified: %s. Using ZIP_STORED (no compression)' % module_compression)
|
||
|
compression_method = zipfile.ZIP_STORED
|
||
|
|
||
|
lookup_path = os.path.join(C.DEFAULT_LOCAL_TMP, 'ansiballz_cache')
|
||
|
cached_module_filename = os.path.join(lookup_path, "%s-%s" % (module_name, module_compression))
|
||
|
|
||
|
zipdata = None
|
||
|
# Optimization -- don't lock if the module has already been cached
|
||
|
if os.path.exists(cached_module_filename):
|
||
|
display.debug('ANSIBALLZ: using cached module: %s' % cached_module_filename)
|
||
|
zipdata = open(cached_module_filename, 'rb').read()
|
||
|
else:
|
||
|
if module_name in action_write_locks.action_write_locks:
|
||
|
display.debug('ANSIBALLZ: Using lock for %s' % module_name)
|
||
|
lock = action_write_locks.action_write_locks[module_name]
|
||
|
else:
|
||
|
# If the action plugin directly invokes the module (instead of
|
||
|
# going through a strategy) then we don't have a cross-process
|
||
|
# Lock specifically for this module. Use the "unexpected
|
||
|
# module" lock instead
|
||
|
display.debug('ANSIBALLZ: Using generic lock for %s' % module_name)
|
||
|
lock = action_write_locks.action_write_locks[None]
|
||
|
|
||
|
display.debug('ANSIBALLZ: Acquiring lock')
|
||
|
with lock:
|
||
|
display.debug('ANSIBALLZ: Lock acquired: %s' % id(lock))
|
||
|
# Check that no other process has created this while we were
|
||
|
# waiting for the lock
|
||
|
if not os.path.exists(cached_module_filename):
|
||
|
display.debug('ANSIBALLZ: Creating module')
|
||
|
# Create the module zip data
|
||
|
zipoutput = BytesIO()
|
||
|
zf = zipfile.ZipFile(zipoutput, mode='w', compression=compression_method)
|
||
|
# Note: If we need to import from release.py first,
|
||
|
# remember to catch all exceptions: https://github.com/ansible/ansible/issues/16523
|
||
|
zf.writestr('ansible/__init__.py',
|
||
|
b'from pkgutil import extend_path\n__path__=extend_path(__path__,__name__)\n__version__="' +
|
||
|
to_bytes(__version__) + b'"\n__author__="' +
|
||
|
to_bytes(__author__) + b'"\n')
|
||
|
zf.writestr('ansible/module_utils/__init__.py', b'from pkgutil import extend_path\n__path__=extend_path(__path__,__name__)\n')
|
||
|
|
||
|
zf.writestr('__main__.py', b_module_data)
|
||
|
|
||
|
py_module_cache = {('__init__',): (b'', '[builtin]')}
|
||
|
recursive_finder(module_name, b_module_data, py_module_names, py_module_cache, zf)
|
||
|
zf.close()
|
||
|
zipdata = base64.b64encode(zipoutput.getvalue())
|
||
|
|
||
|
# Write the assembled module to a temp file (write to temp
|
||
|
# so that no one looking for the file reads a partially
|
||
|
# written file)
|
||
|
if not os.path.exists(lookup_path):
|
||
|
# Note -- if we have a global function to setup, that would
|
||
|
# be a better place to run this
|
||
|
os.makedirs(lookup_path)
|
||
|
display.debug('ANSIBALLZ: Writing module')
|
||
|
with open(cached_module_filename + '-part', 'wb') as f:
|
||
|
f.write(zipdata)
|
||
|
|
||
|
# Rename the file into its final position in the cache so
|
||
|
# future users of this module can read it off the
|
||
|
# filesystem instead of constructing from scratch.
|
||
|
display.debug('ANSIBALLZ: Renaming module')
|
||
|
os.rename(cached_module_filename + '-part', cached_module_filename)
|
||
|
display.debug('ANSIBALLZ: Done creating module')
|
||
|
|
||
|
if zipdata is None:
|
||
|
display.debug('ANSIBALLZ: Reading module after lock')
|
||
|
# Another process wrote the file while we were waiting for
|
||
|
# the write lock. Go ahead and read the data from disk
|
||
|
# instead of re-creating it.
|
||
|
try:
|
||
|
zipdata = open(cached_module_filename, 'rb').read()
|
||
|
except IOError:
|
||
|
raise AnsibleError('A different worker process failed to create module file. '
|
||
|
'Look at traceback for that process for debugging information.')
|
||
|
zipdata = to_text(zipdata, errors='surrogate_or_strict')
|
||
|
|
||
|
shebang, interpreter = _get_shebang(u'/usr/bin/python', task_vars, templar)
|
||
|
if shebang is None:
|
||
|
shebang = u'#!/usr/bin/python'
|
||
|
|
||
|
# Enclose the parts of the interpreter in quotes because we're
|
||
|
# substituting it into the template as a Python string
|
||
|
interpreter_parts = interpreter.split(u' ')
|
||
|
interpreter = u"'{0}'".format(u"', '".join(interpreter_parts))
|
||
|
|
||
|
coverage_config = os.environ.get('_ANSIBLE_COVERAGE_CONFIG')
|
||
|
|
||
|
if coverage_config:
|
||
|
# Enable code coverage analysis of the module.
|
||
|
# This feature is for internal testing and may change without notice.
|
||
|
coverage = ANSIBALLZ_COVERAGE_TEMPLATE % dict(
|
||
|
coverage_config=coverage_config,
|
||
|
coverage_output=os.environ['_ANSIBLE_COVERAGE_OUTPUT']
|
||
|
)
|
||
|
else:
|
||
|
coverage = ''
|
||
|
|
||
|
now = datetime.datetime.utcnow()
|
||
|
output.write(to_bytes(ACTIVE_ANSIBALLZ_TEMPLATE % dict(
|
||
|
zipdata=zipdata,
|
||
|
ansible_module=module_name,
|
||
|
params=python_repred_params,
|
||
|
shebang=shebang,
|
||
|
interpreter=interpreter,
|
||
|
coding=ENCODING_STRING,
|
||
|
year=now.year,
|
||
|
month=now.month,
|
||
|
day=now.day,
|
||
|
hour=now.hour,
|
||
|
minute=now.minute,
|
||
|
second=now.second,
|
||
|
coverage=coverage,
|
||
|
)))
|
||
|
b_module_data = output.getvalue()
|
||
|
|
||
|
elif module_substyle == 'powershell':
|
||
|
# Powershell/winrm don't actually make use of shebang so we can
|
||
|
# safely set this here. If we let the fallback code handle this
|
||
|
# it can fail in the presence of the UTF8 BOM commonly added by
|
||
|
# Windows text editors
|
||
|
shebang = u'#!powershell'
|
||
|
|
||
|
exec_manifest = dict(
|
||
|
module_entry=to_text(base64.b64encode(b_module_data)),
|
||
|
powershell_modules=dict(),
|
||
|
module_args=module_args,
|
||
|
actions=['exec'],
|
||
|
environment=environment
|
||
|
)
|
||
|
|
||
|
exec_manifest['exec'] = to_text(base64.b64encode(to_bytes(leaf_exec)))
|
||
|
|
||
|
if async_timeout > 0:
|
||
|
exec_manifest["actions"].insert(0, 'async_watchdog')
|
||
|
exec_manifest["async_watchdog"] = to_text(base64.b64encode(to_bytes(async_watchdog)))
|
||
|
exec_manifest["actions"].insert(0, 'async_wrapper')
|
||
|
exec_manifest["async_wrapper"] = to_text(base64.b64encode(to_bytes(async_wrapper)))
|
||
|
exec_manifest["async_jid"] = str(random.randint(0, 999999999999))
|
||
|
exec_manifest["async_timeout_sec"] = async_timeout
|
||
|
|
||
|
if become and become_method == 'runas':
|
||
|
exec_manifest["actions"].insert(0, 'become')
|
||
|
exec_manifest["become_user"] = become_user
|
||
|
exec_manifest["become_password"] = become_password
|
||
|
exec_manifest['become_flags'] = become_flags
|
||
|
exec_manifest["become"] = to_text(base64.b64encode(to_bytes(become_wrapper)))
|
||
|
|
||
|
finder = PSModuleDepFinder()
|
||
|
finder.scan_module(b_module_data)
|
||
|
|
||
|
for name, data in finder.modules.items():
|
||
|
b64_data = to_text(base64.b64encode(data))
|
||
|
exec_manifest['powershell_modules'][name] = b64_data
|
||
|
|
||
|
exec_manifest['min_ps_version'] = finder.ps_version
|
||
|
exec_manifest['min_os_version'] = finder.os_version
|
||
|
if finder.become and 'become' not in exec_manifest['actions']:
|
||
|
exec_manifest['actions'].insert(0, 'become')
|
||
|
exec_manifest['become_user'] = 'SYSTEM'
|
||
|
exec_manifest['become_password'] = None
|
||
|
exec_manifest['become_flags'] = None
|
||
|
exec_manifest['become'] = to_text(base64.b64encode(to_bytes(become_wrapper)))
|
||
|
|
||
|
# FUTURE: smuggle this back as a dict instead of serializing here; the connection plugin may need to modify it
|
||
|
module_json = json.dumps(exec_manifest)
|
||
|
|
||
|
# delimit the payload JSON from the wrapper to keep sensitive contents out of scriptblocks (which can be logged)
|
||
|
b_module_data = to_bytes(exec_wrapper) + b'\0\0\0\0' + to_bytes(module_json)
|
||
|
|
||
|
elif module_substyle == 'jsonargs':
|
||
|
module_args_json = to_bytes(json.dumps(module_args))
|
||
|
|
||
|
# these strings could be included in a third-party module but
|
||
|
# officially they were included in the 'basic' snippet for new-style
|
||
|
# python modules (which has been replaced with something else in
|
||
|
# ansiballz) If we remove them from jsonargs-style module replacer
|
||
|
# then we can remove them everywhere.
|
||
|
python_repred_args = to_bytes(repr(module_args_json))
|
||
|
b_module_data = b_module_data.replace(REPLACER_VERSION, to_bytes(repr(__version__)))
|
||
|
b_module_data = b_module_data.replace(REPLACER_COMPLEX, python_repred_args)
|
||
|
b_module_data = b_module_data.replace(REPLACER_SELINUX, to_bytes(','.join(C.DEFAULT_SELINUX_SPECIAL_FS)))
|
||
|
|
||
|
# The main event -- substitute the JSON args string into the module
|
||
|
b_module_data = b_module_data.replace(REPLACER_JSONARGS, module_args_json)
|
||
|
|
||
|
facility = b'syslog.' + to_bytes(task_vars.get('ansible_syslog_facility', C.DEFAULT_SYSLOG_FACILITY), errors='surrogate_or_strict')
|
||
|
b_module_data = b_module_data.replace(b'syslog.LOG_USER', facility)
|
||
|
|
||
|
return (b_module_data, module_style, shebang)
|
||
|
|
||
|
|
||
|
def modify_module(module_name, module_path, module_args, templar, task_vars=None, module_compression='ZIP_STORED', async_timeout=0, become=False,
|
||
|
become_method=None, become_user=None, become_password=None, become_flags=None, environment=None):
|
||
|
"""
|
||
|
Used to insert chunks of code into modules before transfer rather than
|
||
|
doing regular python imports. This allows for more efficient transfer in
|
||
|
a non-bootstrapping scenario by not moving extra files over the wire and
|
||
|
also takes care of embedding arguments in the transferred modules.
|
||
|
|
||
|
This version is done in such a way that local imports can still be
|
||
|
used in the module code, so IDEs don't have to be aware of what is going on.
|
||
|
|
||
|
Example:
|
||
|
|
||
|
from ansible.module_utils.basic import *
|
||
|
|
||
|
... will result in the insertion of basic.py into the module
|
||
|
from the module_utils/ directory in the source tree.
|
||
|
|
||
|
For powershell, this code effectively no-ops, as the exec wrapper requires access to a number of
|
||
|
properties not available here.
|
||
|
|
||
|
"""
|
||
|
task_vars = {} if task_vars is None else task_vars
|
||
|
environment = {} if environment is None else environment
|
||
|
|
||
|
with open(module_path, 'rb') as f:
|
||
|
|
||
|
# read in the module source
|
||
|
b_module_data = f.read()
|
||
|
|
||
|
(b_module_data, module_style, shebang) = _find_module_utils(module_name, b_module_data, module_path, module_args, task_vars, templar, module_compression,
|
||
|
async_timeout=async_timeout, become=become, become_method=become_method,
|
||
|
become_user=become_user, become_password=become_password, become_flags=become_flags,
|
||
|
environment=environment)
|
||
|
|
||
|
if module_style == 'binary':
|
||
|
return (b_module_data, module_style, to_text(shebang, nonstring='passthru'))
|
||
|
elif shebang is None:
|
||
|
b_lines = b_module_data.split(b"\n", 1)
|
||
|
if b_lines[0].startswith(b"#!"):
|
||
|
b_shebang = b_lines[0].strip()
|
||
|
# shlex.split on python-2.6 needs bytes. On python-3.x it needs text
|
||
|
args = shlex.split(to_native(b_shebang[2:], errors='surrogate_or_strict'))
|
||
|
|
||
|
# _get_shebang() takes text strings
|
||
|
args = [to_text(a, errors='surrogate_or_strict') for a in args]
|
||
|
interpreter = args[0]
|
||
|
b_new_shebang = to_bytes(_get_shebang(interpreter, task_vars, templar, args[1:])[0],
|
||
|
errors='surrogate_or_strict', nonstring='passthru')
|
||
|
|
||
|
if b_new_shebang:
|
||
|
b_lines[0] = b_shebang = b_new_shebang
|
||
|
|
||
|
if os.path.basename(interpreter).startswith(u'python'):
|
||
|
b_lines.insert(1, b_ENCODING_STRING)
|
||
|
|
||
|
shebang = to_text(b_shebang, nonstring='passthru', errors='surrogate_or_strict')
|
||
|
else:
|
||
|
# No shebang, assume a binary module?
|
||
|
pass
|
||
|
|
||
|
b_module_data = b"\n".join(b_lines)
|
||
|
|
||
|
return (b_module_data, module_style, shebang)
|