mirror of
https://github.com/valitydev/salt.git
synced 2024-11-07 08:58:59 +00:00
Merge branch '2018.3' into fix-broken-rabbitmq-list-policies
This commit is contained in:
commit
6d7ef27557
2
.gitignore
vendored
2
.gitignore
vendored
@ -11,6 +11,8 @@ MANIFEST
|
||||
*.wpr
|
||||
*.wpu
|
||||
*.DS_Store
|
||||
.pytest_cache
|
||||
Pipfile.lock
|
||||
|
||||
# virtualenv
|
||||
# - ignores directories of a virtualenv when you create it right on
|
||||
|
6
Gemfile
6
Gemfile
@ -2,8 +2,8 @@
|
||||
|
||||
source 'https://rubygems.org'
|
||||
|
||||
gem 'test-kitchen', :git => 'https://github.com/gtmanfred/test-kitchen.git'
|
||||
gem 'kitchen-salt', :git => 'https://github.com/saltstack/kitchen-salt.git'
|
||||
gem 'test-kitchen', '~>1.21'
|
||||
gem 'kitchen-salt', '~>0.2'
|
||||
gem 'kitchen-sync'
|
||||
gem 'git'
|
||||
|
||||
@ -20,7 +20,7 @@ group :windows do
|
||||
gem 'vagrant-wrapper'
|
||||
gem 'kitchen-vagrant'
|
||||
gem 'winrm', '~>2.0'
|
||||
gem 'winrm-fs', :git => 'https://github.com/gtmanfred/winrm-fs.git'
|
||||
gem 'winrm-fs', :git => 'https://github.com/WinRb/winrm-fs.git'
|
||||
end
|
||||
|
||||
group :ec2 do
|
||||
|
40
Pipfile
Normal file
40
Pipfile
Normal file
@ -0,0 +1,40 @@
|
||||
[[source]]
|
||||
url = "https://pypi.org/simple"
|
||||
verify_ssl = true
|
||||
name = "pypi"
|
||||
|
||||
[packages]
|
||||
Jinja2 = "*"
|
||||
msgpack-python = ">0.3,!=0.5.5"
|
||||
PyYAML = "*"
|
||||
MarkupSafe = "*"
|
||||
requests = ">=1.0.0"
|
||||
tornado = ">=4.2.1,<5.0"
|
||||
pycrypto = ">=2.6.1"
|
||||
pyzmq = ">=2.2.0"
|
||||
|
||||
[dev-packages]
|
||||
mock = ">=2.0.0"
|
||||
apache-libcloud = ">=0.14.0"
|
||||
boto = ">=2.32.1"
|
||||
boto3 = ">=1.2.1"
|
||||
moto = ">=0.3.6"
|
||||
SaltPyLint = ">=v2017.3.6"
|
||||
pytest = ">=3.5.0"
|
||||
|
||||
[packages.futures]
|
||||
# Required by Tornado to handle threads stuff.
|
||||
version = ">=2.0"
|
||||
markers = "python_version < '3.0'"
|
||||
|
||||
[dev-packages.pytest-salt]
|
||||
git = "git://github.com/saltstack/pytest-salt.git"
|
||||
ref = "master"
|
||||
|
||||
[dev-packages.httpretty]
|
||||
# httpretty Needs to be here for now even though it's a dependency of boto.
|
||||
# A pip install on a fresh system will decide to target httpretty 0.8.10 to
|
||||
# satisfy other requirements, and httpretty 0.8.10 has bugs in setup.py that
|
||||
# prevent it from being successfully installed (at least on Python 3.4).
|
||||
version = "*"
|
||||
markers = "python_version >= '3.4'"
|
@ -538,6 +538,10 @@
|
||||
# targeted with the normal -N argument to salt-ssh.
|
||||
#ssh_list_nodegroups: {}
|
||||
|
||||
# salt-ssh has the ability to update the flat roster file if a minion is not
|
||||
# found in the roster. Set this to True to enable it.
|
||||
#ssh_update_roster: False
|
||||
|
||||
##### Master Module Management #####
|
||||
##########################################
|
||||
# Manage how master side modules are loaded.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-API" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-API" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-api \- salt-api Command
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-CALL" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-CALL" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-call \- salt-call Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-CLOUD" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-CLOUD" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-cloud \- Salt Cloud Command
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-CP" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-CP" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-cp \- salt-cp Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-KEY" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-KEY" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-key \- salt-key Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-MASTER" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-MASTER" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-master \- salt-master Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-MINION" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-MINION" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-minion \- salt-minion Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-PROXY" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-PROXY" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-proxy \- salt-proxy Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-RUN" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-RUN" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-run \- salt-run Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-SSH" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-SSH" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-ssh \- salt-ssh Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-SYNDIC" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-SYNDIC" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-syndic \- salt-syndic Documentation
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT-UNITY" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT-UNITY" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt-unity \- salt-unity Command
|
||||
.
|
||||
|
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SALT" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SALT" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
salt \- salt
|
||||
.
|
||||
|
12721
doc/man/salt.7
12721
doc/man/salt.7
File diff suppressed because it is too large
Load Diff
@ -1,6 +1,6 @@
|
||||
.\" Man page generated from reStructuredText.
|
||||
.
|
||||
.TH "SPM" "1" "Feb 23, 2018" "2018.3.0" "Salt"
|
||||
.TH "SPM" "1" "May 09, 2018" "2018.3.1" "Salt"
|
||||
.SH NAME
|
||||
spm \- Salt Package Manager Command
|
||||
.
|
||||
|
@ -3219,3 +3219,31 @@ URL of the repository:
|
||||
Replace ``<commit_id>`` with the SHA1 hash of a commit ID. Specifying a commit
|
||||
ID is useful in that it allows one to revert back to a previous version in the
|
||||
event that an error is introduced in the latest revision of the repo.
|
||||
|
||||
``ssh_merge_pillar``
|
||||
--------------------
|
||||
|
||||
.. versionadded:: 2018.3.2
|
||||
|
||||
Default: ``True``
|
||||
|
||||
Merges the compiled pillar data with the pillar data already available globally.
|
||||
This is useful when using ``salt-ssh`` or ``salt-call --local`` and overriding the pillar
|
||||
data in a state file:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
apply_showpillar:
|
||||
module.run:
|
||||
- name: state.apply
|
||||
- mods:
|
||||
- showpillar
|
||||
- kwargs:
|
||||
pillar:
|
||||
test: "foo bar"
|
||||
|
||||
If set to ``True`` the ``showpillar`` state will have access to the
|
||||
global pillar data.
|
||||
|
||||
If set to ``False`` only the overriding pillar data will be available
|
||||
to the ``showpillar`` state.
|
||||
|
@ -1,6 +1,6 @@
|
||||
=====================
|
||||
==================
|
||||
salt.modules.swarm
|
||||
=====================
|
||||
==================
|
||||
|
||||
.. automodule:: salt.modules.swarm
|
||||
:memebers:
|
||||
:members:
|
||||
|
@ -124,7 +124,6 @@ state modules
|
||||
influxdb_database
|
||||
influxdb_retention_policy
|
||||
influxdb_user
|
||||
infoblox
|
||||
infoblox_a
|
||||
infoblox_cname
|
||||
infoblox_host_record
|
||||
|
@ -1,5 +0,0 @@
|
||||
salt.states.infoblox module
|
||||
===========================
|
||||
|
||||
.. automodule:: salt.states.infoblox
|
||||
:members:
|
@ -410,10 +410,11 @@ exactly like the ``require`` requisite (the watching state will execute if
|
||||
service.running:
|
||||
- watch_any:
|
||||
- file: /etc/apache2/sites-available/site1.conf
|
||||
- file: /etc/apache2/sites-available/site2.conf
|
||||
- file: apache2-site2
|
||||
file.managed:
|
||||
- name: /etc/apache2/sites-available/site1.conf
|
||||
- source: salt://apache2/files/site1.conf
|
||||
apache2-site2:
|
||||
file.managed:
|
||||
- name: /etc/apache2/sites-available/site2.conf
|
||||
- source: salt://apache2/files/site2.conf
|
||||
|
@ -40,8 +40,9 @@ Beacons are typically enabled by placing a ``beacons:`` top level block in
|
||||
|
||||
beacons:
|
||||
inotify:
|
||||
/etc/important_file: {}
|
||||
/opt: {}
|
||||
- files:
|
||||
/etc/important_file: {}
|
||||
/opt: {}
|
||||
|
||||
The beacon system, like many others in Salt, can also be configured via the
|
||||
minion pillar, grains, or local config file.
|
||||
@ -50,6 +51,8 @@ minion pillar, grains, or local config file.
|
||||
The `inotify` beacon only works on OSes that have `inotify` kernel support.
|
||||
Currently this excludes FreeBSD, macOS, and Windows.
|
||||
|
||||
All beacon configuration is done using list based configuration.
|
||||
|
||||
Beacon Monitoring Interval
|
||||
--------------------------
|
||||
|
||||
@ -61,21 +64,23 @@ and 10-second intervals:
|
||||
|
||||
beacons:
|
||||
inotify:
|
||||
/etc/important_file: {}
|
||||
/opt: {}
|
||||
interval: 5
|
||||
disable_during_state_run: True
|
||||
- files:
|
||||
/etc/important_file: {}
|
||||
/opt: {}
|
||||
- interval: 5
|
||||
- disable_during_state_run: True
|
||||
load:
|
||||
1m:
|
||||
- 0.0
|
||||
- 2.0
|
||||
5m:
|
||||
- 0.0
|
||||
- 1.5
|
||||
15m:
|
||||
- 0.1
|
||||
- 1.0
|
||||
interval: 10
|
||||
- averages:
|
||||
1m:
|
||||
- 0.0
|
||||
- 2.0
|
||||
5m:
|
||||
- 0.0
|
||||
- 1.5
|
||||
15m:
|
||||
- 0.1
|
||||
- 1.0
|
||||
- interval: 10
|
||||
|
||||
.. _avoid-beacon-event-loops:
|
||||
|
||||
@ -96,8 +101,9 @@ which point the normal beacon interval will resume.
|
||||
|
||||
beacons:
|
||||
inotify:
|
||||
/etc/important_file: {}
|
||||
disable_during_state_run: True
|
||||
- files:
|
||||
/etc/important_file: {}
|
||||
- disable_during_state_run: True
|
||||
|
||||
.. _beacon-example:
|
||||
|
||||
@ -137,10 +143,11 @@ On the Salt minion, add the following configuration to
|
||||
|
||||
beacons:
|
||||
inotify:
|
||||
/etc/important_file:
|
||||
mask:
|
||||
- modify
|
||||
disable_during_state_run: True
|
||||
- files:
|
||||
/etc/important_file:
|
||||
mask:
|
||||
- modify
|
||||
- disable_during_state_run: True
|
||||
|
||||
Save the configuration file and restart the minion service. The beacon is now
|
||||
set up to notify salt upon modifications made to the file.
|
||||
|
@ -351,6 +351,7 @@ This driver can be configured using the ``/etc/openstack/clouds.yml`` file with
|
||||
`os-client-config <https://docs.openstack.org/os-client-config/latest/>`
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
myopenstack:
|
||||
driver: openstack
|
||||
region_name: RegionOne
|
||||
@ -359,6 +360,7 @@ This driver can be configured using the ``/etc/openstack/clouds.yml`` file with
|
||||
Or by just configuring the same auth block directly in the cloud provider config.
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
myopenstack:
|
||||
driver: openstack
|
||||
region_name: RegionOne
|
||||
|
@ -44,7 +44,7 @@ at ``/etc/salt/cloud.profiles`` or in the ``/etc/salt/cloud.profiles.d/`` direct
|
||||
|
||||
linode_1024:
|
||||
provider: my-linode-config
|
||||
size: Linode 2048
|
||||
size: Linode 2GB
|
||||
image: CentOS 7
|
||||
location: London, England, UK
|
||||
|
||||
@ -77,12 +77,14 @@ command:
|
||||
----------
|
||||
linode:
|
||||
----------
|
||||
Linode 1024:
|
||||
Linode 2GB:
|
||||
----------
|
||||
AVAIL:
|
||||
----------
|
||||
10:
|
||||
500
|
||||
11:
|
||||
500
|
||||
2:
|
||||
500
|
||||
3:
|
||||
@ -100,11 +102,19 @@ command:
|
||||
CORES:
|
||||
1
|
||||
DISK:
|
||||
24
|
||||
50
|
||||
HOURLY:
|
||||
0.015
|
||||
LABEL:
|
||||
Linode 1024
|
||||
Linode 2GB
|
||||
PLANID:
|
||||
2
|
||||
PRICE:
|
||||
10.0
|
||||
RAM:
|
||||
2048
|
||||
XFER:
|
||||
2000
|
||||
...SNIP...
|
||||
|
||||
|
||||
|
@ -6,7 +6,7 @@ Debian GNU/Linux / Raspbian
|
||||
|
||||
Debian GNU/Linux distribution and some derivatives such as Raspbian already
|
||||
have included Salt packages to their repositories. However, current stable
|
||||
release codenamed "Jessie" contains old outdated Salt release. It is
|
||||
Debian release contains old outdated Salt releases. It is
|
||||
recommended to use SaltStack repository for Debian as described
|
||||
:ref:`below <installation-debian-repo>`.
|
||||
|
||||
@ -33,11 +33,13 @@ Instructions are at https://repo.saltstack.com/#debian.
|
||||
Installation from the Debian / Raspbian Official Repository
|
||||
===========================================================
|
||||
|
||||
Stretch (Testing) and Sid (Unstable) distributions are already contain mostly
|
||||
up-to-date Salt packages built by Debian Salt Team. You can install Salt
|
||||
components directly from Debian.
|
||||
The Debian distributions contain mostly old Salt packages
|
||||
built by the Debian Salt Team. You can install Salt
|
||||
components directly from Debian but it is recommended to
|
||||
use the instructions above for the packages from the official
|
||||
Salt repository.
|
||||
|
||||
On Jessie (Stable) there is an option to install Salt minion from Stretch with
|
||||
On Jessie there is an option to install Salt minion from Stretch with
|
||||
`python-tornado` dependency from `jessie-backports` repositories.
|
||||
|
||||
To install fresh release of Salt minion on Jessie:
|
||||
@ -79,7 +81,7 @@ To install fresh release of Salt minion on Jessie:
|
||||
apt-get update
|
||||
apt-get install python-zmq python-tornado/stretch salt-common/stretch
|
||||
|
||||
#. Install Salt minion package from Stretch:
|
||||
#. Install Salt minion package from Latest Debian Release:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
|
File diff suppressed because it is too large
Load Diff
17
doc/topics/releases/2017.7.7.rst
Normal file
17
doc/topics/releases/2017.7.7.rst
Normal file
@ -0,0 +1,17 @@
|
||||
===========================
|
||||
In Progress: Salt 2017.7.7 Release Notes
|
||||
===========================
|
||||
|
||||
Version 2017.7.7 is an **unreleased** bugfix release for :ref:`2017.7.0 <release-2017-7-0>`.
|
||||
This release is still in progress and has not been released yet.
|
||||
|
||||
New win_snmp behavior
|
||||
---------------------
|
||||
|
||||
`win_snmp.get_community_names` now returns the SNMP settings actually in effect
|
||||
on the box. If settings are managed via GroupPolicy, those settings will be
|
||||
returned. Otherwise, normal settings are returned.
|
||||
|
||||
`win_snmp.set_community_names` now raises a CommandExecutionError when SNMP
|
||||
settings are being managed by GroupPolicy
|
||||
|
File diff suppressed because it is too large
Load Diff
16
doc/topics/releases/2018.3.2.rst
Normal file
16
doc/topics/releases/2018.3.2.rst
Normal file
@ -0,0 +1,16 @@
|
||||
========================================
|
||||
In Progress: Salt 2018.3.2 Release Notes
|
||||
========================================
|
||||
|
||||
Version 2018.3.2 is an **unreleased** bugfix release for :ref:`2018.3.0 <release-2018-3-0>`.
|
||||
This release is still in progress and has not been released yet.
|
||||
|
||||
Changes to win_timezone
|
||||
=======================
|
||||
|
||||
Improves timezone detection by using the pytz module.
|
||||
|
||||
``timezone.get_offset`` and ``timezone.get_zonecode`` now work properly.
|
||||
|
||||
Adds ``timezone.list`` to list supported timezones in either Windows or Unix
|
||||
format.
|
@ -699,15 +699,24 @@ repository to be served up from the Salt fileserver path
|
||||
Mountpoints can also be configured on a :ref:`per-remote basis
|
||||
<gitfs-per-remote-config>`.
|
||||
|
||||
|
||||
Using gitfs in Masterless Mode
|
||||
==============================
|
||||
|
||||
Since 2014.7.0, gitfs can be used in masterless mode. To do so, simply add the
|
||||
gitfs configuration parameters (and set :conf_master:`fileserver_backend`) in
|
||||
the _minion_ config file instead of the master config file.
|
||||
|
||||
|
||||
Using gitfs Alongside Other Backends
|
||||
====================================
|
||||
|
||||
Sometimes it may make sense to use multiple backends; for instance, if ``sls``
|
||||
files are stored in git but larger files are stored directly on the master.
|
||||
|
||||
The cascading lookup logic used for multiple remotes is also used with
|
||||
multiple backends. If the ``fileserver_backend`` option contains
|
||||
multiple backends:
|
||||
The cascading lookup logic used for multiple remotes is also used with multiple
|
||||
backends. If the :conf_master:`fileserver_backend` option contains multiple
|
||||
backends:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
@ -719,7 +728,6 @@ Then the ``roots`` backend (the default backend of files in ``/srv/salt``) will
|
||||
be searched first for the requested file; then, if it is not found on the
|
||||
master, each configured git remote will be searched.
|
||||
|
||||
|
||||
Branches, Environments, and Top Files
|
||||
=====================================
|
||||
|
||||
|
@ -4,7 +4,7 @@ Documentation=man:salt-master(1) file:///usr/share/doc/salt/html/contents.html h
|
||||
After=network.target
|
||||
|
||||
[Service]
|
||||
LimitNOFILE=16384
|
||||
LimitNOFILE=100000
|
||||
Type=simple
|
||||
ExecStart=/usr/bin/salt-master
|
||||
TasksMax=infinity
|
||||
|
165
pkg/windows/sign.bat
Normal file
165
pkg/windows/sign.bat
Normal file
@ -0,0 +1,165 @@
|
||||
:: ############################################################################
|
||||
::
|
||||
:: FILE: sign.bat
|
||||
::
|
||||
:: DESCRIPTION: Signing and Hashing script for Salt builds on Windows.
|
||||
:: Requires an official Code Signing Certificate and drivers
|
||||
:: installed to sign the files. Generates hashes in MD5 and
|
||||
:: SHA256 in a file of the same name with a `.md5` or
|
||||
:: `.sha256` extension.
|
||||
::
|
||||
:: NOTE: This script is used internally by SaltStack to sign and
|
||||
:: hash Windows Installer builds and uses resources not
|
||||
:: available to the community, such as SaltStack's Code
|
||||
:: Signing Certificate. It is placed here for version
|
||||
:: control.
|
||||
::
|
||||
:: COPYRIGHT: (c) 2012-2018 by the SaltStack Team
|
||||
::
|
||||
:: LICENSE: Apache 2.0
|
||||
:: ORGANIZATION: SaltStack, Inc (saltstack.com)
|
||||
:: CREATED: 2017
|
||||
::
|
||||
:: ############################################################################
|
||||
::
|
||||
:: USAGE: The script must be located in a directory that has the installer
|
||||
:: files in a subfolder named with the major version, ie: `2018.3`.
|
||||
:: Insert the key fob that contains the code signing certificate. Run
|
||||
:: the script passing the full version: `.\sign.bat 2018.3.1`.
|
||||
::
|
||||
:: The script will sign the installers and generate the corresponding
|
||||
:: hash files. These can then be uploaded to the salt repo.
|
||||
::
|
||||
:: The files must be in the following format:
|
||||
:: <Series>\Salt-Minion-<Version>-<Python Version>-<System Architecture>-Setup.exe
|
||||
:: So, for a Salt Minion installer for 2018.3.1 on AMD64 for Python 3
|
||||
:: file would be placed in a subdirectory named `2018.3` and the file
|
||||
:: would be named: `Salt-Minion-2018.3.1-Py3-AMD64-Setup.exe`. This
|
||||
:: is how the file is created by the NSI Script anyway.
|
||||
::
|
||||
:: ############################################################################
|
||||
@ echo off
|
||||
if [%1]==[] (
|
||||
echo You must pass a version
|
||||
goto quit
|
||||
) else (
|
||||
set "Version=%~1"
|
||||
)
|
||||
|
||||
for /F "tokens=1,2 delims=." %%a in ("%Version%") do (set Series=%%a.%%b)
|
||||
|
||||
:: Sign Installer Files
|
||||
echo ===========================================================================
|
||||
echo Signing...
|
||||
echo ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
signtool.exe sign /t http://timestamp.digicert.com ^
|
||||
"%Series%\Salt-Minion-%Version%-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-Minion-%Version%-x86-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-x86-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-Py2-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-Py2-x86-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-Py3-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-%Version%-Py3-x86-Setup.exe" ^
|
||||
"%Series%\Salt-Minion-%Version%-Py2-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-Minion-%Version%-Py2-x86-Setup.exe" ^
|
||||
"%Series%\Salt-Minion-%Version%-Py3-AMD64-Setup.exe" ^
|
||||
"%Series%\Salt-Minion-%Version%-Py3-x86-Setup.exe"
|
||||
echo ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
echo Signing Complete
|
||||
echo ===========================================================================
|
||||
|
||||
:: Create Hash files
|
||||
echo ===========================================================================
|
||||
echo Creating Hashes...
|
||||
echo ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
set "file_name=Salt-Minion-%Version%-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
|
||||
set "file_name=Salt-Minion-%Version%-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-Py2-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-Py2-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-Py3-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-%Version%-Py3-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-Minion-%Version%-Py2-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-Minion-%Version%-Py2-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-Minion-%Version%-Py3-AMD64-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
set "file_name=Salt-Minion-%Version%-Py3-x86-Setup.exe"
|
||||
set "file=.\%Series%\%file_name%"
|
||||
if exist "%file%" (
|
||||
echo - %file_name%
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm MD5 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.md5\""
|
||||
powershell -c "$hash = (Get-FileHash -Algorithm SHA256 \"%file%\").Hash; Out-File -InputObject $hash\" %file_name%\" -FilePath \"%file%.sha256\"")
|
||||
|
||||
echo ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
echo Hashing Complete
|
||||
echo ===========================================================================
|
||||
|
||||
:quit
|
4
pytest.ini
Normal file
4
pytest.ini
Normal file
@ -0,0 +1,4 @@
|
||||
[pytest]
|
||||
addopts = --ssh-tests -ra -sv
|
||||
testpaths = tests
|
||||
norecursedirs = tests/kitchen
|
@ -1,4 +0,0 @@
|
||||
-r base.txt
|
||||
|
||||
# Required by Tornado to handle threads stuff.
|
||||
futures>=2.0
|
@ -1 +0,0 @@
|
||||
-r base.txt
|
@ -1,8 +1,12 @@
|
||||
Jinja2
|
||||
msgpack-python>0.3,!=0.5.5
|
||||
# This should be changed to msgpack-python for Packages
|
||||
# msgpack-python>0.3,!=0.5.5
|
||||
msgpack>=0.5,!=0.5.5
|
||||
PyYAML
|
||||
MarkupSafe
|
||||
requests>=1.0.0
|
||||
tornado>=4.2.1,<5.0
|
||||
tornado>=4.2.1,<6.0; python_version < '3'
|
||||
tornado>=4.2.1,<5.0; python_version >= '3.4'
|
||||
|
||||
# Required by Tornado to handle threads stuff.
|
||||
futures>=2.0
|
||||
futures>=2.0; python_version < '3.0'
|
||||
|
17
requirements/dev.txt
Normal file
17
requirements/dev.txt
Normal file
@ -0,0 +1,17 @@
|
||||
-r base.txt
|
||||
|
||||
mock>=2.0.0
|
||||
apache-libcloud>=0.14.0
|
||||
boto>=2.32.1
|
||||
boto3>=1.2.1
|
||||
moto>=0.3.6
|
||||
SaltPyLint>=v2017.3.6
|
||||
pytest>=3.5.0
|
||||
git+https://github.com/saltstack/pytest-salt.git@master#egg=pytest-salt
|
||||
testinfra>=1.7.0
|
||||
|
||||
# httpretty Needs to be here for now even though it's a dependency of boto.
|
||||
# A pip install on a fresh system will decide to target httpretty 0.8.10 to
|
||||
# satisfy other requirements, and httpretty 0.8.10 has bugs in setup.py that
|
||||
# prevent it from being successfully installed (at least on Python 3.4).
|
||||
httpretty; python_version >= '3.4'
|
@ -1,12 +1,2 @@
|
||||
-r base-py2.txt
|
||||
|
||||
mock>=2.0.0
|
||||
apache-libcloud>=0.14.0
|
||||
boto>=2.32.1
|
||||
boto3>=1.2.1
|
||||
moto>=0.3.6
|
||||
SaltPyLint>=v2017.3.6
|
||||
pytest>=3.5.0
|
||||
git+https://github.com/eisensheng/pytest-catchlog.git@develop#egg=Pytest-catchlog
|
||||
git+https://github.com/saltstack/pytest-salt.git@master#egg=pytest-salt
|
||||
testinfra>=1.7.0
|
||||
# This is a legacy file, use dev.txt
|
||||
-r dev.txt
|
||||
|
@ -1,17 +1,2 @@
|
||||
-r base-py3.txt
|
||||
|
||||
mock>=2.0.0
|
||||
apache-libcloud>=0.14.0
|
||||
boto>=2.32.1
|
||||
boto3>=1.2.1
|
||||
moto>=0.3.6
|
||||
# httpretty Needs to be here for now even though it's a dependency of boto.
|
||||
# A pip install on a fresh system will decide to target httpretty 0.8.10 to
|
||||
# satisfy other requirements, and httpretty 0.8.10 has bugs in setup.py that
|
||||
# prevent it from being successfully installed (at least on Python 3.4).
|
||||
httpretty
|
||||
SaltPyLint>=v2017.2.29
|
||||
pytest>=3.5.0
|
||||
git+https://github.com/saltstack/pytest-salt.git@master#egg=pytest-salt
|
||||
git+https://github.com/eisensheng/pytest-catchlog.git@develop#egg=Pytest-catchlog
|
||||
testinfra>=1.7.0
|
||||
# This is a legacy file, use dev.txt
|
||||
-r dev.txt
|
||||
|
@ -410,7 +410,7 @@ class SSH(object):
|
||||
'host': hostname,
|
||||
'user': user,
|
||||
}
|
||||
if not self.opts.get('ssh_skip_roster'):
|
||||
if self.opts.get('ssh_update_roster'):
|
||||
self._update_roster()
|
||||
|
||||
def get_pubkey(self):
|
||||
|
@ -1916,7 +1916,7 @@ class Map(Cloud):
|
||||
pmap = self.map_providers_parallel(cached=cached)
|
||||
exist = set()
|
||||
defined = set()
|
||||
for profile_name, nodes in six.iteritems(self.rendered_map):
|
||||
for profile_name, nodes in six.iteritems(copy.deepcopy(self.rendered_map)):
|
||||
if profile_name not in self.opts['profiles']:
|
||||
msg = (
|
||||
'The required profile, \'{0}\', defined in the map '
|
||||
@ -1937,13 +1937,13 @@ class Map(Cloud):
|
||||
# Get associated provider data, in case something like size
|
||||
# or image is specified in the provider file. See issue #32510.
|
||||
alias, driver = profile_data.get('provider').split(':')
|
||||
provider_details = self.opts['providers'][alias][driver].copy()
|
||||
provider_details = copy.deepcopy(self.opts['providers'][alias][driver])
|
||||
del provider_details['profiles']
|
||||
|
||||
# Update the provider details information with profile data
|
||||
# Profile data should override provider data, if defined.
|
||||
# This keeps map file data definitions consistent with -p usage.
|
||||
provider_details.update(profile_data)
|
||||
salt.utils.dictupdate.update(provider_details, profile_data)
|
||||
profile_data = provider_details
|
||||
|
||||
for nodename, overrides in six.iteritems(nodes):
|
||||
|
@ -3675,6 +3675,25 @@ def enable_term_protect(name, call=None):
|
||||
return _toggle_term_protect(name, 'true')
|
||||
|
||||
|
||||
def disable_term_protect(name, call=None):
|
||||
'''
|
||||
Disable termination protection on a node
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt-cloud -a disable_term_protect mymachine
|
||||
'''
|
||||
if call != 'action':
|
||||
raise SaltCloudSystemExit(
|
||||
'The enable_term_protect action must be called with '
|
||||
'-a or --action.'
|
||||
)
|
||||
|
||||
return _toggle_term_protect(name, 'false')
|
||||
|
||||
|
||||
def disable_detailed_monitoring(name, call=None):
|
||||
'''
|
||||
Enable/disable detailed monitoring on a node
|
||||
|
@ -993,6 +993,59 @@ def get_password(vm_):
|
||||
)
|
||||
|
||||
|
||||
def _decode_linode_plan_label(label):
|
||||
'''
|
||||
Attempts to decode a user-supplied Linode plan label
|
||||
into the format in Linode API output
|
||||
|
||||
label
|
||||
The label, or name, of the plan to decode.
|
||||
|
||||
Example:
|
||||
`Linode 2048` will decode to `Linode 2GB`
|
||||
'''
|
||||
sizes = avail_sizes()
|
||||
|
||||
if label not in sizes:
|
||||
if 'GB' in label:
|
||||
raise SaltCloudException(
|
||||
'Invalid Linode plan ({}) specified - call avail_sizes() for all available options'.format(label)
|
||||
)
|
||||
else:
|
||||
plan = label.split()
|
||||
|
||||
if len(plan) != 2:
|
||||
raise SaltCloudException(
|
||||
'Invalid Linode plan ({}) specified - call avail_sizes() for all available options'.format(label)
|
||||
)
|
||||
|
||||
plan_type = plan[0]
|
||||
try:
|
||||
plan_size = int(plan[1])
|
||||
except TypeError:
|
||||
plan_size = 0
|
||||
log.debug('Failed to decode Linode plan label in Cloud Profile: %s', label)
|
||||
|
||||
if plan_type == 'Linode' and plan_size == 1024:
|
||||
plan_type = 'Nanode'
|
||||
|
||||
plan_size = plan_size/1024
|
||||
new_label = "{} {}GB".format(plan_type, plan_size)
|
||||
|
||||
if new_label not in sizes:
|
||||
raise SaltCloudException(
|
||||
'Invalid Linode plan ({}) specified - call avail_sizes() for all available options'.format(new_label)
|
||||
)
|
||||
|
||||
log.warning('An outdated Linode plan label was detected in your Cloud Profile ({}).'
|
||||
' Please update the profile to use'
|
||||
' the new label format ({}) for the requested Linode plan size.'.format(label, new_label))
|
||||
|
||||
label = new_label
|
||||
|
||||
return sizes[label]['PLANID']
|
||||
|
||||
|
||||
def get_plan_id(kwargs=None, call=None):
|
||||
'''
|
||||
Returns the Linode Plan ID.
|
||||
@ -1020,7 +1073,9 @@ def get_plan_id(kwargs=None, call=None):
|
||||
'The get_plan_id function requires a \'label\'.'
|
||||
)
|
||||
|
||||
return avail_sizes()[label]['PLANID']
|
||||
label = _decode_linode_plan_label(label)
|
||||
|
||||
return label
|
||||
|
||||
|
||||
def get_private_ip(vm_):
|
||||
|
@ -597,6 +597,7 @@ def _clean_create_kwargs(**kwargs):
|
||||
'volume_size': int,
|
||||
'nat_destination': six.string_types,
|
||||
'group': six.string_types,
|
||||
'userdata': six.string_types,
|
||||
}
|
||||
extra = kwargs.pop('extra', {})
|
||||
for key, value in six.iteritems(kwargs.copy()):
|
||||
|
@ -989,6 +989,7 @@ VALID_OPTS = {
|
||||
'ssh_identities_only': bool,
|
||||
'ssh_log_file': six.string_types,
|
||||
'ssh_config_file': six.string_types,
|
||||
'ssh_merge_pillar': bool,
|
||||
|
||||
# Enable ioflo verbose logging. Warning! Very verbose!
|
||||
'ioflo_verbose': int,
|
||||
@ -1485,6 +1486,7 @@ DEFAULT_MINION_OPTS = {
|
||||
},
|
||||
'discovery': False,
|
||||
'schedule': {},
|
||||
'ssh_merge_pillar': True
|
||||
}
|
||||
|
||||
DEFAULT_MASTER_OPTS = {
|
||||
@ -2088,7 +2090,7 @@ def _validate_ssh_minion_opts(opts):
|
||||
|
||||
for opt_name in list(ssh_minion_opts):
|
||||
if re.match('^[a-z0-9]+fs_', opt_name, flags=re.IGNORECASE) \
|
||||
or 'pillar' in opt_name \
|
||||
or ('pillar' in opt_name and not 'ssh_merge_pillar' == opt_name) \
|
||||
or opt_name in ('fileserver_backend',):
|
||||
log.warning(
|
||||
'\'%s\' is not a valid ssh_minion_opts parameter, ignoring',
|
||||
|
@ -13,6 +13,7 @@ import string
|
||||
import shutil
|
||||
import ftplib
|
||||
from tornado.httputil import parse_response_start_line, HTTPHeaders, HTTPInputError
|
||||
import salt.utils.atomicfile
|
||||
|
||||
# Import salt libs
|
||||
from salt.exceptions import (
|
||||
@ -46,6 +47,7 @@ from salt.ext.six.moves.urllib.parse import urlparse, urlunparse
|
||||
# pylint: enable=no-name-in-module,import-error
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
MAX_FILENAME_LENGTH = 255
|
||||
|
||||
|
||||
def get_file_client(opts, pillar=False):
|
||||
@ -830,6 +832,9 @@ class Client(object):
|
||||
else:
|
||||
file_name = url_data.path
|
||||
|
||||
if len(file_name) > MAX_FILENAME_LENGTH:
|
||||
file_name = salt.utils.hashutils.sha256_digest(file_name)
|
||||
|
||||
return salt.utils.path.join(
|
||||
cachedir,
|
||||
'extrn_files',
|
||||
@ -1166,7 +1171,11 @@ class RemoteClient(Client):
|
||||
destdir = os.path.dirname(dest)
|
||||
if not os.path.isdir(destdir):
|
||||
if makedirs:
|
||||
os.makedirs(destdir)
|
||||
try:
|
||||
os.makedirs(destdir)
|
||||
except OSError as exc:
|
||||
if exc.errno != errno.EEXIST: # ignore if it was there already
|
||||
raise
|
||||
else:
|
||||
return False
|
||||
# We need an open filehandle here, that's why we're not using a
|
||||
@ -1221,7 +1230,7 @@ class RemoteClient(Client):
|
||||
# remove it to avoid a traceback trying to write the file
|
||||
if os.path.isdir(dest):
|
||||
salt.utils.files.rm_rf(dest)
|
||||
fn_ = salt.utils.files.fopen(dest, 'wb+')
|
||||
fn_ = salt.utils.atomicfile.atomic_open(dest, 'wb+')
|
||||
if data.get('gzip', None):
|
||||
data = salt.utils.gzip_util.uncompress(data['data'])
|
||||
else:
|
||||
@ -1391,14 +1400,12 @@ class RemoteClient(Client):
|
||||
'''
|
||||
Return the metadata derived from the master_tops system
|
||||
'''
|
||||
salt.utils.versions.warn_until(
|
||||
'Magnesium',
|
||||
'The _ext_nodes master function has '
|
||||
'been renamed to _master_tops. To ensure '
|
||||
'compatibility when using older Salt masters '
|
||||
'we continue to pass the function as _ext_nodes.'
|
||||
log.debug(
|
||||
'The _ext_nodes master function has been renamed to _master_tops. '
|
||||
'To ensure compatibility when using older Salt masters we will '
|
||||
'continue to invoke the function as _ext_nodes until the '
|
||||
'Magnesium release.'
|
||||
)
|
||||
|
||||
# TODO: Change back to _master_tops
|
||||
# for Magnesium release
|
||||
load = {'cmd': '_ext_nodes',
|
||||
|
@ -850,7 +850,8 @@ class FSChan(object):
|
||||
self.opts['__fs_update'] = True
|
||||
else:
|
||||
self.fs.update()
|
||||
self.cmd_stub = {'master_tops': {}}
|
||||
self.cmd_stub = {'master_tops': {},
|
||||
'ext_nodes': {}}
|
||||
|
||||
def send(self, load, tries=None, timeout=None, raw=False): # pylint: disable=unused-argument
|
||||
'''
|
||||
|
@ -253,7 +253,7 @@ def file_hash(load, fnd):
|
||||
except OSError:
|
||||
pass
|
||||
return file_hash(load, fnd)
|
||||
if os.path.getmtime(path) == mtime:
|
||||
if str(os.path.getmtime(path)) == mtime:
|
||||
# check if mtime changed
|
||||
ret['hsum'] = hsum
|
||||
return ret
|
||||
|
@ -451,7 +451,11 @@ def _bsd_memdata(osdata):
|
||||
|
||||
if osdata['kernel'] in ['OpenBSD', 'NetBSD']:
|
||||
swapctl = salt.utils.path.which('swapctl')
|
||||
swap_total = __salt__['cmd.run']('{0} -sk'.format(swapctl)).split(' ')[1]
|
||||
swap_data = __salt__['cmd.run']('{0} -sk'.format(swapctl))
|
||||
if swap_data == 'no swap devices configured':
|
||||
swap_total = 0
|
||||
else:
|
||||
swap_total = swap_data.split(' ')[1]
|
||||
else:
|
||||
swap_total = __salt__['cmd.run']('{0} -n vm.swap_total'.format(sysctl))
|
||||
grains['swap_total'] = int(swap_total) // 1024 // 1024
|
||||
@ -1213,6 +1217,7 @@ _OS_NAME_MAP = {
|
||||
'synology': 'Synology',
|
||||
'nilrt': 'NILinuxRT',
|
||||
'nilrt-xfce': 'NILinuxRT-XFCE',
|
||||
'poky': 'Poky',
|
||||
'manjaro': 'Manjaro',
|
||||
'manjarolin': 'Manjaro',
|
||||
'antergos': 'Antergos',
|
||||
@ -1766,7 +1771,7 @@ def os_data():
|
||||
osarch = __salt__['cmd.run']('dpkg --print-architecture').strip()
|
||||
elif grains.get('os_family') == 'RedHat':
|
||||
osarch = __salt__['cmd.run']('rpm --eval %{_host_cpu}').strip()
|
||||
elif grains.get('os_family') == 'NILinuxRT':
|
||||
elif grains.get('os_family') in ('NILinuxRT', 'Poky'):
|
||||
archinfo = {}
|
||||
for line in __salt__['cmd.run']('opkg print-architecture').splitlines():
|
||||
if line.startswith('arch'):
|
||||
@ -2304,9 +2309,9 @@ def _hw_data(osdata):
|
||||
|
||||
product_regexes = [
|
||||
re.compile(r) for r in [
|
||||
r'(?im)^\s*System\s+Configuration:\s*.*?sun\d\S+\s(.*)', # prtdiag
|
||||
r'(?im)^\s*banner-name:\s*(.*)', # prtconf
|
||||
r'(?im)^\s*product-name:\s*(.*)', # prtconf
|
||||
r'(?im)^\s*System\s+Configuration:\s*.*?sun\d\S+[^\S\r\n]*(.*)', # prtdiag
|
||||
r'(?im)^[^\S\r\n]*banner-name:[^\S\r\n]*(.*)', # prtconf
|
||||
r'(?im)^[^\S\r\n]*product-name:[^\S\r\n]*(.*)', # prtconf
|
||||
]
|
||||
]
|
||||
|
||||
@ -2373,8 +2378,11 @@ def _hw_data(osdata):
|
||||
for regex in product_regexes:
|
||||
res = regex.search(data)
|
||||
if res and len(res.groups()) >= 1:
|
||||
grains['product'] = res.group(1).strip().replace("'", "")
|
||||
break
|
||||
t_productname = res.group(1).strip().replace("'", "")
|
||||
if t_productname:
|
||||
grains['product'] = t_productname
|
||||
grains['productname'] = t_productname
|
||||
break
|
||||
|
||||
return grains
|
||||
|
||||
|
@ -91,7 +91,7 @@ def _retrieve_device_cache(proxy=None):
|
||||
DEVICE_CACHE = proxy['napalm.get_device']()
|
||||
elif not proxy and salt.utils.napalm.is_minion(__opts__):
|
||||
# if proxy var not passed and is running in a straight minion
|
||||
DEVICE_CACHE = salt.utils.napalm.get_device_opts(__opts__)
|
||||
DEVICE_CACHE = salt.utils.napalm.get_device(__opts__)
|
||||
return DEVICE_CACHE
|
||||
|
||||
|
||||
|
@ -1629,7 +1629,11 @@ class LazyLoader(salt.utils.lazy.LazyDict):
|
||||
return True
|
||||
# if the modulename isn't in the whitelist, don't bother
|
||||
if self.whitelist and mod_name not in self.whitelist:
|
||||
raise KeyError
|
||||
log.error(
|
||||
'Failed to load function %s because its module (%s) is '
|
||||
'not in the whitelist: %s', key, mod_name, self.whitelist
|
||||
)
|
||||
raise KeyError(key)
|
||||
|
||||
def _inner_load(mod_name):
|
||||
for name in self._iter_files(mod_name):
|
||||
|
@ -103,6 +103,20 @@ class SysLogHandler(ExcInfoOnLogLevelFormatMixIn, logging.handlers.SysLogHandler
|
||||
'''
|
||||
Syslog handler which properly handles exc_info on a per handler basis
|
||||
'''
|
||||
def handleError(self, record):
|
||||
'''
|
||||
Override the default error handling mechanism for py3
|
||||
Deal with syslog os errors when the log file does not exist
|
||||
'''
|
||||
handled = False
|
||||
if sys.stderr and sys.version_info >= (3, 5, 4):
|
||||
t, v, tb = sys.exc_info()
|
||||
if t.__name__ in 'FileNotFoundError':
|
||||
sys.stderr.write('[WARNING ] The log_file does not exist. Logging not setup correctly or syslog service not started.\n')
|
||||
handled = True
|
||||
|
||||
if not handled:
|
||||
super(SysLogHandler, self).handleError(record)
|
||||
|
||||
|
||||
class RotatingFileHandler(ExcInfoOnLogLevelFormatMixIn, logging.handlers.RotatingFileHandler, NewStyleClassMixIn):
|
||||
|
@ -5,7 +5,12 @@
|
||||
|
||||
.. versionadded:: 0.17.0
|
||||
|
||||
This module provides a `Sentry`_ logging handler.
|
||||
This module provides a `Sentry`_ logging handler. Sentry is an open source
|
||||
error tracking platform that provides deep context about exceptions that
|
||||
happen in production. Details about stack traces along with the context
|
||||
variables available at the time of the exeption are easily browsable and
|
||||
filterable from the online interface. For more details please see
|
||||
`Sentry`_.
|
||||
|
||||
.. admonition:: Note
|
||||
|
||||
@ -41,6 +46,11 @@
|
||||
- cpuarch
|
||||
- ec2.tags.environment
|
||||
|
||||
.. admonition:: Note
|
||||
|
||||
The ``public_key`` and ``secret_key`` variables are not supported with
|
||||
Sentry > 3.0. The `DSN`_ key should be used instead.
|
||||
|
||||
All the client configuration keys are supported, please see the
|
||||
`Raven client documentation`_.
|
||||
|
||||
|
@ -410,9 +410,9 @@ def list_(name,
|
||||
item.sort()
|
||||
|
||||
if verbose:
|
||||
ret = {'dirs': sorted(dirs),
|
||||
'files': sorted(files),
|
||||
'links': sorted(links)}
|
||||
ret = {'dirs': sorted(salt.utils.data.decode_list(dirs)),
|
||||
'files': sorted(salt.utils.data.decode_list(files)),
|
||||
'links': sorted(salt.utils.data.decode_list(links))}
|
||||
ret['top_level_dirs'] = [x for x in ret['dirs']
|
||||
if x.count('/') == 1]
|
||||
ret['top_level_files'] = [x for x in ret['files']
|
||||
|
@ -53,7 +53,6 @@ import time
|
||||
|
||||
# Import Salt libs
|
||||
from salt.exceptions import SaltInvocationError, CommandExecutionError
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.versions
|
||||
|
||||
|
@ -54,7 +54,6 @@ import logging
|
||||
import time
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.versions
|
||||
from salt.exceptions import SaltInvocationError
|
||||
@ -551,6 +550,10 @@ def associate_vpc_with_hosted_zone(HostedZoneId=None, Name=None, VPCId=None,
|
||||
r = conn.associate_vpc_with_hosted_zone(**args)
|
||||
return _wait_for_sync(r['ChangeInfo']['Id'], conn)
|
||||
except ClientError as e:
|
||||
if e.response.get('Error', {}).get('Code') == 'ConflictingDomainExists':
|
||||
log.debug('VPC Association already exists.')
|
||||
# return True since the current state is the desired one
|
||||
return True
|
||||
if tries and e.response.get('Error', {}).get('Code') == 'Throttling':
|
||||
log.debug('Throttled by AWS API.')
|
||||
time.sleep(3)
|
||||
@ -642,6 +645,10 @@ def disassociate_vpc_from_hosted_zone(HostedZoneId=None, Name=None, VPCId=None,
|
||||
r = conn.disassociate_vpc_from_hosted_zone(**args)
|
||||
return _wait_for_sync(r['ChangeInfo']['Id'], conn)
|
||||
except ClientError as e:
|
||||
if e.response.get('Error', {}).get('Code') == 'VPCAssociationNotFound':
|
||||
log.debug('No VPC Association exists.')
|
||||
# return True since the current state is the desired one
|
||||
return True
|
||||
if tries and e.response.get('Error', {}).get('Code') == 'Throttling':
|
||||
log.debug('Throttled by AWS API.')
|
||||
time.sleep(3)
|
||||
|
@ -82,7 +82,6 @@ import datetime
|
||||
|
||||
# Import Salt libs
|
||||
from salt.ext import six
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.json
|
||||
import salt.utils.versions
|
||||
@ -182,7 +181,7 @@ def _find_apis_by_name(name, description=None,
|
||||
apis = _filter_apis_desc(description, apis)
|
||||
return {'restapi': [_convert_datetime_str(api) for api in apis]}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_apis(name=None, description=None, region=None, key=None, keyid=None, profile=None):
|
||||
@ -251,7 +250,7 @@ def create_api(name, description, cloneFrom=None,
|
||||
api = _convert_datetime_str(api)
|
||||
return {'created': True, 'restapi': api} if api else {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api(name, description=None, region=None, key=None, keyid=None, profile=None):
|
||||
@ -282,7 +281,7 @@ def delete_api(name, description=None, region=None, key=None, keyid=None, profil
|
||||
else:
|
||||
return {'deleted': False}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_resources(restApiId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -303,7 +302,7 @@ def describe_api_resources(restApiId, region=None, key=None, keyid=None, profile
|
||||
|
||||
return {'resources': resources}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_resource(restApiId, path,
|
||||
@ -364,7 +363,7 @@ def create_api_resources(restApiId, path,
|
||||
else:
|
||||
return {'created': False, 'error': 'unexpected error.'}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_resources(restApiId, path,
|
||||
@ -393,7 +392,7 @@ def delete_api_resources(restApiId, path,
|
||||
else:
|
||||
return {'deleted': False, 'error': 'no resource found by {0}'.format(path)}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_resource_method(restApiId, resourcePath, httpMethod,
|
||||
@ -421,7 +420,7 @@ def describe_api_resource_method(restApiId, resourcePath, httpMethod,
|
||||
method = conn.get_method(restApiId=restApiId, resourceId=resource['id'], httpMethod=httpMethod)
|
||||
return {'method': method}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
@ -440,7 +439,7 @@ def describe_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
response = conn.get_api_key(apiKey=apiKey)
|
||||
return {'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_keys(region=None, key=None, keyid=None, profile=None):
|
||||
@ -460,7 +459,7 @@ def describe_api_keys(region=None, key=None, keyid=None, profile=None):
|
||||
|
||||
return {'apiKeys': [_convert_datetime_str(apikey) for apikey in apikeys]}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_key(name, description, enabled=True, stageKeys=None,
|
||||
@ -498,7 +497,7 @@ def create_api_key(name, description, enabled=True, stageKeys=None,
|
||||
|
||||
return {'created': True, 'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
@ -517,7 +516,7 @@ def delete_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
conn.delete_api_key(apiKey=apiKey)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _api_key_patch_replace(conn, apiKey, path, value):
|
||||
@ -570,7 +569,7 @@ def update_api_key_description(apiKey, description, region=None, key=None, keyid
|
||||
response = _api_key_patch_replace(conn, apiKey, '/description', description)
|
||||
return {'updated': True, 'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def enable_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
@ -589,7 +588,7 @@ def enable_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
response = _api_key_patch_replace(conn, apiKey, '/enabled', 'True')
|
||||
return {'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def disable_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
@ -608,7 +607,7 @@ def disable_api_key(apiKey, region=None, key=None, keyid=None, profile=None):
|
||||
response = _api_key_patch_replace(conn, apiKey, '/enabled', 'False')
|
||||
return {'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def associate_api_key_stagekeys(apiKey, stagekeyslist, region=None, key=None, keyid=None, profile=None):
|
||||
@ -629,7 +628,7 @@ def associate_api_key_stagekeys(apiKey, stagekeyslist, region=None, key=None, ke
|
||||
response = _api_key_patch_add(conn, apiKey, pvlist)
|
||||
return {'associated': True, 'apiKey': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'associated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'associated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def disassociate_api_key_stagekeys(apiKey, stagekeyslist, region=None, key=None, keyid=None, profile=None):
|
||||
@ -650,7 +649,7 @@ def disassociate_api_key_stagekeys(apiKey, stagekeyslist, region=None, key=None,
|
||||
response = _api_key_patch_remove(conn, apiKey, pvlist)
|
||||
return {'disassociated': True}
|
||||
except ClientError as e:
|
||||
return {'disassociated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'disassociated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_deployments(restApiId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -678,7 +677,7 @@ def describe_api_deployments(restApiId, region=None, key=None, keyid=None, profi
|
||||
|
||||
return {'deployments': [_convert_datetime_str(deployment) for deployment in deployments]}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_deployment(restApiId, deploymentId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -697,7 +696,7 @@ def describe_api_deployment(restApiId, deploymentId, region=None, key=None, keyi
|
||||
deployment = conn.get_deployment(restApiId=restApiId, deploymentId=deploymentId)
|
||||
return {'deployment': _convert_datetime_str(deployment)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def activate_api_deployment(restApiId, stageName, deploymentId,
|
||||
@ -720,7 +719,7 @@ def activate_api_deployment(restApiId, stageName, deploymentId,
|
||||
'value': deploymentId}])
|
||||
return {'set': True, 'response': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'set': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'set': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_deployment(restApiId, stageName, stageDescription='', description='', cacheClusterEnabled=False,
|
||||
@ -747,7 +746,7 @@ def create_api_deployment(restApiId, stageName, stageDescription='', description
|
||||
variables=variables)
|
||||
return {'created': True, 'deployment': _convert_datetime_str(deployment)}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_deployment(restApiId, deploymentId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -766,7 +765,7 @@ def delete_api_deployment(restApiId, deploymentId, region=None, key=None, keyid=
|
||||
conn.delete_deployment(restApiId=restApiId, deploymentId=deploymentId)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def overwrite_api_stage_variables(restApiId, stageName, variables, region=None, key=None, keyid=None, profile=None):
|
||||
@ -812,7 +811,7 @@ def overwrite_api_stage_variables(restApiId, stageName, variables, region=None,
|
||||
|
||||
return {'overwrite': True, 'stage': _convert_datetime_str(stage)}
|
||||
except ClientError as e:
|
||||
return {'overwrite': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'overwrite': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_stage(restApiId, stageName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -831,7 +830,7 @@ def describe_api_stage(restApiId, stageName, region=None, key=None, keyid=None,
|
||||
stage = conn.get_stage(restApiId=restApiId, stageName=stageName)
|
||||
return {'stage': _convert_datetime_str(stage)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_stages(restApiId, deploymentId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -850,7 +849,7 @@ def describe_api_stages(restApiId, deploymentId, region=None, key=None, keyid=No
|
||||
stages = conn.get_stages(restApiId=restApiId, deploymentId=deploymentId)
|
||||
return {'stages': [_convert_datetime_str(stage) for stage in stages['item']]}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_stage(restApiId, stageName, deploymentId, description='',
|
||||
@ -876,7 +875,7 @@ def create_api_stage(restApiId, stageName, deploymentId, description='',
|
||||
cacheClusterSize=cacheClusterSize, variables=variables)
|
||||
return {'created': True, 'stage': _convert_datetime_str(stage)}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_stage(restApiId, stageName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -895,7 +894,7 @@ def delete_api_stage(restApiId, stageName, region=None, key=None, keyid=None, pr
|
||||
conn.delete_stage(restApiId=restApiId, stageName=stageName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def flush_api_stage_cache(restApiId, stageName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -914,7 +913,7 @@ def flush_api_stage_cache(restApiId, stageName, region=None, key=None, keyid=Non
|
||||
conn.flush_stage_cache(restApiId=restApiId, stageName=stageName)
|
||||
return {'flushed': True}
|
||||
except ClientError as e:
|
||||
return {'flushed': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'flushed': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_method(restApiId, resourcePath, httpMethod, authorizationType,
|
||||
@ -946,7 +945,7 @@ def create_api_method(restApiId, resourcePath, httpMethod, authorizationType,
|
||||
return {'created': False, 'error': 'Failed to create method'}
|
||||
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_method(restApiId, resourcePath, httpMethod, region=None, key=None, keyid=None, profile=None):
|
||||
@ -969,7 +968,7 @@ def describe_api_method(restApiId, resourcePath, httpMethod, region=None, key=No
|
||||
return {'method': _convert_datetime_str(method)}
|
||||
return {'error': 'get API method failed: no such resource'}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_method(restApiId, resourcePath, httpMethod, region=None, key=None, keyid=None, profile=None):
|
||||
@ -992,7 +991,7 @@ def delete_api_method(restApiId, resourcePath, httpMethod, region=None, key=None
|
||||
return {'deleted': True}
|
||||
return {'deleted': False, 'error': 'get API method failed: no such resource'}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_method_response(restApiId, resourcePath, httpMethod, statusCode, responseParameters=None,
|
||||
@ -1022,7 +1021,7 @@ def create_api_method_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
return {'created': True, 'response': response}
|
||||
return {'created': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_method_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
@ -1047,7 +1046,7 @@ def delete_api_method_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
return {'deleted': True}
|
||||
return {'deleted': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_method_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
@ -1072,7 +1071,7 @@ def describe_api_method_response(restApiId, resourcePath, httpMethod, statusCode
|
||||
return {'response': _convert_datetime_str(response)}
|
||||
return {'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_models(restApiId, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1091,7 +1090,7 @@ def describe_api_models(restApiId, region=None, key=None, keyid=None, profile=No
|
||||
models = _multi_call(conn.get_models, 'items', restApiId=restApiId)
|
||||
return {'models': [_convert_datetime_str(model) for model in models]}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_model(restApiId, modelName, flatten=True, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1110,7 +1109,7 @@ def describe_api_model(restApiId, modelName, flatten=True, region=None, key=None
|
||||
model = conn.get_model(restApiId=restApiId, modelName=modelName, flatten=flatten)
|
||||
return {'model': _convert_datetime_str(model)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def api_model_exists(restApiId, modelName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1154,7 +1153,7 @@ def update_api_model_schema(restApiId, modelName, schema, region=None, key=None,
|
||||
response = _api_model_patch_replace(conn, restApiId, modelName, '/schema', schema_json)
|
||||
return {'updated': True, 'model': _convert_datetime_str(response)}
|
||||
except ClientError as e:
|
||||
return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_model(restApiId, modelName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1173,7 +1172,7 @@ def delete_api_model(restApiId, modelName, region=None, key=None, keyid=None, pr
|
||||
conn.delete_model(restApiId=restApiId, modelName=modelName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_model(restApiId, modelName, modelDescription, schema, contentType='application/json',
|
||||
@ -1196,7 +1195,7 @@ def create_api_model(restApiId, modelName, modelDescription, schema, contentType
|
||||
schema=schema_json, contentType=contentType)
|
||||
return {'created': True, 'model': _convert_datetime_str(model)}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_integration(restApiId, resourcePath, httpMethod, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1219,7 +1218,7 @@ def describe_api_integration(restApiId, resourcePath, httpMethod, region=None, k
|
||||
return {'integration': _convert_datetime_str(integration)}
|
||||
return {'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_api_integration_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
@ -1244,7 +1243,7 @@ def describe_api_integration_response(restApiId, resourcePath, httpMethod, statu
|
||||
return {'response': _convert_datetime_str(response)}
|
||||
return {'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_integration(restApiId, resourcePath, httpMethod, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1267,7 +1266,7 @@ def delete_api_integration(restApiId, resourcePath, httpMethod, region=None, key
|
||||
return {'deleted': True}
|
||||
return {'deleted': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_api_integration_response(restApiId, resourcePath, httpMethod, statusCode,
|
||||
@ -1292,7 +1291,7 @@ def delete_api_integration_response(restApiId, resourcePath, httpMethod, statusC
|
||||
return {'deleted': True}
|
||||
return {'deleted': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _get_role_arn(name, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1349,7 +1348,7 @@ def create_api_integration(restApiId, resourcePath, httpMethod, integrationType,
|
||||
return {'created': True, 'integration': integration}
|
||||
return {'created': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_api_integration_response(restApiId, resourcePath, httpMethod, statusCode, selectionPattern,
|
||||
@ -1382,7 +1381,7 @@ def create_api_integration_response(restApiId, resourcePath, httpMethod, statusC
|
||||
return {'created': True, 'response': response}
|
||||
return {'created': False, 'error': 'no such resource'}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _filter_plans(attr, name, plans):
|
||||
@ -1418,7 +1417,7 @@ def describe_usage_plans(name=None, plan_id=None, region=None, key=None, keyid=N
|
||||
return {'plans': [_convert_datetime_str(plan) for plan in plans]}
|
||||
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _validate_throttle(throttle):
|
||||
@ -1497,7 +1496,7 @@ def create_usage_plan(name, description=None, throttle=None, quota=None, region=
|
||||
res = conn.create_usage_plan(**values)
|
||||
return {'created': True, 'result': res}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
except (TypeError, ValueError) as e:
|
||||
return {'error': six.text_type(e)}
|
||||
|
||||
@ -1571,7 +1570,7 @@ def update_usage_plan(plan_id, throttle=None, quota=None, region=None, key=None,
|
||||
return {'updated': False}
|
||||
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
except (TypeError, ValueError) as e:
|
||||
return {'error': six.text_type(e)}
|
||||
|
||||
@ -1600,7 +1599,7 @@ def delete_usage_plan(plan_id, region=None, key=None, keyid=None, profile=None):
|
||||
res = conn.delete_usage_plan(usagePlanId=plan_id)
|
||||
return {'deleted': True, 'usagePlanId': plan_id}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _update_usage_plan_apis(plan_id, apis, op, region=None, key=None, keyid=None, profile=None):
|
||||
@ -1634,7 +1633,7 @@ def _update_usage_plan_apis(plan_id, apis, op, region=None, key=None, keyid=None
|
||||
patchOperations=patchOperations)
|
||||
return {'success': True, 'result': res}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
except Exception as e:
|
||||
return {'error': e}
|
||||
|
||||
|
@ -74,7 +74,6 @@ except ImportError:
|
||||
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.json
|
||||
import salt.utils.odict as odict
|
||||
@ -885,7 +884,7 @@ def enter_standby(name, instance_ids, should_decrement_desired_capacity=False,
|
||||
AutoScalingGroupName=name,
|
||||
ShouldDecrementDesiredCapacity=should_decrement_desired_capacity)
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': err}
|
||||
@ -911,7 +910,7 @@ def exit_standby(name, instance_ids, should_decrement_desired_capacity=False,
|
||||
InstanceIds=instance_ids,
|
||||
AutoScalingGroupName=name)
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': err}
|
||||
|
@ -55,7 +55,6 @@ import logging
|
||||
|
||||
# Import Salt libs
|
||||
from salt.ext import six
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.versions
|
||||
|
||||
@ -117,7 +116,7 @@ def exists(Name,
|
||||
conn.get_trail_status(Name=Name)
|
||||
return {'exists': True}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'TrailNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': err}
|
||||
@ -167,7 +166,7 @@ def create(Name,
|
||||
log.warning('Trail was not created')
|
||||
return {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete(Name,
|
||||
@ -191,7 +190,7 @@ def delete(Name,
|
||||
conn.delete_trail(Name=Name)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe(Name,
|
||||
@ -224,10 +223,10 @@ def describe(Name,
|
||||
else:
|
||||
return {'trail': None}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'TrailNotFoundException':
|
||||
return {'trail': None}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def status(Name,
|
||||
@ -265,10 +264,10 @@ def status(Name,
|
||||
else:
|
||||
return {'trail': None}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'TrailNotFoundException':
|
||||
return {'trail': None}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list(region=None, key=None, keyid=None, profile=None):
|
||||
@ -293,7 +292,7 @@ def list(region=None, key=None, keyid=None, profile=None):
|
||||
log.warning('No trails found')
|
||||
return {'trails': trails.get('trailList', [])}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def update(Name,
|
||||
@ -340,7 +339,7 @@ def update(Name,
|
||||
log.warning('Trail was not created')
|
||||
return {'updated': False}
|
||||
except ClientError as e:
|
||||
return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def start_logging(Name,
|
||||
@ -364,7 +363,7 @@ def start_logging(Name,
|
||||
conn.start_logging(Name=Name)
|
||||
return {'started': True}
|
||||
except ClientError as e:
|
||||
return {'started': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'started': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def stop_logging(Name,
|
||||
@ -388,7 +387,7 @@ def stop_logging(Name,
|
||||
conn.stop_logging(Name=Name)
|
||||
return {'stopped': True}
|
||||
except ClientError as e:
|
||||
return {'stopped': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'stopped': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _get_trail_arn(name, region=None, key=None, keyid=None, profile=None):
|
||||
@ -433,7 +432,7 @@ def add_tags(Name,
|
||||
profile=profile), TagsList=tagslist)
|
||||
return {'tagged': True}
|
||||
except ClientError as e:
|
||||
return {'tagged': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'tagged': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def remove_tags(Name,
|
||||
@ -464,7 +463,7 @@ def remove_tags(Name,
|
||||
profile=profile), TagsList=tagslist)
|
||||
return {'tagged': True}
|
||||
except ClientError as e:
|
||||
return {'tagged': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'tagged': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_tags(Name,
|
||||
@ -497,4 +496,4 @@ def list_tags(Name,
|
||||
tagdict[tag.get('Key')] = tag.get('Value')
|
||||
return {'tags': tagdict}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
@ -81,7 +81,6 @@ from __future__ import absolute_import, print_function, unicode_literals
|
||||
import logging
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.versions
|
||||
|
||||
@ -131,7 +130,7 @@ def _find_identity_pool_ids(name, pool_id, conn):
|
||||
'''
|
||||
ids = []
|
||||
if pool_id is None:
|
||||
for pools in salt.utils.boto3.paged_call(conn.list_identity_pools,
|
||||
for pools in __utils__['boto3.paged_call'](conn.list_identity_pools,
|
||||
marker_flag='NextToken', marker_arg='NextToken', MaxResults=25):
|
||||
for pool in pools['IdentityPools']:
|
||||
if pool['IdentityPoolName'] == name:
|
||||
@ -174,7 +173,7 @@ def describe_identity_pools(IdentityPoolName, IdentityPoolId=None,
|
||||
else:
|
||||
return {'identity_pools': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_identity_pool(IdentityPoolName,
|
||||
@ -216,7 +215,7 @@ def create_identity_pool(IdentityPoolName,
|
||||
|
||||
return {'created': True, 'identity_pool': response}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_identity_pools(IdentityPoolName, IdentityPoolId=None,
|
||||
@ -250,7 +249,7 @@ def delete_identity_pools(IdentityPoolName, IdentityPoolId=None,
|
||||
else:
|
||||
return {'deleted': False, 'count': count}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def get_identity_pool_roles(IdentityPoolName, IdentityPoolId=None,
|
||||
@ -284,7 +283,7 @@ def get_identity_pool_roles(IdentityPoolName, IdentityPoolId=None,
|
||||
else:
|
||||
return {'identity_pool_roles': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _get_role_arn(name, **conn_params):
|
||||
@ -349,7 +348,7 @@ def set_identity_pool_roles(IdentityPoolId, AuthenticatedRole=None, Unauthentica
|
||||
|
||||
return {'set': True, 'roles': Roles}
|
||||
except ClientError as e:
|
||||
return {'set': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'set': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def update_identity_pool(IdentityPoolId,
|
||||
@ -420,4 +419,4 @@ def update_identity_pool(IdentityPoolId,
|
||||
|
||||
return {'updated': True, 'identity_pool': response}
|
||||
except ClientError as e:
|
||||
return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
@ -56,6 +56,7 @@ import salt.utils.data
|
||||
import salt.utils.json
|
||||
import salt.utils.versions
|
||||
from salt.ext import six
|
||||
from salt.ext.six.moves import map
|
||||
from salt.exceptions import SaltInvocationError, CommandExecutionError
|
||||
|
||||
# Import third party libs
|
||||
@ -65,6 +66,7 @@ try:
|
||||
import boto.ec2
|
||||
# pylint: enable=unused-import
|
||||
from boto.ec2.blockdevicemapping import BlockDeviceMapping, BlockDeviceType
|
||||
from boto.ec2.networkinterface import NetworkInterfaceSpecification, NetworkInterfaceCollection
|
||||
HAS_BOTO = True
|
||||
except ImportError:
|
||||
HAS_BOTO = False
|
||||
@ -1003,14 +1005,19 @@ def run(image_id, name=None, tags=None, key_name=None, security_groups=None,
|
||||
return False
|
||||
security_group_ids += [r]
|
||||
|
||||
if all((network_interface_id, network_interface_name)):
|
||||
raise SaltInvocationError('Only one of network_interface_id or '
|
||||
'network_interface_name may be provided.')
|
||||
network_interface_args = list(map(int, [network_interface_id is not None,
|
||||
network_interface_name is not None,
|
||||
network_interfaces is not None]))
|
||||
|
||||
if sum(network_interface_args) > 1:
|
||||
raise SaltInvocationError('Only one of network_interface_id, '
|
||||
'network_interface_name or '
|
||||
'network_interfaces may be provided.')
|
||||
if network_interface_name:
|
||||
result = get_network_interface_id(network_interface_name,
|
||||
region=region, key=key,
|
||||
keyid=keyid,
|
||||
profile=profile)
|
||||
region=region, key=key,
|
||||
keyid=keyid,
|
||||
profile=profile)
|
||||
network_interface_id = result['result']
|
||||
if not network_interface_id:
|
||||
log.warning(
|
||||
@ -1019,17 +1026,20 @@ def run(image_id, name=None, tags=None, key_name=None, security_groups=None,
|
||||
)
|
||||
|
||||
if network_interface_id:
|
||||
interface = boto.ec2.networkinterface.NetworkInterfaceSpecification(
|
||||
interface = NetworkInterfaceSpecification(
|
||||
network_interface_id=network_interface_id,
|
||||
device_index=0
|
||||
)
|
||||
device_index=0)
|
||||
else:
|
||||
interface = boto.ec2.networkinterface.NetworkInterfaceSpecification(
|
||||
interface = NetworkInterfaceSpecification(
|
||||
subnet_id=subnet_id,
|
||||
groups=security_group_ids,
|
||||
device_index=0
|
||||
)
|
||||
interfaces = boto.ec2.networkinterface.NetworkInterfaceCollection(interface)
|
||||
device_index=0)
|
||||
|
||||
if network_interfaces:
|
||||
interfaces_specs = [NetworkInterfaceSpecification(**x) for x in network_interfaces]
|
||||
interfaces = NetworkInterfaceCollection(*interfaces_specs)
|
||||
else:
|
||||
interfaces = NetworkInterfaceCollection(interface)
|
||||
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
|
||||
|
@ -80,7 +80,6 @@ import logging
|
||||
|
||||
# Import Salt libs
|
||||
from salt.ext import six
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.json
|
||||
import salt.utils.versions
|
||||
@ -148,7 +147,7 @@ def exists(DomainName,
|
||||
except ClientError as e:
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def status(DomainName,
|
||||
@ -179,7 +178,7 @@ def status(DomainName,
|
||||
else:
|
||||
return {'domain': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe(DomainName,
|
||||
@ -208,7 +207,7 @@ def describe(DomainName,
|
||||
else:
|
||||
return {'domain': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create(DomainName, ElasticsearchClusterConfig=None, EBSOptions=None,
|
||||
@ -262,7 +261,7 @@ def create(DomainName, ElasticsearchClusterConfig=None, EBSOptions=None,
|
||||
log.warning('Domain was not created')
|
||||
return {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete(DomainName, region=None, key=None, keyid=None, profile=None):
|
||||
@ -285,7 +284,7 @@ def delete(DomainName, region=None, key=None, keyid=None, profile=None):
|
||||
conn.delete_elasticsearch_domain(DomainName=DomainName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def update(DomainName, ElasticsearchClusterConfig=None, EBSOptions=None,
|
||||
@ -335,7 +334,7 @@ def update(DomainName, ElasticsearchClusterConfig=None, EBSOptions=None,
|
||||
return {'updated': False}
|
||||
return {'updated': True}
|
||||
except ClientError as e:
|
||||
return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def add_tags(DomainName=None, ARN=None,
|
||||
@ -378,7 +377,7 @@ def add_tags(DomainName=None, ARN=None,
|
||||
conn.add_tags(ARN=ARN, TagList=tagslist)
|
||||
return {'tagged': True}
|
||||
except ClientError as e:
|
||||
return {'tagged': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'tagged': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def remove_tags(TagKeys, DomainName=None, ARN=None,
|
||||
@ -417,7 +416,7 @@ def remove_tags(TagKeys, DomainName=None, ARN=None,
|
||||
TagKeys=TagKeys)
|
||||
return {'tagged': True}
|
||||
except ClientError as e:
|
||||
return {'tagged': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'tagged': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_tags(DomainName=None, ARN=None,
|
||||
@ -462,4 +461,4 @@ def list_tags(DomainName=None, ARN=None,
|
||||
tagdict[tag.get('Key')] = tag.get('Value')
|
||||
return {'tags': tagdict}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
@ -54,7 +54,6 @@ import salt.utils.versions
|
||||
# Import third-party libs
|
||||
try:
|
||||
# pylint: disable=unused-import
|
||||
import salt.utils.boto3
|
||||
import boto3
|
||||
import botocore
|
||||
# pylint: enable=unused-import
|
||||
|
@ -55,7 +55,6 @@ import logging
|
||||
import datetime
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.json
|
||||
import salt.utils.versions
|
||||
@ -125,7 +124,7 @@ def thing_type_exists(thingTypeName,
|
||||
else:
|
||||
return {'exists': False}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': err}
|
||||
@ -162,7 +161,7 @@ def describe_thing_type(thingTypeName,
|
||||
else:
|
||||
return {'thing_type': None}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'thing_type': None}
|
||||
return {'error': err}
|
||||
@ -207,7 +206,7 @@ def create_thing_type(thingTypeName, thingTypeDescription,
|
||||
log.warning('thing type was not created')
|
||||
return {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def deprecate_thing_type(thingTypeName, undoDeprecate=False,
|
||||
@ -238,7 +237,7 @@ def deprecate_thing_type(thingTypeName, undoDeprecate=False,
|
||||
deprecated = True if undoDeprecate is False else False
|
||||
return {'deprecated': deprecated}
|
||||
except ClientError as e:
|
||||
return {'deprecated': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deprecated': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_thing_type(thingTypeName,
|
||||
@ -264,7 +263,7 @@ def delete_thing_type(thingTypeName,
|
||||
conn.delete_thing_type(thingTypeName=thingTypeName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'deleted': True}
|
||||
return {'deleted': False, 'error': err}
|
||||
@ -291,7 +290,7 @@ def policy_exists(policyName,
|
||||
conn.get_policy(policyName=policyName)
|
||||
return {'exists': True}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': err}
|
||||
@ -331,7 +330,7 @@ def create_policy(policyName, policyDocument,
|
||||
log.warning('Policy was not created')
|
||||
return {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_policy(policyName,
|
||||
@ -355,7 +354,7 @@ def delete_policy(policyName,
|
||||
conn.delete_policy(policyName=policyName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_policy(policyName,
|
||||
@ -383,10 +382,10 @@ def describe_policy(policyName,
|
||||
else:
|
||||
return {'policy': None}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'policy': None}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def policy_version_exists(policyName, policyVersionId,
|
||||
@ -411,10 +410,10 @@ def policy_version_exists(policyName, policyVersionId,
|
||||
policyversionId=policyVersionId)
|
||||
return {'exists': bool(policy)}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'exists': False}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_policy_version(policyName, policyDocument, setAsDefault=False,
|
||||
@ -449,7 +448,7 @@ def create_policy_version(policyName, policyDocument, setAsDefault=False,
|
||||
log.warning('Policy version was not created')
|
||||
return {'created': False}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_policy_version(policyName, policyVersionId,
|
||||
@ -474,7 +473,7 @@ def delete_policy_version(policyName, policyVersionId,
|
||||
policyVersionId=policyVersionId)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_policy_version(policyName, policyVersionId,
|
||||
@ -503,10 +502,10 @@ def describe_policy_version(policyName, policyVersionId,
|
||||
else:
|
||||
return {'policy': None}
|
||||
except ClientError as e:
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'ResourceNotFoundException':
|
||||
return {'policy': None}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_policies(region=None, key=None, keyid=None, profile=None):
|
||||
@ -533,7 +532,7 @@ def list_policies(region=None, key=None, keyid=None, profile=None):
|
||||
try:
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
policies = []
|
||||
for ret in salt.utils.boto3.paged_call(conn.list_policies,
|
||||
for ret in __utils__['boto3.paged_call'](conn.list_policies,
|
||||
marker_flag='nextMarker',
|
||||
marker_arg='marker'):
|
||||
policies.extend(ret['policies'])
|
||||
@ -541,7 +540,7 @@ def list_policies(region=None, key=None, keyid=None, profile=None):
|
||||
log.warning('No policies found')
|
||||
return {'policies': policies}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_policy_versions(policyName,
|
||||
@ -567,7 +566,7 @@ def list_policy_versions(policyName,
|
||||
try:
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
vers = []
|
||||
for ret in salt.utils.boto3.paged_call(conn.list_policy_versions,
|
||||
for ret in __utils__['boto3.paged_call'](conn.list_policy_versions,
|
||||
marker_flag='nextMarker',
|
||||
marker_arg='marker',
|
||||
policyName=policyName):
|
||||
@ -576,7 +575,7 @@ def list_policy_versions(policyName,
|
||||
log.warning('No versions found')
|
||||
return {'policyVersions': vers}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def set_default_policy_version(policyName, policyVersionId,
|
||||
@ -603,7 +602,7 @@ def set_default_policy_version(policyName, policyVersionId,
|
||||
policyVersionId=str(policyVersionId)) # future lint: disable=blacklisted-function
|
||||
return {'changed': True}
|
||||
except ClientError as e:
|
||||
return {'changed': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'changed': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_principal_policies(principal,
|
||||
@ -629,7 +628,7 @@ def list_principal_policies(principal,
|
||||
try:
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
vers = []
|
||||
for ret in salt.utils.boto3.paged_call(conn.list_principal_policies,
|
||||
for ret in __utils__['boto3.paged_call'](conn.list_principal_policies,
|
||||
principal=principal,
|
||||
marker_flag='nextMarker',
|
||||
marker_arg='marker'):
|
||||
@ -638,7 +637,7 @@ def list_principal_policies(principal,
|
||||
log.warning('No policies found')
|
||||
return {'policies': vers}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def attach_principal_policy(policyName, principal,
|
||||
@ -664,7 +663,7 @@ def attach_principal_policy(policyName, principal,
|
||||
principal=principal)
|
||||
return {'attached': True}
|
||||
except ClientError as e:
|
||||
return {'attached': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'attached': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def detach_principal_policy(policyName, principal,
|
||||
@ -689,7 +688,7 @@ def detach_principal_policy(policyName, principal,
|
||||
principal=principal)
|
||||
return {'detached': True}
|
||||
except ClientError as e:
|
||||
return {'detached': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'detached': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def topic_rule_exists(ruleName,
|
||||
@ -718,10 +717,10 @@ def topic_rule_exists(ruleName,
|
||||
# use, it's more useful to assume lack of existence than to assume a
|
||||
# genuine authorization problem; authorization problems should not be
|
||||
# the common case.
|
||||
err = salt.utils.boto3.get_error(e)
|
||||
err = __utils__['boto3.get_error'](e)
|
||||
if e.response.get('Error', {}).get('Code') == 'UnauthorizedException':
|
||||
return {'exists': False}
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_topic_rule(ruleName, sql, actions, description,
|
||||
@ -754,7 +753,7 @@ def create_topic_rule(ruleName, sql, actions, description,
|
||||
})
|
||||
return {'created': True}
|
||||
except ClientError as e:
|
||||
return {'created': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def replace_topic_rule(ruleName, sql, actions, description,
|
||||
@ -787,7 +786,7 @@ def replace_topic_rule(ruleName, sql, actions, description,
|
||||
})
|
||||
return {'replaced': True}
|
||||
except ClientError as e:
|
||||
return {'replaced': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'replaced': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_topic_rule(ruleName,
|
||||
@ -811,7 +810,7 @@ def delete_topic_rule(ruleName,
|
||||
conn.delete_topic_rule(ruleName=ruleName)
|
||||
return {'deleted': True}
|
||||
except ClientError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_topic_rule(ruleName,
|
||||
@ -840,7 +839,7 @@ def describe_topic_rule(ruleName,
|
||||
else:
|
||||
return {'rule': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def list_topic_rules(topic=None, ruleDisabled=None,
|
||||
@ -873,7 +872,7 @@ def list_topic_rules(topic=None, ruleDisabled=None,
|
||||
if ruleDisabled is not None:
|
||||
kwargs['ruleDisabled'] = ruleDisabled
|
||||
rules = []
|
||||
for ret in salt.utils.boto3.paged_call(conn.list_topic_rules,
|
||||
for ret in __utils__['boto3.paged_call'](conn.list_topic_rules,
|
||||
marker_flag='nextToken',
|
||||
marker_arg='nextToken',
|
||||
**kwargs):
|
||||
@ -882,4 +881,4 @@ def list_topic_rules(topic=None, ruleDisabled=None,
|
||||
log.warning('No rules found')
|
||||
return {'rules': rules}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
@ -53,7 +53,6 @@ import logging
|
||||
import time
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.odict as odict
|
||||
import salt.utils.versions
|
||||
@ -153,7 +152,7 @@ def exists(name, tags=None, region=None, key=None, keyid=None, profile=None):
|
||||
rds = conn.describe_db_instances(DBInstanceIdentifier=name)
|
||||
return {'exists': bool(rds)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def option_group_exists(name, tags=None, region=None, key=None, keyid=None,
|
||||
@ -171,7 +170,7 @@ def option_group_exists(name, tags=None, region=None, key=None, keyid=None,
|
||||
rds = conn.describe_option_groups(OptionGroupName=name)
|
||||
return {'exists': bool(rds)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def parameter_group_exists(name, tags=None, region=None, key=None, keyid=None,
|
||||
@ -193,7 +192,7 @@ def parameter_group_exists(name, tags=None, region=None, key=None, keyid=None,
|
||||
resp = {}
|
||||
if e.response['Error']['Code'] == 'DBParameterGroupNotFound':
|
||||
resp['exists'] = False
|
||||
resp['error'] = salt.utils.boto3.get_error(e)
|
||||
resp['error'] = __utils__['boto3.get_error'](e)
|
||||
return resp
|
||||
|
||||
|
||||
@ -218,7 +217,7 @@ def subnet_group_exists(name, tags=None, region=None, key=None, keyid=None,
|
||||
if "DBSubnetGroupNotFoundFault" in e.message:
|
||||
return {'exists': False}
|
||||
else:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create(name, allocated_storage, db_instance_class, engine,
|
||||
@ -317,7 +316,7 @@ def create(name, allocated_storage, db_instance_class, engine,
|
||||
log.info('Instance status after 10 seconds is: %s', stat)
|
||||
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_read_replica(name, source_name, db_instance_class=None,
|
||||
@ -375,7 +374,7 @@ def create_read_replica(name, source_name, db_instance_class=None,
|
||||
|
||||
return {'exists': bool(rds_replica)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_option_group(name, engine_name, major_engine_version,
|
||||
@ -408,7 +407,7 @@ def create_option_group(name, engine_name, major_engine_version,
|
||||
|
||||
return {'exists': bool(rds)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_parameter_group(name, db_parameter_group_family, description,
|
||||
@ -444,7 +443,7 @@ def create_parameter_group(name, db_parameter_group_family, description,
|
||||
return {'exists': bool(rds), 'message':
|
||||
'Created RDS parameter group {0}'.format(name)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def create_subnet_group(name, description, subnet_ids, tags=None,
|
||||
@ -475,7 +474,7 @@ def create_subnet_group(name, description, subnet_ids, tags=None,
|
||||
|
||||
return {'created': bool(rds)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def update_parameter_group(name, parameters, apply_method="pending-reboot",
|
||||
@ -520,7 +519,7 @@ def update_parameter_group(name, parameters, apply_method="pending-reboot",
|
||||
Parameters=param_list)
|
||||
return {'results': bool(res)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe(name, tags=None, region=None, key=None, keyid=None,
|
||||
@ -568,7 +567,7 @@ def describe(name, tags=None, region=None, key=None, keyid=None,
|
||||
else:
|
||||
return {'rds': None}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
except IndexError:
|
||||
return {'rds': None}
|
||||
|
||||
@ -597,7 +596,7 @@ def describe_db_instances(name=None, filters=None, jmespath='DBInstances',
|
||||
except ClientError as e:
|
||||
code = getattr(e, 'response', {}).get('Error', {}).get('Code')
|
||||
if code != 'DBInstanceNotFound':
|
||||
log.error(salt.utils.boto3.get_error(e))
|
||||
log.error(__utils__['boto3.get_error'](e))
|
||||
return []
|
||||
|
||||
|
||||
@ -647,7 +646,7 @@ def get_endpoint(name, tags=None, region=None, key=None, keyid=None,
|
||||
return endpoint
|
||||
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
return endpoint
|
||||
|
||||
@ -706,7 +705,7 @@ def delete(name, skip_final_snapshot=None, final_db_snapshot_identifier=None,
|
||||
'deleted.', timeout, name)
|
||||
time.sleep(10)
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_option_group(name, region=None, key=None, keyid=None, profile=None):
|
||||
@ -731,7 +730,7 @@ def delete_option_group(name, region=None, key=None, keyid=None, profile=None):
|
||||
return {'deleted': bool(res), 'message':
|
||||
'Deleted RDS option group {0}.'.format(name)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_parameter_group(name, region=None, key=None, keyid=None,
|
||||
@ -753,7 +752,7 @@ def delete_parameter_group(name, region=None, key=None, keyid=None,
|
||||
return {'deleted': bool(r), 'message':
|
||||
'Deleted RDS parameter group {0}.'.format(name)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def delete_subnet_group(name, region=None, key=None, keyid=None,
|
||||
@ -775,7 +774,7 @@ def delete_subnet_group(name, region=None, key=None, keyid=None,
|
||||
return {'deleted': bool(r), 'message':
|
||||
'Deleted RDS subnet group {0}.'.format(name)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_parameter_group(name, Filters=None, MaxRecords=None, Marker=None,
|
||||
@ -817,7 +816,7 @@ def describe_parameter_group(name, Filters=None, MaxRecords=None, Marker=None,
|
||||
return {'results': bool(info), 'message':
|
||||
'Got RDS descrition for group {0}.'.format(name)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def describe_parameters(name, Source=None, MaxRecords=None, Marker=None,
|
||||
@ -873,7 +872,7 @@ def describe_parameters(name, Source=None, MaxRecords=None, Marker=None,
|
||||
ret['parameters'] = parameters
|
||||
return ret
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def modify_db_instance(name,
|
||||
@ -950,7 +949,7 @@ def modify_db_instance(name,
|
||||
'Modified RDS db instance {0}.'.format(name),
|
||||
'results': dict(info)}
|
||||
except ClientError as e:
|
||||
return {'error': salt.utils.boto3.get_error(e)}
|
||||
return {'error': __utils__['boto3.get_error'](e)}
|
||||
|
||||
|
||||
def _tag_doc(tags):
|
||||
|
@ -473,6 +473,9 @@ def authorize(name=None, source_group_name=None,
|
||||
log.error(msg)
|
||||
return False
|
||||
except boto.exception.EC2ResponseError as e:
|
||||
# if we are trying to add the same rule then we are already in the desired state, return true
|
||||
if e.error_code == 'InvalidPermission.Duplicate':
|
||||
return True
|
||||
msg = ('Failed to add rule to security group {0} with id {1}.'
|
||||
.format(group.name, group.id))
|
||||
log.error(msg)
|
||||
|
@ -134,8 +134,6 @@ import time
|
||||
import random
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.boto
|
||||
import salt.utils.boto3
|
||||
import salt.utils.compat
|
||||
import salt.utils.versions
|
||||
from salt.exceptions import SaltInvocationError, CommandExecutionError
|
||||
@ -279,7 +277,7 @@ def _create_resource(resource, name=None, tags=None, region=None, key=None,
|
||||
log.warning(e)
|
||||
return {'created': False, 'error': {'message': e}}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _delete_resource(resource, name=None, resource_id=None, region=None,
|
||||
@ -323,7 +321,7 @@ def _delete_resource(resource, name=None, resource_id=None, region=None,
|
||||
e = '{0} was not deleted.'.format(resource)
|
||||
return {'deleted': False, 'error': {'message': e}}
|
||||
except BotoServerError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _get_resource(resource, name=None, resource_id=None, region=None,
|
||||
@ -451,7 +449,7 @@ def get_resource_id(resource, name=None, resource_id=None, region=None,
|
||||
return {'id': _get_resource_id(resource, name, region=region, key=key,
|
||||
keyid=keyid, profile=profile)}
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def resource_exists(resource, name=None, resource_id=None, tags=None,
|
||||
@ -478,7 +476,7 @@ def resource_exists(resource, name=None, resource_id=None, tags=None,
|
||||
key=key, keyid=keyid,
|
||||
profile=profile))}
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _find_vpcs(vpc_id=None, vpc_name=None, cidr=None, tags=None,
|
||||
@ -570,7 +568,7 @@ def get_id(name=None, cidr=None, tags=None, region=None, key=None, keyid=None,
|
||||
return {'id': _get_id(vpc_name=name, cidr=cidr, tags=tags, region=region,
|
||||
key=key, keyid=keyid, profile=profile)}
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def exists(vpc_id=None, name=None, cidr=None, tags=None, region=None, key=None,
|
||||
@ -593,7 +591,7 @@ def exists(vpc_id=None, name=None, cidr=None, tags=None, region=None, key=None,
|
||||
vpc_ids = _find_vpcs(vpc_id=vpc_id, vpc_name=name, cidr=cidr, tags=tags,
|
||||
region=region, key=key, keyid=keyid, profile=profile)
|
||||
except BotoServerError as err:
|
||||
boto_err = salt.utils.boto.get_error(err)
|
||||
boto_err = __utils__['boto.get_error'](err)
|
||||
if boto_err.get('aws', {}).get('code') == 'InvalidVpcID.NotFound':
|
||||
# VPC was not found: handle the error and return False.
|
||||
return {'exists': False}
|
||||
@ -643,7 +641,7 @@ def create(cidr_block, instance_tenancy=None, vpc_name=None,
|
||||
log.warning('VPC was not created')
|
||||
return {'created': False}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def delete(vpc_id=None, name=None, vpc_name=None, tags=None,
|
||||
@ -693,7 +691,7 @@ def delete(vpc_id=None, name=None, vpc_name=None, tags=None,
|
||||
log.warning('VPC %s was not deleted.', vpc_id)
|
||||
return {'deleted': False}
|
||||
except BotoServerError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def describe(vpc_id=None, vpc_name=None, region=None, key=None,
|
||||
@ -722,7 +720,7 @@ def describe(vpc_id=None, vpc_name=None, region=None, key=None,
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
vpc_id = check_vpc(vpc_id, vpc_name, region, key, keyid, profile)
|
||||
except BotoServerError as err:
|
||||
boto_err = salt.utils.boto.get_error(err)
|
||||
boto_err = __utils__['boto.get_error'](err)
|
||||
if boto_err.get('aws', {}).get('code') == 'InvalidVpcID.NotFound':
|
||||
# VPC was not found: handle the error and return None.
|
||||
return {'vpc': None}
|
||||
@ -736,7 +734,7 @@ def describe(vpc_id=None, vpc_name=None, region=None, key=None,
|
||||
try:
|
||||
vpcs = conn.get_all_vpcs(**filter_parameters)
|
||||
except BotoServerError as err:
|
||||
return {'error': salt.utils.boto.get_error(err)}
|
||||
return {'error': __utils__['boto.get_error'](err)}
|
||||
|
||||
if vpcs:
|
||||
vpc = vpcs[0] # Found!
|
||||
@ -806,7 +804,7 @@ def describe_vpcs(vpc_id=None, name=None, cidr=None, tags=None,
|
||||
return {'vpcs': []}
|
||||
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _find_subnets(subnet_name=None, vpc_id=None, cidr=None, tags=None, conn=None):
|
||||
@ -871,7 +869,7 @@ def create_subnet(vpc_id=None, cidr_block=None, vpc_name=None,
|
||||
if not vpc_id:
|
||||
return {'created': False, 'error': {'message': 'VPC {0} does not exist.'.format(vpc_name or vpc_id)}}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
return _create_resource('subnet', name=subnet_name, tags=tags, vpc_id=vpc_id,
|
||||
availability_zone=availability_zone,
|
||||
@ -934,7 +932,7 @@ def subnet_exists(subnet_id=None, name=None, subnet_name=None, cidr=None,
|
||||
try:
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
except BotoServerError as err:
|
||||
return {'error': salt.utils.boto.get_error(err)}
|
||||
return {'error': __utils__['boto.get_error'](err)}
|
||||
|
||||
filter_parameters = {'filters': {}}
|
||||
if subnet_id:
|
||||
@ -952,7 +950,7 @@ def subnet_exists(subnet_id=None, name=None, subnet_name=None, cidr=None,
|
||||
try:
|
||||
subnets = conn.get_all_subnets(**filter_parameters)
|
||||
except BotoServerError as err:
|
||||
boto_err = salt.utils.boto.get_error(err)
|
||||
boto_err = __utils__['boto.get_error'](err)
|
||||
if boto_err.get('aws', {}).get('code') == 'InvalidSubnetID.NotFound':
|
||||
# Subnet was not found: handle the error and return False.
|
||||
return {'exists': False}
|
||||
@ -995,7 +993,7 @@ def get_subnet_association(subnets, region=None, key=None, keyid=None,
|
||||
# subnet_ids=subnets can accept either a string or a list
|
||||
subnets = conn.get_all_subnets(subnet_ids=subnets)
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
# using a set to store vpc_ids - the use of set prevents duplicate
|
||||
# vpc_id values
|
||||
@ -1035,7 +1033,7 @@ def describe_subnet(subnet_id=None, subnet_name=None, region=None,
|
||||
subnet = _get_resource('subnet', name=subnet_name, resource_id=subnet_id,
|
||||
region=region, key=key, keyid=keyid, profile=profile)
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
if not subnet:
|
||||
return {'subnet': None}
|
||||
@ -1116,7 +1114,7 @@ def describe_subnets(subnet_ids=None, subnet_names=None, vpc_id=None, cidr=None,
|
||||
return {'subnets': subnets_list}
|
||||
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def create_internet_gateway(internet_gateway_name=None, vpc_id=None,
|
||||
@ -1158,7 +1156,7 @@ def create_internet_gateway(internet_gateway_name=None, vpc_id=None,
|
||||
)
|
||||
return r
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def delete_internet_gateway(internet_gateway_id=None,
|
||||
@ -1212,7 +1210,7 @@ def delete_internet_gateway(internet_gateway_id=None,
|
||||
region=region, key=key, keyid=keyid,
|
||||
profile=profile)
|
||||
except BotoServerError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _find_nat_gateways(nat_gateway_id=None, subnet_id=None, subnet_name=None, vpc_id=None, vpc_name=None,
|
||||
@ -1253,7 +1251,7 @@ def _find_nat_gateways(nat_gateway_id=None, subnet_id=None, subnet_name=None, vp
|
||||
|
||||
conn3 = _get_conn3(region=region, key=key, keyid=keyid, profile=profile)
|
||||
nat_gateways = []
|
||||
for ret in salt.utils.boto3.paged_call(conn3.describe_nat_gateways,
|
||||
for ret in __utils__['boto3.paged_call'](conn3.describe_nat_gateways,
|
||||
marker_flag='NextToken', marker_arg='NextToken',
|
||||
**filter_parameters):
|
||||
for gw in ret.get('NatGateways', []):
|
||||
@ -1376,7 +1374,7 @@ def create_nat_gateway(subnet_id=None,
|
||||
r = conn3.create_nat_gateway(SubnetId=subnet_id, AllocationId=allocation_id)
|
||||
return {'created': True, 'id': r.get('NatGateway', {}).get('NatGatewayId')}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def delete_nat_gateway(nat_gateway_id,
|
||||
@ -1452,7 +1450,7 @@ def delete_nat_gateway(nat_gateway_id,
|
||||
conn3.release_address(AllocationId=addr.get('AllocationId'))
|
||||
return {'deleted': True}
|
||||
except BotoServerError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def create_customer_gateway(vpn_connection_type, ip_address, bgp_asn,
|
||||
@ -1573,7 +1571,7 @@ def create_dhcp_options(domain_name=None, domain_name_servers=None, ntp_servers=
|
||||
)
|
||||
return r
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def get_dhcp_options(dhcp_options_name=None, dhcp_options_id=None,
|
||||
@ -1604,7 +1602,7 @@ def get_dhcp_options(dhcp_options_name=None, dhcp_options_id=None,
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
r = conn.get_all_dhcp_options(dhcp_options_ids=[dhcp_options_id])
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
if not r:
|
||||
return {'dhcp_options': None}
|
||||
@ -1667,7 +1665,7 @@ def associate_dhcp_options_to_vpc(dhcp_options_id, vpc_id=None, vpc_name=None,
|
||||
dhcp_options_id, vpc_id)
|
||||
return {'associated': False, 'error': {'message': 'DHCP options could not be associated.'}}
|
||||
except BotoServerError as e:
|
||||
return {'associated': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'associated': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def dhcp_options_exists(dhcp_options_id=None, name=None, dhcp_options_name=None,
|
||||
@ -1720,7 +1718,7 @@ def create_network_acl(vpc_id=None, vpc_name=None, network_acl_name=None,
|
||||
try:
|
||||
vpc_id = check_vpc(vpc_id, vpc_name, region, key, keyid, profile)
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
if not vpc_id:
|
||||
return {'created': False,
|
||||
@ -1751,7 +1749,7 @@ def create_network_acl(vpc_id=None, vpc_name=None, network_acl_name=None,
|
||||
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
|
||||
association_id = conn.associate_network_acl(r['id'], subnet_id)
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
r['association_id'] = association_id
|
||||
return r
|
||||
|
||||
@ -1866,7 +1864,7 @@ def associate_network_acl_to_subnet(network_acl_id=None, subnet_id=None,
|
||||
network_acl_id, subnet_id)
|
||||
return {'associated': False, 'error': {'message': 'ACL could not be assocaited.'}}
|
||||
except BotoServerError as e:
|
||||
return {'associated': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'associated': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def disassociate_network_acl(subnet_id=None, vpc_id=None, subnet_name=None, vpc_name=None,
|
||||
@ -1905,7 +1903,7 @@ def disassociate_network_acl(subnet_id=None, vpc_id=None, subnet_name=None, vpc_
|
||||
association_id = conn.disassociate_network_acl(subnet_id, vpc_id=vpc_id)
|
||||
return {'disassociated': True, 'association_id': association_id}
|
||||
except BotoServerError as e:
|
||||
return {'disassociated': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'disassociated': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _create_network_acl_entry(network_acl_id=None, rule_number=None, protocol=None,
|
||||
@ -1958,7 +1956,7 @@ def _create_network_acl_entry(network_acl_id=None, rule_number=None, protocol=No
|
||||
log.warning('Network ACL entry was not %s', rkey)
|
||||
return {rkey: created}
|
||||
except BotoServerError as e:
|
||||
return {rkey: False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {rkey: False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def create_network_acl_entry(network_acl_id=None, rule_number=None, protocol=None,
|
||||
@ -2041,7 +2039,7 @@ def delete_network_acl_entry(network_acl_id=None, rule_number=None, egress=None,
|
||||
log.warning('Network ACL was not deleted')
|
||||
return {'deleted': deleted}
|
||||
except BotoServerError as e:
|
||||
return {'deleted': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'deleted': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def create_route_table(vpc_id=None, vpc_name=None, route_table_name=None,
|
||||
@ -2177,7 +2175,7 @@ def route_exists(destination_cidr_block, route_table_name=None, route_table_id=N
|
||||
log.warning('Route %s does not exist.', destination_cidr_block)
|
||||
return {'exists': False}
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def associate_route_table(route_table_id=None, subnet_id=None,
|
||||
@ -2229,7 +2227,7 @@ def associate_route_table(route_table_id=None, subnet_id=None,
|
||||
route_table_id, subnet_id)
|
||||
return {'association_id': association_id}
|
||||
except BotoServerError as e:
|
||||
return {'associated': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'associated': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def disassociate_route_table(association_id, region=None, key=None, keyid=None, profile=None):
|
||||
@ -2256,7 +2254,7 @@ def disassociate_route_table(association_id, region=None, key=None, keyid=None,
|
||||
log.warning('Route table with association id %s has not been disassociated.', association_id)
|
||||
return {'disassociated': False}
|
||||
except BotoServerError as e:
|
||||
return {'disassociated': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'disassociated': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def replace_route_table_association(association_id, route_table_id, region=None, key=None, keyid=None, profile=None):
|
||||
@ -2278,7 +2276,7 @@ def replace_route_table_association(association_id, route_table_id, region=None,
|
||||
route_table_id, association_id)
|
||||
return {'replaced': True, 'association_id': association_id}
|
||||
except BotoServerError as e:
|
||||
return {'replaced': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'replaced': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def create_route(route_table_id=None, destination_cidr_block=None,
|
||||
@ -2359,7 +2357,7 @@ def create_route(route_table_id=None, destination_cidr_block=None,
|
||||
nat_gateway_id = gws[0]['NatGatewayId']
|
||||
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
if not nat_gateway_id:
|
||||
return _create_resource('route', route_table_id=route_table_id,
|
||||
@ -2375,7 +2373,7 @@ def create_route(route_table_id=None, destination_cidr_block=None,
|
||||
NatGatewayId=nat_gateway_id)
|
||||
return {'created': True, 'id': ret.get('NatGatewayId')}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def delete_route(route_table_id=None, destination_cidr_block=None,
|
||||
@ -2408,7 +2406,7 @@ def delete_route(route_table_id=None, destination_cidr_block=None,
|
||||
return {'created': False,
|
||||
'error': {'message': 'route table {0} does not exist.'.format(route_table_name)}}
|
||||
except BotoServerError as e:
|
||||
return {'created': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'created': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
return _delete_resource(resource='route', resource_id=route_table_id,
|
||||
destination_cidr_block=destination_cidr_block,
|
||||
@ -2464,7 +2462,7 @@ def replace_route(route_table_id=None, destination_cidr_block=None,
|
||||
)
|
||||
return {'replaced': False}
|
||||
except BotoServerError as e:
|
||||
return {'replaced': False, 'error': salt.utils.boto.get_error(e)}
|
||||
return {'replaced': False, 'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def describe_route_table(route_table_id=None, route_table_name=None,
|
||||
@ -2525,7 +2523,7 @@ def describe_route_table(route_table_id=None, route_table_name=None,
|
||||
return route_table
|
||||
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def describe_route_tables(route_table_id=None, route_table_name=None,
|
||||
@ -2611,7 +2609,7 @@ def describe_route_tables(route_table_id=None, route_table_name=None,
|
||||
return tables
|
||||
|
||||
except BotoServerError as e:
|
||||
return {'error': salt.utils.boto.get_error(e)}
|
||||
return {'error': __utils__['boto.get_error'](e)}
|
||||
|
||||
|
||||
def _create_dhcp_options(conn, domain_name=None, domain_name_servers=None, ntp_servers=None, netbios_name_servers=None,
|
||||
@ -2821,7 +2819,7 @@ def request_vpc_peering_connection(requester_vpc_id=None, requester_vpc_name=Non
|
||||
return {'msg': msg}
|
||||
except botocore.exceptions.ClientError as err:
|
||||
log.error('Got an error while trying to request vpc peering')
|
||||
return {'error': salt.utils.boto.get_error(err)}
|
||||
return {'error': __utils__['boto.get_error'](err)}
|
||||
|
||||
|
||||
def _get_peering_connection_ids(name, conn):
|
||||
@ -2948,7 +2946,7 @@ def accept_vpc_peering_connection( # pylint: disable=too-many-arguments
|
||||
return {'msg': 'VPC peering connection accepted.'}
|
||||
except botocore.exceptions.ClientError as err:
|
||||
log.error('Got an error while trying to accept vpc peering')
|
||||
return {'error': salt.utils.boto.get_error(err)}
|
||||
return {'error': __utils__['boto.get_error'](err)}
|
||||
|
||||
|
||||
def _vpc_peering_conn_id_for_name(name, conn):
|
||||
@ -3026,7 +3024,7 @@ def delete_vpc_peering_connection(conn_id=None, conn_name=None, region=None,
|
||||
conn.delete_vpc_peering_connection(DryRun=dry_run, VpcPeeringConnectionId=conn_id)
|
||||
return {'msg': 'VPC peering connection deleted.'}
|
||||
except botocore.exceptions.ClientError as err:
|
||||
e = salt.utils.boto.get_error(err)
|
||||
e = __utils__['boto.get_error'](err)
|
||||
log.error('Failed to delete VPC peering %s: %s', conn_name or conn_id, e)
|
||||
return {'error': e}
|
||||
|
||||
|
@ -43,7 +43,7 @@ from salt.ext import six
|
||||
from salt.exceptions import CommandExecutionError, TimedProcTimeoutError, \
|
||||
SaltInvocationError
|
||||
from salt.log import LOG_LEVELS
|
||||
from salt.ext.six.moves import range, zip
|
||||
from salt.ext.six.moves import range, zip, map
|
||||
|
||||
# Only available on POSIX systems, nonfatal on windows
|
||||
try:
|
||||
@ -405,6 +405,19 @@ def _run(cmd,
|
||||
|
||||
return win_runas(cmd, runas, password, cwd)
|
||||
|
||||
if runas and salt.utils.platform.is_darwin():
|
||||
# we need to insert the user simulation into the command itself and not
|
||||
# just run it from the environment on macOS as that
|
||||
# method doesn't work properly when run as root for certain commands.
|
||||
if isinstance(cmd, (list, tuple)):
|
||||
cmd = ' '.join(map(_cmd_quote, cmd))
|
||||
|
||||
cmd = 'su -l {0} -c "{1}"'.format(runas, cmd)
|
||||
# set runas to None, because if you try to run `su -l` as well as
|
||||
# simulate the environment macOS will prompt for the password of the
|
||||
# user and will cause salt to hang.
|
||||
runas = None
|
||||
|
||||
if runas:
|
||||
# Save the original command before munging it
|
||||
try:
|
||||
@ -478,10 +491,18 @@ def _run(cmd,
|
||||
for k, v in six.iteritems(env_runas)
|
||||
)
|
||||
env_runas.update(env)
|
||||
|
||||
# Fix platforms like Solaris that don't set a USER env var in the
|
||||
# user's default environment as obtained above.
|
||||
if env_runas.get('USER') != runas:
|
||||
env_runas['USER'] = runas
|
||||
|
||||
# Fix some corner cases where shelling out to get the user's
|
||||
# environment returns the wrong home directory.
|
||||
runas_home = os.path.expanduser('~{0}'.format(runas))
|
||||
if env_runas.get('HOME') != runas_home:
|
||||
env_runas['HOME'] = runas_home
|
||||
|
||||
env = env_runas
|
||||
except ValueError as exc:
|
||||
log.exception('Error raised retrieving environment for user %s', runas)
|
||||
|
@ -3210,6 +3210,7 @@ def run_container(image,
|
||||
CLI Examples:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt myminion docker.run_container myuser/myimage command=/usr/local/bin/myscript.sh
|
||||
# Run container in the background
|
||||
salt myminion docker.run_container myuser/myimage command=/usr/local/bin/myscript.sh bg=True
|
||||
|
@ -14,8 +14,8 @@ from __future__ import absolute_import, print_function, unicode_literals
|
||||
import datetime
|
||||
import difflib
|
||||
import errno
|
||||
import fileinput
|
||||
import fnmatch
|
||||
import io
|
||||
import itertools
|
||||
import logging
|
||||
import operator
|
||||
@ -2535,10 +2535,16 @@ def blockreplace(path,
|
||||
if not os.path.exists(path):
|
||||
raise SaltInvocationError('File not found: {0}'.format(path))
|
||||
|
||||
if not __utils__['files.is_text'](path):
|
||||
raise SaltInvocationError(
|
||||
'Cannot perform string replacements on a binary file: {0}'
|
||||
.format(path)
|
||||
try:
|
||||
file_encoding = __utils__['files.get_encoding'](path)
|
||||
except CommandExecutionError:
|
||||
file_encoding = None
|
||||
|
||||
if __utils__['files.is_binary'](path):
|
||||
if not file_encoding:
|
||||
raise SaltInvocationError(
|
||||
'Cannot perform string replacements on a binary file: {0}'
|
||||
.format(path)
|
||||
)
|
||||
|
||||
if append_newline is None and not content.endswith((os.linesep, '\n')):
|
||||
@ -2593,18 +2599,9 @@ def blockreplace(path,
|
||||
# We do not use in-place editing to avoid file attrs modifications when
|
||||
# no changes are required and to avoid any file access on a partially
|
||||
# written file.
|
||||
#
|
||||
# We could also use salt.utils.filebuffer.BufferedReader
|
||||
try:
|
||||
fi_file = fileinput.input(
|
||||
path,
|
||||
inplace=False,
|
||||
backup=False,
|
||||
bufsize=1,
|
||||
mode='rb')
|
||||
|
||||
fi_file = io.open(path, mode='r', encoding=file_encoding, newline='')
|
||||
for line in fi_file:
|
||||
line = salt.utils.stringutils.to_unicode(line)
|
||||
write_line_to_new_file = True
|
||||
|
||||
if linesep is None:
|
||||
@ -2709,7 +2706,7 @@ def blockreplace(path,
|
||||
try:
|
||||
fh_ = salt.utils.atomicfile.atomic_open(path, 'wb')
|
||||
for line in new_file:
|
||||
fh_.write(salt.utils.stringutils.to_bytes(line))
|
||||
fh_.write(salt.utils.stringutils.to_bytes(line, encoding=file_encoding))
|
||||
finally:
|
||||
fh_.close()
|
||||
|
||||
|
@ -9,10 +9,12 @@ Glance module for interacting with OpenStack Glance
|
||||
Example configuration
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
glance:
|
||||
cloud: default
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
glance:
|
||||
auth:
|
||||
username: admin
|
||||
|
@ -9,10 +9,12 @@ Keystone module for interacting with OpenStack Keystone
|
||||
Example configuration
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
keystone:
|
||||
cloud: default
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
keystone:
|
||||
auth:
|
||||
username: admin
|
||||
|
@ -35,7 +35,7 @@ In case both are provided the `file` entry is preferred.
|
||||
|
||||
.. warning::
|
||||
|
||||
Configuration options will change in Flourine. All options above will be replaced by:
|
||||
Configuration options will change in Fluorine. All options above will be replaced by:
|
||||
|
||||
- kubernetes.kubeconfig or kubernetes.kubeconfig-data
|
||||
- kubernetes.context
|
||||
|
@ -867,7 +867,7 @@ def _network_conf(conf_tuples=None, **kwargs):
|
||||
# on old versions of lxc, still support the gateway auto mode
|
||||
# if we didn't explicitly say no to
|
||||
# (lxc.network.ipv4.gateway: auto)
|
||||
if _LooseVersion(version()) <= '1.0.7' and \
|
||||
if _LooseVersion(version()) <= _LooseVersion('1.0.7') and \
|
||||
True not in ['lxc.network.ipv4.gateway' in a for a in ret] and \
|
||||
True in ['lxc.network.ipv4' in a for a in ret]:
|
||||
ret.append({'lxc.network.ipv4.gateway': 'auto'})
|
||||
|
@ -14,7 +14,6 @@ import salt.utils.files
|
||||
import salt.utils.path
|
||||
import salt.utils.platform
|
||||
import salt.utils.stringutils
|
||||
import salt.utils.mac_utils
|
||||
from salt.exceptions import CommandExecutionError
|
||||
from salt.utils.versions import LooseVersion as _LooseVersion
|
||||
|
||||
@ -62,7 +61,7 @@ def _get_service(name):
|
||||
:return: The service information for the service, otherwise an Error
|
||||
:rtype: dict
|
||||
'''
|
||||
services = salt.utils.mac_utils.available_services()
|
||||
services = __utils__['mac_utils.available_services']()
|
||||
name = name.lower()
|
||||
|
||||
if name in services:
|
||||
@ -127,7 +126,7 @@ def launchctl(sub_cmd, *args, **kwargs):
|
||||
|
||||
salt '*' service.launchctl debug org.cups.cupsd
|
||||
'''
|
||||
return salt.utils.mac_utils.launchctl(sub_cmd, *args, **kwargs)
|
||||
return __utils__['mac_utils.launchctl'](sub_cmd, *args, **kwargs)
|
||||
|
||||
|
||||
def list_(name=None, runas=None):
|
||||
@ -158,13 +157,11 @@ def list_(name=None, runas=None):
|
||||
return launchctl('list',
|
||||
label,
|
||||
return_stdout=True,
|
||||
output_loglevel='trace',
|
||||
runas=runas)
|
||||
|
||||
# Collect information on all services: will raise an error if it fails
|
||||
return launchctl('list',
|
||||
return_stdout=True,
|
||||
output_loglevel='trace',
|
||||
runas=runas)
|
||||
|
||||
|
||||
@ -412,7 +409,7 @@ def enabled(name, runas=None):
|
||||
return False
|
||||
|
||||
|
||||
def disabled(name, runas=None):
|
||||
def disabled(name, runas=None, domain='system'):
|
||||
'''
|
||||
Check if the specified service is not enabled. This is the opposite of
|
||||
``service.enabled``
|
||||
@ -421,6 +418,8 @@ def disabled(name, runas=None):
|
||||
|
||||
:param str runas: User to run launchctl commands
|
||||
|
||||
:param str domain: domain to check for disabled services. Default is system.
|
||||
|
||||
:return: True if the specified service is NOT enabled, otherwise False
|
||||
:rtype: bool
|
||||
|
||||
@ -430,8 +429,22 @@ def disabled(name, runas=None):
|
||||
|
||||
salt '*' service.disabled org.cups.cupsd
|
||||
'''
|
||||
# A service is disabled if it is not enabled
|
||||
return not enabled(name, runas=runas)
|
||||
ret = False
|
||||
disabled = launchctl('print-disabled',
|
||||
domain,
|
||||
return_stdout=True,
|
||||
output_loglevel='trace',
|
||||
runas=runas)
|
||||
for service in disabled.split("\n"):
|
||||
if name in service:
|
||||
srv_name = service.split("=>")[0].split("\"")[1]
|
||||
status = service.split("=>")[1]
|
||||
if name != srv_name:
|
||||
pass
|
||||
else:
|
||||
return True if 'true' in status.lower() else False
|
||||
|
||||
return False
|
||||
|
||||
|
||||
def get_all(runas=None):
|
||||
@ -454,7 +467,7 @@ def get_all(runas=None):
|
||||
enabled = get_enabled(runas=runas)
|
||||
|
||||
# Get list of all services
|
||||
available = list(salt.utils.mac_utils.available_services().keys())
|
||||
available = list(__utils__['mac_utils.available_services']().keys())
|
||||
|
||||
# Return composite list
|
||||
return sorted(set(enabled + available))
|
||||
|
@ -74,8 +74,7 @@ def _atrun_enabled():
|
||||
# Collect information on service: will raise an error if it fails
|
||||
salt.utils.mac_utils.launchctl('list',
|
||||
label,
|
||||
return_stdout=True,
|
||||
output_loglevel='quiet')
|
||||
return_stdout=True)
|
||||
return True
|
||||
except CommandExecutionError:
|
||||
return False
|
||||
@ -111,9 +110,8 @@ def _enable_atrun():
|
||||
return False
|
||||
|
||||
salt.utils.mac_utils.launchctl('enable',
|
||||
'system/{0}'.format(label),
|
||||
output_loglevel='quiet')
|
||||
salt.utils.mac_utils.launchctl('load', path, output_loglevel='quiet')
|
||||
'system/{0}'.format(label))
|
||||
salt.utils.mac_utils.launchctl('load', path)
|
||||
return _atrun_enabled()
|
||||
|
||||
|
||||
|
@ -9,10 +9,12 @@ Neutron module for interacting with OpenStack Neutron
|
||||
Example configuration
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
neutron:
|
||||
cloud: default
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
neutron:
|
||||
auth:
|
||||
username: admin
|
||||
|
@ -60,9 +60,9 @@ def __virtual__():
|
||||
'''
|
||||
Confirm this module is on a nilrt based system
|
||||
'''
|
||||
if __grains__.get('os_family', False) == 'NILinuxRT':
|
||||
if os.path.isdir(OPKG_CONFDIR):
|
||||
return __virtualname__
|
||||
return (False, "Module opkg only works on nilrt based systems")
|
||||
return False, "Module opkg only works on OpenEmbedded based systems"
|
||||
|
||||
|
||||
def latest_version(*names, **kwargs):
|
||||
|
@ -460,13 +460,10 @@ def install(pkgs=None, # pylint: disable=R0912,R0913,R0914
|
||||
Path to requirements
|
||||
|
||||
bin_env
|
||||
Path to pip bin or path to virtualenv. If doing a system install,
|
||||
and want to use a specific pip bin (pip-2.7, pip-2.6, etc..) just
|
||||
specify the pip bin you want.
|
||||
|
||||
.. note::
|
||||
If installing into a virtualenv, just use the path to the
|
||||
virtualenv (e.g. ``/home/code/path/to/virtualenv/``)
|
||||
Path to pip (or to a virtualenv). This can be used to specify the path
|
||||
to the pip to use when more than one Python release is installed (e.g.
|
||||
``/usr/bin/pip-2.7`` or ``/usr/bin/pip-2.6``. If a directory path is
|
||||
specified, it is assumed to be a virtualenv.
|
||||
|
||||
use_wheel
|
||||
Prefer wheel archives (requires pip>=1.4)
|
||||
@ -569,7 +566,7 @@ def install(pkgs=None, # pylint: disable=R0912,R0913,R0914
|
||||
The user under which to run pip
|
||||
|
||||
cwd
|
||||
Current working directory to run pip from
|
||||
Directory from which to run pip
|
||||
|
||||
pre_releases
|
||||
Include pre-releases in the available versions
|
||||
@ -630,7 +627,7 @@ def install(pkgs=None, # pylint: disable=R0912,R0913,R0914
|
||||
'''
|
||||
if 'no_chown' in kwargs:
|
||||
salt.utils.versions.warn_until(
|
||||
'Flourine',
|
||||
'Fluorine',
|
||||
'The no_chown argument has been deprecated and is no longer used. '
|
||||
'Its functionality was removed in Boron.')
|
||||
kwargs.pop('no_chown')
|
||||
@ -941,36 +938,38 @@ def uninstall(pkgs=None,
|
||||
saltenv='base',
|
||||
use_vt=False):
|
||||
'''
|
||||
Uninstall packages with pip
|
||||
|
||||
Uninstall packages individually or from a pip requirements file. Uninstall
|
||||
packages globally or from a virtualenv.
|
||||
Uninstall packages individually or from a pip requirements file
|
||||
|
||||
pkgs
|
||||
comma separated list of packages to install
|
||||
|
||||
requirements
|
||||
path to requirements.
|
||||
Path to requirements file
|
||||
|
||||
bin_env
|
||||
path to pip bin or path to virtualenv. If doing an uninstall from
|
||||
the system python and want to use a specific pip bin (pip-2.7,
|
||||
pip-2.6, etc..) just specify the pip bin you want.
|
||||
If uninstalling from a virtualenv, just use the path to the virtualenv
|
||||
(/home/code/path/to/virtualenv/)
|
||||
Path to pip (or to a virtualenv). This can be used to specify the path
|
||||
to the pip to use when more than one Python release is installed (e.g.
|
||||
``/usr/bin/pip-2.7`` or ``/usr/bin/pip-2.6``. If a directory path is
|
||||
specified, it is assumed to be a virtualenv.
|
||||
|
||||
log
|
||||
Log file where a complete (maximum verbosity) record will be kept
|
||||
|
||||
proxy
|
||||
Specify a proxy in the form
|
||||
user:passwd@proxy.server:port. Note that the
|
||||
user:password@ is optional and required only if you
|
||||
are behind an authenticated proxy. If you provide
|
||||
user@proxy.server:port then you will be prompted for a
|
||||
password.
|
||||
Specify a proxy in the format ``user:passwd@proxy.server:port``. Note
|
||||
that the ``user:password@`` is optional and required only if you are
|
||||
behind an authenticated proxy. If you provide
|
||||
``user@proxy.server:port`` then you will be prompted for a password.
|
||||
|
||||
timeout
|
||||
Set the socket timeout (default 15 seconds)
|
||||
|
||||
user
|
||||
The user under which to run pip
|
||||
|
||||
cwd
|
||||
Current working directory to run pip from
|
||||
Directory from which to run pip
|
||||
|
||||
use_vt
|
||||
Use VT terminal emulation (see output while installing)
|
||||
|
||||
@ -982,7 +981,6 @@ def uninstall(pkgs=None,
|
||||
salt '*' pip.uninstall requirements=/path/to/requirements.txt
|
||||
salt '*' pip.uninstall <package name> bin_env=/path/to/virtualenv
|
||||
salt '*' pip.uninstall <package name> bin_env=/path/to/pip_bin
|
||||
|
||||
'''
|
||||
cmd = _get_pip_bin(bin_env)
|
||||
cmd.extend(['uninstall', '-y'])
|
||||
@ -1065,32 +1063,27 @@ def freeze(bin_env=None,
|
||||
virtualenv
|
||||
|
||||
bin_env
|
||||
path to pip bin or path to virtualenv. If doing an uninstall from
|
||||
the system python and want to use a specific pip bin (pip-2.7,
|
||||
pip-2.6, etc..) just specify the pip bin you want.
|
||||
If uninstalling from a virtualenv, just use the path to the virtualenv
|
||||
(/home/code/path/to/virtualenv/)
|
||||
Path to pip (or to a virtualenv). This can be used to specify the path
|
||||
to the pip to use when more than one Python release is installed (e.g.
|
||||
``/usr/bin/pip-2.7`` or ``/usr/bin/pip-2.6``. If a directory path is
|
||||
specified, it is assumed to be a virtualenv.
|
||||
|
||||
user
|
||||
The user under which to run pip
|
||||
|
||||
cwd
|
||||
Current working directory to run pip from
|
||||
Directory from which to run pip
|
||||
|
||||
.. note::
|
||||
|
||||
If the version of pip available is older than 8.0.3, the list will not
|
||||
include the packages pip, wheel, setuptools, or distribute even if they
|
||||
are installed.
|
||||
include the packages ``pip``, ``wheel``, ``setuptools``, or
|
||||
``distribute`` even if they are installed.
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' pip.freeze /home/code/path/to/virtualenv/
|
||||
|
||||
.. versionchanged:: 2016.11.2
|
||||
|
||||
The packages pip, wheel, setuptools, and distribute are included if the
|
||||
installed pip is new enough.
|
||||
salt '*' pip.freeze bin_env=/home/code/path/to/virtualenv
|
||||
'''
|
||||
cmd = _get_pip_bin(bin_env)
|
||||
cmd.append('freeze')
|
||||
@ -1135,21 +1128,16 @@ def list_(prefix=None,
|
||||
.. note::
|
||||
|
||||
If the version of pip available is older than 8.0.3, the packages
|
||||
wheel, setuptools, and distribute will not be reported by this function
|
||||
even if they are installed. Unlike
|
||||
:py:func:`pip.freeze <salt.modules.pip.freeze>`, this function always
|
||||
reports the version of pip which is installed.
|
||||
``wheel``, ``setuptools``, and ``distribute`` will not be reported by
|
||||
this function even if they are installed. Unlike :py:func:`pip.freeze
|
||||
<salt.modules.pip.freeze>`, this function always reports the version of
|
||||
pip which is installed.
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' pip.list salt
|
||||
|
||||
.. versionchanged:: 2016.11.2
|
||||
|
||||
The packages wheel, setuptools, and distribute are included if the
|
||||
installed pip is new enough.
|
||||
'''
|
||||
packages = {}
|
||||
|
||||
@ -1457,9 +1445,10 @@ def list_all_versions(pkg,
|
||||
The package to check
|
||||
|
||||
bin_env
|
||||
Path to pip bin or path to virtualenv. If doing a system install,
|
||||
and want to use a specific pip bin (pip-2.7, pip-2.6, etc..) just
|
||||
specify the pip bin you want.
|
||||
Path to pip (or to a virtualenv). This can be used to specify the path
|
||||
to the pip to use when more than one Python release is installed (e.g.
|
||||
``/usr/bin/pip-2.7`` or ``/usr/bin/pip-2.6``. If a directory path is
|
||||
specified, it is assumed to be a virtualenv.
|
||||
|
||||
include_alpha
|
||||
Include alpha versions in the list
|
||||
@ -1474,7 +1463,7 @@ def list_all_versions(pkg,
|
||||
The user under which to run pip
|
||||
|
||||
cwd
|
||||
Current working directory to run pip from
|
||||
Directory from which to run pip
|
||||
|
||||
CLI Example:
|
||||
|
||||
|
@ -181,7 +181,9 @@ def latest_version(*names, **kwargs):
|
||||
|
||||
out = __salt__['cmd.run'](cmd, output_loglevel='trace')
|
||||
for line in out.splitlines():
|
||||
p = line.split(',' if _supports_parsing() else None)
|
||||
if line.startswith('No results found for'):
|
||||
return pkglist
|
||||
p = line.split(';' if _supports_parsing() else None)
|
||||
|
||||
if p and p[0] in ('=:', '<:', '>:', ''):
|
||||
# These are explanation comments
|
||||
@ -190,7 +192,7 @@ def latest_version(*names, **kwargs):
|
||||
s = _splitpkg(p[0])
|
||||
if s:
|
||||
if not s[0] in pkglist:
|
||||
if len(p) > 1 and p[1] == '<':
|
||||
if len(p) > 1 and p[1] in ('<', '', '='):
|
||||
pkglist[s[0]] = s[1]
|
||||
else:
|
||||
pkglist[s[0]] = ''
|
||||
@ -669,7 +671,6 @@ def file_dict(*packages):
|
||||
for package in packages:
|
||||
cmd = ['pkg_info', '-qL', package]
|
||||
ret = __salt__['cmd.run_all'](cmd, output_loglevel='trace')
|
||||
|
||||
files[package] = []
|
||||
for line in ret['stderr'].splitlines():
|
||||
errors.append(line)
|
||||
@ -681,7 +682,7 @@ def file_dict(*packages):
|
||||
continue # unexpected string
|
||||
|
||||
ret = {'errors': errors, 'files': files}
|
||||
for field in ret:
|
||||
for field in list(ret):
|
||||
if not ret[field] or ret[field] == '':
|
||||
del ret[field]
|
||||
return ret
|
||||
|
@ -65,21 +65,22 @@ def __virtual__():
|
||||
# The module will be exposed as `rpmbuild` on non-RPM based systems
|
||||
return 'rpmbuild'
|
||||
else:
|
||||
return False, 'The rpmbuild module could not be loaded: requires python-gnupg, gpg, rpm, rpmbuild, mock and createrepo utilities to be installed'
|
||||
return False, 'The rpmbuild module could not be loaded: requires python-gnupg, ' \
|
||||
'gpg, rpm, rpmbuild, mock and createrepo utilities to be installed'
|
||||
|
||||
|
||||
def _create_rpmmacros():
|
||||
def _create_rpmmacros(runas='root'):
|
||||
'''
|
||||
Create the .rpmmacros file in user's home directory
|
||||
'''
|
||||
home = os.path.expanduser('~')
|
||||
rpmbuilddir = os.path.join(home, 'rpmbuild')
|
||||
if not os.path.isdir(rpmbuilddir):
|
||||
os.makedirs(rpmbuilddir)
|
||||
__salt__['file.makedirs_perms'](name=rpmbuilddir, user=runas, group='mock')
|
||||
|
||||
mockdir = os.path.join(home, 'mock')
|
||||
if not os.path.isdir(mockdir):
|
||||
os.makedirs(mockdir)
|
||||
__salt__['file.makedirs_perms'](name=mockdir, user=runas, group='mock')
|
||||
|
||||
rpmmacros = os.path.join(home, '.rpmmacros')
|
||||
with salt.utils.files.fopen(rpmmacros, 'w') as afile:
|
||||
@ -92,7 +93,7 @@ def _create_rpmmacros():
|
||||
afile.write('%_gpg_name packaging@saltstack.com\n')
|
||||
|
||||
|
||||
def _mk_tree():
|
||||
def _mk_tree(runas='root'):
|
||||
'''
|
||||
Create the rpm build tree
|
||||
'''
|
||||
@ -100,7 +101,7 @@ def _mk_tree():
|
||||
paths = ['BUILD', 'RPMS', 'SOURCES', 'SPECS', 'SRPMS']
|
||||
for path in paths:
|
||||
full = os.path.join(basedir, path)
|
||||
os.makedirs(full)
|
||||
__salt__['file.makedirs_perms'](name=full, user=runas, group='mock')
|
||||
return basedir
|
||||
|
||||
|
||||
@ -116,7 +117,7 @@ def _get_spec(tree_base, spec, template, saltenv='base'):
|
||||
saltenv=saltenv)
|
||||
|
||||
|
||||
def _get_src(tree_base, source, saltenv='base'):
|
||||
def _get_src(tree_base, source, saltenv='base', runas='root'):
|
||||
'''
|
||||
Get the named sources and place them into the tree_base
|
||||
'''
|
||||
@ -127,6 +128,7 @@ def _get_src(tree_base, source, saltenv='base'):
|
||||
lsrc = __salt__['cp.get_url'](source, dest, saltenv=saltenv)
|
||||
else:
|
||||
shutil.copy(source, dest)
|
||||
__salt__['file.chown'](path=dest, user=runas, group='mock')
|
||||
|
||||
|
||||
def _get_distset(tgt):
|
||||
@ -171,7 +173,7 @@ def _get_deps(deps, tree_base, saltenv='base'):
|
||||
return deps_list
|
||||
|
||||
|
||||
def make_src_pkg(dest_dir, spec, sources, env=None, template=None, saltenv='base'):
|
||||
def make_src_pkg(dest_dir, spec, sources, env=None, template=None, saltenv='base', runas='root'):
|
||||
'''
|
||||
Create a source rpm from the given spec file and sources
|
||||
|
||||
@ -179,33 +181,74 @@ def make_src_pkg(dest_dir, spec, sources, env=None, template=None, saltenv='base
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' pkgbuild.make_src_pkg /var/www/html/ https://raw.githubusercontent.com/saltstack/libnacl/master/pkg/rpm/python-libnacl.spec https://pypi.python.org/packages/source/l/libnacl/libnacl-1.3.5.tar.gz
|
||||
salt '*' pkgbuild.make_src_pkg /var/www/html/
|
||||
https://raw.githubusercontent.com/saltstack/libnacl/master/pkg/rpm/python-libnacl.spec
|
||||
https://pypi.python.org/packages/source/l/libnacl/libnacl-1.3.5.tar.gz
|
||||
|
||||
This example command should build the libnacl SOURCE package and place it in
|
||||
/var/www/html/ on the minion
|
||||
|
||||
.. versionchanged:: 2017.7.0
|
||||
|
||||
dest_dir
|
||||
The directory on the minion to place the built package(s)
|
||||
|
||||
spec
|
||||
The location of the spec file (used for rpms)
|
||||
|
||||
sources
|
||||
The list of package sources
|
||||
|
||||
env
|
||||
A dictionary of environment variables to be set prior to execution.
|
||||
|
||||
template
|
||||
Run the spec file through a templating engine
|
||||
Optional arguement, allows for no templating engine used to be
|
||||
if none is desired.
|
||||
|
||||
saltenv
|
||||
The saltenv to use for files downloaded from the salt filesever
|
||||
|
||||
runas
|
||||
The user to run the build process as
|
||||
|
||||
.. versionadded:: 2018.3.2
|
||||
|
||||
|
||||
.. note::
|
||||
|
||||
using SHA256 as digest and minimum level dist el6
|
||||
|
||||
'''
|
||||
_create_rpmmacros()
|
||||
tree_base = _mk_tree()
|
||||
_create_rpmmacros(runas)
|
||||
tree_base = _mk_tree(runas)
|
||||
spec_path = _get_spec(tree_base, spec, template, saltenv)
|
||||
__salt__['file.chown'](path=spec_path, user=runas, group='mock')
|
||||
__salt__['file.chown'](path=tree_base, user=runas, group='mock')
|
||||
|
||||
if isinstance(sources, six.string_types):
|
||||
sources = sources.split(',')
|
||||
for src in sources:
|
||||
_get_src(tree_base, src, saltenv)
|
||||
_get_src(tree_base, src, saltenv, runas)
|
||||
|
||||
# make source rpms for dist el6 with SHA256, usable with mock on other dists
|
||||
cmd = 'rpmbuild --verbose --define "_topdir {0}" -bs --define "dist .el6" {1}'.format(tree_base, spec_path)
|
||||
__salt__['cmd.run'](cmd)
|
||||
retrc = __salt__['cmd.retcode'](cmd, runas=runas)
|
||||
if retrc != 0:
|
||||
raise SaltInvocationError(
|
||||
'Make source package for destination directory {0}, spec {1}, sources {2}, failed '
|
||||
'with return error {3}, check logs for further details'.format(
|
||||
dest_dir,
|
||||
spec,
|
||||
sources,
|
||||
retrc)
|
||||
)
|
||||
|
||||
srpms = os.path.join(tree_base, 'SRPMS')
|
||||
ret = []
|
||||
if not os.path.isdir(dest_dir):
|
||||
os.makedirs(dest_dir)
|
||||
__salt__['file.makedirs_perms'](name=dest_dir, user=runas, group='mock')
|
||||
for fn_ in os.listdir(srpms):
|
||||
full = os.path.join(srpms, fn_)
|
||||
tgt = os.path.join(dest_dir, fn_)
|
||||
@ -232,14 +275,16 @@ def build(runas,
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' pkgbuild.build mock epel-7-x86_64 /var/www/html https://raw.githubusercontent.com/saltstack/libnacl/master/pkg/rpm/python-libnacl.spec https://pypi.python.org/packages/source/l/libnacl/libnacl-1.3.5.tar.gz
|
||||
salt '*' pkgbuild.build mock epel-7-x86_64 /var/www/html
|
||||
https://raw.githubusercontent.com/saltstack/libnacl/master/pkg/rpm/python-libnacl.spec
|
||||
https://pypi.python.org/packages/source/l/libnacl/libnacl-1.3.5.tar.gz
|
||||
|
||||
This example command should build the libnacl package for rhel 7 using user
|
||||
mock and place it in /var/www/html/ on the minion
|
||||
'''
|
||||
ret = {}
|
||||
try:
|
||||
os.makedirs(dest_dir)
|
||||
__salt__['file.chown'](path=dest_dir, user=runas, group='mock')
|
||||
except OSError as exc:
|
||||
if exc.errno != errno.EEXIST:
|
||||
raise
|
||||
@ -247,7 +292,7 @@ def build(runas,
|
||||
srpm_build_dir = tempfile.mkdtemp()
|
||||
try:
|
||||
srpms = make_src_pkg(srpm_build_dir, spec, sources,
|
||||
env, template, saltenv)
|
||||
env, template, saltenv, runas)
|
||||
except Exception as exc:
|
||||
shutil.rmtree(srpm_build_dir)
|
||||
log.error('Failed to make src package')
|
||||
@ -259,17 +304,18 @@ def build(runas,
|
||||
deps_dir = tempfile.mkdtemp()
|
||||
deps_list = _get_deps(deps, deps_dir, saltenv)
|
||||
|
||||
retrc = 0
|
||||
for srpm in srpms:
|
||||
dbase = os.path.dirname(srpm)
|
||||
results_dir = tempfile.mkdtemp()
|
||||
try:
|
||||
__salt__['cmd.run']('chown {0} -R {1}'.format(runas, dbase))
|
||||
__salt__['cmd.run']('chown {0} -R {1}'.format(runas, results_dir))
|
||||
__salt__['file.chown'](path=dbase, user=runas, group='mock')
|
||||
__salt__['file.chown'](path=results_dir, user=runas, group='mock')
|
||||
cmd = 'mock --root={0} --resultdir={1} --init'.format(tgt, results_dir)
|
||||
__salt__['cmd.run'](cmd, runas=runas)
|
||||
retrc |= __salt__['cmd.retcode'](cmd, runas=runas)
|
||||
if deps_list and not deps_list.isspace():
|
||||
cmd = 'mock --root={0} --resultdir={1} --install {2} {3}'.format(tgt, results_dir, deps_list, noclean)
|
||||
__salt__['cmd.run'](cmd, runas=runas)
|
||||
retrc |= __salt__['cmd.retcode'](cmd, runas=runas)
|
||||
noclean += ' --no-clean'
|
||||
|
||||
cmd = 'mock --root={0} --resultdir={1} {2} {3} {4}'.format(
|
||||
@ -278,17 +324,20 @@ def build(runas,
|
||||
distset,
|
||||
noclean,
|
||||
srpm)
|
||||
__salt__['cmd.run'](cmd, runas=runas)
|
||||
cmd = ['rpm', '-qp', '--queryformat',
|
||||
'{0}/%{{name}}/%{{version}}-%{{release}}'.format(log_dir),
|
||||
srpm]
|
||||
log_dest = __salt__['cmd.run_stdout'](cmd, python_shell=False)
|
||||
retrc |= __salt__['cmd.retcode'](cmd, runas=runas)
|
||||
cmdlist = [
|
||||
'rpm',
|
||||
'-qp',
|
||||
'--queryformat',
|
||||
'{0}/%{{name}}/%{{version}}-%{{release}}'.format(log_dir),
|
||||
srpm]
|
||||
log_dest = __salt__['cmd.run_stdout'](cmdlist, python_shell=False)
|
||||
for filename in os.listdir(results_dir):
|
||||
full = os.path.join(results_dir, filename)
|
||||
if filename.endswith('src.rpm'):
|
||||
sdest = os.path.join(srpm_dir, filename)
|
||||
try:
|
||||
os.makedirs(srpm_dir)
|
||||
__salt__['file.makedirs_perms'](name=srpm_dir, user=runas, group='mock')
|
||||
except OSError as exc:
|
||||
if exc.errno != errno.EEXIST:
|
||||
raise
|
||||
@ -301,7 +350,7 @@ def build(runas,
|
||||
else:
|
||||
log_file = os.path.join(log_dest, filename)
|
||||
try:
|
||||
os.makedirs(log_dest)
|
||||
__salt__['file.makedirs_perms'](name=log_dest, user=runas, group='mock')
|
||||
except OSError as exc:
|
||||
if exc.errno != errno.EEXIST:
|
||||
raise
|
||||
@ -311,6 +360,15 @@ def build(runas,
|
||||
log.error('Error building from %s: %s', srpm, exc)
|
||||
finally:
|
||||
shutil.rmtree(results_dir)
|
||||
if retrc != 0:
|
||||
raise SaltInvocationError(
|
||||
'Building packages for destination directory {0}, spec {1}, sources {2}, failed '
|
||||
'with return error {3}, check logs for further details'.format(
|
||||
dest_dir,
|
||||
spec,
|
||||
sources,
|
||||
retrc)
|
||||
)
|
||||
shutil.rmtree(deps_dir)
|
||||
shutil.rmtree(srpm_build_dir)
|
||||
return ret
|
||||
@ -433,7 +491,7 @@ def make_repo(repodir,
|
||||
phrase = ''
|
||||
|
||||
if keyid is not None:
|
||||
## import_keys
|
||||
# import_keys
|
||||
pkg_pub_key_file = '{0}/{1}'.format(gnupghome, __salt__['pillar.get']('gpg_pkg_pub_keyname', None))
|
||||
pkg_priv_key_file = '{0}/{1}'.format(gnupghome, __salt__['pillar.get']('gpg_pkg_priv_keyname', None))
|
||||
|
||||
@ -477,14 +535,21 @@ def make_repo(repodir,
|
||||
|
||||
# need to update rpm with public key
|
||||
cmd = 'rpm --import {0}'.format(pkg_pub_key_file)
|
||||
__salt__['cmd.run'](cmd, runas=runas, use_vt=True)
|
||||
retrc = __salt__['cmd.retcode'](cmd, runas=runas, use_vt=True)
|
||||
if retrc != 0:
|
||||
raise SaltInvocationError(
|
||||
'Failed to import public key from file {0} with return '
|
||||
'error {1}, check logs for further details'.format(
|
||||
pkg_pub_key_file,
|
||||
retrc)
|
||||
)
|
||||
|
||||
## sign_it_here
|
||||
# sign_it_here
|
||||
# interval of 0.125 is really too fast on some systems
|
||||
interval = 0.5
|
||||
for file in os.listdir(repodir):
|
||||
if file.endswith('.rpm'):
|
||||
abs_file = os.path.join(repodir, file)
|
||||
for fileused in os.listdir(repodir):
|
||||
if fileused.endswith('.rpm'):
|
||||
abs_file = os.path.join(repodir, fileused)
|
||||
number_retries = timeout / interval
|
||||
times_looped = 0
|
||||
error_msg = 'Failed to sign file {0}'.format(abs_file)
|
||||
|
@ -989,7 +989,7 @@ def diskusage(*args):
|
||||
elif __grains__['kernel'] in ('FreeBSD', 'SunOS'):
|
||||
ifile = __salt__['cmd.run']('mount -p').splitlines()
|
||||
else:
|
||||
ifile = []
|
||||
raise CommandExecutionError('status.diskusage not yet supported on this platform')
|
||||
|
||||
for line in ifile:
|
||||
comps = line.split()
|
||||
|
@ -1,6 +1,5 @@
|
||||
|
||||
# -*- coding: utf-8 -*-
|
||||
"""
|
||||
'''
|
||||
Docker Swarm Module using Docker's Python SDK
|
||||
=============================================
|
||||
|
||||
@ -18,12 +17,17 @@ Dependencies
|
||||
|
||||
Docker Python SDK
|
||||
-----------------
|
||||
pip install -U docker
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
pip install -U docker
|
||||
|
||||
More information: https://docker-py.readthedocs.io/en/stable/
|
||||
"""
|
||||
'''
|
||||
# Import python libraries
|
||||
from __future__ import absolute_import, unicode_literals, print_function
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.json
|
||||
|
||||
try:
|
||||
@ -41,7 +45,7 @@ def __virtual__():
|
||||
'''
|
||||
if HAS_DOCKER:
|
||||
return __virtualname__
|
||||
return False, 'The swarm module failed to load: Docker python module is not avaialble.'
|
||||
return False, 'The swarm module failed to load: Docker python module is not available.'
|
||||
|
||||
|
||||
def __init__(self):
|
||||
@ -58,7 +62,7 @@ def swarm_tokens():
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' swarm.swarm_tokens
|
||||
salt '*' swarm.swarm_tokens
|
||||
'''
|
||||
client = docker.APIClient(base_url='unix://var/run/docker.sock')
|
||||
service = client.inspect_swarm()
|
||||
@ -72,37 +76,39 @@ def swarm_init(advertise_addr=str,
|
||||
Initalize Docker on Minion as a Swarm Manager
|
||||
|
||||
advertise_addr
|
||||
The ip of the manager
|
||||
The ip of the manager
|
||||
|
||||
listen_addr
|
||||
Listen address used for inter-manager communication,
|
||||
as well as determining the networking interface used
|
||||
for the VXLAN Tunnel Endpoint (VTEP).
|
||||
This can either be an address/port combination in
|
||||
the form 192.168.1.1:4567,
|
||||
or an interface followed by a port number,
|
||||
like eth0:4567
|
||||
Listen address used for inter-manager communication,
|
||||
as well as determining the networking interface used
|
||||
for the VXLAN Tunnel Endpoint (VTEP).
|
||||
This can either be an address/port combination in
|
||||
the form 192.168.1.1:4567,
|
||||
or an interface followed by a port number,
|
||||
like eth0:4567
|
||||
|
||||
force_new_cluster
|
||||
Force a new cluster if True is passed
|
||||
Force a new cluster if True is passed
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' swarm.swarm_init advertise_addr='192.168.50.10' listen_addr='0.0.0.0' force_new_cluster=False
|
||||
salt '*' swarm.swarm_init advertise_addr='192.168.50.10' listen_addr='0.0.0.0' force_new_cluster=False
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
__context__['client'].swarm.init(advertise_addr,
|
||||
listen_addr,
|
||||
force_new_cluster)
|
||||
output = 'Docker swarm has been Initalized on '+ __context__['server_name'] + ' and the worker/manager Join token is below'
|
||||
output = 'Docker swarm has been initialized on {0} ' \
|
||||
'and the worker/manager Join token is below'.format(__context__['server_name'])
|
||||
salt_return.update({'Comment': output,
|
||||
'Tokens': swarm_tokens()})
|
||||
except TypeError:
|
||||
salt_return = {}
|
||||
salt_return.update({'Error': 'Please make sure your passing advertise_addr, listen_addr and force_new_cluster correctly.'})
|
||||
salt_return.update({'Error': 'Please make sure you are passing advertise_addr, '
|
||||
'listen_addr and force_new_cluster correctly.'})
|
||||
return salt_return
|
||||
|
||||
|
||||
@ -113,21 +119,22 @@ def joinswarm(remote_addr=int,
|
||||
Join a Swarm Worker to the cluster
|
||||
|
||||
remote_addr
|
||||
The manager node you want to connect to for the swarm
|
||||
The manager node you want to connect to for the swarm
|
||||
|
||||
listen_addr
|
||||
Listen address used for inter-manager communication if the node gets promoted to manager,
|
||||
as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP)
|
||||
Listen address used for inter-manager communication if the node gets promoted to manager,
|
||||
as well as determining the networking interface used for the VXLAN Tunnel Endpoint (VTEP)
|
||||
|
||||
token
|
||||
Either the manager join token or the worker join token.
|
||||
You can get the worker or manager token via `salt '*' swarm.swarm_tokens`
|
||||
Either the manager join token or the worker join token.
|
||||
You can get the worker or manager token via ``salt '*' swarm.swarm_tokens``
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' swarm.joinswarm remote_addr=192.168.50.10 listen_addr='0.0.0.0' token='SWMTKN-1-64tux2g0701r84ofq93zppcih0pe081akq45owe9ts61f30x4t-06trjugdu7x2z47j938s54il'
|
||||
salt '*' swarm.joinswarm remote_addr=192.168.50.10 listen_addr='0.0.0.0' \
|
||||
token='SWMTKN-1-64tux2g0701r84ofq93zppcih0pe081akq45owe9ts61f30x4t-06trjugdu7x2z47j938s54il'
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
@ -138,7 +145,8 @@ def joinswarm(remote_addr=int,
|
||||
salt_return.update({'Comment': output, 'Manager_Addr': remote_addr})
|
||||
except TypeError:
|
||||
salt_return = {}
|
||||
salt_return.update({'Error': 'Please make sure this minion is not part of a swarm and your passing remote_addr, listen_addr and token correctly.'})
|
||||
salt_return.update({'Error': 'Please make sure this minion is not part of a swarm and you are '
|
||||
'passing remote_addr, listen_addr and token correctly.'})
|
||||
return salt_return
|
||||
|
||||
|
||||
@ -147,13 +155,13 @@ def leave_swarm(force=bool):
|
||||
Force the minion to leave the swarm
|
||||
|
||||
force
|
||||
Will force the minion/worker/manager to leave the swarm
|
||||
Will force the minion/worker/manager to leave the swarm
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' swarm.leave_swarm force=False
|
||||
salt '*' swarm.leave_swarm force=False
|
||||
'''
|
||||
salt_return = {}
|
||||
__context__['client'].swarm.leave(force=force)
|
||||
@ -173,30 +181,32 @@ def service_create(image=str,
|
||||
Create Docker Swarm Service Create
|
||||
|
||||
image
|
||||
The docker image
|
||||
The docker image
|
||||
|
||||
name
|
||||
Is the service name
|
||||
Is the service name
|
||||
|
||||
command
|
||||
The docker command to run in the container at launch
|
||||
The docker command to run in the container at launch
|
||||
|
||||
hostname
|
||||
The hostname of the containers
|
||||
The hostname of the containers
|
||||
|
||||
replicas
|
||||
How many replicas you want running in the swarm
|
||||
How many replicas you want running in the swarm
|
||||
|
||||
target_port
|
||||
The target port on the container
|
||||
The target port on the container
|
||||
|
||||
published_port
|
||||
port thats published on the host/os
|
||||
port thats published on the host/os
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' swarm.service_create image=httpd name=Test_Service command=None hostname=salthttpd replicas=6 target_port=80 published_port=80
|
||||
|
||||
salt '*' swarm.service_create image=httpd name=Test_Service \
|
||||
command=None hostname=salthttpd replicas=6 target_port=80 published_port=80
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
@ -219,7 +229,8 @@ def service_create(image=str,
|
||||
'Published_Port': published_port})
|
||||
except TypeError:
|
||||
salt_return = {}
|
||||
salt_return.update({'Error': 'Please make sure your passing arguments correctly [image, name, command, hostname, replicas, target_port and published_port]'})
|
||||
salt_return.update({'Error': 'Please make sure you are passing arguments correctly '
|
||||
'[image, name, command, hostname, replicas, target_port and published_port]'})
|
||||
return salt_return
|
||||
|
||||
|
||||
@ -228,12 +239,13 @@ def swarm_service_info(service_name=str):
|
||||
Swarm Service Information
|
||||
|
||||
service_name
|
||||
The name of the service that you want information on about the service
|
||||
The name of the service that you want information on about the service
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' swarm.swarm_service_info service_name=Test_Service
|
||||
|
||||
salt '*' swarm.swarm_service_info service_name=Test_Service
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
@ -282,12 +294,13 @@ def remove_service(service=str):
|
||||
Remove Swarm Service
|
||||
|
||||
service
|
||||
The name of the service
|
||||
The name of the service
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' swarm.remove_service service=Test_Service
|
||||
|
||||
salt '*' swarm.remove_service service=Test_Service
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
@ -306,12 +319,13 @@ def node_ls(server=str):
|
||||
Displays Information about Swarm Nodes with passing in the server
|
||||
|
||||
server
|
||||
The minion/server name
|
||||
The minion/server name
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' swarm.node_ls server=minion1
|
||||
|
||||
salt '*' swarm.node_ls server=minion1
|
||||
'''
|
||||
try:
|
||||
salt_return = {}
|
||||
@ -327,7 +341,7 @@ def node_ls(server=str):
|
||||
role = items['Spec']['Role']
|
||||
availability = items['Spec']['Availability']
|
||||
status = items['Status']
|
||||
Version = items['Version']['Index']
|
||||
version = items['Version']['Index']
|
||||
salt_return.update({'Docker Version': docker_version,
|
||||
'Platform': platform,
|
||||
'Hostname': hostnames,
|
||||
@ -335,7 +349,7 @@ def node_ls(server=str):
|
||||
'Roles': role,
|
||||
'Availability': availability,
|
||||
'Status': status,
|
||||
'Version': Version})
|
||||
'Version': version})
|
||||
except TypeError:
|
||||
salt_return = {}
|
||||
salt_return.update({'Error': 'The server arg is missing or you not targeting a Manager node?'})
|
||||
@ -347,15 +361,16 @@ def remove_node(node_id=str, force=bool):
|
||||
Remove a node from a swarm and the target needs to be a swarm manager
|
||||
|
||||
node_id
|
||||
The node id from the return of swarm.node_ls
|
||||
The node id from the return of swarm.node_ls
|
||||
|
||||
force
|
||||
Forcefully remove the node/minion from the service
|
||||
Forcefully remove the node/minion from the service
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' swarm.remove_node node_id=z4gjbe9rwmqahc2a91snvolm5 force=false
|
||||
|
||||
salt '*' swarm.remove_node node_id=z4gjbe9rwmqahc2a91snvolm5 force=false
|
||||
'''
|
||||
client = docker.APIClient(base_url='unix://var/run/docker.sock')
|
||||
try:
|
||||
@ -380,24 +395,26 @@ def update_node(availability=str,
|
||||
Updates docker swarm nodes/needs to target a manager node/minion
|
||||
|
||||
availability
|
||||
Drain or Active
|
||||
Drain or Active
|
||||
|
||||
node_name
|
||||
minion/node
|
||||
minion/node
|
||||
|
||||
role
|
||||
role of manager or worker
|
||||
role of manager or worker
|
||||
|
||||
node_id
|
||||
The Id and that can be obtained via swarm.node_ls
|
||||
The Id and that can be obtained via swarm.node_ls
|
||||
|
||||
version
|
||||
Is obtained by swarm.node_ls
|
||||
Is obtained by swarm.node_ls
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
salt '*' docker_util.update_node availability=drain node_name=minion2 role=worker node_id=3k9x7t8m4pel9c0nqr3iajnzp version=19
|
||||
|
||||
salt '*' docker_util.update_node availability=drain node_name=minion2 \
|
||||
role=worker node_id=3k9x7t8m4pel9c0nqr3iajnzp version=19
|
||||
'''
|
||||
client = docker.APIClient(base_url='unix://var/run/docker.sock')
|
||||
try:
|
||||
|
@ -60,17 +60,18 @@ def _get_gecos(name):
|
||||
Retrieve GECOS field info and return it in dictionary form
|
||||
'''
|
||||
gecos_field = salt.utils.stringutils.to_unicode(
|
||||
pwd.getpwnam(_quote_username(name)).pw_gecos).split(',', 3)
|
||||
pwd.getpwnam(_quote_username(name)).pw_gecos).split(',', 4)
|
||||
if not gecos_field:
|
||||
return {}
|
||||
else:
|
||||
# Assign empty strings for any unspecified trailing GECOS fields
|
||||
while len(gecos_field) < 4:
|
||||
while len(gecos_field) < 5:
|
||||
gecos_field.append('')
|
||||
return {'fullname': salt.utils.locales.sdecode(gecos_field[0]),
|
||||
'roomnumber': salt.utils.locales.sdecode(gecos_field[1]),
|
||||
'workphone': salt.utils.locales.sdecode(gecos_field[2]),
|
||||
'homephone': salt.utils.locales.sdecode(gecos_field[3])}
|
||||
'homephone': salt.utils.locales.sdecode(gecos_field[3]),
|
||||
'other': salt.utils.locales.sdecode(gecos_field[4])}
|
||||
|
||||
|
||||
def _build_gecos(gecos_dict):
|
||||
@ -78,10 +79,11 @@ def _build_gecos(gecos_dict):
|
||||
Accepts a dictionary entry containing GECOS field names and their values,
|
||||
and returns a full GECOS comment string, to be used with usermod.
|
||||
'''
|
||||
return '{0},{1},{2},{3}'.format(gecos_dict.get('fullname', ''),
|
||||
gecos_dict.get('roomnumber', ''),
|
||||
gecos_dict.get('workphone', ''),
|
||||
gecos_dict.get('homephone', ''))
|
||||
return '{0},{1},{2},{3},{4}'.format(gecos_dict.get('fullname', ''),
|
||||
gecos_dict.get('roomnumber', ''),
|
||||
gecos_dict.get('workphone', ''),
|
||||
gecos_dict.get('homephone', ''),
|
||||
gecos_dict.get('other', ''),).rstrip(',')
|
||||
|
||||
|
||||
def _update_gecos(name, key, value, root=None):
|
||||
@ -124,6 +126,7 @@ def add(name,
|
||||
roomnumber='',
|
||||
workphone='',
|
||||
homephone='',
|
||||
other='',
|
||||
createhome=True,
|
||||
loginclass=None,
|
||||
root=None,
|
||||
@ -237,6 +240,8 @@ def add(name,
|
||||
chworkphone(name, workphone)
|
||||
if homephone:
|
||||
chhomephone(name, homephone)
|
||||
if other:
|
||||
chother(name, other)
|
||||
return True
|
||||
|
||||
|
||||
@ -507,6 +512,19 @@ def chhomephone(name, homephone):
|
||||
return _update_gecos(name, 'homephone', homephone)
|
||||
|
||||
|
||||
def chother(name, other):
|
||||
'''
|
||||
Change the user's other GECOS attribute
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' user.chother foobar
|
||||
'''
|
||||
return _update_gecos(name, 'other', other)
|
||||
|
||||
|
||||
def chloginclass(name, loginclass, root=None):
|
||||
'''
|
||||
Change the default login class of the user
|
||||
@ -588,9 +606,9 @@ def _format_info(data):
|
||||
Return user information in a pretty way
|
||||
'''
|
||||
# Put GECOS info into a list
|
||||
gecos_field = salt.utils.stringutils.to_unicode(data.pw_gecos).split(',', 3)
|
||||
# Make sure our list has at least four elements
|
||||
while len(gecos_field) < 4:
|
||||
gecos_field = salt.utils.stringutils.to_unicode(data.pw_gecos).split(',', 4)
|
||||
# Make sure our list has at least five elements
|
||||
while len(gecos_field) < 5:
|
||||
gecos_field.append('')
|
||||
|
||||
return {'gid': data.pw_gid,
|
||||
@ -603,7 +621,8 @@ def _format_info(data):
|
||||
'fullname': gecos_field[0],
|
||||
'roomnumber': gecos_field[1],
|
||||
'workphone': gecos_field[2],
|
||||
'homephone': gecos_field[3]}
|
||||
'homephone': gecos_field[3],
|
||||
'other': gecos_field[4]}
|
||||
|
||||
|
||||
@salt.utils.decorators.path.which('id')
|
||||
|
@ -30,7 +30,7 @@ import shutil # do not remove, used in imported file.py functions
|
||||
import re # do not remove, used in imported file.py functions
|
||||
import string # do not remove, used in imported file.py functions
|
||||
import sys # do not remove, used in imported file.py functions
|
||||
import fileinput # do not remove, used in imported file.py functions
|
||||
import io # do not remove, used in imported file.py functions
|
||||
import fnmatch # do not remove, used in imported file.py functions
|
||||
import mmap # do not remove, used in imported file.py functions
|
||||
import glob # do not remove, used in imported file.py functions
|
||||
|
@ -35,7 +35,7 @@ Current known limitations
|
||||
- lxml
|
||||
- uuid
|
||||
- struct
|
||||
- salt.modules.reg
|
||||
- salt.utils.win_reg
|
||||
'''
|
||||
# Import Python libs
|
||||
from __future__ import absolute_import, unicode_literals, print_function
|
||||
@ -98,7 +98,7 @@ try:
|
||||
import lxml
|
||||
import struct
|
||||
from lxml import etree
|
||||
from salt.modules.reg import Registry as Registry
|
||||
from salt.utils.win_reg import Registry
|
||||
HAS_WINDOWS_MODULES = True
|
||||
TRUE_VALUE_XPATH = etree.XPath('.//*[local-name() = "trueValue"]')
|
||||
FALSE_VALUE_XPATH = etree.XPath('.//*[local-name() = "falseValue"]')
|
||||
@ -2672,9 +2672,11 @@ def __virtual__():
|
||||
'''
|
||||
Only works on Windows systems
|
||||
'''
|
||||
if salt.utils.platform.is_windows() and HAS_WINDOWS_MODULES:
|
||||
return __virtualname__
|
||||
return False
|
||||
if not salt.utils.platform.is_windows():
|
||||
return False, 'win_lgpo: Not a Windows System'
|
||||
if not HAS_WINDOWS_MODULES:
|
||||
return False, 'win_lgpo: Required modules failed to load'
|
||||
return __virtualname__
|
||||
|
||||
|
||||
def _updateNamespace(item, new_namespace):
|
||||
@ -5372,7 +5374,7 @@ def set_(computer_policy=None, user_policy=None,
|
||||
else:
|
||||
raise SaltInvocationError(msg)
|
||||
if policy_namespace and policy_name in _admTemplateData[policy_namespace] and the_policy is not None:
|
||||
log.debug('setting == %s', _admTemplateData[policy_namespace][policy_name].lower())
|
||||
log.debug('setting == %s', six.text_type(_admTemplateData[policy_namespace][policy_name]).lower())
|
||||
log.debug(six.text_type(_admTemplateData[policy_namespace][policy_name]).lower())
|
||||
if six.text_type(_admTemplateData[policy_namespace][policy_name]).lower() != 'disabled' \
|
||||
and six.text_type(_admTemplateData[policy_namespace][policy_name]).lower() != 'not configured':
|
||||
|
@ -331,6 +331,7 @@ def version(*names, **kwargs):
|
||||
dict: The package name(s) with the installed versions.
|
||||
|
||||
.. code-block:: cfg
|
||||
|
||||
{['<version>', '<version>', ]} OR
|
||||
{'<package name>': ['<version>', '<version>', ]}
|
||||
|
||||
|
@ -9,16 +9,20 @@ import logging
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.platform
|
||||
from salt.exceptions import SaltInvocationError
|
||||
from salt.exceptions import SaltInvocationError, CommandExecutionError
|
||||
|
||||
# Import 3rd party libs
|
||||
from salt.ext import six
|
||||
|
||||
_HKEY = 'HKLM'
|
||||
|
||||
_SNMP_KEY = r'SYSTEM\CurrentControlSet\Services\SNMP\Parameters'
|
||||
_AGENT_KEY = r'{0}\RFC1156Agent'.format(_SNMP_KEY)
|
||||
_COMMUNITIES_KEY = r'{0}\ValidCommunities'.format(_SNMP_KEY)
|
||||
|
||||
_SNMP_GPO_KEY = r'SOFTWARE\Policies\SNMP\Parameters'
|
||||
_COMMUNITIES_GPO_KEY = r'{0}\ValidCommunities'.format(_SNMP_GPO_KEY)
|
||||
|
||||
_PERMISSION_TYPES = {'None': 1,
|
||||
'Notify': 2,
|
||||
'Read Only': 4,
|
||||
@ -284,6 +288,21 @@ def get_community_names():
|
||||
'''
|
||||
Get the current accepted SNMP community names and their permissions.
|
||||
|
||||
If community names are being managed by Group Policy, those values will be
|
||||
returned instead like this:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
TestCommunity:
|
||||
Managed by GPO
|
||||
|
||||
Community names managed normally will denote the permission instead:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
TestCommunity:
|
||||
Read Only
|
||||
|
||||
Returns:
|
||||
dict: A dictionary of community names and permissions.
|
||||
|
||||
@ -294,25 +313,69 @@ def get_community_names():
|
||||
salt '*' win_snmp.get_community_names
|
||||
'''
|
||||
ret = dict()
|
||||
current_values = __salt__['reg.list_values'](
|
||||
_HKEY, _COMMUNITIES_KEY, include_default=False)
|
||||
|
||||
# The communities are stored as the community name with a numeric permission
|
||||
# value. Convert the numeric value to the text equivalent, as present in the
|
||||
# Windows SNMP service GUI.
|
||||
if isinstance(current_values, list):
|
||||
for current_value in current_values:
|
||||
# Look in GPO settings first
|
||||
if __salt__['reg.key_exists'](_HKEY, _COMMUNITIES_GPO_KEY):
|
||||
|
||||
# Ignore error values
|
||||
if not isinstance(current_value, dict):
|
||||
continue
|
||||
_LOG.debug('Loading communities from Group Policy settings')
|
||||
|
||||
permissions = six.text_type()
|
||||
for permission_name in _PERMISSION_TYPES:
|
||||
if current_value['vdata'] == _PERMISSION_TYPES[permission_name]:
|
||||
permissions = permission_name
|
||||
break
|
||||
ret[current_value['vname']] = permissions
|
||||
current_values = __salt__['reg.list_values'](
|
||||
_HKEY, _COMMUNITIES_GPO_KEY, include_default=False)
|
||||
|
||||
# GPO settings are different in that they do not designate permissions
|
||||
# They are a numbered list of communities like so:
|
||||
#
|
||||
# {1: "community 1",
|
||||
# 2: "community 2"}
|
||||
#
|
||||
# Denote that it is being managed by Group Policy.
|
||||
#
|
||||
# community 1:
|
||||
# Managed by GPO
|
||||
# community 2:
|
||||
# Managed by GPO
|
||||
if isinstance(current_values, list):
|
||||
for current_value in current_values:
|
||||
|
||||
# Ignore error values
|
||||
if not isinstance(current_value, dict):
|
||||
continue
|
||||
|
||||
ret[current_value['vdata']] = 'Managed by GPO'
|
||||
|
||||
if not ret:
|
||||
|
||||
_LOG.debug('Loading communities from SNMP settings')
|
||||
|
||||
current_values = __salt__['reg.list_values'](
|
||||
_HKEY, _COMMUNITIES_KEY, include_default=False)
|
||||
|
||||
# The communities are stored as the community name with a numeric
|
||||
# permission value. Like this (4 = Read Only):
|
||||
#
|
||||
# {"community 1": 4,
|
||||
# "community 2": 4}
|
||||
#
|
||||
# Convert the numeric value to the text equivalent, as present in the
|
||||
# Windows SNMP service GUI.
|
||||
#
|
||||
# community 1:
|
||||
# Read Only
|
||||
# community 2:
|
||||
# Read Only
|
||||
if isinstance(current_values, list):
|
||||
for current_value in current_values:
|
||||
|
||||
# Ignore error values
|
||||
if not isinstance(current_value, dict):
|
||||
continue
|
||||
|
||||
permissions = six.text_type()
|
||||
for permission_name in _PERMISSION_TYPES:
|
||||
if current_value['vdata'] == _PERMISSION_TYPES[permission_name]:
|
||||
permissions = permission_name
|
||||
break
|
||||
ret[current_value['vname']] = permissions
|
||||
|
||||
if not ret:
|
||||
_LOG.debug('Unable to find existing communities.')
|
||||
@ -323,6 +386,11 @@ def set_community_names(communities):
|
||||
'''
|
||||
Manage the SNMP accepted community names and their permissions.
|
||||
|
||||
.. note::
|
||||
Settings managed by Group Policy will always take precedence over those
|
||||
set using the SNMP interface. Therefore if this function finds Group
|
||||
Policy settings it will raise a CommandExecutionError
|
||||
|
||||
Args:
|
||||
communities (dict): A dictionary of SNMP community names and
|
||||
permissions. The possible permissions can be found via
|
||||
@ -331,6 +399,10 @@ def set_community_names(communities):
|
||||
Returns:
|
||||
bool: True if successful, otherwise False
|
||||
|
||||
Raises:
|
||||
CommandExecutionError:
|
||||
If SNMP settings are being managed by Group Policy
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
@ -339,6 +411,11 @@ def set_community_names(communities):
|
||||
'''
|
||||
values = dict()
|
||||
|
||||
if __salt__['reg.key_exists'](_HKEY, _COMMUNITIES_GPO_KEY):
|
||||
_LOG.debug('Communities on this system are managed by Group Policy')
|
||||
raise CommandExecutionError(
|
||||
'Communities on this system are managed by Group Policy')
|
||||
|
||||
current_communities = get_community_names()
|
||||
|
||||
if communities == current_communities:
|
||||
|
@ -511,7 +511,7 @@ def get_system_info():
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt 'minion-id' system.get_info
|
||||
salt 'minion-id' system.get_system_info
|
||||
'''
|
||||
os_type = {1: 'Work Station',
|
||||
2: 'Domain Controller',
|
||||
|
@ -6,487 +6,224 @@ from __future__ import absolute_import, unicode_literals, print_function
|
||||
|
||||
# Import Python libs
|
||||
import logging
|
||||
import re
|
||||
from datetime import datetime
|
||||
|
||||
# Import Salt libs
|
||||
import salt.utils.path
|
||||
import salt.utils.platform
|
||||
from salt.exceptions import CommandExecutionError
|
||||
|
||||
# Import 3rd party libs
|
||||
try:
|
||||
import pytz
|
||||
HAS_PYTZ = True
|
||||
except ImportError:
|
||||
HAS_PYTZ = False
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
# Maybe put in a different file ... ? %-0
|
||||
# http://unicode.org/repos/cldr/trunk/common/supplemental/windowsZones.xml
|
||||
LINTOWIN = {
|
||||
'Africa/Abidjan': 'Greenwich Standard Time',
|
||||
'Africa/Accra': 'Greenwich Standard Time',
|
||||
'Africa/Addis_Ababa': 'E. Africa Standard Time',
|
||||
'Africa/Algiers': 'W. Central Africa Standard Time',
|
||||
'Africa/Asmera': 'E. Africa Standard Time',
|
||||
'Africa/Bamako': 'Greenwich Standard Time',
|
||||
'Africa/Bangui': 'W. Central Africa Standard Time',
|
||||
'Africa/Banjul': 'Greenwich Standard Time',
|
||||
'Africa/Bissau': 'Greenwich Standard Time',
|
||||
'Africa/Blantyre': 'South Africa Standard Time',
|
||||
'Africa/Brazzaville': 'W. Central Africa Standard Time',
|
||||
'Africa/Bujumbura': 'South Africa Standard Time',
|
||||
'Africa/Cairo': 'Egypt Standard Time',
|
||||
'Africa/Casablanca': 'Morocco Standard Time',
|
||||
'Africa/Conakry': 'Greenwich Standard Time',
|
||||
'Africa/Dakar': 'Greenwich Standard Time',
|
||||
'Africa/Dar_es_Salaam': 'E. Africa Standard Time',
|
||||
'Africa/Djibouti': 'E. Africa Standard Time',
|
||||
'Africa/Douala': 'W. Central Africa Standard Time',
|
||||
'Africa/El_Aaiun': 'Greenwich Standard Time',
|
||||
'Africa/Freetown': 'Greenwich Standard Time',
|
||||
'Africa/Gaborone': 'South Africa Standard Time',
|
||||
'Africa/Harare': 'South Africa Standard Time',
|
||||
'Africa/Johannesburg': 'South Africa Standard Time',
|
||||
'Africa/Juba': 'E. Africa Standard Time',
|
||||
'Africa/Kampala': 'E. Africa Standard Time',
|
||||
'Africa/Khartoum': 'E. Africa Standard Time',
|
||||
'Africa/Kigali': 'South Africa Standard Time',
|
||||
'Africa/Kinshasa': 'W. Central Africa Standard Time',
|
||||
'Africa/Lagos': 'W. Central Africa Standard Time',
|
||||
'Africa/Libreville': 'W. Central Africa Standard Time',
|
||||
'Africa/Lome': 'Greenwich Standard Time',
|
||||
'Africa/Luanda': 'W. Central Africa Standard Time',
|
||||
'Africa/Lubumbashi': 'South Africa Standard Time',
|
||||
'Africa/Lusaka': 'South Africa Standard Time',
|
||||
'Africa/Malabo': 'W. Central Africa Standard Time',
|
||||
'Africa/Maputo': 'South Africa Standard Time',
|
||||
'Africa/Maseru': 'South Africa Standard Time',
|
||||
'Africa/Mbabane': 'South Africa Standard Time',
|
||||
'Africa/Mogadishu': 'E. Africa Standard Time',
|
||||
'Africa/Monrovia': 'Greenwich Standard Time',
|
||||
'Africa/Nairobi': 'E. Africa Standard Time',
|
||||
'Africa/Ndjamena': 'W. Central Africa Standard Time',
|
||||
'Africa/Niamey': 'W. Central Africa Standard Time',
|
||||
'Africa/Nouakchott': 'Greenwich Standard Time',
|
||||
'Africa/Ouagadougou': 'Greenwich Standard Time',
|
||||
'Africa/Porto-Novo': 'W. Central Africa Standard Time',
|
||||
'Africa/Sao_Tome': 'Greenwich Standard Time',
|
||||
'Africa/Tripoli': 'W. Europe Standard Time',
|
||||
'Africa/Tunis': 'W. Central Africa Standard Time',
|
||||
'Africa/Windhoek': 'Namibia Standard Time',
|
||||
'America/Anchorage': 'Alaskan Standard Time',
|
||||
'America/Juneau': 'Alaskan Standard Time',
|
||||
'America/Nome': 'Alaskan Standard Time',
|
||||
'America/Sitka': 'Alaskan Standard Time',
|
||||
'America/Yakutat': 'Alaskan Standard Time',
|
||||
'America/Anguilla': 'SA Western Standard Time',
|
||||
'America/Antigua': 'SA Western Standard Time',
|
||||
'America/Aruba': 'SA Western Standard Time',
|
||||
'America/Asuncion': 'Paraguay Standard Time',
|
||||
'America/Bahia': 'Bahia Standard Time',
|
||||
'America/Barbados': 'SA Western Standard Time',
|
||||
'America/Belize': 'Central America Standard Time',
|
||||
'America/Blanc-Sablon': 'SA Western Standard Time',
|
||||
'America/Bogota': 'SA Pacific Standard Time',
|
||||
'America/Buenos_Aires': 'Argentina Standard Time',
|
||||
'America/Argentina/La_Rioja': 'Argentina Standard Time',
|
||||
'America/Argentina/Rio_Gallegos': 'Argentina Standard Time',
|
||||
'America/Argentina/Salta': 'Argentina Standard Time',
|
||||
'America/Argentina/San_Juan': 'Argentina Standard Time',
|
||||
'America/Argentina/San_Luis': 'Argentina Standard Time',
|
||||
'America/Argentina/Tucuman': 'Argentina Standard Time',
|
||||
'America/Argentina/Ushuaia': 'Argentina Standard Time',
|
||||
'America/Catamarca': 'Argentina Standard Time',
|
||||
'America/Cordoba': 'Argentina Standard Time',
|
||||
'America/Jujuy': 'Argentina Standard Time',
|
||||
'America/Mendoza': 'Argentina Standard Time',
|
||||
'America/Caracas': 'Venezuela Standard Time',
|
||||
'America/Cayenne': 'SA Eastern Standard Time',
|
||||
'America/Cayman': 'SA Pacific Standard Time',
|
||||
'America/Chicago': 'Central Standard Time',
|
||||
'America/Indiana/Knox': 'Central Standard Time',
|
||||
'America/Indiana/Tell_City': 'Central Standard Time',
|
||||
'America/Menominee': 'Central Standard Time',
|
||||
'America/North_Dakota/Beulah': 'Central Standard Time',
|
||||
'America/North_Dakota/Center': 'Central Standard Time',
|
||||
'America/North_Dakota/New_Salem': 'Central Standard Time',
|
||||
'America/Chihuahua': 'Mountain Standard Time (Mexico)',
|
||||
'America/Mazatlan': 'Mountain Standard Time (Mexico)',
|
||||
'America/Coral_Harbour': 'SA Pacific Standard Time',
|
||||
'America/Costa_Rica': 'Central America Standard Time',
|
||||
'America/Cuiaba': 'Central Brazilian Standard Time',
|
||||
'America/Campo_Grande': 'Central Brazilian Standard Time',
|
||||
'America/Curacao': 'SA Western Standard Time',
|
||||
'America/Danmarkshavn': 'UTC',
|
||||
'America/Dawson_Creek': 'US Mountain Standard Time',
|
||||
'America/Creston': 'US Mountain Standard Time',
|
||||
'America/Denver': 'Mountain Standard Time',
|
||||
'America/Boise': 'Mountain Standard Time',
|
||||
'America/Shiprock': 'Mountain Standard Time',
|
||||
'America/Dominica': 'SA Western Standard Time',
|
||||
'America/Edmonton': 'Mountain Standard Time',
|
||||
'America/Cambridge_Bay': 'Mountain Standard Time',
|
||||
'America/Inuvik': 'Mountain Standard Time',
|
||||
'America/Yellowknife': 'Mountain Standard Time',
|
||||
'America/El_Salvador': 'Central America Standard Time',
|
||||
'America/Fortaleza': 'SA Eastern Standard Time',
|
||||
'America/Belem': 'SA Eastern Standard Time',
|
||||
'America/Maceio': 'SA Eastern Standard Time',
|
||||
'America/Recife': 'SA Eastern Standard Time',
|
||||
'America/Santarem': 'SA Eastern Standard Time',
|
||||
'America/Godthab': 'Greenland Standard Time',
|
||||
'America/Grand_Turk': 'Eastern Standard Time',
|
||||
'America/Grenada': 'SA Western Standard Time',
|
||||
'America/Guadeloupe': 'SA Western Standard Time',
|
||||
'America/Guatemala': 'Central America Standard Time',
|
||||
'America/Guayaquil': 'SA Pacific Standard Time',
|
||||
'America/Guyana': 'SA Western Standard Time',
|
||||
'America/Halifax': 'Atlantic Standard Time',
|
||||
'America/Glace_Bay': 'Atlantic Standard Time',
|
||||
'America/Goose_Bay': 'Atlantic Standard Time',
|
||||
'America/Moncton': 'Atlantic Standard Time',
|
||||
'America/Hermosillo': 'US Mountain Standard Time',
|
||||
'America/Indianapolis': 'US Eastern Standard Time',
|
||||
'America/Indiana/Marengo': 'US Eastern Standard Time',
|
||||
'America/Indiana/Vevay': 'US Eastern Standard Time',
|
||||
'America/Jamaica': 'SA Pacific Standard Time',
|
||||
'America/Kralendijk': 'SA Western Standard Time',
|
||||
'America/La_Paz': 'SA Western Standard Time',
|
||||
'America/Lima': 'SA Pacific Standard Time',
|
||||
'America/Los_Angeles': 'Pacific Standard Time',
|
||||
'America/Lower_Princes': 'SA Western Standard Time',
|
||||
'America/Managua': 'Central America Standard Time',
|
||||
'America/Manaus': 'SA Western Standard Time',
|
||||
'America/Boa_Vista': 'SA Western Standard Time',
|
||||
'America/Eirunepe': 'SA Western Standard Time',
|
||||
'America/Porto_Velho': 'SA Western Standard Time',
|
||||
'America/Rio_Branco': 'SA Western Standard Time',
|
||||
'America/Marigot': 'SA Western Standard Time',
|
||||
'America/Martinique': 'SA Western Standard Time',
|
||||
'America/Matamoros': 'Central Standard Time',
|
||||
'America/Mexico_City': 'Central Standard Time (Mexico)',
|
||||
'America/Bahia_Banderas': 'Central Standard Time (Mexico)',
|
||||
'America/Cancun': 'Central Standard Time (Mexico)',
|
||||
'America/Merida': 'Central Standard Time (Mexico)',
|
||||
'America/Monterrey': 'Central Standard Time (Mexico)',
|
||||
'America/Montevideo': 'Montevideo Standard Time',
|
||||
'America/Montserrat': 'SA Western Standard Time',
|
||||
'America/Nassau': 'Eastern Standard Time',
|
||||
'America/New_York': 'Eastern Standard Time',
|
||||
'America/Detroit': 'Eastern Standard Time',
|
||||
'America/Indiana/Petersburg': 'Eastern Standard Time',
|
||||
'America/Indiana/Vincennes': 'Eastern Standard Time',
|
||||
'America/Indiana/Winamac': 'Eastern Standard Time',
|
||||
'America/Kentucky/Monticello': 'Eastern Standard Time',
|
||||
'America/Louisville': 'Eastern Standard Time',
|
||||
'America/Noronha': 'UTC-02',
|
||||
'America/Ojinaga': 'Mountain Standard Time',
|
||||
'America/Panama': 'SA Pacific Standard Time',
|
||||
'America/Paramaribo': 'SA Eastern Standard Time',
|
||||
'America/Phoenix': 'US Mountain Standard Time',
|
||||
'America/Port-au-Prince': 'SA Pacific Standard Time',
|
||||
'America/Port_of_Spain': 'SA Western Standard Time',
|
||||
'America/Puerto_Rico': 'SA Western Standard Time',
|
||||
'America/Regina': 'Canada Central Standard Time',
|
||||
'America/Swift_Current': 'Canada Central Standard Time',
|
||||
'America/Santa_Isabel': 'Pacific Standard Time (Mexico)',
|
||||
'America/Santiago': 'Pacific SA Standard Time',
|
||||
'America/Santo_Domingo': 'SA Western Standard Time',
|
||||
'America/Sao_Paulo': 'E. South America Standard Time',
|
||||
'America/Araguaina': 'E. South America Standard Time',
|
||||
'America/Scoresbysund': 'Azores Standard Time',
|
||||
'America/St_Barthelemy': 'SA Western Standard Time',
|
||||
'America/St_Johns': 'Newfoundland Standard Time',
|
||||
'America/St_Kitts': 'SA Western Standard Time',
|
||||
'America/St_Lucia': 'SA Western Standard Time',
|
||||
'America/St_Thomas': 'SA Western Standard Time',
|
||||
'America/St_Vincent': 'SA Western Standard Time',
|
||||
'America/Tegucigalpa': 'Central America Standard Time',
|
||||
'America/Thule': 'Atlantic Standard Time',
|
||||
'America/Tijuana': 'Pacific Standard Time',
|
||||
'America/Toronto': 'Eastern Standard Time',
|
||||
'America/Iqaluit': 'Eastern Standard Time',
|
||||
'America/Montreal': 'Eastern Standard Time',
|
||||
'America/Nipigon': 'Eastern Standard Time',
|
||||
'America/Pangnirtung': 'Eastern Standard Time',
|
||||
'America/Thunder_Bay': 'Eastern Standard Time',
|
||||
'America/Tortola': 'SA Western Standard Time',
|
||||
'America/Whitehorse': 'Pacific Standard Time',
|
||||
'America/Vancouver': 'Pacific Standard Time',
|
||||
'America/Dawson': 'Pacific Standard Time',
|
||||
'America/Winnipeg': 'Central Standard Time',
|
||||
'America/Rainy_River': 'Central Standard Time',
|
||||
'America/Rankin_Inlet': 'Central Standard Time',
|
||||
'America/Resolute': 'Central Standard Time',
|
||||
'Antarctica/Casey': 'W. Australia Standard Time',
|
||||
'Antarctica/Davis': 'SE Asia Standard Time',
|
||||
'Antarctica/DumontDUrville': 'West Pacific Standard Time',
|
||||
'Antarctica/Macquarie': 'Central Pacific Standard Time',
|
||||
'Antarctica/Mawson': 'West Asia Standard Time',
|
||||
'Antarctica/Palmer': 'Pacific SA Standard Time',
|
||||
'Antarctica/Rothera': 'SA Eastern Standard Time',
|
||||
'Antarctica/South_Pole': 'New Zealand Standard Time',
|
||||
'Antarctica/McMurdo': 'New Zealand Standard Time',
|
||||
'Antarctica/Syowa': 'E. Africa Standard Time',
|
||||
'Antarctica/Vostok': 'Central Asia Standard Time',
|
||||
'Arctic/Longyearbyen': 'W. Europe Standard Time',
|
||||
'Asia/Aden': 'Arab Standard Time',
|
||||
'Asia/Almaty': 'Central Asia Standard Time',
|
||||
'Asia/Qyzylorda': 'Central Asia Standard Time',
|
||||
'Asia/Amman': 'Jordan Standard Time',
|
||||
'Asia/Ashgabat': 'West Asia Standard Time',
|
||||
'Asia/Baghdad': 'Arabic Standard Time',
|
||||
'Asia/Bahrain': 'Arab Standard Time',
|
||||
'Asia/Baku': 'Azerbaijan Standard Time',
|
||||
'Asia/Bangkok': 'SE Asia Standard Time',
|
||||
'Asia/Beirut': 'Middle East Standard Time',
|
||||
'Asia/Bishkek': 'Central Asia Standard Time',
|
||||
'Asia/Brunei': 'Singapore Standard Time',
|
||||
'Asia/Calcutta': 'India Standard Time',
|
||||
'Asia/Colombo': 'Sri Lanka Standard Time',
|
||||
'Asia/Damascus': 'Syria Standard Time',
|
||||
'Asia/Dhaka': 'Bangladesh Standard Time',
|
||||
'Asia/Dili': 'Tokyo Standard Time',
|
||||
'Asia/Dubai': 'Arabian Standard Time',
|
||||
'Asia/Dushanbe': 'West Asia Standard Time',
|
||||
'Asia/Gaza': 'Egypt Standard Time',
|
||||
'Asia/Hebron': 'Egypt Standard Time',
|
||||
'Asia/Hong_Kong': 'China Standard Time',
|
||||
'Asia/Hovd': 'SE Asia Standard Time',
|
||||
'Asia/Irkutsk': 'North Asia East Standard Time',
|
||||
'Asia/Jakarta': 'SE Asia Standard Time',
|
||||
'Asia/Pontianak': 'SE Asia Standard Time',
|
||||
'Asia/Jayapura': 'Tokyo Standard Time',
|
||||
'Asia/Jerusalem': 'Israel Standard Time',
|
||||
'Asia/Kabul': 'Afghanistan Standard Time',
|
||||
'Asia/Karachi': 'Pakistan Standard Time',
|
||||
'Asia/Kathmandu': 'Nepal Standard Time',
|
||||
'Asia/Katmandu': 'Nepal Standard Time',
|
||||
'Asia/Krasnoyarsk': 'North Asia Standard Time',
|
||||
'Asia/Kuala_Lumpur': 'Singapore Standard Time',
|
||||
'Asia/Kuching': 'Singapore Standard Time',
|
||||
'Asia/Kuwait': 'Arab Standard Time',
|
||||
'Asia/Macau': 'China Standard Time',
|
||||
'Asia/Magadan': 'Magadan Standard Time',
|
||||
'Asia/Anadyr Asia/Kamchatka': 'Magadan Standard Time',
|
||||
'Asia/Kamchatka': 'Magadan Standard Time',
|
||||
'Asia/Makassar': 'Singapore Standard Time',
|
||||
'Asia/Manila': 'Singapore Standard Time',
|
||||
'Asia/Muscat': 'Arabian Standard Time',
|
||||
'Asia/Nicosia': 'E. Europe Standard Time',
|
||||
'Asia/Novosibirsk': 'N. Central Asia Standard Time',
|
||||
'Asia/Novokuznetsk': 'N. Central Asia Standard Time',
|
||||
'Asia/Omsk': 'N. Central Asia Standard Time',
|
||||
'Asia/Oral': 'West Asia Standard Time',
|
||||
'Asia/Aqtau': 'West Asia Standard Time',
|
||||
'Asia/Aqtobe': 'West Asia Standard Time',
|
||||
'Asia/Phnom_Penh': 'SE Asia Standard Time',
|
||||
'Asia/Pyongyang': 'Korea Standard Time',
|
||||
'Asia/Qatar': 'Arab Standard Time',
|
||||
'Asia/Rangoon': 'Myanmar Standard Time',
|
||||
'Asia/Riyadh': 'Arab Standard Time',
|
||||
'Asia/Saigon': 'SE Asia Standard Time',
|
||||
'Asia/Seoul': 'Korea Standard Time',
|
||||
'Asia/Shanghai': 'China Standard Time',
|
||||
'Asia/Chongqing': 'China Standard Time',
|
||||
'Asia/Harbin': 'China Standard Time',
|
||||
'Asia/Kashgar': 'China Standard Time',
|
||||
'Asia/Urumqi': 'China Standard Time',
|
||||
'Asia/Singapore': 'Singapore Standard Time',
|
||||
'Asia/Taipei': 'Taipei Standard Time',
|
||||
'Asia/Tashkent': 'West Asia Standard Time',
|
||||
'Asia/Samarkand': 'West Asia Standard Time',
|
||||
'Asia/Tbilisi': 'Georgian Standard Time',
|
||||
'Asia/Tehran': 'Iran Standard Time',
|
||||
'Asia/Thimphu': 'Bangladesh Standard Time',
|
||||
'Asia/Tokyo': 'Tokyo Standard Time',
|
||||
'Asia/Ulaanbaatar': 'Ulaanbaatar Standard Time',
|
||||
'Asia/Choibalsan': 'Ulaanbaatar Standard Time',
|
||||
'Asia/Vientiane': 'SE Asia Standard Time',
|
||||
'Asia/Vladivostok': 'Vladivostok Standard Time',
|
||||
'Asia/Ust-Nera': 'Vladivostok Standard Time',
|
||||
'Asia/Sakhalin': 'Vladivostok Standard Time',
|
||||
'Asia/Yakutsk': 'Yakutsk Standard Time',
|
||||
'Asia/Khandyga': 'Yakutsk Standard Time',
|
||||
'Asia/Yekaterinburg': 'Ekaterinburg Standard Time',
|
||||
'Asia/Yerevan': 'Caucasus Standard Time',
|
||||
'Atlantic/Azores': 'Azores Standard Time',
|
||||
'Atlantic/Bermuda': 'Atlantic Standard Time',
|
||||
'Atlantic/Canary': 'GMT Standard Time',
|
||||
'Atlantic/Cape_Verde': 'Cape Verde Standard Time',
|
||||
'Atlantic/Faeroe': 'GMT Standard Time',
|
||||
'Atlantic/Reykjavik': 'Greenwich Standard Time',
|
||||
'Atlantic/South_Georgia': 'UTC-02',
|
||||
'Atlantic/St_Helena': 'Greenwich Standard Time',
|
||||
'Atlantic/Stanley': 'SA Eastern Standard Time',
|
||||
'Australia/Adelaide': 'Cen. Australia Standard Time',
|
||||
'Australia/Broken_Hill': 'Cen. Australia Standard Time',
|
||||
'Australia/Brisbane': 'E. Australia Standard Time',
|
||||
'Australia/Lindeman': 'E. Australia Standard Time',
|
||||
'Australia/Darwin': 'AUS Central Standard Time',
|
||||
'Australia/Hobart': 'Tasmania Standard Time',
|
||||
'Australia/Currie': 'Tasmania Standard Time',
|
||||
'Australia/Perth': 'W. Australia Standard Time',
|
||||
'Australia/Sydney': 'AUS Eastern Standard Time',
|
||||
'Australia/Melbourne': 'AUS Eastern Standard Time',
|
||||
'CST6CDT': 'Central Standard Time',
|
||||
'EST5EDT': 'Eastern Standard Time',
|
||||
'Etc/UTC': 'UTC',
|
||||
'Etc/GMT': 'UTC',
|
||||
'Etc/GMT+1': 'Cape Verde Standard Time',
|
||||
'Etc/GMT+10': 'Hawaiian Standard Time',
|
||||
'Etc/GMT+11': 'UTC-11',
|
||||
'Etc/GMT+12': 'Dateline Standard Time',
|
||||
'Etc/GMT+2': 'UTC-02',
|
||||
'Etc/GMT+3': 'SA Eastern Standard Time',
|
||||
'Etc/GMT+4': 'SA Western Standard Time',
|
||||
'Etc/GMT+5': 'SA Pacific Standard Time',
|
||||
'Etc/GMT+6': 'Central America Standard Time',
|
||||
'Etc/GMT+7': 'US Mountain Standard Time',
|
||||
'Etc/GMT-1': 'W. Central Africa Standard Time',
|
||||
'Etc/GMT-10': 'West Pacific Standard Time',
|
||||
'Etc/GMT-11': 'Central Pacific Standard Time',
|
||||
'Etc/GMT-12': 'UTC+12',
|
||||
'Etc/GMT-13': 'Tonga Standard Time',
|
||||
'Etc/GMT-2': 'South Africa Standard Time',
|
||||
'Etc/GMT-3': 'E. Africa Standard Time',
|
||||
'Etc/GMT-4': 'Arabian Standard Time',
|
||||
'Etc/GMT-5': 'West Asia Standard Time',
|
||||
'Etc/GMT-6': 'Central Asia Standard Time',
|
||||
'Etc/GMT-7': 'SE Asia Standard Time',
|
||||
'Etc/GMT-8': 'Singapore Standard Time',
|
||||
'Etc/GMT-9': 'Tokyo Standard Time',
|
||||
'Europe/Amsterdam': 'W. Europe Standard Time',
|
||||
'Europe/Andorra': 'W. Europe Standard Time',
|
||||
'Europe/Athens': 'GTB Standard Time',
|
||||
'Europe/Belgrade': 'Central Europe Standard Time',
|
||||
'Europe/Berlin': 'W. Europe Standard Time',
|
||||
'Europe/Busingen': 'W. Europe Standard Time',
|
||||
'Europe/Bratislava': 'Central Europe Standard Time',
|
||||
'Europe/Brussels': 'Romance Standard Time',
|
||||
'Europe/Bucharest': 'GTB Standard Time',
|
||||
'Europe/Budapest': 'Central Europe Standard Time',
|
||||
'Europe/Chisinau': 'GTB Standard Time',
|
||||
'Europe/Copenhagen': 'Romance Standard Time',
|
||||
'Europe/Dublin': 'GMT Standard Time',
|
||||
'Europe/Gibraltar': 'W. Europe Standard Time',
|
||||
'Europe/Guernsey': 'GMT Standard Time',
|
||||
'Europe/Helsinki': 'FLE Standard Time',
|
||||
'Europe/Isle_of_Man': 'GMT Standard Time',
|
||||
'Europe/Istanbul': 'Turkey Standard Time',
|
||||
'Europe/Jersey': 'GMT Standard Time',
|
||||
'Europe/Kaliningrad': 'Kaliningrad Standard Time',
|
||||
'Europe/Kiev': 'FLE Standard Time',
|
||||
'Europe/Simferopol': 'FLE Standard Time',
|
||||
'Europe/Uzhgorod': 'FLE Standard Time',
|
||||
'Europe/Zaporozhye': 'FLE Standard Time',
|
||||
'Europe/Lisbon': 'GMT Standard Time',
|
||||
'Atlantic/Madeira': 'GMT Standard Time',
|
||||
'Europe/Ljubljana': 'Central Europe Standard Time',
|
||||
'Europe/London': 'GMT Standard Time',
|
||||
'Europe/Luxembourg': 'W. Europe Standard Time',
|
||||
'Europe/Madrid': 'Romance Standard Time',
|
||||
'Africa/Ceuta': 'Romance Standard Time',
|
||||
'Europe/Malta': 'W. Europe Standard Time',
|
||||
'Europe/Mariehamn': 'FLE Standard Time',
|
||||
'Europe/Minsk': 'Kaliningrad Standard Time',
|
||||
'Europe/Monaco': 'W. Europe Standard Time',
|
||||
'Europe/Moscow': 'Russian Standard Time',
|
||||
'Europe/Volgograd': 'Russian Standard Time',
|
||||
'Europe/Samara': 'Russian Standard Time',
|
||||
'Europe/Oslo': 'W. Europe Standard Time',
|
||||
'Europe/Paris': 'Romance Standard Time',
|
||||
'Europe/Podgorica': 'Central Europe Standard Time',
|
||||
'Europe/Prague': 'Central Europe Standard Time',
|
||||
'Europe/Riga': 'FLE Standard Time',
|
||||
'Europe/Rome': 'W. Europe Standard Time',
|
||||
'Europe/San_Marino': 'W. Europe Standard Time',
|
||||
'Europe/Sarajevo': 'Central European Standard Time',
|
||||
'Europe/Skopje': 'Central European Standard Time',
|
||||
'Europe/Sofia': 'FLE Standard Time',
|
||||
'Europe/Stockholm': 'W. Europe Standard Time',
|
||||
'Europe/Tallinn': 'FLE Standard Time',
|
||||
'Europe/Tirane': 'Central Europe Standard Time',
|
||||
'Europe/Vaduz': 'W. Europe Standard Time',
|
||||
'Europe/Vatican': 'W. Europe Standard Time',
|
||||
'Europe/Vienna': 'W. Europe Standard Time',
|
||||
'Europe/Vilnius': 'FLE Standard Time',
|
||||
'Europe/Warsaw': 'Central European Standard Time',
|
||||
'Europe/Zagreb': 'Central European Standard Time',
|
||||
'Europe/Zurich': 'W. Europe Standard Time',
|
||||
'Indian/Antananarivo': 'E. Africa Standard Time',
|
||||
'Indian/Chagos': 'Central Asia Standard Time',
|
||||
'Indian/Christmas': 'SE Asia Standard Time',
|
||||
'Indian/Cocos': 'Myanmar Standard Time',
|
||||
'Indian/Comoro': 'E. Africa Standard Time',
|
||||
'Indian/Kerguelen': 'West Asia Standard Time',
|
||||
'Indian/Mahe': 'Mauritius Standard Time',
|
||||
'Indian/Maldives': 'West Asia Standard Time',
|
||||
'Indian/Mauritius': 'Mauritius Standard Time',
|
||||
'Indian/Mayotte': 'E. Africa Standard Time',
|
||||
'Indian/Reunion': 'Mauritius Standard Time',
|
||||
'MST7MDT': 'Mountain Standard Time',
|
||||
'PST8PDT': 'Pacific Standard Time',
|
||||
'Pacific/Apia': 'Samoa Standard Time',
|
||||
'Pacific/Auckland': 'New Zealand Standard Time',
|
||||
'Pacific/Efate': 'Central Pacific Standard Time',
|
||||
'Pacific/Enderbury': 'Tonga Standard Time',
|
||||
'Pacific/Fakaofo': 'Tonga Standard Time',
|
||||
'Pacific/Fiji': 'Fiji Standard Time',
|
||||
'Pacific/Funafuti': 'UTC+12',
|
||||
'Pacific/Galapagos': 'Central America Standard Time',
|
||||
'Pacific/Guadalcanal': 'Central Pacific Standard Time',
|
||||
'Pacific/Guam': 'West Pacific Standard Time',
|
||||
'Pacific/Honolulu': 'Hawaiian Standard Time',
|
||||
'Pacific/Johnston': 'Hawaiian Standard Time',
|
||||
'Pacific/Majuro Pacific/Kwajalein': 'UTC+12',
|
||||
'Pacific/Midway': 'UTC-11',
|
||||
'Pacific/Nauru': 'UTC+12',
|
||||
'Pacific/Niue': 'UTC-11',
|
||||
'Pacific/Noumea': 'Central Pacific Standard Time',
|
||||
'Pacific/Pago_Pago': 'UTC-11',
|
||||
'Pacific/Palau': 'Tokyo Standard Time',
|
||||
'Pacific/Ponape': 'Central Pacific Standard Time',
|
||||
'Pacific/Kosrae': 'Central Pacific Standard Time',
|
||||
'Pacific/Port_Moresby': 'West Pacific Standard Time',
|
||||
'Pacific/Rarotonga': 'Hawaiian Standard Time',
|
||||
'Pacific/Saipan': 'West Pacific Standard Time',
|
||||
'Pacific/Tahiti': 'Hawaiian Standard Time',
|
||||
'Pacific/Tarawa': 'UTC+12',
|
||||
'Pacific/Tongatapu': 'Tonga Standard Time',
|
||||
'Pacific/Truk': 'West Pacific Standard Time',
|
||||
'Pacific/Wake': 'UTC+12',
|
||||
'Pacific/Wallis': 'UTC+12'
|
||||
}
|
||||
|
||||
# Define the module's virtual name
|
||||
__virtualname__ = 'timezone'
|
||||
|
||||
|
||||
class TzMapper(object):
|
||||
def __init__(self, unix_to_win):
|
||||
self.win_to_unix = {k.lower(): v for k, v in unix_to_win.items()}
|
||||
self.unix_to_win = {v.lower(): k for k, v in unix_to_win.items()}
|
||||
|
||||
def add(self, k, v):
|
||||
self.unix_to_win[k.lower()] = v
|
||||
self.win_to_unix[v.lower()] = k
|
||||
|
||||
def remove(self, k):
|
||||
self.win_to_unix.pop(self.unix_to_win.pop(k.lower()).lower())
|
||||
|
||||
def get_win(self, key, default=None):
|
||||
return self.unix_to_win.get(key.lower(), default)
|
||||
|
||||
def get_unix(self, key, default=None):
|
||||
return self.win_to_unix.get(key.lower(), default)
|
||||
|
||||
def list_win(self):
|
||||
return sorted(self.unix_to_win.values())
|
||||
|
||||
def list_unix(self):
|
||||
return sorted(self.win_to_unix.values())
|
||||
|
||||
|
||||
mapper = TzMapper({
|
||||
'AUS Central Standard Time': 'Australia/Darwin',
|
||||
'AUS Eastern Standard Time': 'Australia/Sydney',
|
||||
'Afghanistan Standard Time': 'Asia/Kabul',
|
||||
'Alaskan Standard Time': 'America/Anchorage',
|
||||
'Aleutian Standard Time': 'America/Adak',
|
||||
'Altai Standard Time': 'Asia/Barnaul',
|
||||
'Arab Standard Time': 'Asia/Riyadh',
|
||||
'Arabian Standard Time': 'Asia/Dubai',
|
||||
'Arabic Standard Time': 'Asia/Baghdad',
|
||||
'Argentina Standard Time': 'America/Buenos_Aires',
|
||||
'Astrakhan Standard Time': 'Europe/Astrakhan',
|
||||
'Atlantic Standard Time': 'America/Halifax',
|
||||
'Aus Central W. Standard Time': 'Australia/Eucla',
|
||||
'Azerbaijan Standard Time': 'Asia/Baku',
|
||||
'Azores Standard Time': 'Atlantic/Azores',
|
||||
'Bahia Standard Time': 'America/Bahia',
|
||||
'Bangladesh Standard Time': 'Asia/Dhaka',
|
||||
'Belarus Standard Time': 'Europe/Minsk',
|
||||
'Bougainville Standard Time': 'Pacific/Bougainville',
|
||||
'Canada Central Standard Time': 'America/Regina',
|
||||
'Cape Verde Standard Time': 'Atlantic/Cape_Verde',
|
||||
'Caucasus Standard Time': 'Asia/Yerevan',
|
||||
'Cen. Australia Standard Time': 'Australia/Adelaide',
|
||||
'Central America Standard Time': 'America/Guatemala',
|
||||
'Central Asia Standard Time': 'Asia/Almaty',
|
||||
'Central Brazilian Standard Time': 'America/Cuiaba',
|
||||
'Central Europe Standard Time': 'Europe/Budapest',
|
||||
'Central European Standard Time': 'Europe/Warsaw',
|
||||
'Central Pacific Standard Time': 'Pacific/Guadalcanal',
|
||||
'Central Standard Time': 'America/Chicago',
|
||||
'Central Standard Time (Mexico)': 'America/Mexico_City',
|
||||
'Chatham Islands Standard Time': 'Pacific/Chatham',
|
||||
'China Standard Time': 'Asia/Shanghai',
|
||||
'Cuba Standard Time': 'America/Havana',
|
||||
'Dateline Standard Time': 'Etc/GMT+12',
|
||||
'E. Africa Standard Time': 'Africa/Nairobi',
|
||||
'E. Australia Standard Time': 'Australia/Brisbane',
|
||||
'E. Europe Standard Time': 'Europe/Chisinau',
|
||||
'E. South America Standard Time': 'America/Sao_Paulo',
|
||||
'Easter Island Standard Time': 'Pacific/Easter',
|
||||
'Eastern Standard Time': 'America/New_York',
|
||||
'Eastern Standard Time (Mexico)': 'America/Cancun',
|
||||
'Egypt Standard Time': 'Africa/Cairo',
|
||||
'Ekaterinburg Standard Time': 'Asia/Yekaterinburg',
|
||||
'FLE Standard Time': 'Europe/Kiev',
|
||||
'Fiji Standard Time': 'Pacific/Fiji',
|
||||
'GMT Standard Time': 'Europe/London',
|
||||
'GTB Standard Time': 'Europe/Bucharest',
|
||||
'Georgian Standard Time': 'Asia/Tbilisi',
|
||||
'Greenland Standard Time': 'America/Godthab',
|
||||
'Greenwich Standard Time': 'Atlantic/Reykjavik',
|
||||
'Haiti Standard Time': 'America/Port-au-Prince',
|
||||
'Hawaiian Standard Time': 'Pacific/Honolulu',
|
||||
'India Standard Time': 'Asia/Calcutta',
|
||||
'Iran Standard Time': 'Asia/Tehran',
|
||||
'Israel Standard Time': 'Asia/Jerusalem',
|
||||
'Jordan Standard Time': 'Asia/Amman',
|
||||
'Kaliningrad Standard Time': 'Europe/Kaliningrad',
|
||||
'Korea Standard Time': 'Asia/Seoul',
|
||||
'Libya Standard Time': 'Africa/Tripoli',
|
||||
'Line Islands Standard Time': 'Pacific/Kiritimati',
|
||||
'Lord Howe Standard Time': 'Australia/Lord_Howe',
|
||||
'Magadan Standard Time': 'Asia/Magadan',
|
||||
'Magallanes Standard Time': 'America/Punta_Arenas',
|
||||
'Marquesas Standard Time': 'Pacific/Marquesas',
|
||||
'Mauritius Standard Time': 'Indian/Mauritius',
|
||||
'Middle East Standard Time': 'Asia/Beirut',
|
||||
'Montevideo Standard Time': 'America/Montevideo',
|
||||
'Morocco Standard Time': 'Africa/Casablanca',
|
||||
'Mountain Standard Time': 'America/Denver',
|
||||
'Mountain Standard Time (Mexico)': 'America/Chihuahua',
|
||||
'Myanmar Standard Time': 'Asia/Rangoon',
|
||||
'N. Central Asia Standard Time': 'Asia/Novosibirsk',
|
||||
'Namibia Standard Time': 'Africa/Windhoek',
|
||||
'Nepal Standard Time': 'Asia/Katmandu',
|
||||
'New Zealand Standard Time': 'Pacific/Auckland',
|
||||
'Newfoundland Standard Time': 'America/St_Johns',
|
||||
'Norfolk Standard Time': 'Pacific/Norfolk',
|
||||
'North Asia East Standard Time': 'Asia/Irkutsk',
|
||||
'North Asia Standard Time': 'Asia/Krasnoyarsk',
|
||||
'North Korea Standard Time': 'Asia/Pyongyang',
|
||||
'Omsk Standard Time': 'Asia/Omsk',
|
||||
'Pacific SA Standard Time': 'America/Santiago',
|
||||
'Pacific Standard Time': 'America/Los_Angeles',
|
||||
'Pacific Standard Time (Mexico)': 'America/Tijuana',
|
||||
'Pakistan Standard Time': 'Asia/Karachi',
|
||||
'Paraguay Standard Time': 'America/Asuncion',
|
||||
'Romance Standard Time': 'Europe/Paris',
|
||||
'Russia Time Zone 10': 'Asia/Srednekolymsk',
|
||||
'Russia Time Zone 11': 'Asia/Kamchatka',
|
||||
'Russia Time Zone 3': 'Europe/Samara',
|
||||
'Russian Standard Time': 'Europe/Moscow',
|
||||
'SA Eastern Standard Time': 'America/Cayenne',
|
||||
'SA Pacific Standard Time': 'America/Bogota',
|
||||
'SA Western Standard Time': 'America/La_Paz',
|
||||
'SE Asia Standard Time': 'Asia/Bangkok',
|
||||
'Saint Pierre Standard Time': 'America/Miquelon',
|
||||
'Sakhalin Standard Time': 'Asia/Sakhalin',
|
||||
'Samoa Standard Time': 'Pacific/Apia',
|
||||
'Saratov Standard Time': 'Europe/Saratov',
|
||||
'Singapore Standard Time': 'Asia/Singapore',
|
||||
'South Africa Standard Time': 'Africa/Johannesburg',
|
||||
'Sri Lanka Standard Time': 'Asia/Colombo',
|
||||
'Syria Standard Time': 'Asia/Damascus',
|
||||
'Taipei Standard Time': 'Asia/Taipei',
|
||||
'Tasmania Standard Time': 'Australia/Hobart',
|
||||
'Tocantins Standard Time': 'America/Araguaina',
|
||||
'Tokyo Standard Time': 'Asia/Tokyo',
|
||||
'Tomsk Standard Time': 'Asia/Tomsk',
|
||||
'Tonga Standard Time': 'Pacific/Tongatapu',
|
||||
'Transbaikal Standard Time': 'Asia/Chita',
|
||||
'Turkey Standard Time': 'Europe/Istanbul',
|
||||
'Turks And Caicos Standard Time': 'America/Grand_Turk',
|
||||
'US Eastern Standard Time': 'America/Indianapolis',
|
||||
'US Mountain Standard Time': 'America/Phoenix',
|
||||
'UTC': 'Etc/GMT',
|
||||
'UTC+12': 'Etc/GMT-12',
|
||||
'UTC+13': 'Etc/GMT-13',
|
||||
'UTC-02': 'Etc/GMT+2',
|
||||
'UTC-08': 'Etc/GMT+8',
|
||||
'UTC-09': 'Etc/GMT+9',
|
||||
'UTC-11': 'Etc/GMT+11',
|
||||
'Ulaanbaatar Standard Time': 'Asia/Ulaanbaatar',
|
||||
'Venezuela Standard Time': 'America/Caracas',
|
||||
'Vladivostok Standard Time': 'Asia/Vladivostok',
|
||||
'W. Australia Standard Time': 'Australia/Perth',
|
||||
'W. Central Africa Standard Time': 'Africa/Lagos',
|
||||
'W. Europe Standard Time': 'Europe/Berlin',
|
||||
'W. Mongolia Standard Time': 'Asia/Hovd',
|
||||
'West Asia Standard Time': 'Asia/Tashkent',
|
||||
'West Bank Standard Time': 'Asia/Hebron',
|
||||
'West Pacific Standard Time': 'Pacific/Port_Moresby',
|
||||
'Yakutsk Standard Time': 'Asia/Yakutsk'})
|
||||
|
||||
|
||||
def __virtual__():
|
||||
'''
|
||||
Only load on windows
|
||||
'''
|
||||
if salt.utils.platform.is_windows() and salt.utils.path.which('tzutil'):
|
||||
return __virtualname__
|
||||
return (False, "Module win_timezone: tzutil not found or is not on Windows client")
|
||||
if not __utils__['platform.is_windows']():
|
||||
return False, "Module win_timezone: Not on Windows client"
|
||||
if not HAS_PYTZ:
|
||||
return False, "Module win_timezone: pytz not found"
|
||||
if not __utils__['path.which']('tzutil'):
|
||||
return False, "Module win_timezone: tzutil not found"
|
||||
return __virtualname__
|
||||
|
||||
|
||||
def get_zone():
|
||||
'''
|
||||
Get current timezone (i.e. America/Denver)
|
||||
|
||||
Returns:
|
||||
str: Timezone in unix format
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' timezone.get_zone
|
||||
'''
|
||||
winzone = __salt__['cmd.run'](['tzutil', '/g'], python_shell=False)
|
||||
for key in LINTOWIN:
|
||||
if LINTOWIN[key] == winzone:
|
||||
return key
|
||||
|
||||
return False
|
||||
win_zone = __utils__['reg.read_value'](
|
||||
hive='HKLM',
|
||||
key='SYSTEM\\CurrentControlSet\\Control\\TimeZoneInformation',
|
||||
vname='TimeZoneKeyName')['vdata']
|
||||
return mapper.get_unix(win_zone.lower(), 'Unknown')
|
||||
|
||||
|
||||
def get_offset():
|
||||
'''
|
||||
Get current numeric timezone offset from UCT (i.e. -0700)
|
||||
Get current numeric timezone offset from UTC (i.e. -0700)
|
||||
|
||||
Returns:
|
||||
str: Offset from UTC
|
||||
|
||||
CLI Example:
|
||||
|
||||
@ -494,52 +231,44 @@ def get_offset():
|
||||
|
||||
salt '*' timezone.get_offset
|
||||
'''
|
||||
string = False
|
||||
zone = __salt__['cmd.run'](['tzutil', '/g'], python_shell=False)
|
||||
prev = ''
|
||||
zone_list = __salt__['cmd.run'](['tzutil', '/l'],
|
||||
python_shell=False,
|
||||
output_loglevel='trace').splitlines()
|
||||
for line in zone_list:
|
||||
if zone == line:
|
||||
string = prev
|
||||
break
|
||||
else:
|
||||
prev = line
|
||||
|
||||
if not string:
|
||||
return False
|
||||
|
||||
reg = re.search(r"\(UTC(.\d\d:\d\d)\) .*", string, re.M)
|
||||
if not reg:
|
||||
ret = '0000'
|
||||
else:
|
||||
ret = reg.group(1).replace(':', '')
|
||||
|
||||
return ret
|
||||
# http://craigglennie.com/programming/python/2013/07/21/working-with-timezones-using-Python-and-pytz-localize-vs-normalize/
|
||||
tz_object = pytz.timezone(get_zone())
|
||||
utc_time = pytz.utc.localize(datetime.utcnow())
|
||||
loc_time = utc_time.astimezone(tz_object)
|
||||
norm_time = tz_object.normalize(loc_time)
|
||||
return norm_time.strftime('%z')
|
||||
|
||||
|
||||
def get_zonecode():
|
||||
'''
|
||||
Get current timezone (i.e. PST, MDT, etc)
|
||||
|
||||
Returns:
|
||||
str: An abbreviated timezone code
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' timezone.get_zonecode
|
||||
'''
|
||||
# Still not implemented on windows
|
||||
return False
|
||||
tz_object = pytz.timezone(get_zone())
|
||||
loc_time = tz_object.localize(datetime.utcnow())
|
||||
return loc_time.tzname()
|
||||
|
||||
|
||||
def set_zone(timezone):
|
||||
'''
|
||||
Unlinks, then symlinks /etc/localtime to the set timezone.
|
||||
Sets the timezone using the tzutil.
|
||||
|
||||
The timezone is crucial to several system processes, each of which SHOULD
|
||||
be restarted (for instance, whatever you system uses as its cron and
|
||||
syslog daemons). This will not be magically done for you!
|
||||
Args:
|
||||
timezone (str): A valid timezone
|
||||
|
||||
Returns:
|
||||
bool: ``True`` if successful, otherwise ``False``
|
||||
|
||||
Raises:
|
||||
CommandExecutionError: If invalid timezone is passed
|
||||
|
||||
CLI Example:
|
||||
|
||||
@ -547,15 +276,40 @@ def set_zone(timezone):
|
||||
|
||||
salt '*' timezone.set_zone 'America/Denver'
|
||||
'''
|
||||
cmd = ['tzutil', '/s', LINTOWIN[timezone]]
|
||||
return __salt__['cmd.retcode'](cmd, python_shell=False) == 0
|
||||
# if it's one of the key's just use it
|
||||
if timezone.lower() in mapper.win_to_unix:
|
||||
win_zone = timezone
|
||||
|
||||
elif timezone.lower() in mapper.unix_to_win:
|
||||
# if it's one of the values, use the key
|
||||
win_zone = mapper.get_win(timezone)
|
||||
|
||||
else:
|
||||
# Raise error because it's neither key nor value
|
||||
raise CommandExecutionError('Invalid timezone passed: {0}'.format(timezone))
|
||||
|
||||
# Set the value
|
||||
cmd = ['tzutil', '/s', win_zone]
|
||||
res = __salt__['cmd.run_all'](cmd, python_shell=False)
|
||||
if res['retcode']:
|
||||
raise CommandExecutionError('tzutil encountered an error setting '
|
||||
'timezone: {0}'.format(timezone),
|
||||
info=res)
|
||||
return zone_compare(timezone)
|
||||
|
||||
|
||||
def zone_compare(timezone):
|
||||
'''
|
||||
Checks the md5sum between the given timezone, and the one set in
|
||||
/etc/localtime. Returns True if they match, and False if not. Mostly useful
|
||||
for running state checks.
|
||||
Compares the given timezone with the machine timezone. Mostly useful for
|
||||
running state checks.
|
||||
|
||||
Args:
|
||||
timezone (str):
|
||||
The timezone to compare. This can be in Windows or Unix format. Can
|
||||
be any of the values returned by the ``timezone.list`` function
|
||||
|
||||
Returns:
|
||||
bool: ``True`` if they match, otherwise ``False``
|
||||
|
||||
Example:
|
||||
|
||||
@ -563,21 +317,68 @@ def zone_compare(timezone):
|
||||
|
||||
salt '*' timezone.zone_compare 'America/Denver'
|
||||
'''
|
||||
cmd = ['tzutil', '/g']
|
||||
return __salt__['cmd.run'](cmd, python_shell=False) == LINTOWIN[timezone]
|
||||
# if it's one of the key's just use it
|
||||
if timezone.lower() in mapper.win_to_unix:
|
||||
check_zone = timezone
|
||||
|
||||
elif timezone.lower() in mapper.unix_to_win:
|
||||
# if it's one of the values, use the key
|
||||
check_zone = mapper.get_win(timezone)
|
||||
|
||||
else:
|
||||
# Raise error because it's neither key nor value
|
||||
raise CommandExecutionError('Invalid timezone passed: {0}'
|
||||
''.format(timezone))
|
||||
|
||||
return get_zone() == mapper.get_unix(check_zone, 'Unknown')
|
||||
|
||||
|
||||
def list(unix_style=True):
|
||||
'''
|
||||
Return a list of Timezones that this module supports. These can be in either
|
||||
Unix or Windows format.
|
||||
|
||||
.. version-added:: 2018.3.2
|
||||
|
||||
Args:
|
||||
unix_style (bool):
|
||||
``True`` returns Unix-style timezones. ``False`` returns
|
||||
Windows-style timezones. Default is ``True``
|
||||
|
||||
Returns:
|
||||
list: A list of supported timezones
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
# Unix-style timezones
|
||||
salt '*' timezone.list
|
||||
|
||||
# Windows-style timezones
|
||||
salt '*' timezone.list unix_style=False
|
||||
'''
|
||||
if unix_style:
|
||||
return mapper.list_unix()
|
||||
else:
|
||||
return mapper.list_win()
|
||||
|
||||
|
||||
def get_hwclock():
|
||||
'''
|
||||
Get current hardware clock setting (UTC or localtime)
|
||||
|
||||
.. note::
|
||||
The hardware clock is always local time on Windows so this will always
|
||||
return "localtime"
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' timezone.get_hwclock
|
||||
'''
|
||||
# Need to search for a way to figure it out ...
|
||||
# The hardware clock is always localtime on Windows
|
||||
return 'localtime'
|
||||
|
||||
|
||||
@ -585,11 +386,15 @@ def set_hwclock(clock):
|
||||
'''
|
||||
Sets the hardware clock to be either UTC or localtime
|
||||
|
||||
.. note::
|
||||
The hardware clock is always local time on Windows so this will always
|
||||
return ``False``
|
||||
|
||||
CLI Example:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
salt '*' timezone.set_hwclock UTC
|
||||
'''
|
||||
# Need to search for a way to figure it out ...
|
||||
# The hardware clock is always localtime on Windows
|
||||
return False
|
||||
|
@ -92,6 +92,7 @@ def _strip_headers(output, *args):
|
||||
if not args:
|
||||
args_lc = ('installed packages',
|
||||
'available packages',
|
||||
'available upgrades',
|
||||
'updated packages',
|
||||
'upgraded packages')
|
||||
else:
|
||||
|
@ -470,14 +470,13 @@ def get(zpool, prop=None, show_source=False, parsable=True):
|
||||
|
||||
'''
|
||||
ret = OrderedDict()
|
||||
value_properties = ['property', 'value', 'source']
|
||||
value_properties = ['name', 'property', 'value', 'source']
|
||||
|
||||
## collect get output
|
||||
res = __salt__['cmd.run_all'](
|
||||
__utils__['zfs.zpool_command'](
|
||||
command='get',
|
||||
flags=['-H'],
|
||||
opts={'-o': ','.join(value_properties)},
|
||||
property_name=prop if prop else 'all',
|
||||
target=zpool,
|
||||
),
|
||||
@ -503,6 +502,9 @@ def get(zpool, prop=None, show_source=False, parsable=True):
|
||||
[x for x in line.strip().split('\t') if x not in ['']],
|
||||
)))
|
||||
|
||||
# NOTE: older zfs does not have -o, fall back to manually stipping the name field
|
||||
del prop_data['name']
|
||||
|
||||
# NOTE: normalize values
|
||||
if parsable:
|
||||
# NOTE: raw numbers and pythonic types
|
||||
@ -634,6 +636,8 @@ def scrub(zpool, stop=False, pause=False):
|
||||
|
||||
.. note::
|
||||
|
||||
Pause is only available on recent versions of ZFS.
|
||||
|
||||
If both ``pause`` and ``stop`` are ``True``, then ``stop`` will
|
||||
win.
|
||||
|
||||
@ -706,6 +710,9 @@ def create(zpool, *vdevs, **kwargs):
|
||||
|
||||
.. versionadded:: 2018.3.0
|
||||
|
||||
.. warning:
|
||||
This is only available on illumos and Solaris
|
||||
|
||||
CLI Examples:
|
||||
|
||||
.. code-block:: bash
|
||||
|
@ -853,7 +853,7 @@ def _get_configured_repos():
|
||||
'''
|
||||
|
||||
repos_cfg = configparser.ConfigParser()
|
||||
repos_cfg.read([REPOS + '/' + fname for fname in os.listdir(REPOS)])
|
||||
repos_cfg.read([REPOS + '/' + fname for fname in os.listdir(REPOS) if fname.endswith(".repo")])
|
||||
|
||||
return repos_cfg
|
||||
|
||||
|
@ -6,8 +6,10 @@ for managing outputters.
|
||||
|
||||
# Import Python libs
|
||||
from __future__ import absolute_import, print_function, unicode_literals
|
||||
|
||||
import errno
|
||||
import logging
|
||||
import io
|
||||
import os
|
||||
import re
|
||||
import sys
|
||||
@ -168,7 +170,7 @@ def get_printout(out, opts=None, **kwargs):
|
||||
'''
|
||||
try:
|
||||
fileno = sys.stdout.fileno()
|
||||
except AttributeError:
|
||||
except (AttributeError, io.UnsupportedOperation):
|
||||
fileno = -1 # sys.stdout is StringIO or fake
|
||||
return not os.isatty(fileno)
|
||||
|
||||
|
@ -43,6 +43,7 @@ from functools import reduce # pylint: disable=redefined-builtin
|
||||
# Import Salt libs
|
||||
import salt.output
|
||||
import salt.utils.color
|
||||
import salt.utils.data
|
||||
import salt.utils.locales
|
||||
|
||||
# Import 3rd-party libs
|
||||
@ -62,9 +63,9 @@ class TableDisplay(object):
|
||||
'''
|
||||
|
||||
_JUSTIFY_MAP = {
|
||||
'center': str.center,
|
||||
'right': str.rjust,
|
||||
'left': str.ljust
|
||||
'center': six.text_type.center,
|
||||
'right': six.text_type.rjust,
|
||||
'left': six.text_type.ljust
|
||||
}
|
||||
|
||||
def __init__(self,
|
||||
@ -147,7 +148,7 @@ class TableDisplay(object):
|
||||
for item in row
|
||||
]
|
||||
rows = []
|
||||
for item in map(None, *new_rows):
|
||||
for item in map(lambda *args: args, *new_rows):
|
||||
if isinstance(item, (tuple, list)):
|
||||
rows.append([substr or '' for substr in item])
|
||||
else:
|
||||
@ -159,7 +160,7 @@ class TableDisplay(object):
|
||||
for row in rows
|
||||
]
|
||||
|
||||
columns = map(None, *reduce(operator.add, logical_rows))
|
||||
columns = map(lambda *args: args, *reduce(operator.add, logical_rows))
|
||||
|
||||
max_widths = [
|
||||
max([len(six.text_type(item)) for item in column])
|
||||
@ -363,7 +364,7 @@ def output(ret, **kwargs):
|
||||
)
|
||||
)
|
||||
|
||||
return '\n'.join(table.display(ret,
|
||||
return '\n'.join(table.display(salt.utils.data.decode(ret),
|
||||
base_indent,
|
||||
out,
|
||||
rows_key=rows_key,
|
||||
|
@ -1014,6 +1014,13 @@ class Pillar(object):
|
||||
mopts['file_roots'] = self.actual_file_roots
|
||||
mopts['saltversion'] = __version__
|
||||
pillar['master'] = mopts
|
||||
if 'pillar' in self.opts and self.opts.get('ssh_merge_pillar', False):
|
||||
pillar = merge(
|
||||
self.opts['pillar'],
|
||||
pillar,
|
||||
self.merge_strategy,
|
||||
self.opts.get('renderer', 'yaml'),
|
||||
self.opts.get('pillar_merge_lists', False))
|
||||
if errors:
|
||||
for error in errors:
|
||||
log.critical('Pillar render error: %s', error)
|
||||
|
@ -236,16 +236,7 @@ def get_grains():
|
||||
'''
|
||||
Retrieve facts from the network device.
|
||||
'''
|
||||
refresh_needed = False
|
||||
refresh_needed = refresh_needed or (not DETAILS.get('grains_cache', {}))
|
||||
refresh_needed = refresh_needed or (not DETAILS.get('grains_cache', {}).get('result', False))
|
||||
refresh_needed = refresh_needed or (not DETAILS.get('grains_cache', {}).get('out', {}))
|
||||
|
||||
if refresh_needed:
|
||||
facts = call('get_facts', **{})
|
||||
DETAILS['grains_cache'] = facts
|
||||
|
||||
return DETAILS.get('grains_cache', {})
|
||||
return call('get_facts', **{})
|
||||
|
||||
|
||||
def grains_refresh():
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user