diff --git "a/codeparrot-valid_1000.txt" "b/codeparrot-valid_1000.txt" new file mode 100644--- /dev/null +++ "b/codeparrot-valid_1000.txt" @@ -0,0 +1,10000 @@ + "__package__": mod_name.rpartition(".")[0], + }) + def create_ns(init_globals): + return run_module(mod_name, init_globals, run_name) + try: + self.check_code_execution(create_ns, expected_ns) + finally: + self._del_pkg(pkg_dir, depth, mod_name) + + +class RunPathTestCase(unittest.TestCase, CodeExecutionMixin): + """Unit tests for runpy.run_path""" + + def _make_test_script(self, script_dir, script_basename, source=None): + if source is None: + source = example_source + return make_script(script_dir, script_basename, source) + + def _check_script(self, script_name, expected_name, expected_file, + expected_argv0): + # First check is without run_name + def create_ns(init_globals): + return run_path(script_name, init_globals) + expected_ns = example_namespace.copy() + expected_ns.update({ + "__name__": expected_name, + "__file__": expected_file, + "__package__": "", + "run_argv0": expected_argv0, + "run_name_in_sys_modules": True, + "module_in_sys_modules": True, + }) + self.check_code_execution(create_ns, expected_ns) + # Second check makes sure run_name works in all cases + run_name = "prove.issue15230.is.fixed" + def create_ns(init_globals): + return run_path(script_name, init_globals, run_name) + expected_ns["__name__"] = run_name + expected_ns["__package__"] = run_name.rpartition(".")[0] + self.check_code_execution(create_ns, expected_ns) + + def _check_import_error(self, script_name, msg): + msg = re.escape(msg) + self.assertRaisesRegex(ImportError, msg, run_path, script_name) + + def test_basic_script(self): + with temp_dir() as script_dir: + mod_name = 'script' + script_name = self._make_test_script(script_dir, mod_name) + self._check_script(script_name, "", script_name, + script_name) + + def test_script_compiled(self): + with temp_dir() as script_dir: + mod_name = 'script' + script_name = self._make_test_script(script_dir, mod_name) + compiled_name = py_compile.compile(script_name, doraise=True) + os.remove(script_name) + self._check_script(compiled_name, "", compiled_name, + compiled_name) + + def test_directory(self): + with temp_dir() as script_dir: + mod_name = '__main__' + script_name = self._make_test_script(script_dir, mod_name) + self._check_script(script_dir, "", script_name, + script_dir) + + def test_directory_compiled(self): + with temp_dir() as script_dir: + mod_name = '__main__' + script_name = self._make_test_script(script_dir, mod_name) + compiled_name = py_compile.compile(script_name, doraise=True) + os.remove(script_name) + if not sys.dont_write_bytecode: + legacy_pyc = make_legacy_pyc(script_name) + self._check_script(script_dir, "", legacy_pyc, + script_dir) + + def test_directory_error(self): + with temp_dir() as script_dir: + mod_name = 'not_main' + script_name = self._make_test_script(script_dir, mod_name) + msg = "can't find '__main__' module in %r" % script_dir + self._check_import_error(script_dir, msg) + + def test_zipfile(self): + with temp_dir() as script_dir: + mod_name = '__main__' + script_name = self._make_test_script(script_dir, mod_name) + zip_name, fname = make_zip_script(script_dir, 'test_zip', script_name) + self._check_script(zip_name, "", fname, zip_name) + + def test_zipfile_compiled(self): + with temp_dir() as script_dir: + mod_name = '__main__' + script_name = self._make_test_script(script_dir, mod_name) + compiled_name = py_compile.compile(script_name, doraise=True) + zip_name, fname = make_zip_script(script_dir, 'test_zip', + compiled_name) + self._check_script(zip_name, "", fname, zip_name) + + def test_zipfile_error(self): + with temp_dir() as script_dir: + mod_name = 'not_main' + script_name = self._make_test_script(script_dir, mod_name) + zip_name, fname = make_zip_script(script_dir, 'test_zip', script_name) + msg = "can't find '__main__' module in %r" % zip_name + self._check_import_error(zip_name, msg) + + def test_main_recursion_error(self): + with temp_dir() as script_dir, temp_dir() as dummy_dir: + mod_name = '__main__' + source = ("import runpy\n" + "runpy.run_path(%r)\n") % dummy_dir + script_name = self._make_test_script(script_dir, mod_name, source) + zip_name, fname = make_zip_script(script_dir, 'test_zip', script_name) + msg = "recursion depth exceeded" + self.assertRaisesRegex(RuntimeError, msg, run_path, zip_name) + + def test_encoding(self): + with temp_dir() as script_dir: + filename = os.path.join(script_dir, 'script.py') + with open(filename, 'w', encoding='latin1') as f: + f.write(""" +#coding:latin1 +s = "non-ASCII: h\xe9" +""") + result = run_path(filename) + self.assertEqual(result['s'], "non-ASCII: h\xe9") + + +def test_main(): + run_unittest( + ExecutionLayerTestCase, + RunModuleTestCase, + RunPathTestCase + ) + +if __name__ == "__main__": + test_main() + +import sys + +import pytest + +from setuptools_scm.utils import do +from setuptools_scm import PRETEND_KEY, PRETEND_KEY_NAMED + + +@pytest.fixture +def wd(wd): + wd("git init") + wd("git config user.email test@example.com") + wd('git config user.name "a test"') + wd.add_command = "git add ." + wd.commit_command = "git commit -m test-{reason}" + return wd + + +def test_pyproject_support(tmpdir, monkeypatch): + pytest.importorskip("toml") + monkeypatch.delenv("SETUPTOOLS_SCM_DEBUG") + pkg = tmpdir.ensure("package", dir=42) + pkg.join("pyproject.toml").write( + """[tool.setuptools_scm] +fallback_version = "12.34" +""" + ) + pkg.join("setup.py").write("__import__('setuptools').setup()") + res = do((sys.executable, "setup.py", "--version"), pkg) + assert res == "12.34" + + +def test_pyproject_support_with_git(tmpdir, monkeypatch, wd): + pytest.importorskip("toml") + pkg = tmpdir.join("wd") + pkg.join("pyproject.toml").write("""[tool.setuptools_scm]""") + pkg.join("setup.py").write( + "__import__('setuptools').setup(name='setuptools_scm_example')" + ) + res = do((sys.executable, "setup.py", "--version"), pkg) + assert res.endswith("0.1.dev0") + + +def test_pretend_version(tmpdir, monkeypatch, wd): + monkeypatch.setenv(PRETEND_KEY, "1.0.0") + + assert wd.get_version() == "1.0.0" + assert wd.get_version(dist_name="ignored") == "1.0.0" + + +def test_pretend_version_named_pyproject_integration(tmpdir, monkeypatch, wd): + test_pyproject_support_with_git(tmpdir, monkeypatch, wd) + monkeypatch.setenv( + PRETEND_KEY_NAMED.format(name="setuptools_scm_example".upper()), "3.2.1" + ) + res = do((sys.executable, "setup.py", "--version"), tmpdir / "wd") + assert res.endswith("3.2.1") + + +def test_pretend_version_named(tmpdir, monkeypatch, wd): + monkeypatch.setenv(PRETEND_KEY_NAMED.format(name="test".upper()), "1.0.0") + monkeypatch.setenv(PRETEND_KEY_NAMED.format(name="test2".upper()), "2.0.0") + assert wd.get_version(dist_name="test") == "1.0.0" + assert wd.get_version(dist_name="test2") == "2.0.0" + + +def test_pretend_version_name_takes_precedence(tmpdir, monkeypatch, wd): + monkeypatch.setenv(PRETEND_KEY_NAMED.format(name="test".upper()), "1.0.0") + monkeypatch.setenv(PRETEND_KEY, "2.0.0") + assert wd.get_version(dist_name="test") == "1.0.0" + +# -*- coding: utf-8 -*- +## +## +## This file is part of Indico. +## Copyright (C) 2002 - 2014 European Organization for Nuclear Research (CERN). +## +## Indico is free software; you can redistribute it and/or +## modify it under the terms of the GNU General Public License as +## published by the Free Software Foundation; either version 3 of the +## License, or (at your option) any later version. +## +## Indico is distributed in the hope that it will be useful, but +## WITHOUT ANY WARRANTY; without even the implied warranty of +## MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +## General Public License for more details. +## +## You should have received a copy of the GNU General Public License +## along with Indico;if not, see . + +import copy +import os +import logging +import logging.handlers +import logging.config +import ConfigParser +from flask import request, session +from ZODB.POSException import POSError + +from indico.core.config import Config +from MaKaC.common.contextManager import ContextManager + + +class AddIDFilter(logging.Filter): + def filter(self, record): + if not logging.Filter.filter(self, record): + return False + # Add request ID if available + try: + record.request_id = request.id + except RuntimeError: + record.request_id = '0' * 12 + return True + + +class ExtraIndicoFilter(AddIDFilter): + def filter(self, record): + if record.name.split('.')[0] == 'indico': + return False + return AddIDFilter.filter(self, record) + + +class IndicoMailFormatter(logging.Formatter): + def format(self, record): + s = logging.Formatter.format(self, record) + if isinstance(s, unicode): + s = s.encode('utf-8') + return s + self._getRequestInfo() + + def _getRequestInfo(self): + rh = ContextManager.get('currentRH', None) + info = ['Additional information:'] + + try: + info.append('Request: %s' % request.id) + info.append('URL: %s' % request.url) + + if request.url_rule: + info.append('Endpoint: {0}'.format(request.url_rule.endpoint)) + + info.append('Method: %s' % request.method) + if rh: + info.append('Params: %s' % rh._getTruncatedParams()) + + if session: + try: + info.append('User: {0}'.format(session.user)) + except POSError: + # If the DB connection is closed getting the avatar may fail + info.append('User id: {0}'.format(session.get('_avatarId'))) + info.append('IP: %s' % request.remote_addr) + info.append('User Agent: %s' % request.user_agent) + info.append('Referer: %s' % (request.referrer or 'n/a')) + except RuntimeError, e: + info.append('Not available: %s' % e) + return '\n\n%s' % '\n'.join(x.encode('utf-8') if isinstance(x, unicode) else x for x in info) + + +class LoggerUtils: + + @classmethod + def _bootstrap_cp(cls, cp, defaultArgs): + """ + Creates a very basic logging config for cases in which + logging.conf does not yet exist + """ + if not cp.has_section('loggers'): + cp.add_section('loggers') + cp.add_section('logger_root') + cp.add_section('handlers') + cp.set('loggers', 'keys', 'root') + cp.set('logger_root', 'handlers', ','.join(defaultArgs)) + cp.set('handlers', 'keys', ','.join(defaultArgs)) + for handler_name in defaultArgs: + section_name = 'handler_' + handler_name + cp.add_section(section_name) + cp.set(section_name, 'formatter', 'defaultFormatter') + + @classmethod + def configFromFile(cls, fname, defaultArgs, filters): + """ + Read the logging configuration from the logging.conf file. + Fetch default values if the logging.conf file is not set. + """ + cp = ConfigParser.ConfigParser() + parsed_files = cp.read(fname) + + if cp.has_section('formatters'): + formatters = logging.config._create_formatters(cp) + else: + formatters = {} + + # Really ugly.. but logging fails to import MaKaC.common.logger.IndicoMailFormatter + # when using it in the class= option... + if 'mailFormatter' in formatters: + f = formatters.get('mailFormatter') + if f: + formatters['mailFormatter'] = IndicoMailFormatter(f._fmt, f.datefmt) + + # if there is a problem with the config file, set some sane defaults + if not parsed_files: + formatters['defaultFormatter'] = logging.Formatter( + '%(asctime)s %(levelname)-7s %(request_id)s %(name)-25s %(message)s') + cls._bootstrap_cp(cp, defaultArgs) + + logging._acquireLock() + try: + logging._handlers.clear() + del logging._handlerList[:] + + handlers = cls._install_handlers(cp, defaultArgs, formatters, filters) + logging.config._install_loggers(cp, handlers, False) + + finally: + logging._releaseLock() + return handlers + + @classmethod + def _install_handlers(cls, cp, defaultArgs, formatters, filters=None): + """ + Install and return handlers. If a handler configuration + is missing its args, fetches the default values from the + indico.conf file + """ + hlist = cp.get("handlers", "keys") + hlist = hlist.split(",") + handlers = {} + fixups = [] # for inter-handler references + + for hand in hlist: + sectname = "handler_%s" % hand.strip() + opts = cp.options(sectname) + if "class" in opts: + klass = cp.get(sectname, "class") + else: + klass = defaultArgs[hand.strip()][0] + if "formatter" in opts: + fmt = cp.get(sectname, "formatter") + else: + fmt = "" + klass = eval(klass, vars(logging)) + if "args" in opts: + # if the args are not present in the file, + # take default values + args = cp.get(sectname, "args") + else: + try: + args = defaultArgs[hand.strip()][1] + except KeyError: + continue + args = eval(args, vars(logging)) + h = apply(klass, args) + if "level" in opts: + level = cp.get(sectname, "level") + h.setLevel(logging._levelNames[level]) + else: + h.setLevel(logging._levelNames[defaultArgs[hand.strip()][2]]) + if len(fmt): + h.setFormatter(formatters[fmt]) + if filters and hand.strip() in filters: + for fltr in filters[hand.strip()]: + h.addFilter(fltr) + #temporary hack for FileHandler and MemoryHandler. + if klass == logging.handlers.MemoryHandler: + if "target" in opts: + target = cp.get(sectname,"target") + else: + target = "" + if len(target): #the target handler may not be loaded yet, so keep for later... + fixups.append((h, target)) + handlers[hand] = h + #now all handlers are loaded, fixup inter-handler references... + for h, t in fixups: + h.setTarget(handlers[t]) + return handlers + + +class Logger: + """ + Encapsulates the features provided by the standard logging module + """ + + handlers = {} + + @classmethod + def initialize(cls): + # Lists of filters for each handler + filters = {'indico': [AddIDFilter('indico')], + 'other': [ExtraIndicoFilter()], + 'smtp': [AddIDFilter('indico')]} + + config = Config.getInstance() + + if 'files' in config.getLoggers(): + logConfFilepath = os.path.join(config.getConfigurationDir(), 'logging.conf') + smtpServer = config.getSmtpServer() + serverName = config.getWorkerName() + if not serverName: + serverName = config.getHostNameURL() + + # Default arguments for the handlers, taken mostly for the configuration + defaultArgs = { + 'indico': ("FileHandler", "('%s', 'a')" % cls._log_path('indico.log'), 'DEBUG'), + 'other': ("FileHandler", "('%s', 'a')" % cls._log_path('other.log'), 'DEBUG'), + 'smtp': ( + "handlers.SMTPHandler", "(%s, 'logger@%s', ['%s'], 'Unexpected Exception occurred at %s')" + % (smtpServer, serverName, config.getSupportEmail(), serverName), "ERROR") + } + + cls.handlers.update(LoggerUtils.configFromFile(logConfFilepath, defaultArgs, filters)) + + @classmethod + def init_app(cls, app): + """ + Initialize Flask app logging (add Sentry if needed) + """ + config = Config.getInstance() + + if 'sentry' in config.getLoggers(): + from raven.contrib.flask import Sentry + app.config['SENTRY_DSN'] = config.getSentryDSN() + + # Plug into both Flask and `logging` + Sentry(app, logging=True, level=getattr(logging, config.getSentryLoggingLevel())) + + @classmethod + def reset(cls): + """ + Reset the config, using new paths, etc (useful for testing) + """ + if cls.handlers: + for handler in copy.copy(cls.handlers): + cls.removeHandler(handler) + + cls.initialize() + + @classmethod + def removeHandler(cls, handlerName): + if cls.handlers: + handler = cls.handlers.get(handlerName) + + if handler and handler in cls.handlers: + del cls.handlers[handlerName] + logging.root.handlers.remove(handler) + + @classmethod + def get(cls, module=None): + return logging.getLogger('indico' if module is None else 'indico.' + module) + + @classmethod + def _log_path(cls, fname): + config = Config.getInstance() + configDir = config.getLogDir() + fpath = os.path.join(configDir, fname) + + if not os.access(os.path.dirname(fpath), os.W_OK): + # if the file in the config is not accessible, use a "local" one + fpath = os.path.join(os.getcwd(), '.indico.log') + + return fpath.replace('\\', '\\\\') + + +Logger.initialize() + +# Licensed to Cloudera, Inc. under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. Cloudera, Inc. licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +import sys +sys.path.append("../if/gen-py") +from thrift.transport import TSocket +from thrift.transport import TTransport +from thrift.protocol import TBinaryProtocol + +host = "localhost" +port = 9290 + +from hadoop.api import Jobtracker +from hadoop.api.ttypes import * + +#print dir(ThriftTaskPhase) +#exit() + +socket = TSocket.TSocket(host,port) +transport = TTransport.TBufferedTransport(socket) +protocol = TBinaryProtocol.TBinaryProtocol(transport) + +client = Jobtracker.Client(protocol) +transport.open() + +#print client.trackerName(None,) +#print client.getClusterStatus(None,) +#print dir(client) +#print client.getQueues(None,) +jobs = client.getCompletedJobs(None,) +print jobs +if jobs and len(jobs) > 0: + counters = client.getJobCounters(None, jobs[0].job_id) + + for c in counters: + print "--------------------------------------------" + print "CounterGroup:: ", c.displayName + for name in c.counters: + print "Counter '%s':: %s" % (name,c.counters[name].value) + +# Copyright 2015 The TensorFlow Authors. All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# ============================================================================== + +"""Tests for tensorflow.python.client.session.Session's list_devices API.""" + +from __future__ import absolute_import +from __future__ import division +from __future__ import print_function + +from tensorflow.core.protobuf import cluster_pb2 +from tensorflow.core.protobuf import config_pb2 +from tensorflow.python.client import pywrap_tf_session as tf_session +from tensorflow.python.client import session +from tensorflow.python.framework import errors +from tensorflow.python.framework import ops +from tensorflow.python.framework import test_util +from tensorflow.python.platform import googletest +from tensorflow.python.training import server_lib + + +class SessionListDevicesTest(test_util.TensorFlowTestCase): + + def testListDevices(self): + with session.Session() as sess: + devices = sess.list_devices() + self.assertTrue('/job:localhost/replica:0/task:0/device:CPU:0' in set( + [d.name for d in devices]), devices) + # All valid device incarnations must be non-zero. + self.assertTrue(all(d.incarnation != 0 for d in devices)) + + def testInvalidDeviceNumber(self): + opts = tf_session.TF_NewSessionOptions() + c_session = tf_session.TF_NewSession(ops.get_default_graph()._c_graph, opts) + raw_device_list = tf_session.TF_SessionListDevices(c_session) + size = tf_session.TF_DeviceListCount(raw_device_list) + with self.assertRaises(errors.InvalidArgumentError): + tf_session.TF_DeviceListMemoryBytes(raw_device_list, size) + tf_session.TF_DeleteDeviceList(raw_device_list) + tf_session.TF_CloseSession(c_session) + + def testListDevicesGrpcSession(self): + server = server_lib.Server.create_local_server() + with session.Session(server.target) as sess: + devices = sess.list_devices() + self.assertTrue( + '/job:localhost/replica:0/task:0/device:CPU:0' in set( + [d.name for d in devices]), devices) + # All valid device incarnations must be non-zero. + self.assertTrue(all(d.incarnation != 0 for d in devices)) + + def testListDevicesClusterSpecPropagation(self): + server1 = server_lib.Server.create_local_server() + server2 = server_lib.Server.create_local_server() + + cluster_def = cluster_pb2.ClusterDef() + job = cluster_def.job.add() + job.name = 'worker' + job.tasks[0] = server1.target[len('grpc://'):] + job.tasks[1] = server2.target[len('grpc://'):] + config = config_pb2.ConfigProto(cluster_def=cluster_def) + with session.Session(server1.target, config=config) as sess: + devices = sess.list_devices() + device_names = set(d.name for d in devices) + self.assertTrue( + '/job:worker/replica:0/task:0/device:CPU:0' in device_names) + self.assertTrue( + '/job:worker/replica:0/task:1/device:CPU:0' in device_names) + # All valid device incarnations must be non-zero. + self.assertTrue(all(d.incarnation != 0 for d in devices)) + + +if __name__ == '__main__': + googletest.main() + +# Copyright (c) 2012 The Chromium Authors. All rights reserved. +# Use of this source code is governed by a BSD-style license that can be +# found in the LICENSE file. + +import collections + +import easy_template + +def CmpByName(x, y): + return cmp(x['NAME'], y['NAME']) + +class LandingPage(object): + def __init__(self): + self.section_list = ['Getting Started', 'API', 'Demo', 'Tutorial'] + self.section_map = collections.defaultdict(list) + + def GeneratePage(self, template_path): + with open(template_path) as template_file: + template = template_file.read() + + sec_map = {} + for section_name in self.section_map: + items = self.section_map[section_name] + items = sorted(items, cmp=CmpByName) + sec_map[section_name] = items + print 'Add section ' + section_name + + template_dict = { 'section_map': sec_map } + return easy_template.RunTemplateString(template, template_dict) + + def AddDesc(self, desc): + group = desc['GROUP'] + assert group in self.section_list + self.section_map[group].append(desc) + +# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors +# MIT License. See license.txt + +from __future__ import unicode_literals +"""docfield utililtes""" + +import webnotes + +def rename(doctype, fieldname, newname): + """rename docfield""" + df = webnotes.conn.sql("""select * from tabDocField where parent=%s and fieldname=%s""", + (doctype, fieldname), as_dict=1) + if not df: + return + + df = df[0] + + if webnotes.conn.get_value('DocType', doctype, 'issingle'): + update_single(df, newname) + else: + update_table(df, newname) + update_parent_field(df, newname) + +def update_single(f, new): + """update in tabSingles""" + webnotes.conn.begin() + webnotes.conn.sql("""update tabSingles set field=%s where doctype=%s and field=%s""", + (new, f['parent'], f['fieldname'])) + webnotes.conn.commit() + +def update_table(f, new): + """update table""" + query = get_change_column_query(f, new) + if query: + webnotes.conn.sql(query) + +def update_parent_field(f, new): + """update 'parentfield' in tables""" + if f['fieldtype']=='Table': + webnotes.conn.begin() + webnotes.conn.sql("""update `tab%s` set parentfield=%s where parentfield=%s""" \ + % (f['options'], '%s', '%s'), (new, f['fieldname'])) + webnotes.conn.commit() + +def get_change_column_query(f, new): + """generate change fieldname query""" + desc = webnotes.conn.sql("desc `tab%s`" % f['parent']) + for d in desc: + if d[0]== f['fieldname']: + return 'alter table `tab%s` change `%s` `%s` %s' % \ + (f['parent'], f['fieldname'], new, d[1]) +""" + + >>> from pybrain.tools.shortcuts import buildNetwork + >>> from test_recurrent_network import buildRecurrentNetwork + >>> from test_peephole_lstm import buildMinimalLSTMNetwork + >>> from test_peephole_mdlstm import buildMinimalMDLSTMNetwork + >>> from test_nested_network import buildNestedNetwork + >>> from test_simple_lstm_network import buildSimpleLSTMNetwork + >>> from test_simple_mdlstm import buildSimpleMDLSTMNetwork + >>> from test_swiping_network import buildSwipingNetwork + >>> from test_shared_connections import buildSharedCrossedNetwork + >>> from test_sliced_connections import buildSlicedNetwork + >>> from test_borderswipingnetwork import buildSimpleBorderSwipingNet + +Test a number of network architectures, and compare if they produce the same output, +whether the Python implementation is used, or CTYPES. + +Use the network construction scripts in other test files to build a number of networks, +and then test the equivalence of each. + +Simple net + >>> testEquivalence(buildNetwork(2,2)) + True + +A lot of layers + >>> net = buildNetwork(2,3,4,3,2,3,4,3,2) + >>> testEquivalence(net) + True + +Nonstandard components + >>> from pybrain.structure import TanhLayer + >>> net = buildNetwork(2,3,2, bias = True, outclass = TanhLayer) + >>> testEquivalence(net) + True + +Shared connections + >>> net = buildSharedCrossedNetwork() + >>> testEquivalence(net) + True + +Sliced connections + >>> net = buildSlicedNetwork() + >>> testEquivalence(net) + True + +Nested networks (not supposed to work yet!) + >>> net = buildNestedNetwork() + >>> testEquivalence(net) + Network cannot be converted. + +Recurrent networks + >>> net = buildRecurrentNetwork() + >>> net.name = '22' + >>> net.params[:] = [1,1,0.5] + >>> testEquivalence(net) + True + +Swiping networks + >>> net = buildSwipingNetwork() + >>> testEquivalence(net) + True + +Border-swiping networks + >>> net = buildSimpleBorderSwipingNet() + >>> testEquivalence(net) + True + +Lstm + >>> net = buildSimpleLSTMNetwork() + >>> testEquivalence(net) + True + +Mdlstm + >>> net = buildSimpleMDLSTMNetwork() + >>> testEquivalence(net) + True + +Lstm with peepholes + >>> net = buildMinimalLSTMNetwork(True) + >>> testEquivalence(net) + True + +Mdlstm with peepholes + >>> net = buildMinimalMDLSTMNetwork(True) + >>> testEquivalence(net) + True + + +TODO: +- heavily nested +- exotic module use + +""" + +__author__ = 'Tom Schaul, tom@idsia.ch' +_dependencies = ['arac'] + +from pybrain.tests.helpers import buildAppropriateDataset, epsilonCheck +from pybrain.tests import runModuleTestSuite + +def testEquivalence(net): + cnet = net.convertToFastNetwork() + if cnet == None: + return None + ds = buildAppropriateDataset(net) + if net.sequential: + for seq in ds: + net.reset() + cnet.reset() + for input, _ in seq: + res = net.activate(input) + cres = cnet.activate(input) + if net.name == '22': + h = net['hidden0'] + ch = cnet['hidden0'] + print(('ni', input, net.inputbuffer.T)) + print(('ci', input, cnet.inputbuffer.T)) + print(('hni', h.inputbuffer.T[0])) + print(('hci', ch.inputbuffer.T[0])) + print(('hnout', h.outputbuffer.T[0])) + print(('hcout', ch.outputbuffer.T[0])) + print() + + else: + for input, _ in ds: + res = net.activate(input) + cres = cnet.activate(input) + if epsilonCheck(sum(res - cres), 0.001): + return True + else: + print(('in-net', net.inputbuffer.T)) + print(('in-arac', cnet.inputbuffer.T)) + print(('out-net', net.outputbuffer.T)) + print(('out-arac', cnet.outputbuffer.T)) + return (res, cres) + + +if __name__ == "__main__": + runModuleTestSuite(__import__('__main__')) + + +# Copyright 2014 The Chromium Authors. All rights reserved. +# Use of this source code is governed by a BSD-style license that can be +# found in the LICENSE file. +from telemetry.page import page as page_module +from telemetry.page import page_set as page_set_module + + +class GpuRasterizationTestsPage(page_module.Page): + + def __init__(self, page_set): + super(GpuRasterizationTestsPage, self).__init__( + url='file://../../data/gpu/pixel_background.html', + page_set=page_set, + name='GpuRasterization.BlueBox') + + self.expectations = [ + {'comment': 'body-t', + 'color': [255, 255, 255], + 'tolerance': 0, + 'location': [5, 5]}, + {'comment': 'body-r', + 'color': [255, 255, 255], + 'tolerance': 0, + 'location': [215, 5]}, + {'comment': 'body-b', + 'color': [255, 255, 255], + 'tolerance': 0, + 'location': [215, 215]}, + {'comment': 'body-l', + 'color': [255, 255, 255], + 'tolerance': 0, + 'location': [5, 215]}, + {'comment': 'background-t', + 'color': [0, 0, 0], + 'tolerance': 0, + 'location': [30, 30]}, + {'comment': 'background-r', + 'color': [0, 0, 0], + 'tolerance': 0, + 'location': [170, 30]}, + {'comment': 'background-b', + 'color': [0, 0, 0], + 'tolerance': 0, + 'location': [170, 170]}, + {'comment': 'background-l', + 'color': [0, 0, 0], + 'tolerance': 0, + 'location': [30, 170]}, + {'comment': 'box-t', + 'color': [0, 0, 255], + 'tolerance': 0, + 'location': [70, 70]}, + {'comment': 'box-r', + 'color': [0, 0, 255], + 'tolerance': 0, + 'location': [140, 70]}, + {'comment': 'box-b', + 'color': [0, 0, 255], + 'tolerance': 0, + 'location': [140, 140]}, + {'comment': 'box-l', + 'color': [0, 0, 255], + 'tolerance': 0, + 'location': [70, 140]} + ] + self.test_rect = [0, 0, 220, 220] + + def RunNavigateSteps(self, action_runner): + action_runner.NavigateToPage(self) + action_runner.WaitForJavaScriptCondition( + 'domAutomationController._finished', timeout_in_seconds=30) + + +class GpuRasterizationTestsPageSet(page_set_module.PageSet): + + """ Basic test cases for GPU rasterization. """ + + def __init__(self): + super(GpuRasterizationTestsPageSet, self).__init__() + + self.AddPage(GpuRasterizationTestsPage(self)) + +"""Temporary files. + +This module provides generic, low- and high-level interfaces for +creating temporary files and directories. The interfaces listed +as "safe" just below can be used without fear of race conditions. +Those listed as "unsafe" cannot, and are provided for backward +compatibility only. + +This module also provides some data items to the user: + + TMP_MAX - maximum number of names that will be tried before + giving up. + tempdir - If this is set to a string before the first use of + any routine from this module, it will be considered as + another candidate location to store temporary files. +""" + +__all__ = [ + "NamedTemporaryFile", "TemporaryFile", # high level safe interfaces + "SpooledTemporaryFile", "TemporaryDirectory", + "mkstemp", "mkdtemp", # low level safe interfaces + "mktemp", # deprecated unsafe interface + "TMP_MAX", "gettempprefix", # constants + "tempdir", "gettempdir" + ] + + +# Imports. + +import functools as _functools +import warnings as _warnings +import io as _io +import os as _os +import shutil as _shutil +import errno as _errno +from random import Random as _Random +import weakref as _weakref + +try: + import _thread +except ImportError: + import _dummy_thread as _thread +_allocate_lock = _thread.allocate_lock + +_text_openflags = _os.O_RDWR | _os.O_CREAT | _os.O_EXCL +if hasattr(_os, 'O_NOFOLLOW'): + _text_openflags |= _os.O_NOFOLLOW + +_bin_openflags = _text_openflags +if hasattr(_os, 'O_BINARY'): + _bin_openflags |= _os.O_BINARY + +if hasattr(_os, 'TMP_MAX'): + TMP_MAX = _os.TMP_MAX +else: + TMP_MAX = 10000 + +# Although it does not have an underscore for historical reasons, this +# variable is an internal implementation detail (see issue 10354). +template = "tmp" + +# Internal routines. + +_once_lock = _allocate_lock() + +if hasattr(_os, "lstat"): + _stat = _os.lstat +elif hasattr(_os, "stat"): + _stat = _os.stat +else: + # Fallback. All we need is something that raises OSError if the + # file doesn't exist. + def _stat(fn): + fd = _os.open(fn, _os.O_RDONLY) + _os.close(fd) + +def _exists(fn): + try: + _stat(fn) + except OSError: + return False + else: + return True + +class _RandomNameSequence: + """An instance of _RandomNameSequence generates an endless + sequence of unpredictable strings which can safely be incorporated + into file names. Each string is six characters long. Multiple + threads can safely use the same instance at the same time. + + _RandomNameSequence is an iterator.""" + + characters = "abcdefghijklmnopqrstuvwxyz0123456789_" + + @property + def rng(self): + cur_pid = _os.getpid() + if cur_pid != getattr(self, '_rng_pid', None): + self._rng = _Random() + self._rng_pid = cur_pid + return self._rng + + def __iter__(self): + return self + + def __next__(self): + c = self.characters + choose = self.rng.choice + letters = [choose(c) for dummy in range(8)] + return ''.join(letters) + +def _candidate_tempdir_list(): + """Generate a list of candidate temporary directories which + _get_default_tempdir will try.""" + + dirlist = [] + + # First, try the environment. + for envname in 'TMPDIR', 'TEMP', 'TMP': + dirname = _os.getenv(envname) + if dirname: dirlist.append(dirname) + + # Failing that, try OS-specific locations. + if _os.name == 'nt': + dirlist.extend([ r'c:\temp', r'c:\tmp', r'\temp', r'\tmp' ]) + else: + dirlist.extend([ '/tmp', '/var/tmp', '/usr/tmp' ]) + + # As a last resort, the current directory. + try: + dirlist.append(_os.getcwd()) + except (AttributeError, OSError): + dirlist.append(_os.curdir) + + return dirlist + +def _get_default_tempdir(): + """Calculate the default directory to use for temporary files. + This routine should be called exactly once. + + We determine whether or not a candidate temp dir is usable by + trying to create and write to a file in that directory. If this + is successful, the test file is deleted. To prevent denial of + service, the name of the test file must be randomized.""" + + namer = _RandomNameSequence() + dirlist = _candidate_tempdir_list() + + for dir in dirlist: + if dir != _os.curdir: + dir = _os.path.abspath(dir) + # Try only a few names per directory. + for seq in range(100): + name = next(namer) + filename = _os.path.join(dir, name) + try: + fd = _os.open(filename, _bin_openflags, 0o600) + try: + try: + with _io.open(fd, 'wb', closefd=False) as fp: + fp.write(b'blat') + finally: + _os.close(fd) + finally: + _os.unlink(filename) + return dir + except FileExistsError: + pass + except OSError: + break # no point trying more names in this directory + raise FileNotFoundError(_errno.ENOENT, + "No usable temporary directory found in %s" % + dirlist) + +_name_sequence = None + +def _get_candidate_names(): + """Common setup sequence for all user-callable interfaces.""" + + global _name_sequence + if _name_sequence is None: + _once_lock.acquire() + try: + if _name_sequence is None: + _name_sequence = _RandomNameSequence() + finally: + _once_lock.release() + return _name_sequence + + +def _mkstemp_inner(dir, pre, suf, flags): + """Code common to mkstemp, TemporaryFile, and NamedTemporaryFile.""" + + names = _get_candidate_names() + + for seq in range(TMP_MAX): + name = next(names) + file = _os.path.join(dir, pre + name + suf) + try: + fd = _os.open(file, flags, 0o600) + return (fd, _os.path.abspath(file)) + except FileExistsError: + continue # try again + except PermissionError: + # This exception is thrown when a directory with the chosen name + # already exists on windows. + if _os.name == 'nt': + continue + else: + raise + + raise FileExistsError(_errno.EEXIST, + "No usable temporary file name found") + + +# User visible interfaces. + +def gettempprefix(): + """Accessor for tempdir.template.""" + return template + +tempdir = None + +def gettempdir(): + """Accessor for tempfile.tempdir.""" + global tempdir + if tempdir is None: + _once_lock.acquire() + try: + if tempdir is None: + tempdir = _get_default_tempdir() + finally: + _once_lock.release() + return tempdir + +def mkstemp(suffix="", prefix=template, dir=None, text=False): + """User-callable function to create and return a unique temporary + file. The return value is a pair (fd, name) where fd is the + file descriptor returned by os.open, and name is the filename. + + If 'suffix' is specified, the file name will end with that suffix, + otherwise there will be no suffix. + + If 'prefix' is specified, the file name will begin with that prefix, + otherwise a default prefix is used. + + If 'dir' is specified, the file will be created in that directory, + otherwise a default directory is used. + + If 'text' is specified and true, the file is opened in text + mode. Else (the default) the file is opened in binary mode. On + some operating systems, this makes no difference. + + The file is readable and writable only by the creating user ID. + If the operating system uses permission bits to indicate whether a + file is executable, the file is executable by no one. The file + descriptor is not inherited by children of this process. + + Caller is responsible for deleting the file when done with it. + """ + + if dir is None: + dir = gettempdir() + + if text: + flags = _text_openflags + else: + flags = _bin_openflags + + return _mkstemp_inner(dir, prefix, suffix, flags) + + +def mkdtemp(suffix="", prefix=template, dir=None): + """User-callable function to create and return a unique temporary + directory. The return value is the pathname of the directory. + + Arguments are as for mkstemp, except that the 'text' argument is + not accepted. + + The directory is readable, writable, and searchable only by the + creating user. + + Caller is responsible for deleting the directory when done with it. + """ + + if dir is None: + dir = gettempdir() + + names = _get_candidate_names() + + for seq in range(TMP_MAX): + name = next(names) + file = _os.path.join(dir, prefix + name + suffix) + try: + _os.mkdir(file, 0o700) + return file + except FileExistsError: + continue # try again + + raise FileExistsError(_errno.EEXIST, + "No usable temporary directory name found") + +def mktemp(suffix="", prefix=template, dir=None): + """User-callable function to return a unique temporary file name. The + file is not created. + + Arguments are as for mkstemp, except that the 'text' argument is + not accepted. + + This function is unsafe and should not be used. The file name + refers to a file that did not exist at some point, but by the time + you get around to creating it, someone else may have beaten you to + the punch. + """ + +## from warnings import warn as _warn +## _warn("mktemp is a potential security risk to your program", +## RuntimeWarning, stacklevel=2) + + if dir is None: + dir = gettempdir() + + names = _get_candidate_names() + for seq in range(TMP_MAX): + name = next(names) + file = _os.path.join(dir, prefix + name + suffix) + if not _exists(file): + return file + + raise FileExistsError(_errno.EEXIST, + "No usable temporary filename found") + + +class _TemporaryFileCloser: + """A separate object allowing proper closing of a temporary file's + underlying file object, without adding a __del__ method to the + temporary file.""" + + file = None # Set here since __del__ checks it + close_called = False + + def __init__(self, file, name, delete=True): + self.file = file + self.name = name + self.delete = delete + + # NT provides delete-on-close as a primitive, so we don't need + # the wrapper to do anything special. We still use it so that + # file.name is useful (i.e. not "(fdopen)") with NamedTemporaryFile. + if _os.name != 'nt': + # Cache the unlinker so we don't get spurious errors at + # shutdown when the module-level "os" is None'd out. Note + # that this must be referenced as self.unlink, because the + # name TemporaryFileWrapper may also get None'd out before + # __del__ is called. + + def close(self, unlink=_os.unlink): + if not self.close_called and self.file is not None: + self.close_called = True + self.file.close() + if self.delete: + unlink(self.name) + + # Need to ensure the file is deleted on __del__ + def __del__(self): + self.close() + + else: + def close(self): + if not self.close_called: + self.close_called = True + self.file.close() + + +class _TemporaryFileWrapper: + """Temporary file wrapper + + This class provides a wrapper around files opened for + temporary use. In particular, it seeks to automatically + remove the file when it is no longer needed. + """ + + def __init__(self, file, name, delete=True): + self.file = file + self.name = name + self.delete = delete + self._closer = _TemporaryFileCloser(file, name, delete) + + def __getattr__(self, name): + # Attribute lookups are delegated to the underlying file + # and cached for non-numeric results + # (i.e. methods are cached, closed and friends are not) + file = self.__dict__['file'] + a = getattr(file, name) + if hasattr(a, '__call__'): + func = a + @_functools.wraps(func) + def func_wrapper(*args, **kwargs): + return func(*args, **kwargs) + # Avoid closing the file as long as the wrapper is alive, + # see issue #18879. + func_wrapper._closer = self._closer + a = func_wrapper + if not isinstance(a, int): + setattr(self, name, a) + return a + + # The underlying __enter__ method returns the wrong object + # (self.file) so override it to return the wrapper + def __enter__(self): + self.file.__enter__() + return self + + # Need to trap __exit__ as well to ensure the file gets + # deleted when used in a with statement + def __exit__(self, exc, value, tb): + result = self.file.__exit__(exc, value, tb) + self.close() + return result + + def close(self): + """ + Close the temporary file, possibly deleting it. + """ + self._closer.close() + + # iter() doesn't use __getattr__ to find the __iter__ method + def __iter__(self): + return iter(self.file) + + +def NamedTemporaryFile(mode='w+b', buffering=-1, encoding=None, + newline=None, suffix="", prefix=template, + dir=None, delete=True): + """Create and return a temporary file. + Arguments: + 'prefix', 'suffix', 'dir' -- as for mkstemp. + 'mode' -- the mode argument to io.open (default "w+b"). + 'buffering' -- the buffer size argument to io.open (default -1). + 'encoding' -- the encoding argument to io.open (default None) + 'newline' -- the newline argument to io.open (default None) + 'delete' -- whether the file is deleted on close (default True). + The file is created as mkstemp() would do it. + + Returns an object with a file-like interface; the name of the file + is accessible as file.name. The file will be automatically deleted + when it is closed unless the 'delete' argument is set to False. + """ + + if dir is None: + dir = gettempdir() + + flags = _bin_openflags + + # Setting O_TEMPORARY in the flags causes the OS to delete + # the file when it is closed. This is only supported by Windows. + if _os.name == 'nt' and delete: + flags |= _os.O_TEMPORARY + + (fd, name) = _mkstemp_inner(dir, prefix, suffix, flags) + try: + file = _io.open(fd, mode, buffering=buffering, + newline=newline, encoding=encoding) + + return _TemporaryFileWrapper(file, name, delete) + except Exception: + _os.close(fd) + raise + +if _os.name != 'posix' or _os.sys.platform == 'cygwin': + # On non-POSIX and Cygwin systems, assume that we cannot unlink a file + # while it is open. + TemporaryFile = NamedTemporaryFile + +else: + def TemporaryFile(mode='w+b', buffering=-1, encoding=None, + newline=None, suffix="", prefix=template, + dir=None): + """Create and return a temporary file. + Arguments: + 'prefix', 'suffix', 'dir' -- as for mkstemp. + 'mode' -- the mode argument to io.open (default "w+b"). + 'buffering' -- the buffer size argument to io.open (default -1). + 'encoding' -- the encoding argument to io.open (default None) + 'newline' -- the newline argument to io.open (default None) + The file is created as mkstemp() would do it. + + Returns an object with a file-like interface. The file has no + name, and will cease to exist when it is closed. + """ + + if dir is None: + dir = gettempdir() + + flags = _bin_openflags + + (fd, name) = _mkstemp_inner(dir, prefix, suffix, flags) + try: + _os.unlink(name) + return _io.open(fd, mode, buffering=buffering, + newline=newline, encoding=encoding) + except: + _os.close(fd) + raise + +class SpooledTemporaryFile: + """Temporary file wrapper, specialized to switch from BytesIO + or StringIO to a real file when it exceeds a certain size or + when a fileno is needed. + """ + _rolled = False + + def __init__(self, max_size=0, mode='w+b', buffering=-1, + encoding=None, newline=None, + suffix="", prefix=template, dir=None): + if 'b' in mode: + self._file = _io.BytesIO() + else: + # Setting newline="\n" avoids newline translation; + # this is important because otherwise on Windows we'd + # hget double newline translation upon rollover(). + self._file = _io.StringIO(newline="\n") + self._max_size = max_size + self._rolled = False + self._TemporaryFileArgs = {'mode': mode, 'buffering': buffering, + 'suffix': suffix, 'prefix': prefix, + 'encoding': encoding, 'newline': newline, + 'dir': dir} + + def _check(self, file): + if self._rolled: return + max_size = self._max_size + if max_size and file.tell() > max_size: + self.rollover() + + def rollover(self): + if self._rolled: return + file = self._file + newfile = self._file = TemporaryFile(**self._TemporaryFileArgs) + del self._TemporaryFileArgs + + newfile.write(file.getvalue()) + newfile.seek(file.tell(), 0) + + self._rolled = True + + # The method caching trick from NamedTemporaryFile + # won't work here, because _file may change from a + # BytesIO/StringIO instance to a real file. So we list + # all the methods directly. + + # Context management protocol + def __enter__(self): + if self._file.closed: + raise ValueError("Cannot enter context with closed file") + return self + + def __exit__(self, exc, value, tb): + self._file.close() + + # file protocol + def __iter__(self): + return self._file.__iter__() + + def close(self): + self._file.close() + + @property + def closed(self): + return self._file.closed + + @property + def encoding(self): + try: + return self._file.encoding + except AttributeError: + if 'b' in self._TemporaryFileArgs['mode']: + raise + return self._TemporaryFileArgs['encoding'] + + def fileno(self): + self.rollover() + return self._file.fileno() + + def flush(self): + self._file.flush() + + def isatty(self): + return self._file.isatty() + + @property + def mode(self): + try: + return self._file.mode + except AttributeError: + return self._TemporaryFileArgs['mode'] + + @property + def name(self): + try: + return self._file.name + except AttributeError: + return None + + @property + def newlines(self): + try: + return self._file.newlines + except AttributeError: + if 'b' in self._TemporaryFileArgs['mode']: + raise + return self._TemporaryFileArgs['newline'] + + def read(self, *args): + return self._file.read(*args) + + def readline(self, *args): + return self._file.readline(*args) + + def readlines(self, *args): + return self._file.readlines(*args) + + def seek(self, *args): + self._file.seek(*args) + + @property + def softspace(self): + return self._file.softspace + + def tell(self): + return self._file.tell() + + def truncate(self, size=None): + if size is None: + self._file.truncate() + else: + if size > self._max_size: + self.rollover() + self._file.truncate(size) + + def write(self, s): + file = self._file + rv = file.write(s) + self._check(file) + return rv + + def writelines(self, iterable): + file = self._file + rv = file.writelines(iterable) + self._check(file) + return rv + + +class TemporaryDirectory(object): + """Create and return a temporary directory. This has the same + behavior as mkdtemp but can be used as a context manager. For + example: + + with TemporaryDirectory() as tmpdir: + ... + + Upon exiting the context, the directory and everything contained + in it are removed. + """ + + # Handle mkdtemp raising an exception + name = None + _finalizer = None + _closed = False + + def __init__(self, suffix="", prefix=template, dir=None): + self.name = mkdtemp(suffix, prefix, dir) + self._finalizer = _weakref.finalize( + self, self._cleanup, self.name, + warn_message="Implicitly cleaning up {!r}".format(self)) + + @classmethod + def _cleanup(cls, name, warn_message=None): + _shutil.rmtree(name) + if warn_message is not None: + _warnings.warn(warn_message, ResourceWarning) + + + def __repr__(self): + return "<{} {!r}>".format(self.__class__.__name__, self.name) + + def __enter__(self): + return self.name + + def __exit__(self, exc, value, tb): + self.cleanup() + + def cleanup(self): + if self._finalizer is not None: + self._finalizer.detach() + if self.name is not None and not self._closed: + _shutil.rmtree(self.name) + self._closed = True + +"""A simple log mechanism styled after PEP 282.""" + +# This module should be kept compatible with Python 2.1. + +# The class here is styled after PEP 282 so that it could later be +# replaced with a standard Python logging implementation. + +DEBUG = 1 +INFO = 2 +WARN = 3 +ERROR = 4 +FATAL = 5 + +import sys + +class Log: + + def __init__(self, threshold=WARN): + self.threshold = threshold + + def _log(self, level, msg, args): + if level >= self.threshold: + if not args: + # msg may contain a '%'. If args is empty, + # don't even try to string-format + print msg + else: + print msg % args + sys.stdout.flush() + + def log(self, level, msg, *args): + self._log(level, msg, args) + + def debug(self, msg, *args): + self._log(DEBUG, msg, args) + + def info(self, msg, *args): + self._log(INFO, msg, args) + + def warn(self, msg, *args): + self._log(WARN, msg, args) + + def error(self, msg, *args): + self._log(ERROR, msg, args) + + def fatal(self, msg, *args): + self._log(FATAL, msg, args) + +_global_log = Log() +log = _global_log.log +debug = _global_log.debug +info = _global_log.info +warn = _global_log.warn +error = _global_log.error +fatal = _global_log.fatal + +def set_threshold(level): + # return the old threshold for use from tests + old = _global_log.threshold + _global_log.threshold = level + return old + +def set_verbosity(v): + if v <= 0: + set_threshold(WARN) + elif v == 1: + set_threshold(INFO) + elif v >= 2: + set_threshold(DEBUG) + +from __future__ import unicode_literals + +from django.http import HttpResponse, Http404 +from django.template import loader +from django.contrib.sites.models import get_current_site +from django.core import urlresolvers +from django.core.paginator import EmptyPage, PageNotAnInteger +from django.contrib.gis.db.models.fields import GeometryField +from django.db import connections, DEFAULT_DB_ALIAS +from django.db.models import get_model +from django.utils import six +from django.utils.translation import ugettext as _ + +from django.contrib.gis.shortcuts import render_to_kml, render_to_kmz + +def index(request, sitemaps): + """ + This view generates a sitemap index that uses the proper view + for resolving geographic section sitemap URLs. + """ + current_site = get_current_site(request) + sites = [] + protocol = 'https' if request.is_secure() else 'http' + for section, site in sitemaps.items(): + if callable(site): + pages = site().paginator.num_pages + else: + pages = site.paginator.num_pages + sitemap_url = urlresolvers.reverse('django.contrib.gis.sitemaps.views.sitemap', kwargs={'section': section}) + sites.append('%s://%s%s' % (protocol, current_site.domain, sitemap_url)) + + if pages > 1: + for page in range(2, pages+1): + sites.append('%s://%s%s?p=%s' % (protocol, current_site.domain, sitemap_url, page)) + xml = loader.render_to_string('sitemap_index.xml', {'sitemaps': sites}) + return HttpResponse(xml, content_type='application/xml') + +def sitemap(request, sitemaps, section=None): + """ + This view generates a sitemap with additional geographic + elements defined by Google. + """ + maps, urls = [], [] + if section is not None: + if section not in sitemaps: + raise Http404(_("No sitemap available for section: %r") % section) + maps.append(sitemaps[section]) + else: + maps = list(six.itervalues(sitemaps)) + + page = request.GET.get("p", 1) + current_site = get_current_site(request) + for site in maps: + try: + if callable(site): + urls.extend(site().get_urls(page=page, site=current_site)) + else: + urls.extend(site.get_urls(page=page, site=current_site)) + except EmptyPage: + raise Http404(_("Page %s empty") % page) + except PageNotAnInteger: + raise Http404(_("No page '%s'") % page) + xml = loader.render_to_string('gis/sitemaps/geo_sitemap.xml', {'urlset': urls}) + return HttpResponse(xml, content_type='application/xml') + +def kml(request, label, model, field_name=None, compress=False, using=DEFAULT_DB_ALIAS): + """ + This view generates KML for the given app label, model, and field name. + + The model's default manager must be GeoManager, and the field name + must be that of a geographic field. + """ + placemarks = [] + klass = get_model(label, model) + if not klass: + raise Http404('You must supply a valid app label and module name. Got "%s.%s"' % (label, model)) + + if field_name: + try: + info = klass._meta.get_field_by_name(field_name) + if not isinstance(info[0], GeometryField): + raise Exception + except: + raise Http404('Invalid geometry field.') + + connection = connections[using] + + if connection.ops.postgis: + # PostGIS will take care of transformation. + placemarks = klass._default_manager.using(using).kml(field_name=field_name) + else: + # There's no KML method on Oracle or MySQL, so we use the `kml` + # attribute of the lazy geometry instead. + placemarks = [] + if connection.ops.oracle: + qs = klass._default_manager.using(using).transform(4326, field_name=field_name) + else: + qs = klass._default_manager.using(using).all() + for mod in qs: + mod.kml = getattr(mod, field_name).kml + placemarks.append(mod) + + # Getting the render function and rendering to the correct. + if compress: + render = render_to_kmz + else: + render = render_to_kml + return render('gis/kml/placemarks.kml', {'places' : placemarks}) + +def kmz(request, label, model, field_name=None, using=DEFAULT_DB_ALIAS): + """ + This view returns KMZ for the given app label, model, and field name. + """ + return kml(request, label, model, field_name, compress=True, using=using) + +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# (c) 2015, Kevin Brebanov +# Based on pacman (Afterburn , Aaron Bull Schaefer ) +# and apt (Matthew Williams ) modules. +# +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function +__metaclass__ = type + + +ANSIBLE_METADATA = {'metadata_version': '1.1', + 'status': ['stableinterface'], + 'supported_by': 'community'} + + +DOCUMENTATION = ''' +--- +module: apk +short_description: Manages apk packages +description: + - Manages I(apk) packages for Alpine Linux. +author: "Kevin Brebanov (@kbrebanov)" +version_added: "2.0" +options: + available: + description: + - During upgrade, reset versioned world dependencies and change logic to prefer replacing or downgrading packages (instead of holding them) + if the currently installed package is no longer available from any repository. + required: false + default: no + choices: [ "yes", "no" ] + version_added: "2.4" + name: + description: + - A package name, like C(foo), or multiple packages, like C(foo, bar). + required: false + default: null + repository: + description: + - A package repository or multiple repositories + required: false + default: null + version_added: "2.4" + state: + description: + - Indicates the desired package(s) state. + - C(present) ensures the package(s) is/are present. + - C(absent) ensures the package(s) is/are absent. + - C(latest) ensures the package(s) is/are present and the latest version(s). + required: false + default: present + choices: [ "present", "absent", "latest" ] + update_cache: + description: + - Update repository indexes. Can be run with other steps or on it's own. + required: false + default: no + choices: [ "yes", "no" ] + upgrade: + description: + - Upgrade all installed packages to their latest version. + required: false + default: no + choices: [ "yes", "no" ] +notes: + - '"name" and "upgrade" are mutually exclusive.' +''' + +EXAMPLES = ''' +# Update repositories and install "foo" package +- apk: + name: foo + update_cache: yes + +# Update repositories and install "foo" and "bar" packages +- apk: + name: foo,bar + update_cache: yes + +# Remove "foo" package +- apk: + name: foo + state: absent + +# Remove "foo" and "bar" packages +- apk: + name: foo,bar + state: absent + +# Install the package "foo" +- apk: + name: foo + state: present + +# Install the packages "foo" and "bar" +- apk: + name: foo,bar + state: present + +# Update repositories and update package "foo" to latest version +- apk: + name: foo + state: latest + update_cache: yes + +# Update repositories and update packages "foo" and "bar" to latest versions +- apk: + name: foo,bar + state: latest + update_cache: yes + +# Update all installed packages to the latest versions +- apk: + upgrade: yes + +# Upgrade / replace / downgrade / uninstall all installed packages to the latest versions available +- apk: + available: yes + upgrade: yes + +# Update repositories as a separate step +- apk: + update_cache: yes + +# Install package from a specific repository +- apk: + name: foo + state: latest + update_cache: yes + repository: http://dl-3.alpinelinux.org/alpine/edge/main +''' + +RETURN = ''' +packages: + description: a list of packages that have been changed + returned: when packages have changed + type: list + sample: ['package', 'other-package'] +''' + +import re +# Import module snippets. +from ansible.module_utils.basic import AnsibleModule + +def parse_for_packages(stdout): + packages = [] + data = stdout.split('\n') + regex = re.compile('^\(\d+/\d+\)\s+\S+\s+(\S+)') + for l in data: + p = regex.search(l) + if p: + packages.append(p.group(1)) + return packages + +def update_package_db(module, exit): + cmd = "%s update" % (APK_PATH) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + if rc != 0: + module.fail_json(msg="could not update package db", stdout=stdout, stderr=stderr) + elif exit: + module.exit_json(changed=True, msg='updated repository indexes', stdout=stdout, stderr=stderr) + else: + return True + +def query_package(module, name): + cmd = "%s -v info --installed %s" % (APK_PATH, name) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + if rc == 0: + return True + else: + return False + +def query_latest(module, name): + cmd = "%s version %s" % (APK_PATH, name) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + search_pattern = r"(%s)-[\d\.\w]+-[\d\w]+\s+(.)\s+[\d\.\w]+-[\d\w]+\s+" % (re.escape(name)) + match = re.search(search_pattern, stdout) + if match and match.group(2) == "<": + return False + return True + +def query_virtual(module, name): + cmd = "%s -v info --description %s" % (APK_PATH, name) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + search_pattern = r"^%s: virtual meta package" % (re.escape(name)) + if re.search(search_pattern, stdout): + return True + return False + +def get_dependencies(module, name): + cmd = "%s -v info --depends %s" % (APK_PATH, name) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + dependencies = stdout.split() + if len(dependencies) > 1: + return dependencies[1:] + else: + return [] + +def upgrade_packages(module, available): + if module.check_mode: + cmd = "%s upgrade --simulate" % (APK_PATH) + else: + cmd = "%s upgrade" % (APK_PATH) + if available: + cmd = "%s --available" % cmd + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + packagelist = parse_for_packages(stdout) + if rc != 0: + module.fail_json(msg="failed to upgrade packages", stdout=stdout, stderr=stderr, packages=packagelist) + if re.search(r'^OK', stdout): + module.exit_json(changed=False, msg="packages already upgraded", stdout=stdout, stderr=stderr, packages=packagelist) + module.exit_json(changed=True, msg="upgraded packages", stdout=stdout, stderr=stderr, packages=packagelist) + +def install_packages(module, names, state): + upgrade = False + to_install = [] + to_upgrade = [] + for name in names: + # Check if virtual package + if query_virtual(module, name): + # Get virtual package dependencies + dependencies = get_dependencies(module, name) + for dependency in dependencies: + if state == 'latest' and not query_latest(module, dependency): + to_upgrade.append(dependency) + else: + if not query_package(module, name): + to_install.append(name) + elif state == 'latest' and not query_latest(module, name): + to_upgrade.append(name) + if to_upgrade: + upgrade = True + if not to_install and not upgrade: + module.exit_json(changed=False, msg="package(s) already installed") + packages = " ".join(to_install + to_upgrade) + if upgrade: + if module.check_mode: + cmd = "%s add --upgrade --simulate %s" % (APK_PATH, packages) + else: + cmd = "%s add --upgrade %s" % (APK_PATH, packages) + else: + if module.check_mode: + cmd = "%s add --simulate %s" % (APK_PATH, packages) + else: + cmd = "%s add %s" % (APK_PATH, packages) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + packagelist = parse_for_packages(stdout) + if rc != 0: + module.fail_json(msg="failed to install %s" % (packages), stdout=stdout, stderr=stderr, packages=packagelist) + module.exit_json(changed=True, msg="installed %s package(s)" % (packages), stdout=stdout, stderr=stderr, packages=packagelist) + +def remove_packages(module, names): + installed = [] + for name in names: + if query_package(module, name): + installed.append(name) + if not installed: + module.exit_json(changed=False, msg="package(s) already removed") + names = " ".join(installed) + if module.check_mode: + cmd = "%s del --purge --simulate %s" % (APK_PATH, names) + else: + cmd = "%s del --purge %s" % (APK_PATH, names) + rc, stdout, stderr = module.run_command(cmd, check_rc=False) + packagelist = parse_for_packages(stdout) + if rc != 0: + module.fail_json(msg="failed to remove %s package(s)" % (names), stdout=stdout, stderr=stderr, packages=packagelist) + module.exit_json(changed=True, msg="removed %s package(s)" % (names), stdout=stdout, stderr=stderr, packages=packagelist) + +# ========================================== +# Main control flow. + +def main(): + module = AnsibleModule( + argument_spec=dict( + state=dict(default='present', choices=['present', 'installed', 'absent', 'removed', 'latest']), + name=dict(type='list'), + repository=dict(type='list'), + update_cache=dict(default='no', type='bool'), + upgrade=dict(default='no', type='bool'), + available=dict(default='no', type='bool'), + ), + required_one_of=[['name', 'update_cache', 'upgrade']], + mutually_exclusive=[['name', 'upgrade']], + supports_check_mode=True + ) + + # Set LANG env since we parse stdout + module.run_command_environ_update = dict(LANG='C', LC_ALL='C', LC_MESSAGES='C', LC_CTYPE='C') + + global APK_PATH + APK_PATH = module.get_bin_path('apk', required=True) + + p = module.params + + # add repositories to the APK_PATH + if p['repository']: + for r in p['repository']: + APK_PATH = "%s --repository %s" % (APK_PATH, r) + + # normalize the state parameter + if p['state'] in ['present', 'installed']: + p['state'] = 'present' + if p['state'] in ['absent', 'removed']: + p['state'] = 'absent' + + if p['update_cache']: + update_package_db(module, not p['name'] and not p['upgrade']) + + if p['upgrade']: + upgrade_packages(module, p['available']) + + if p['state'] in ['present', 'latest']: + install_packages(module, p['name'], p['state']) + elif p['state'] == 'absent': + remove_packages(module, p['name']) + + +if __name__ == '__main__': + main() + +import time, sys +import h2o, h2o_hosts, h2o_import as h2i + +class SetupUnitTest(object): + def tearDown(self): + h2o.check_sandbox_for_errors() + + @classmethod + def setUpClass(cls): + global localhost + localhost = h2o.decide_if_localhost() + if (localhost): + params = collectConf(cls) + h2o.build_cloud(**params) + else: + h2o_hosts.build_cloud_with_hosts() + + @classmethod + def tearDownClass(cls): + # if we got here by time out exception waiting for a job, we should clear + # all jobs, if we're leaving h2o cloud up, and going to run another test + #h2o.cancelAllJobs() + h2o.tear_down_cloud() + + +def collectConf(cls): + result = { } + if hasattr(cls, 'nodes'): result['node_count'] = cls.nodes + if hasattr(cls, 'java_xmx'): result['java_heap_GB'] = cls.java_xmx + + return result + +# typical use in a unittest: +# class releaseTest(h2o_common.ReleaseCommon, unittest.TestCase): +# see multiple inheritance at http://docs.python.org/release/1.5/tut/node66.html +#************************************************************************************ +class SetupOneJVM14(object): + def tearDown(self): + h2o.check_sandbox_for_errors() + + @classmethod + def setUpClass(cls): + global localhost + localhost = h2o.decide_if_localhost() + if (localhost): + h2o.build_cloud(node_count=1, java_heap_GB=14) + else: + h2o_hosts.build_cloud_with_hosts() + + @classmethod + def tearDownClass(cls): + # if we got here by time out exception waiting for a job, we should clear + # all jobs, if we're leaving h2o cloud up, and going to run another test + #h2o.cancelAllJobs() + h2o.tear_down_cloud() + +#************************************************************************************ +class SetupThreeJVM4(object): + def tearDown(self): + h2o.check_sandbox_for_errors() + + @classmethod + def setUpClass(cls): + global localhost + localhost = h2o.decide_if_localhost() + if (localhost): + h2o.build_cloud(node_count=3, java_heap_GB=4) + else: + h2o_hosts.build_cloud_with_hosts() + + @classmethod + def tearDownClass(cls): + # if we got here by time out exception waiting for a job, we should clear + # all jobs, if we're leaving h2o cloud up, and going to run another test + #h2o.cancelAllJobs() + h2o.tear_down_cloud() + +#************************************************************************************ +class ReleaseCommon(object): + def tearDown(self): + print "tearDown" + # try to download the logs...may fail again! If we have no logs, h2o_sandbox will complain about not being able to look at anything + h2o.nodes[0].log_download() + h2o.check_sandbox_for_errors() + + @classmethod + def setUpClass(cls): + print "setUpClass" + # standard method for being able to reproduce the random.* seed + h2o.setup_random_seed() + # do a hack so we can run release tests with a -cj arg. so we don't have to build the cloud separately + # those tests will always want to run non-local (big machien) so detecting -cj is fine + if h2o.config_json: + h2o_hosts.build_cloud_with_hosts() + else: + # this is the normal thing for release tests (separate cloud was built. clone it) + h2o.build_cloud_with_json() + # if you're fast with a test and cloud building, you may need to wait for cloud to stabilize + # normally this is baked into build_cloud, but let's leave it here for now + h2o.stabilize_cloud(h2o.nodes[0], h2o.nodes, timeoutSecs=90) + # this ?should? work although not guaranteed all will agree on the cloud size + # unless we do the conservative stabilize above + h2o.verify_cloud_size() + + @classmethod + def tearDownClass(cls): + print "tearDownClass" + # DON"T + ### h2o.tear_down_cloud() + + # this could fail too + if h2o.nodes[0].delete_keys_at_teardown: + start = time.time() + h2i.delete_keys_at_all_nodes(timeoutSecs=300) + elapsed = time.time() - start + print "delete_keys_at_all_nodes(): took", elapsed, "secs" + +#************************************************************************************ +# no log download or key delete. Used for the test_shutdown.py +class ReleaseCommon2(object): + def tearDown(self): + print "tearDown" + h2o.check_sandbox_for_errors() + + @classmethod + def setUpClass(cls): + print "setUpClass" + h2o.build_cloud_with_json() + # normally this shouldn't be necessary? + h2o.stabilize_cloud(h2o.nodes[0], h2o.nodes, timeoutSecs=90) + + +#************************************************************************************ +# Notes: +# http://stackoverflow.com/questions/1323455/python-unit-test-with-base-and-sub-class +# +# This method only works for setUp and tearDown methods if you reverse the order of the base classes. +# Because the methods are defined in unittest.TestCase, and they don't call super(), +# then any setUp and tearDown methods in CommonTests need to be first in the MRO, +# or they won't be called at all. - Ian Clelland Oct 11 '10 +# +# If you add setUp and tearDown methods to CommonTests class, and you want them to be called for +# each test in derived classes, you have to reverse the order of the base classes, +# so that it will be: class SubTest1(CommonTests, unittest.TestCase). +# - Denis Golomazov July 17 +#************************************************************************************ + +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# Copyright: (c) 2018, Dag Wieers (@dagwieers) +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function +__metaclass__ = type + +ANSIBLE_METADATA = {'metadata_version': '1.1', + 'status': ['preview'], + 'supported_by': 'community'} + +DOCUMENTATION = r''' +--- +module: mso_tenant +short_description: Manage tenants +description: +- Manage tenants on Cisco ACI Multi-Site. +author: +- Dag Wieers (@dagwieers) +version_added: '2.8' +options: + tenant: + description: + - The name of the tenant. + type: str + required: yes + aliases: [ name ] + display_name: + description: + - The name of the tenant to be displayed in the web UI. + type: str + required: yes + description: + description: + - The description for this tenant. + type: str + users: + description: + - A list of associated users for this tenant. + - Using this property will replace any existing associated users. + type: list + sites: + description: + - A list of associated sites for this tenant. + - Using this property will replace any existing associated sites. + type: list + state: + description: + - Use C(present) or C(absent) for adding or removing. + - Use C(query) for listing an object or multiple objects. + type: str + choices: [ absent, present, query ] + default: present +extends_documentation_fragment: mso +''' + +EXAMPLES = r''' +- name: Add a new tenant + mso_tenant: + host: mso_host + username: admin + password: SomeSecretPassword + tenant: north_europe + display_name: North European Datacenter + description: This tenant manages the NEDC environment. + state: present + delegate_to: localhost + +- name: Remove a tenant + mso_tenant: + host: mso_host + username: admin + password: SomeSecretPassword + tenant: north_europe + state: absent + delegate_to: localhost + +- name: Query a tenant + mso_tenant: + host: mso_host + username: admin + password: SomeSecretPassword + tenant: north_europe + state: query + delegate_to: localhost + register: query_result + +- name: Query all tenants + mso_tenant: + host: mso_host + username: admin + password: SomeSecretPassword + state: query + delegate_to: localhost + register: query_result +''' + +RETURN = r''' +''' + +from ansible.module_utils.basic import AnsibleModule +from ansible.module_utils.network.aci.mso import MSOModule, mso_argument_spec, issubset + + +def main(): + argument_spec = mso_argument_spec() + argument_spec.update( + description=dict(type='str'), + display_name=dict(type='str'), + tenant=dict(type='str', aliases=['name']), + users=dict(type='list'), + sites=dict(type='list'), + state=dict(type='str', default='present', choices=['absent', 'present', 'query']), + ) + + module = AnsibleModule( + argument_spec=argument_spec, + supports_check_mode=True, + required_if=[ + ['state', 'absent', ['tenant']], + ['state', 'present', ['tenant']], + ], + ) + + description = module.params['description'] + display_name = module.params['display_name'] + tenant = module.params['tenant'] + state = module.params['state'] + + mso = MSOModule(module) + + # Convert sites and users + sites = mso.lookup_sites(module.params['sites']) + users = mso.lookup_users(module.params['users']) + + tenant_id = None + path = 'tenants' + + # Query for existing object(s) + if tenant: + mso.existing = mso.get_obj(path, name=tenant) + if mso.existing: + tenant_id = mso.existing['id'] + # If we found an existing object, continue with it + path = 'tenants/{id}'.format(id=tenant_id) + else: + mso.existing = mso.query_objs(path) + + if state == 'query': + pass + + elif state == 'absent': + mso.previous = mso.existing + if mso.existing: + if module.check_mode: + mso.existing = {} + else: + mso.existing = mso.request(path, method='DELETE') + + elif state == 'present': + mso.previous = mso.existing + + payload = dict( + description=description, + id=tenant_id, + name=tenant, + displayName=display_name, + siteAssociations=sites, + userAssociations=users, + ) + + mso.sanitize(payload, collate=True) + + # Ensure displayName is not undefined + if mso.sent.get('displayName') is None: + mso.sent['displayName'] = tenant + + # Ensure tenant has at least admin user + if mso.sent.get('userAssociations') is None: + mso.sent['userAssociations'] = [dict(userId="0000ffff0000000000000020")] + + if mso.existing: + if not issubset(mso.sent, mso.existing): + if module.check_mode: + mso.existing = mso.proposed + else: + mso.existing = mso.request(path, method='PUT', data=mso.sent) + else: + if module.check_mode: + mso.existing = mso.proposed + else: + mso.existing = mso.request(path, method='POST', data=mso.sent) + + mso.exit_json() + + +if __name__ == "__main__": + main() + +# -*- coding: utf-8 -*- +""" + pygments.styles + ~~~~~~~~~~~~~~~ + + Contains built-in styles. + + :copyright: Copyright 2006-2013 by the Pygments team, see AUTHORS. + :license: BSD, see LICENSE for details. +""" + +from pygments.plugin import find_plugin_styles +from pygments.util import ClassNotFound + + +#: Maps style names to 'submodule::classname'. +STYLE_MAP = { + 'default': 'default::DefaultStyle', + 'emacs': 'emacs::EmacsStyle', + 'friendly': 'friendly::FriendlyStyle', + 'colorful': 'colorful::ColorfulStyle', + 'autumn': 'autumn::AutumnStyle', + 'murphy': 'murphy::MurphyStyle', + 'manni': 'manni::ManniStyle', + 'monokai': 'monokai::MonokaiStyle', + 'perldoc': 'perldoc::PerldocStyle', + 'pastie': 'pastie::PastieStyle', + 'borland': 'borland::BorlandStyle', + 'trac': 'trac::TracStyle', + 'native': 'native::NativeStyle', + 'fruity': 'fruity::FruityStyle', + 'bw': 'bw::BlackWhiteStyle', + 'vim': 'vim::VimStyle', + 'vs': 'vs::VisualStudioStyle', + 'tango': 'tango::TangoStyle', + 'rrt': 'rrt::RrtStyle', +} + + +def get_style_by_name(name): + if name in STYLE_MAP: + mod, cls = STYLE_MAP[name].split('::') + builtin = "yes" + else: + for found_name, style in find_plugin_styles(): + if name == found_name: + return style + # perhaps it got dropped into our styles package + builtin = "" + mod = name + cls = name.title() + "Style" + + try: + mod = __import__('pygments.styles.' + mod, None, None, [cls]) + except ImportError: + raise ClassNotFound("Could not find style module %r" % mod + + (builtin and ", though it should be builtin") + ".") + try: + return getattr(mod, cls) + except AttributeError: + raise ClassNotFound("Could not find style class %r in style module." % cls) + + +def get_all_styles(): + """Return an generator for all styles by name, + both builtin and plugin.""" + for name in STYLE_MAP: + yield name + for name, _ in find_plugin_styles(): + yield name + +""" +Badge Awarding backend for Badgr-Server. +""" +from __future__ import absolute_import + +import hashlib +import logging +import mimetypes + +import requests +import six +from django.conf import settings +from django.core.exceptions import ImproperlyConfigured +from eventtracking import tracker +from lazy import lazy +from requests.packages.urllib3.exceptions import HTTPError + +from badges.backends.base import BadgeBackend +from badges.models import BadgeAssertion + +MAX_SLUG_LENGTH = 255 +LOGGER = logging.getLogger(__name__) + + +class BadgrBackend(BadgeBackend): + """ + Backend for Badgr-Server by Concentric Sky. http://info.badgr.io/ + """ + badges = [] + + def __init__(self): + super(BadgrBackend, self).__init__() + if not settings.BADGR_API_TOKEN: + raise ImproperlyConfigured("BADGR_API_TOKEN not set.") + + @lazy + def _base_url(self): + """ + Base URL for all API requests. + """ + return "{}/v1/issuer/issuers/{}".format(settings.BADGR_BASE_URL, settings.BADGR_ISSUER_SLUG) + + @lazy + def _badge_create_url(self): + """ + URL for generating a new Badge specification + """ + return "{}/badges".format(self._base_url) + + def _badge_url(self, slug): + """ + Get the URL for a course's badge in a given mode. + """ + return "{}/{}".format(self._badge_create_url, slug) + + def _assertion_url(self, slug): + """ + URL for generating a new assertion. + """ + return "{}/assertions".format(self._badge_url(slug)) + + def _slugify(self, badge_class): + """ + Get a compatible badge slug from the specification. + """ + slug = badge_class.issuing_component + badge_class.slug + if badge_class.issuing_component and badge_class.course_id: + # Make this unique to the course, and down to 64 characters. + # We don't do this to badges without issuing_component set for backwards compatibility. + slug = hashlib.sha256(slug + six.text_type(badge_class.course_id)).hexdigest() + if len(slug) > MAX_SLUG_LENGTH: + # Will be 64 characters. + slug = hashlib.sha256(slug).hexdigest() + return slug + + def _log_if_raised(self, response, data): + """ + Log server response if there was an error. + """ + try: + response.raise_for_status() + except HTTPError: + LOGGER.error( + u"Encountered an error when contacting the Badgr-Server. Request sent to %r with headers %r.\n" + u"and data values %r\n" + u"Response status was %s.\n%s", + response.request.url, response.request.headers, + data, + response.status_code, response.content + ) + raise + + def _create_badge(self, badge_class): + """ + Create the badge class on Badgr. + """ + image = badge_class.image + # We don't want to bother validating the file any further than making sure we can detect its MIME type, + # for HTTP. The Badgr-Server should tell us if there's anything in particular wrong with it. + content_type, __ = mimetypes.guess_type(image.name) + if not content_type: + raise ValueError( + u"Could not determine content-type of image! Make sure it is a properly named .png file. " + u"Filename was: {}".format(image.name) + ) + files = {'image': (image.name, image, content_type)} + data = { + 'name': badge_class.display_name, + 'criteria': badge_class.criteria, + 'slug': self._slugify(badge_class), + 'description': badge_class.description, + } + result = requests.post( + self._badge_create_url, headers=self._get_headers(), data=data, files=files, + timeout=settings.BADGR_TIMEOUT + ) + self._log_if_raised(result, data) + + def _send_assertion_created_event(self, user, assertion): + """ + Send an analytics event to record the creation of a badge assertion. + """ + tracker.emit( + 'edx.badge.assertion.created', { + 'user_id': user.id, + 'badge_slug': assertion.badge_class.slug, + 'badge_name': assertion.badge_class.display_name, + 'issuing_component': assertion.badge_class.issuing_component, + 'course_id': six.text_type(assertion.badge_class.course_id), + 'enrollment_mode': assertion.badge_class.mode, + 'assertion_id': assertion.id, + 'assertion_image_url': assertion.image_url, + 'assertion_json_url': assertion.assertion_url, + 'issuer': assertion.data.get('issuer'), + } + ) + + def _create_assertion(self, badge_class, user, evidence_url): + """ + Register an assertion with the Badgr server for a particular user for a specific class. + """ + data = { + 'email': user.email, + 'evidence': evidence_url, + } + response = requests.post( + self._assertion_url(self._slugify(badge_class)), headers=self._get_headers(), data=data, + timeout=settings.BADGR_TIMEOUT + ) + self._log_if_raised(response, data) + assertion, __ = BadgeAssertion.objects.get_or_create(user=user, badge_class=badge_class) + assertion.data = response.json() + assertion.backend = 'BadgrBackend' + assertion.image_url = assertion.data['image'] + assertion.assertion_url = assertion.data['json']['id'] + assertion.save() + self._send_assertion_created_event(user, assertion) + return assertion + + @staticmethod + def _get_headers(): + """ + Headers to send along with the request-- used for authentication. + """ + return {'Authorization': u'Token {}'.format(settings.BADGR_API_TOKEN)} + + def _ensure_badge_created(self, badge_class): + """ + Verify a badge has been created for this badge class, and create it if not. + """ + slug = self._slugify(badge_class) + if slug in BadgrBackend.badges: + return + response = requests.get(self._badge_url(slug), headers=self._get_headers(), timeout=settings.BADGR_TIMEOUT) + if response.status_code != 200: + self._create_badge(badge_class) + BadgrBackend.badges.append(slug) + + def award(self, badge_class, user, evidence_url=None): + """ + Make sure the badge class has been created on the backend, and then award the badge class to the user. + """ + self._ensure_badge_created(badge_class) + return self._create_assertion(badge_class, user, evidence_url) + +import datetime +from django.contrib.auth.models import User +from django.contrib.comments.managers import CommentManager +from django.contrib.contenttypes import generic +from django.contrib.contenttypes.models import ContentType +from django.contrib.sites.models import Site +from django.db import models +from django.core import urlresolvers +from django.utils.translation import ugettext_lazy as _ +from django.conf import settings + +COMMENT_MAX_LENGTH = getattr(settings,'COMMENT_MAX_LENGTH',3000) + +class BaseCommentAbstractModel(models.Model): + """ + An abstract base class that any custom comment models probably should + subclass. + """ + + # Content-object field + content_type = models.ForeignKey(ContentType, + verbose_name=_('content type'), + related_name="content_type_set_for_%(class)s") + object_pk = models.TextField(_('object ID')) + content_object = generic.GenericForeignKey(ct_field="content_type", fk_field="object_pk") + + # Metadata about the comment + site = models.ForeignKey(Site) + + class Meta: + abstract = True + + def get_content_object_url(self): + """ + Get a URL suitable for redirecting to the content object. + """ + return urlresolvers.reverse( + "comments-url-redirect", + args=(self.content_type_id, self.object_pk) + ) + +class Comment(BaseCommentAbstractModel): + """ + A user comment about some object. + """ + + # Who posted this comment? If ``user`` is set then it was an authenticated + # user; otherwise at least user_name should have been set and the comment + # was posted by a non-authenticated user. + user = models.ForeignKey(User, verbose_name=_('user'), + blank=True, null=True, related_name="%(class)s_comments") + user_name = models.CharField(_("user's name"), max_length=50, blank=True) + user_email = models.EmailField(_("user's email address"), blank=True) + user_url = models.URLField(_("user's URL"), blank=True) + + comment = models.TextField(_('comment'), max_length=COMMENT_MAX_LENGTH) + + # Metadata about the comment + submit_date = models.DateTimeField(_('date/time submitted'), default=None) + ip_address = models.IPAddressField(_('IP address'), blank=True, null=True) + is_public = models.BooleanField(_('is public'), default=True, + help_text=_('Uncheck this box to make the comment effectively ' \ + 'disappear from the site.')) + is_removed = models.BooleanField(_('is removed'), default=False, + help_text=_('Check this box if the comment is inappropriate. ' \ + 'A "This comment has been removed" message will ' \ + 'be displayed instead.')) + + # Manager + objects = CommentManager() + + class Meta: + db_table = "django_comments" + ordering = ('submit_date',) + permissions = [("can_moderate", "Can moderate comments")] + verbose_name = _('comment') + verbose_name_plural = _('comments') + + def __unicode__(self): + return "%s: %s..." % (self.name, self.comment[:50]) + + def save(self, *args, **kwargs): + if self.submit_date is None: + self.submit_date = datetime.datetime.now() + super(Comment, self).save(*args, **kwargs) + + def _get_userinfo(self): + """ + Get a dictionary that pulls together information about the poster + safely for both authenticated and non-authenticated comments. + + This dict will have ``name``, ``email``, and ``url`` fields. + """ + if not hasattr(self, "_userinfo"): + self._userinfo = { + "name" : self.user_name, + "email" : self.user_email, + "url" : self.user_url + } + if self.user_id: + u = self.user + if u.email: + self._userinfo["email"] = u.email + + # If the user has a full name, use that for the user name. + # However, a given user_name overrides the raw user.username, + # so only use that if this comment has no associated name. + if u.get_full_name(): + self._userinfo["name"] = self.user.get_full_name() + elif not self.user_name: + self._userinfo["name"] = u.username + return self._userinfo + userinfo = property(_get_userinfo, doc=_get_userinfo.__doc__) + + def _get_name(self): + return self.userinfo["name"] + def _set_name(self, val): + if self.user_id: + raise AttributeError(_("This comment was posted by an authenticated "\ + "user and thus the name is read-only.")) + self.user_name = val + name = property(_get_name, _set_name, doc="The name of the user who posted this comment") + + def _get_email(self): + return self.userinfo["email"] + def _set_email(self, val): + if self.user_id: + raise AttributeError(_("This comment was posted by an authenticated "\ + "user and thus the email is read-only.")) + self.user_email = val + email = property(_get_email, _set_email, doc="The email of the user who posted this comment") + + def _get_url(self): + return self.userinfo["url"] + def _set_url(self, val): + self.user_url = val + url = property(_get_url, _set_url, doc="The URL given by the user who posted this comment") + + def get_absolute_url(self, anchor_pattern="#c%(id)s"): + return self.get_content_object_url() + (anchor_pattern % self.__dict__) + + def get_as_text(self): + """ + Return this comment as plain text. Useful for emails. + """ + d = { + 'user': self.user or self.name, + 'date': self.submit_date, + 'comment': self.comment, + 'domain': self.site.domain, + 'url': self.get_absolute_url() + } + return _('Posted by %(user)s at %(date)s\n\n%(comment)s\n\nhttp://%(domain)s%(url)s') % d + +class CommentFlag(models.Model): + """ + Records a flag on a comment. This is intentionally flexible; right now, a + flag could be: + + * A "removal suggestion" -- where a user suggests a comment for (potential) removal. + + * A "moderator deletion" -- used when a moderator deletes a comment. + + You can (ab)use this model to add other flags, if needed. However, by + design users are only allowed to flag a comment with a given flag once; + if you want rating look elsewhere. + """ + user = models.ForeignKey(User, verbose_name=_('user'), related_name="comment_flags") + comment = models.ForeignKey(Comment, verbose_name=_('comment'), related_name="flags") + flag = models.CharField(_('flag'), max_length=30, db_index=True) + flag_date = models.DateTimeField(_('date'), default=None) + + # Constants for flag types + SUGGEST_REMOVAL = "removal suggestion" + MODERATOR_DELETION = "moderator deletion" + MODERATOR_APPROVAL = "moderator approval" + + class Meta: + db_table = 'django_comment_flags' + unique_together = [('user', 'comment', 'flag')] + verbose_name = _('comment flag') + verbose_name_plural = _('comment flags') + + def __unicode__(self): + return "%s flag of comment ID %s by %s" % \ + (self.flag, self.comment_id, self.user.username) + + def save(self, *args, **kwargs): + if self.flag_date is None: + self.flag_date = datetime.datetime.now() + super(CommentFlag, self).save(*args, **kwargs) + +#===--- protocol_graph.py -------------------------*- coding: utf-8 -*----===# +# +# This source file is part of the Swift.org open source project +# +# Copyright (c) 2014 - 2015 Apple Inc. and the Swift project authors +# Licensed under Apache License v2.0 with Runtime Library Exception +# +# See http://swift.org/LICENSE.txt for license information +# See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors +# +#===----------------------------------------------------------------------===# +# +# Create a graph of the protocol refinement relationships, associated +# types, operator requirements, and defaulted generic operators. +# +# run as follows to view the Nth-largest connected component in a web browser: +# +# N=0 && rm -f /tmp/protocols.dot && \ +# python protocol_graph.py stdlib.swift > /tmp/p0.dot && \ +# (ccomps -zX#$N -o /tmp/protocols.dot /tmp/p0.dot || true) \ +# && dot -Tsvg /tmp/protocols.dot > /tmp/protocols.svg \ +# && open /tmp/protocols.svg +#===----------------------------------------------------------------------===# + +import re +import sys +import os +import cgi + +# Open 'stdlib.swift' in this directory if no path specified. +args = list(sys.argv) + [os.path.join(os.path.dirname(__file__), 'stdlib.swift')] + +reFlags = re.MULTILINE | re.VERBOSE + +# Pattern to recognize stdlib identifiers (FIXME: doesn't handle Unicode). +identifier = '[A-Za-z_][A-Za-z0-9_]*' + +# Pattern to recognize a (possibly-generic) operator decl. +operator = r''' +(?:(?:prefix|postfix).*)? func \s* +(?=\S)[^A-Za-z_] # non-space, non-identifier: begins an operator name +(?:(?=\S)[^(){])* # rest of operator name +\s* +(<[^>{]+>)? # generic parameter list +\s* +\([^)]*\) # function parameter list +''' + +# substitute local variables into the string +def interpolate(string): + import inspect + frame = inspect.currentframe() + return string % frame.f_back.f_locals + +# Given the bodyText of a protocol definition, return a list of +# associated type and operator requirements. +def bodyLines(bodyText): + return [ + cgi.escape(b.group(0)) for b in + re.finditer( + r'(typealias\s*'+identifier+r'(\s*[:,]\s*'+identifier + ')?|' + operator + '.*)', + bodyText, reFlags) + ] + +# Mapping from protocol to associated type / operator requirements +body = {} + +# Mapping from a parent protocol to set of children. +graph = {} + +# Mapping from protocol to generic operators taking instances as arguments +genericOperators = {} + +comments = r'//.* | /[*] (.|\n)*? [*]/' # FIXME: doesn't respect strings or comment nesting) + +# read source, stripping all comments +sourceSansComments = re.sub(comments, '', open(args[1]).read(), flags=reFlags) + +genericParameterConstraint = interpolate(r' (%(identifier)s) \s* : \s* (%(identifier)s) ') + +def parseGenericOperator(m): + genericParams = m.group(5) + genericOperator = cgi.escape(m.group(0).strip()) + functionParamStart = m.end(5) - m.start(0) + functionParams = genericOperator[functionParamStart:] + for m2 in re.finditer(genericParameterConstraint, genericParams, reFlags): + typeParameter = m2.group(1) + protocol = m2.group(2) + + # we're only interested if we can find a function parameter of that type + if not re.search(r':\s*%s\s*[,)]' % typeParameter, functionParams): continue + + # Make some replacements in the signature to limit the graph size + letterTau = 'τ' + letterPi = 'π' + abbreviatedSignature = re.sub( + r'\b%s\b' % protocol, letterPi, + re.sub(r'\b%s\b' % typeParameter, letterTau, genericOperator)) + + genericOperators.setdefault(protocol, set()).add(abbreviatedSignature) + +def parseProtocol(m): + child = m.group(1) + # skip irrelevant protocols + if re.match(r'_Builtin.*Convertible', child): return + graph.setdefault(child, set()) + body[child] = bodyLines(m.group(3)) + if m.group(2): + for parent in m.group(2).strip().split(","): + if re.match(r'_Builtin.*Convertible', parent): return + graph.setdefault(parent.strip(), set()).add(child) + +protocolsAndOperators = interpolate(r''' +\bprotocol \s+ (%(identifier)s) \s* + (?::\s*([^{]+))? # refinements + {([^{}\n]*(.*\n)*?)} # body +| +%(operator)s [^{]*(?={) # operator definition up to the open brace +''') + +# Main parsing loop +for m in re.finditer(protocolsAndOperators, sourceSansComments, reFlags): + if m.group(1): parseProtocol(m) + elif m.group(5): parseGenericOperator(m) + # otherwise we matched some non-generic operator + +# Find clusters of protocols that have the same name when underscores +# are stripped +clusterBuilder = {} # map from potential cluster name to nodes in the cluster +for n in graph: + clusterBuilder.setdefault(n.translate(None, '_'), set()).add(n) + +# Grab the clusters with more than one member. +clusters = dict((c, nodes) for (c, nodes) in clusterBuilder.items() if len(nodes) > 1) + +# A set of all intra-cluster edges +clusterEdges = set( + (s, t) for (c, elements) in clusters.items() + for s in elements + for t in graph[s] if t in elements) + +print 'digraph ProtocolHierarchies {' +print ' mclimit = 100; ranksep=1.5; ' # ; packmode="array1" +print ' edge [dir="back"];' +print ' node [shape = box, fontname = Helvetica, fontsize = 10];' + +for c in sorted(clusters): + print ' subgraph "cluster_%s" {' % c + for (s, t) in sorted(clusterEdges): + if s in clusters[c]: + print '%s -> %s [weight=100];' % (s, t) + print '}' + +for node in sorted(graph.keys()): + requirements = body.get(node, []) + generics = sorted(genericOperators.get(node, set())) + style = 'solid' if node.startswith('_') else 'bold' + divider = '
\n' if len(requirements) != 0 and len(generics) != 0 else '' + + label = node if len(requirements + generics) == 0 else ( + '\n\n
\n%s%s%s
\n%s\n
\n' % ( + node, + '\n'.join('%s' % r for r in requirements), + divider, + '\n'.join('%s' % g for g in generics))) + + print interpolate(' %(node)s [style = %(style)s, label=<%(label)s>]') + +for (parent, children) in sorted(graph.items()): + print ' %s -> {' % parent, + print '; '.join( + sorted(child for child in children if not (parent, child) in clusterEdges)), + print '}' + +print '}' + +# (c) 2014, Maciej Delmanowski +# +# This file is part of Ansible +# +# Ansible is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Ansible is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with Ansible. If not, see . + +from functools import partial +import types + +try: + import netaddr +except ImportError: + # in this case, we'll make the filters return error messages (see bottom) + netaddr = None +else: + class mac_linux(netaddr.mac_unix): + pass + mac_linux.word_fmt = '%.2x' + +from ansible import errors + + +# ---- IP address and network query helpers ---- + +def _empty_ipaddr_query(v, vtype): + # We don't have any query to process, so just check what type the user + # expects, and return the IP address in a correct format + if v: + if vtype == 'address': + return str(v.ip) + elif vtype == 'network': + return str(v) + +def _6to4_query(v, vtype, value): + if v.version == 4: + + if v.size == 1: + ipconv = str(v.ip) + elif v.size > 1: + if v.ip != v.network: + ipconv = str(v.ip) + else: + ipconv = False + + if ipaddr(ipconv, 'public'): + numbers = list(map(int, ipconv.split('.'))) + + try: + return '2002:{:02x}{:02x}:{:02x}{:02x}::1/48'.format(*numbers) + except: + return False + + elif v.version == 6: + if vtype == 'address': + if ipaddr(str(v), '2002::/16'): + return value + elif vtype == 'network': + if v.ip != v.network: + if ipaddr(str(v.ip), '2002::/16'): + return value + else: + return False + +def _ip_query(v): + if v.size == 1: + return str(v.ip) + if v.size > 1: + if v.ip != v.network: + return str(v.ip) + +def _gateway_query(v): + if v.size > 1: + if v.ip != v.network: + return str(v.ip) + '/' + str(v.prefixlen) + +def _bool_ipaddr_query(v): + if v: + return True + +def _broadcast_query(v): + if v.size > 1: + return str(v.broadcast) + +def _cidr_query(v): + return str(v) + +def _cidr_lookup_query(v, iplist, value): + try: + if v in iplist: + return value + except: + return False + +def _host_query(v): + if v.size == 1: + return str(v) + elif v.size > 1: + if v.ip != v.network: + return str(v.ip) + '/' + str(v.prefixlen) + +def _hostmask_query(v): + return str(v.hostmask) + +def _int_query(v, vtype): + if vtype == 'address': + return int(v.ip) + elif vtype == 'network': + return str(int(v.ip)) + '/' + str(int(v.prefixlen)) + +def _ipv4_query(v, value): + if v.version == 6: + try: + return str(v.ipv4()) + except: + return False + else: + return value + +def _ipv6_query(v, value): + if v.version == 4: + return str(v.ipv6()) + else: + return value + +def _link_local_query(v, value): + v_ip = netaddr.IPAddress(str(v.ip)) + if v.version == 4: + if ipaddr(str(v_ip), '169.254.0.0/24'): + return value + + elif v.version == 6: + if ipaddr(str(v_ip), 'fe80::/10'): + return value + +def _loopback_query(v, value): + v_ip = netaddr.IPAddress(str(v.ip)) + if v_ip.is_loopback(): + return value + +def _multicast_query(v, value): + if v.is_multicast(): + return value + +def _net_query(v): + if v.size > 1: + if v.ip == v.network: + return str(v.network) + '/' + str(v.prefixlen) + +def _netmask_query(v): + if v.size > 1: + return str(v.netmask) + +def _network_query(v): + if v.size > 1: + return str(v.network) + +def _prefix_query(v): + return int(v.prefixlen) + +def _private_query(v, value): + if v.is_private(): + return value + +def _public_query(v, value): + v_ip = netaddr.IPAddress(str(v.ip)) + if v_ip.is_unicast() and not v_ip.is_private() and \ + not v_ip.is_loopback() and not v_ip.is_netmask() and \ + not v_ip.is_hostmask(): + return value + +def _revdns_query(v): + v_ip = netaddr.IPAddress(str(v.ip)) + return v_ip.reverse_dns + +def _size_query(v): + return v.size + +def _subnet_query(v): + return str(v.cidr) + +def _type_query(v): + if v.size == 1: + return 'address' + if v.size > 1: + if v.ip != v.network: + return 'address' + else: + return 'network' + +def _unicast_query(v, value): + if v.is_unicast(): + return value + +def _version_query(v): + return v.version + +def _wrap_query(v, vtype, value): + if v.version == 6: + if vtype == 'address': + return '[' + str(v.ip) + ']' + elif vtype == 'network': + return '[' + str(v.ip) + ']/' + str(v.prefixlen) + else: + return value + + +# ---- HWaddr query helpers ---- +def _bare_query(v): + v.dialect = netaddr.mac_bare + return str(v) + +def _bool_hwaddr_query(v): + if v: + return True + +def _cisco_query(v): + v.dialect = netaddr.mac_cisco + return str(v) + +def _empty_hwaddr_query(v, value): + if v: + return value + +def _linux_query(v): + v.dialect = mac_linux + return str(v) + +def _postgresql_query(v): + v.dialect = netaddr.mac_pgsql + return str(v) + +def _unix_query(v): + v.dialect = netaddr.mac_unix + return str(v) + +def _win_query(v): + v.dialect = netaddr.mac_eui48 + return str(v) + + +# ---- IP address and network filters ---- + +def ipaddr(value, query = '', version = False, alias = 'ipaddr'): + ''' Check if string is an IP address or network and filter it ''' + + query_func_extra_args = { + '': ('vtype',), + '6to4': ('vtype', 'value'), + 'cidr_lookup': ('iplist', 'value'), + 'int': ('vtype',), + 'ipv4': ('value',), + 'ipv6': ('value',), + 'link-local': ('value',), + 'loopback': ('value',), + 'lo': ('value',), + 'multicast': ('value',), + 'private': ('value',), + 'public': ('value',), + 'unicast': ('value',), + 'wrap': ('vtype', 'value'), + } + query_func_map = { + '': _empty_ipaddr_query, + '6to4': _6to4_query, + 'address': _ip_query, + 'address/prefix': _gateway_query, + 'bool': _bool_ipaddr_query, + 'broadcast': _broadcast_query, + 'cidr': _cidr_query, + 'cidr_lookup': _cidr_lookup_query, + 'gateway': _gateway_query, + 'gw': _gateway_query, + 'host': _host_query, + 'host/prefix': _gateway_query, + 'hostmask': _hostmask_query, + 'hostnet': _gateway_query, + 'int': _int_query, + 'ip': _ip_query, + 'ipv4': _ipv4_query, + 'ipv6': _ipv6_query, + 'link-local': _link_local_query, + 'lo': _loopback_query, + 'loopback': _loopback_query, + 'multicast': _multicast_query, + 'net': _net_query, + 'netmask': _netmask_query, + 'network': _network_query, + 'prefix': _prefix_query, + 'private': _private_query, + 'public': _public_query, + 'revdns': _revdns_query, + 'router': _gateway_query, + 'size': _size_query, + 'subnet': _subnet_query, + 'type': _type_query, + 'unicast': _unicast_query, + 'v4': _ipv4_query, + 'v6': _ipv6_query, + 'version': _version_query, + 'wrap': _wrap_query, + } + + vtype = None + + if not value: + return False + + elif value == True: + return False + + # Check if value is a list and parse each element + elif isinstance(value, (list, tuple, types.GeneratorType)): + + _ret = [] + for element in value: + if ipaddr(element, str(query), version): + _ret.append(ipaddr(element, str(query), version)) + + if _ret: + return _ret + else: + return list() + + # Check if value is a number and convert it to an IP address + elif str(value).isdigit(): + + # We don't know what IP version to assume, so let's check IPv4 first, + # then IPv6 + try: + if ((not version) or (version and version == 4)): + v = netaddr.IPNetwork('0.0.0.0/0') + v.value = int(value) + v.prefixlen = 32 + elif version and version == 6: + v = netaddr.IPNetwork('::/0') + v.value = int(value) + v.prefixlen = 128 + + # IPv4 didn't work the first time, so it definitely has to be IPv6 + except: + try: + v = netaddr.IPNetwork('::/0') + v.value = int(value) + v.prefixlen = 128 + + # The value is too big for IPv6. Are you a nanobot? + except: + return False + + # We got an IP address, let's mark it as such + value = str(v) + vtype = 'address' + + # value has not been recognized, check if it's a valid IP string + else: + try: + v = netaddr.IPNetwork(value) + + # value is a valid IP string, check if user specified + # CIDR prefix or just an IP address, this will indicate default + # output format + try: + address, prefix = value.split('/') + vtype = 'network' + except: + vtype = 'address' + + # value hasn't been recognized, maybe it's a numerical CIDR? + except: + try: + address, prefix = value.split('/') + address.isdigit() + address = int(address) + prefix.isdigit() + prefix = int(prefix) + + # It's not numerical CIDR, give up + except: + return False + + # It is something, so let's try and build a CIDR from the parts + try: + v = netaddr.IPNetwork('0.0.0.0/0') + v.value = address + v.prefixlen = prefix + + # It's not a valid IPv4 CIDR + except: + try: + v = netaddr.IPNetwork('::/0') + v.value = address + v.prefixlen = prefix + + # It's not a valid IPv6 CIDR. Give up. + except: + return False + + # We have a valid CIDR, so let's write it in correct format + value = str(v) + vtype = 'network' + + # We have a query string but it's not in the known query types. Check if + # that string is a valid subnet, if so, we can check later if given IP + # address/network is inside that specific subnet + try: + ### ?? 6to4 and link-local were True here before. Should they still? + if query and (query not in query_func_map or query == 'cidr_lookup') and ipaddr(query, 'network'): + iplist = netaddr.IPSet([netaddr.IPNetwork(query)]) + query = 'cidr_lookup' + except: + pass + + # This code checks if value maches the IP version the user wants, ie. if + # it's any version ("ipaddr()"), IPv4 ("ipv4()") or IPv6 ("ipv6()") + # If version does not match, return False + if version and v.version != version: + return False + + extras = [] + for arg in query_func_extra_args.get(query, tuple()): + extras.append(locals()[arg]) + try: + return query_func_map[query](v, *extras) + except KeyError: + try: + float(query) + if v.size == 1: + if vtype == 'address': + return str(v.ip) + elif vtype == 'network': + return str(v) + + elif v.size > 1: + try: + return str(v[query]) + '/' + str(v.prefixlen) + except: + return False + + else: + return value + + except: + raise errors.AnsibleFilterError(alias + ': unknown filter type: %s' % query) + + return False + + +def ipwrap(value, query = ''): + try: + if isinstance(value, (list, tuple, types.GeneratorType)): + _ret = [] + for element in value: + if ipaddr(element, query, version = False, alias = 'ipwrap'): + _ret.append(ipaddr(element, 'wrap')) + else: + _ret.append(element) + + return _ret + else: + _ret = ipaddr(value, query, version = False, alias = 'ipwrap') + if _ret: + return ipaddr(_ret, 'wrap') + else: + return value + + except: + return value + + +def ipv4(value, query = ''): + return ipaddr(value, query, version = 4, alias = 'ipv4') + + +def ipv6(value, query = ''): + return ipaddr(value, query, version = 6, alias = 'ipv6') + + +# Split given subnet into smaller subnets or find out the biggest subnet of +# a given IP address with given CIDR prefix +# Usage: +# +# - address or address/prefix | ipsubnet +# returns CIDR subnet of a given input +# +# - address/prefix | ipsubnet(cidr) +# returns number of possible subnets for given CIDR prefix +# +# - address/prefix | ipsubnet(cidr, index) +# returns new subnet with given CIDR prefix +# +# - address | ipsubnet(cidr) +# returns biggest subnet with given CIDR prefix that address belongs to +# +# - address | ipsubnet(cidr, index) +# returns next indexed subnet which contains given address +def ipsubnet(value, query = '', index = 'x'): + ''' Manipulate IPv4/IPv6 subnets ''' + + try: + vtype = ipaddr(value, 'type') + if vtype == 'address': + v = ipaddr(value, 'cidr') + elif vtype == 'network': + v = ipaddr(value, 'subnet') + + value = netaddr.IPNetwork(v) + except: + return False + + if not query: + return str(value) + + elif str(query).isdigit(): + vsize = ipaddr(v, 'size') + query = int(query) + + try: + float(index) + index = int(index) + + if vsize > 1: + try: + return str(list(value.subnet(query))[index]) + except: + return False + + elif vsize == 1: + try: + return str(value.supernet(query)[index]) + except: + return False + + except: + if vsize > 1: + try: + return str(len(list(value.subnet(query)))) + except: + return False + + elif vsize == 1: + try: + return str(value.supernet(query)[0]) + except: + return False + + return False + +# Returns the nth host within a network described by value. +# Usage: +# +# - address or address/prefix | nthhost(nth) +# returns the nth host within the given network +def nthhost(value, query=''): + ''' Get the nth host within a given network ''' + try: + vtype = ipaddr(value, 'type') + if vtype == 'address': + v = ipaddr(value, 'cidr') + elif vtype == 'network': + v = ipaddr(value, 'subnet') + + value = netaddr.IPNetwork(v) + except: + return False + + if not query: + return False + + try: + vsize = ipaddr(v, 'size') + nth = int(query) + if value.size > nth: + return value[nth] + + except ValueError: + return False + + return False + +# Returns the SLAAC address within a network for a given HW/MAC address. +# Usage: +# +# - prefix | slaac(mac) +def slaac(value, query = ''): + ''' Get the SLAAC address within given network ''' + try: + vtype = ipaddr(value, 'type') + if vtype == 'address': + v = ipaddr(value, 'cidr') + elif vtype == 'network': + v = ipaddr(value, 'subnet') + + if v.version != 6: + return False + + value = netaddr.IPNetwork(v) + except: + return False + + if not query: + return False + + try: + mac = hwaddr(query, alias = 'slaac') + + eui = netaddr.EUI(mac) + except: + return False + + return eui.ipv6(value.network) + + +# ---- HWaddr / MAC address filters ---- + +def hwaddr(value, query = '', alias = 'hwaddr'): + ''' Check if string is a HW/MAC address and filter it ''' + + query_func_extra_args = { + '': ('value',), + } + query_func_map = { + '': _empty_hwaddr_query, + 'bare': _bare_query, + 'bool': _bool_hwaddr_query, + 'cisco': _cisco_query, + 'eui48': _win_query, + 'linux': _linux_query, + 'pgsql': _postgresql_query, + 'postgresql': _postgresql_query, + 'psql': _postgresql_query, + 'unix': _unix_query, + 'win': _win_query, + } + + try: + v = netaddr.EUI(value) + except: + if query and query != 'bool': + raise errors.AnsibleFilterError(alias + ': not a hardware address: %s' % value) + + extras = [] + for arg in query_func_extra_args.get(query, tuple()): + extras.append(locals()[arg]) + try: + return query_func_map[query](v, *extras) + except KeyError: + raise errors.AnsibleFilterError(alias + ': unknown filter type: %s' % query) + + return False + +def macaddr(value, query = ''): + return hwaddr(value, query, alias = 'macaddr') + +def _need_netaddr(f_name, *args, **kwargs): + raise errors.AnsibleFilterError('The {0} filter requires python-netaddr be' + ' installed on the ansible controller'.format(f_name)) + +# ---- Ansible filters ---- + +class FilterModule(object): + ''' IP address and network manipulation filters ''' + filter_map = { + # IP addresses and networks + 'ipaddr': ipaddr, + 'ipwrap': ipwrap, + 'ipv4': ipv4, + 'ipv6': ipv6, + 'ipsubnet': ipsubnet, + 'nthhost': nthhost, + 'slaac': slaac, + + # MAC / HW addresses + 'hwaddr': hwaddr, + 'macaddr': macaddr + } + + def filters(self): + if netaddr: + return self.filter_map + else: + # Need to install python-netaddr for these filters to work + return dict((f, partial(_need_netaddr, f)) for f in self.filter_map) + +## features.py +## +## Copyright (C) 2003-2004 Alexey "Snake" Nezhdanov +## +## This program is free software; you can redistribute it and/or modify +## it under the terms of the GNU General Public License as published by +## the Free Software Foundation; either version 2, or (at your option) +## any later version. +## +## This program is distributed in the hope that it will be useful, +## but WITHOUT ANY WARRANTY; without even the implied warranty of +## MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +## GNU General Public License for more details. + +# $Id: features.py,v 1.25 2009/04/07 07:11:48 snakeru Exp $ + +""" +This module contains variable stuff that is not worth splitting into separate modules. +Here is: + DISCO client and agents-to-DISCO and browse-to-DISCO emulators. + IBR and password manager. + jabber:iq:privacy methods +All these methods takes 'disp' first argument that should be already connected +(and in most cases already authorised) dispatcher instance. +""" + +from protocol import * + +REGISTER_DATA_RECEIVED='REGISTER DATA RECEIVED' + +### DISCO ### http://jabber.org/protocol/disco ### JEP-0030 #################### +### Browse ### jabber:iq:browse ### JEP-0030 ################################### +### Agents ### jabber:iq:agents ### JEP-0030 ################################### +def _discover(disp,ns,jid,node=None,fb2b=0,fb2a=1): + """ Try to obtain info from the remote object. + If remote object doesn't support disco fall back to browse (if fb2b is true) + and if it doesnt support browse (or fb2b is not true) fall back to agents protocol + (if gb2a is true). Returns obtained info. Used internally. """ + iq=Iq(to=jid,typ='get',queryNS=ns) + if node: iq.setQuerynode(node) + rep=disp.SendAndWaitForResponse(iq) + if fb2b and not isResultNode(rep): rep=disp.SendAndWaitForResponse(Iq(to=jid,typ='get',queryNS=NS_BROWSE)) # Fallback to browse + if fb2a and not isResultNode(rep): rep=disp.SendAndWaitForResponse(Iq(to=jid,typ='get',queryNS=NS_AGENTS)) # Fallback to agents + if isResultNode(rep): return [n for n in rep.getQueryPayload() if isinstance(n, Node)] + return [] + +def discoverItems(disp,jid,node=None): + """ Query remote object about any items that it contains. Return items list. """ + """ According to JEP-0030: + query MAY have node attribute + item: MUST HAVE jid attribute and MAY HAVE name, node, action attributes. + action attribute of item can be either of remove or update value.""" + ret=[] + for i in _discover(disp,NS_DISCO_ITEMS,jid,node): + if i.getName()=='agent' and i.getTag('name'): i.setAttr('name',i.getTagData('name')) + ret.append(i.attrs) + return ret + +def discoverInfo(disp,jid,node=None): + """ Query remote object about info that it publishes. Returns identities and features lists.""" + """ According to JEP-0030: + query MAY have node attribute + identity: MUST HAVE category and name attributes and MAY HAVE type attribute. + feature: MUST HAVE var attribute""" + identities , features = [] , [] + for i in _discover(disp,NS_DISCO_INFO,jid,node): + if i.getName()=='identity': identities.append(i.attrs) + elif i.getName()=='feature': features.append(i.getAttr('var')) + elif i.getName()=='agent': + if i.getTag('name'): i.setAttr('name',i.getTagData('name')) + if i.getTag('description'): i.setAttr('name',i.getTagData('description')) + identities.append(i.attrs) + if i.getTag('groupchat'): features.append(NS_GROUPCHAT) + if i.getTag('register'): features.append(NS_REGISTER) + if i.getTag('search'): features.append(NS_SEARCH) + return identities , features + +### Registration ### jabber:iq:register ### JEP-0077 ########################### +def getRegInfo(disp,host,info={},sync=True): + """ Gets registration form from remote host. + You can pre-fill the info dictionary. + F.e. if you are requesting info on registering user joey than specify + info as {'username':'joey'}. See JEP-0077 for details. + 'disp' must be connected dispatcher instance.""" + iq=Iq('get',NS_REGISTER,to=host) + for i in info.keys(): iq.setTagData(i,info[i]) + if sync: + resp=disp.SendAndWaitForResponse(iq) + _ReceivedRegInfo(disp.Dispatcher,resp, host) + return resp + else: disp.SendAndCallForResponse(iq,_ReceivedRegInfo, {'agent': host}) + +def _ReceivedRegInfo(con, resp, agent): + iq=Iq('get',NS_REGISTER,to=agent) + if not isResultNode(resp): return + df=resp.getTag('query',namespace=NS_REGISTER).getTag('x',namespace=NS_DATA) + if df: + con.Event(NS_REGISTER,REGISTER_DATA_RECEIVED,(agent, DataForm(node=df))) + return + df=DataForm(typ='form') + for i in resp.getQueryPayload(): + if type(i)<>type(iq): pass + elif i.getName()=='instructions': df.addInstructions(i.getData()) + else: df.setField(i.getName()).setValue(i.getData()) + con.Event(NS_REGISTER,REGISTER_DATA_RECEIVED,(agent, df)) + +def register(disp,host,info): + """ Perform registration on remote server with provided info. + disp must be connected dispatcher instance. + Returns true or false depending on registration result. + If registration fails you can get additional info from the dispatcher's owner + attributes lastErrNode, lastErr and lastErrCode. + """ + iq=Iq('set',NS_REGISTER,to=host) + if type(info)<>type({}): info=info.asDict() + for i in info.keys(): iq.setTag('query').setTagData(i,info[i]) + resp=disp.SendAndWaitForResponse(iq) + if isResultNode(resp): return 1 + +def unregister(disp,host): + """ Unregisters with host (permanently removes account). + disp must be connected and authorized dispatcher instance. + Returns true on success.""" + resp=disp.SendAndWaitForResponse(Iq('set',NS_REGISTER,to=host,payload=[Node('remove')])) + if isResultNode(resp): return 1 + +def changePasswordTo(disp,newpassword,host=None): + """ Changes password on specified or current (if not specified) server. + disp must be connected and authorized dispatcher instance. + Returns true on success.""" + if not host: host=disp._owner.Server + resp=disp.SendAndWaitForResponse(Iq('set',NS_REGISTER,to=host,payload=[Node('username',payload=[disp._owner.Server]),Node('password',payload=[newpassword])])) + if isResultNode(resp): return 1 + +### Privacy ### jabber:iq:privacy ### draft-ietf-xmpp-im-19 #################### +#type=[jid|group|subscription] +#action=[allow|deny] + +def getPrivacyLists(disp): + """ Requests privacy lists from connected server. + Returns dictionary of existing lists on success.""" + try: + dict={'lists':[]} + resp=disp.SendAndWaitForResponse(Iq('get',NS_PRIVACY)) + if not isResultNode(resp): return + for list in resp.getQueryPayload(): + if list.getName()=='list': dict['lists'].append(list.getAttr('name')) + else: dict[list.getName()]=list.getAttr('name') + return dict + except: pass + +def getPrivacyList(disp,listname): + """ Requests specific privacy list listname. Returns list of XML nodes (rules) + taken from the server responce.""" + try: + resp=disp.SendAndWaitForResponse(Iq('get',NS_PRIVACY,payload=[Node('list',{'name':listname})])) + if isResultNode(resp): return resp.getQueryPayload()[0] + except: pass + +def setActivePrivacyList(disp,listname=None,typ='active'): + """ Switches privacy list 'listname' to specified type. + By default the type is 'active'. Returns true on success.""" + if listname: attrs={'name':listname} + else: attrs={} + resp=disp.SendAndWaitForResponse(Iq('set',NS_PRIVACY,payload=[Node(typ,attrs)])) + if isResultNode(resp): return 1 + +def setDefaultPrivacyList(disp,listname=None): + """ Sets the default privacy list as 'listname'. Returns true on success.""" + return setActivePrivacyList(disp,listname,'default') + +def setPrivacyList(disp,list): + """ Set the ruleset. 'list' should be the simpleXML node formatted + according to RFC 3921 (XMPP-IM) (I.e. Node('list',{'name':listname},payload=[...]) ) + Returns true on success.""" + resp=disp.SendAndWaitForResponse(Iq('set',NS_PRIVACY,payload=[list])) + if isResultNode(resp): return 1 + +def delPrivacyList(disp,listname): + """ Deletes privacy list 'listname'. Returns true on success.""" + resp=disp.SendAndWaitForResponse(Iq('set',NS_PRIVACY,payload=[Node('list',{'name':listname})])) + if isResultNode(resp): return 1 + +# -*- coding: utf-8 -*- +# +# Copyright (C) 2007 Edgewall Software +# All rights reserved. +# +# This software is licensed as described in the file COPYING, which +# you should have received as part of this distribution. The terms +# are also available at http://babel.edgewall.org/wiki/License. +# +# This software consists of voluntary contributions made by many +# individuals. For the exact contribution history, see the revision +# history and logs, available at http://babel.edgewall.org/log/. + +"""Low-level locale data access. + +:note: The `Locale` class, which uses this module under the hood, provides a + more convenient interface for accessing the locale data. +""" + +import os +import pickle +try: + import threading +except ImportError: + import dummy_threading as threading +from UserDict import DictMixin + +__all__ = ['exists', 'list', 'load'] +__docformat__ = 'restructuredtext en' + +_cache = {} +_cache_lock = threading.RLock() +_dirname = os.path.join(os.path.dirname(__file__), 'localedata') + + +def exists(name): + """Check whether locale data is available for the given locale. + + :param name: the locale identifier string + :return: `True` if the locale data exists, `False` otherwise + :rtype: `bool` + """ + if name in _cache: + return True + return os.path.exists(os.path.join(_dirname, '%s.dat' % name)) + + +def list(): + """Return a list of all locale identifiers for which locale data is + available. + + :return: a list of locale identifiers (strings) + :rtype: `list` + :since: version 0.8.1 + """ + return [stem for stem, extension in [ + os.path.splitext(filename) for filename in os.listdir(_dirname) + ] if extension == '.dat' and stem != 'root'] + + +def load(name, merge_inherited=True): + """Load the locale data for the given locale. + + The locale data is a dictionary that contains much of the data defined by + the Common Locale Data Repository (CLDR). This data is stored as a + collection of pickle files inside the ``babel`` package. + + >>> d = load('en_US') + >>> d['languages']['sv'] + u'Swedish' + + Note that the results are cached, and subsequent requests for the same + locale return the same dictionary: + + >>> d1 = load('en_US') + >>> d2 = load('en_US') + >>> d1 is d2 + True + + :param name: the locale identifier string (or "root") + :param merge_inherited: whether the inherited data should be merged into + the data of the requested locale + :return: the locale data + :rtype: `dict` + :raise `IOError`: if no locale data file is found for the given locale + identifer, or one of the locales it inherits from + """ + _cache_lock.acquire() + try: + data = _cache.get(name) + if not data: + # Load inherited data + if name == 'root' or not merge_inherited: + data = {} + else: + parts = name.split('_') + if len(parts) == 1: + parent = 'root' + else: + parent = '_'.join(parts[:-1]) + data = load(parent).copy() + filename = os.path.join(_dirname, '%s.dat' % name) + fileobj = open(filename, 'rb') + try: + if name != 'root' and merge_inherited: + merge(data, pickle.load(fileobj)) + else: + data = pickle.load(fileobj) + _cache[name] = data + finally: + fileobj.close() + return data + finally: + _cache_lock.release() + + +def merge(dict1, dict2): + """Merge the data from `dict2` into the `dict1` dictionary, making copies + of nested dictionaries. + + >>> d = {1: 'foo', 3: 'baz'} + >>> merge(d, {1: 'Foo', 2: 'Bar'}) + >>> items = d.items(); items.sort(); items + [(1, 'Foo'), (2, 'Bar'), (3, 'baz')] + + :param dict1: the dictionary to merge into + :param dict2: the dictionary containing the data that should be merged + """ + for key, val2 in dict2.items(): + if val2 is not None: + val1 = dict1.get(key) + if isinstance(val2, dict): + if val1 is None: + val1 = {} + if isinstance(val1, Alias): + val1 = (val1, val2) + elif isinstance(val1, tuple): + alias, others = val1 + others = others.copy() + merge(others, val2) + val1 = (alias, others) + else: + val1 = val1.copy() + merge(val1, val2) + else: + val1 = val2 + dict1[key] = val1 + + +class Alias(object): + """Representation of an alias in the locale data. + + An alias is a value that refers to some other part of the locale data, + as specified by the `keys`. + """ + + def __init__(self, keys): + self.keys = tuple(keys) + + def __repr__(self): + return '<%s %r>' % (type(self).__name__, self.keys) + + def resolve(self, data): + """Resolve the alias based on the given data. + + This is done recursively, so if one alias resolves to a second alias, + that second alias will also be resolved. + + :param data: the locale data + :type data: `dict` + """ + base = data + for key in self.keys: + data = data[key] + if isinstance(data, Alias): + data = data.resolve(base) + elif isinstance(data, tuple): + alias, others = data + data = alias.resolve(base) + return data + + +class LocaleDataDict(DictMixin, dict): + """Dictionary wrapper that automatically resolves aliases to the actual + values. + """ + + def __init__(self, data, base=None): + dict.__init__(self, data) + if base is None: + base = data + self.base = base + + def __getitem__(self, key): + orig = val = dict.__getitem__(self, key) + if isinstance(val, Alias): # resolve an alias + val = val.resolve(self.base) + if isinstance(val, tuple): # Merge a partial dict with an alias + alias, others = val + val = alias.resolve(self.base).copy() + merge(val, others) + if type(val) is dict: # Return a nested alias-resolving dict + val = LocaleDataDict(val, base=self.base) + if val is not orig: + self[key] = val + return val + + def copy(self): + return LocaleDataDict(dict.copy(self), base=self.base) + +import os + +# toolchains options +ARCH='arm' +CPU='cortex-m4' +CROSS_TOOL='gcc' + +if os.getenv('RTT_CC'): + CROSS_TOOL = os.getenv('RTT_CC') + +# cross_tool provides the cross compiler +# EXEC_PATH is the compiler execute path, for example, CodeSourcery, Keil MDK, IAR + +if CROSS_TOOL == 'gcc': + PLATFORM = 'gcc' + EXEC_PATH = 'C:/Program Files (x86)/GNU Tools ARM Embedded/6 2017-q2-update/bin' +elif CROSS_TOOL == 'keil': + PLATFORM = 'armcc' + EXEC_PATH = 'C:\Keil_v5' +elif CROSS_TOOL == 'iar': + PLATFORM = 'iar' + EXEC_PATH = 'C:/Program Files (x86)/IAR Systems/Embedded Workbench 7.5' + +if os.getenv('RTT_EXEC_PATH'): + EXEC_PATH = os.getenv('RTT_EXEC_PATH') + +BUILD = 'debug' + +if PLATFORM == 'gcc': + # toolchains + PREFIX = 'arm-none-eabi-' + CC = PREFIX + 'gcc' + AS = PREFIX + 'gcc' + AR = PREFIX + 'ar' + LINK = PREFIX + 'g++' + TARGET_EXT = 'elf' + SIZE = PREFIX + 'size' + OBJDUMP = PREFIX + 'objdump' + OBJCPY = PREFIX + 'objcopy' + + DEVICE = ' -mcpu=cortex-m4 -mthumb -ffunction-sections -fdata-sections' + CFLAGS = DEVICE + AFLAGS = ' -c' + DEVICE + ' -x assembler-with-cpp' + LFLAGS = DEVICE + ' -nostartfiles -Wl,--gc-sections,-Map=rtthread-nuc472.map -T ldscripts/sections.ld ' + + CPATH = '' + LPATH = '' + + if BUILD == 'debug': + CFLAGS += ' -O0 -gdwarf-2' + AFLAGS += ' -gdwarf-2' + else: + CFLAGS += ' -O2' + + POST_ACTION = OBJCPY + ' -O binary $TARGET rtthread.bin\n' + SIZE + ' $TARGET \n' + +elif PLATFORM == 'armcc': + # toolchains + CC = 'armcc' + AS = 'armasm' + AR = 'armar' + LINK = 'armlink' + TARGET_EXT = 'axf' + + DEVICE = ' --cpu=cortex-m4.fp' + CFLAGS = DEVICE + ' --apcs=interwork' + AFLAGS = DEVICE + LFLAGS = DEVICE + ' --info sizes --info totals --info unused --info veneers --list rtthread-nuc472.map --scatter nuc472_flash.sct' + + CFLAGS += ' --c99' + CFLAGS += ' -I' + EXEC_PATH + '/ARM/RV31/INC' + LFLAGS += ' --libpath ' + EXEC_PATH + '/ARM/RV31/LIB' + + EXEC_PATH += '/arm/bin40/' + + if BUILD == 'debug': + CFLAGS += ' -g -O0' + AFLAGS += ' -g' + else: + CFLAGS += ' -O2' + + POST_ACTION = 'fromelf --bin $TARGET --output rtthread.bin \nfromelf -z $TARGET' + +elif PLATFORM == 'iar': + # toolchains + CC = 'iccarm' + AS = 'iasmarm' + AR = 'iarchive' + LINK = 'ilinkarm' + TARGET_EXT = 'out' + + DEVICE = ' ' + + CFLAGS = DEVICE + CFLAGS += ' --diag_suppress Pa050' + CFLAGS += ' --no_cse' + CFLAGS += ' --no_unroll' + CFLAGS += ' --no_inline' + CFLAGS += ' --no_code_motion' + CFLAGS += ' --no_tbaa' + CFLAGS += ' --no_clustering' + CFLAGS += ' --no_scheduling' + CFLAGS += ' --debug' + CFLAGS += ' --endian=little' + CFLAGS += ' --cpu=Cortex-M4' + CFLAGS += ' -e' + CFLAGS += ' --fpu=None' + CFLAGS += ' --dlib_config "' + EXEC_PATH + '/arm/INC/c/DLib_Config_Normal.h"' + CFLAGS += ' -Ol' + CFLAGS += ' --use_c++_inline' + + AFLAGS = '' + AFLAGS += ' -s+' + AFLAGS += ' -w+' + AFLAGS += ' -r' + AFLAGS += ' --cpu Cortex-M4' + AFLAGS += ' --fpu None' + + LFLAGS = ' --config nuc472_flash.icf' + LFLAGS += ' --redirect _Printf=_PrintfTiny' + LFLAGS += ' --redirect _Scanf=_ScanfSmall' + LFLAGS += ' --entry __iar_program_start' + + EXEC_PATH = EXEC_PATH + '/arm/bin/' + POST_ACTION = '' + +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +import os +from datetime import datetime + +from django.test import SimpleTestCase, ignore_warnings +from django.utils import html, safestring, six +from django.utils._os import upath +from django.utils.deprecation import RemovedInDjango110Warning +from django.utils.encoding import force_text + + +class TestUtilsHtml(SimpleTestCase): + + def check_output(self, function, value, output=None): + """ + Check that function(value) equals output. If output is None, + check that function(value) equals value. + """ + if output is None: + output = value + self.assertEqual(function(value), output) + + def test_escape(self): + f = html.escape + items = ( + ('&', '&'), + ('<', '<'), + ('>', '>'), + ('"', '"'), + ("'", '''), + ) + # Substitution patterns for testing the above items. + patterns = ("%s", "asdf%sfdsa", "%s1", "1%sb") + for value, output in items: + for pattern in patterns: + self.check_output(f, pattern % value, pattern % output) + # Check repeated values. + self.check_output(f, value * 2, output * 2) + # Verify it doesn't double replace &. + self.check_output(f, '<&', '<&') + + def test_format_html(self): + self.assertEqual( + html.format_html("{} {} {third} {fourth}", + "< Dangerous >", + html.mark_safe("safe"), + third="< dangerous again", + fourth=html.mark_safe("safe again") + ), + "< Dangerous > safe < dangerous again safe again" + ) + + def test_linebreaks(self): + f = html.linebreaks + items = ( + ("para1\n\npara2\r\rpara3", "

para1

\n\n

para2

\n\n

para3

"), + ("para1\nsub1\rsub2\n\npara2", "

para1
sub1
sub2

\n\n

para2

"), + ("para1\r\n\r\npara2\rsub1\r\rpara4", "

para1

\n\n

para2
sub1

\n\n

para4

"), + ("para1\tmore\n\npara2", "

para1\tmore

\n\n

para2

"), + ) + for value, output in items: + self.check_output(f, value, output) + + def test_strip_tags(self): + f = html.strip_tags + items = ( + ('

See: 'é is an apostrophe followed by e acute

', + 'See: 'é is an apostrophe followed by e acute'), + ('a', 'a'), + ('a', 'a'), + ('e', 'e'), + ('hi, b2!', 'b7>b2!'), + ('b', 'b'), + ('a

b

c', 'abc'), + ('a

b

c', 'abc'), + ('de

f', 'def'), + ('foobar', 'foobar'), + # caused infinite loop on Pythons not patched with + # http://bugs.python.org/issue20288 + ('&gotcha&#;<>', '&gotcha&#;<>'), + ) + for value, output in items: + self.check_output(f, value, output) + + # Some convoluted syntax for which parsing may differ between python versions + output = html.strip_tags('ript>test</script>') + self.assertNotIn('&h') + self.assertNotIn('', '\\u003Cscript\\u003Eand this\\u003C/script\\u003E'), + ( + 'paragraph separator:\u2029and line separator:\u2028', + 'paragraph separator:\\u2029and line separator:\\u2028' + ), + ) + for value, output in items: + self.check_output(f, value, output) + + @ignore_warnings(category=RemovedInDjango110Warning) + def test_remove_tags(self): + f = html.remove_tags + items = ( + ("Yes", "b i", "Yes"), + ("x

y

", "a b", "x

y

"), + ) + for value, tags, output in items: + self.assertEqual(f(value, tags), output) + + def test_smart_urlquote(self): + quote = html.smart_urlquote + # Ensure that IDNs are properly quoted + self.assertEqual(quote('http://Ă¶Ă¤Ă¼.com/'), 'http://xn--4ca9at.com/') + self.assertEqual(quote('http://Ă¶Ă¤Ă¼.com/Ă¶Ă¤Ă¼/'), 'http://xn--4ca9at.com/%C3%B6%C3%A4%C3%BC/') + # Ensure that everything unsafe is quoted, !*'();:@&=+$,/?#[]~ is considered safe as per RFC + self.assertEqual(quote('http://example.com/path/Ă¶Ă¤Ă¼/'), 'http://example.com/path/%C3%B6%C3%A4%C3%BC/') + self.assertEqual(quote('http://example.com/%C3%B6/ä/'), 'http://example.com/%C3%B6/%C3%A4/') + self.assertEqual(quote('http://example.com/?x=1&y=2+3&z='), 'http://example.com/?x=1&y=2+3&z=') + self.assertEqual(quote('http://example.com/?x=<>"\''), 'http://example.com/?x=%3C%3E%22%27') + self.assertEqual(quote('http://example.com/?q=http://example.com/?x=1%26q=django'), + 'http://example.com/?q=http%3A%2F%2Fexample.com%2F%3Fx%3D1%26q%3Ddjango') + self.assertEqual(quote('http://example.com/?q=http%3A%2F%2Fexample.com%2F%3Fx%3D1%26q%3Ddjango'), + 'http://example.com/?q=http%3A%2F%2Fexample.com%2F%3Fx%3D1%26q%3Ddjango') + + def test_conditional_escape(self): + s = '

interop

' + self.assertEqual(html.conditional_escape(s), + '<h1>interop</h1>') + self.assertEqual(html.conditional_escape(safestring.mark_safe(s)), s) + + def test_html_safe(self): + @html.html_safe + class HtmlClass(object): + if six.PY2: + def __unicode__(self): + return "

I'm a html class!

" + else: + def __str__(self): + return "

I'm a html class!

" + + html_obj = HtmlClass() + self.assertTrue(hasattr(HtmlClass, '__html__')) + self.assertTrue(hasattr(html_obj, '__html__')) + self.assertEqual(force_text(html_obj), html_obj.__html__()) + + def test_html_safe_subclass(self): + if six.PY2: + class BaseClass(object): + def __html__(self): + # defines __html__ on its own + return 'some html content' + + def __unicode__(self): + return 'some non html content' + + @html.html_safe + class Subclass(BaseClass): + def __unicode__(self): + # overrides __unicode__ and is marked as html_safe + return 'some html safe content' + else: + class BaseClass(object): + def __html__(self): + # defines __html__ on its own + return 'some html content' + + def __str__(self): + return 'some non html content' + + @html.html_safe + class Subclass(BaseClass): + def __str__(self): + # overrides __str__ and is marked as html_safe + return 'some html safe content' + + subclass_obj = Subclass() + self.assertEqual(force_text(subclass_obj), subclass_obj.__html__()) + + def test_html_safe_defines_html_error(self): + msg = "can't apply @html_safe to HtmlClass because it defines __html__()." + with self.assertRaisesMessage(ValueError, msg): + @html.html_safe + class HtmlClass(object): + def __html__(self): + return "

I'm a html class!

" + + def test_html_safe_doesnt_define_str(self): + method_name = '__unicode__()' if six.PY2 else '__str__()' + msg = "can't apply @html_safe to HtmlClass because it doesn't define %s." % method_name + with self.assertRaisesMessage(ValueError, msg): + @html.html_safe + class HtmlClass(object): + pass + +#!/usr/bin/python +# -*- coding: UTF-8 -*- + +######################################################## +# __Author__: Xueer Chen # +# Kaggle competition "Display Advertising Challenge": # +# https://www.kaggle.com/c/avazu-ctr-prediction # +# Credit: Triskelion # +######################################################## + +from datetime import datetime +from csv import DictReader +import sys + +def csv_to_vw(loc_csv, loc_output, train=True): + """ + Munges a CSV file (loc_csv) to a VW file (loc_output). Set "train" + to False when munging a test set. + TODO: Too slow for a daily cron job. Try optimize, Pandas or Go. + """ + start = datetime.now() + print("\nTurning %s into %s. Is_train_set? %s"%(loc_csv,loc_output,train)) + + with open(loc_output,"wb") as outfile: + for e, row in enumerate( DictReader(open(loc_csv)) ): + + #Creating the features + numerical_features = "" + categorical_features = "" + for k,v in row.items(): + if k not in ["id","click"]: + if len(str(v)) > 0: + categorical_features += " %s" % v + + #Creating the labels + if train: #we care about labels + if row['click'] == "1": + label = 1 + else: + label = -1 #we set negative label to -1 + outfile.write( "%s '%s |i%s |c%s\n" % (label,row['id'],numerical_features,categorical_features) ) + + else: #we dont care about labels + outfile.write( "1 '%s |i%s |c%s\n" % (row['id'],numerical_features,categorical_features) ) + + #Reporting progress + if e % 100000 == 0: + print("%s\t%s"%(e, str(datetime.now() - start))) + + print("\n %s Task execution time:\n\t%s"%(e, str(datetime.now() - start))) + +def main(): + return 0 + +if __name__ == '__main__': + # main should return 0 for success, something else (usually 1) for error. + sys.exit(main()) + +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/trainOriginal.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/trainOriginal.vw", train=True) +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/testOriginal.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/testOriginal.vw", train=False) +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/trainProbs.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/trainProbs.vw", train=True) +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/testProbs.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/testProbs.vw", train=False) +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/ProbsTfidf.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/ProbsTfidf.vw", train=True) +#csv_to_vw("/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/testProbsTfidf.csv", +# "/home/wacax/Wacax/Kaggle/Avazu-CTR-Prediction/AvazuCTR/Data/vw/testProbsTfidf.vw", train=False) + +# Copyright (C) 2010-2014 GRNET S.A. +# +# This program is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program. If not, see . + +from django.http import HttpResponseRedirect +from django.shortcuts import render_to_response +from django.template import RequestContext +from django.contrib import messages +from django.utils.translation import ugettext as _ +from django.views.decorators.csrf import csrf_exempt +from django.views.decorators.http import require_http_methods +from django.core.urlresolvers import reverse + +import django.contrib.auth.views as django_auth_views + +from astakos.im.util import prepare_response, get_query +from astakos.im.models import PendingThirdPartyUser +from astakos.im.forms import LoginForm, ExtendedPasswordChangeForm, \ + ExtendedSetPasswordForm +from astakos.im import settings +import astakos.im.messages as astakos_messages +from astakos.im import auth_providers as auth +from astakos.im.views.decorators import cookie_fix, requires_anonymous, \ + signed_terms_required, requires_auth_provider, login_required + +from ratelimit.decorators import ratelimit + +retries = settings.RATELIMIT_RETRIES_ALLOWED - 1 +rate = str(retries) + '/m' + + +@requires_auth_provider('local') +@require_http_methods(["GET", "POST"]) +@csrf_exempt +@requires_anonymous +@cookie_fix +@ratelimit(field='username', method='POST', rate=rate) +def login(request, on_failure='im/login.html'): + """ + on_failure: the template name to render on login failure + """ + if request.method == 'GET': + return HttpResponseRedirect(reverse('login')) + + was_limited = getattr(request, 'limited', False) + form = LoginForm(data=request.POST, + was_limited=was_limited, + request=request) + next = get_query(request).get('next', '') + third_party_token = get_query(request).get('key', False) + provider = auth.get_provider('local') + + if not form.is_valid(): + if third_party_token: + messages.info(request, provider.get_login_to_add_msg) + + return render_to_response( + on_failure, + {'login_form': form, + 'next': next, + 'key': third_party_token}, + context_instance=RequestContext(request)) + + # get the user from the cache + user = form.user_cache + provider = auth.get_provider('local', user) + + if not provider.get_login_policy: + message = provider.get_login_disabled_msg + messages.error(request, message) + return HttpResponseRedirect(reverse('login')) + + message = None + if not user: + message = provider.get_authentication_failed_msg + elif not user.is_active: + message = user.get_inactive_message('local') + + elif not user.has_auth_provider('local'): + # valid user logged in with no auth providers set, add local provider + # and let him log in + if not user.get_available_auth_providers(): + user.add_auth_provider('local') + else: + message = _(astakos_messages.NO_LOCAL_AUTH) + + if message: + messages.error(request, message) + return render_to_response(on_failure, + {'login_form': form}, + context_instance=RequestContext(request)) + + response = prepare_response(request, user, next) + if third_party_token: + # use requests to assign the account he just authenticated with with + # a third party provider account + try: + request.user.add_pending_auth_provider(third_party_token) + except PendingThirdPartyUser.DoesNotExist: + provider = auth.get_provider('local', request.user) + messages.error(request, provider.get_add_failed_msg) + + provider = user.get_auth_provider('local') + messages.success(request, provider.get_login_success_msg) + response.set_cookie('astakos_last_login_method', 'local') + provider.update_last_login_at() + + return response + + +@require_http_methods(["GET"]) +@cookie_fix +def password_reset_done(request, *args, **kwargs): + messages.success(request, _(astakos_messages.PASSWORD_RESET_DONE)) + return HttpResponseRedirect(reverse('index')) + + +@require_http_methods(["GET"]) +@cookie_fix +def password_reset_confirm_done(request, *args, **kwargs): + messages.success(request, _(astakos_messages.PASSWORD_RESET_CONFIRM_DONE)) + return HttpResponseRedirect(reverse('index')) + + +@cookie_fix +def password_reset(request, *args, **kwargs): + kwargs['post_reset_redirect'] = reverse( + 'astakos.im.views.target.local.password_reset_done') + return django_auth_views.password_reset(request, *args, **kwargs) + + +@cookie_fix +def password_reset_confirm(request, *args, **kwargs): + kwargs['post_reset_redirect'] = reverse( + 'astakos.im.views.target.local.password_reset_complete') + return django_auth_views.password_reset_confirm(request, *args, **kwargs) + + +@cookie_fix +def password_reset_complete(request, *args, **kwargs): + return django_auth_views.password_reset_complete(request, *args, **kwargs) + + +@require_http_methods(["GET", "POST"]) +@signed_terms_required +@login_required +@cookie_fix +@requires_auth_provider('local', login=True) +def password_change(request, + template_name='registration/password_change_form.html', + post_change_redirect=None, + password_change_form=ExtendedPasswordChangeForm): + + create_password = False + + provider = auth.get_provider('local', request.user) + + # no local backend user wants to create a password + if not request.user.has_auth_provider('local'): + if not provider.get_add_policy: + messages.error(request, provider.get_add_disabled_msg) + return HttpResponseRedirect(reverse('edit_profile')) + + create_password = True + password_change_form = ExtendedSetPasswordForm + + if post_change_redirect is None: + post_change_redirect = reverse('edit_profile') + + if request.method == "POST": + form_kwargs = dict( + user=request.user, + data=request.POST, + ) + if not create_password: + form_kwargs['session_key'] = request.session.session_key + + form = password_change_form(**form_kwargs) + if form.is_valid(): + form.save() + if create_password: + provider = auth.get_provider('local', request.user) + messages.success(request, provider.get_added_msg) + else: + messages.success(request, + astakos_messages.PASSWORD_RESET_CONFIRM_DONE) + return HttpResponseRedirect(post_change_redirect) + else: + form = password_change_form(user=request.user) + return render_to_response(template_name, { + 'form': form, + }, context_instance=RequestContext(request, {'create_password': + create_password})) + +# -*- coding: utf-8 -*- +from __future__ import with_statement + +from djangocms_text_ckeditor.models import Text +from django.contrib.admin.sites import site +from django.contrib.admin.util import unquote +from django.contrib.auth import get_user_model +from django.contrib.auth.models import AnonymousUser, Group, Permission +from django.contrib.sites.models import Site +from django.core.management import call_command +from django.core.urlresolvers import reverse +from django.db.models import Q +from django.test.client import RequestFactory +from django.test.utils import override_settings + +from cms.api import (add_plugin, assign_user_to_page, create_page, + create_page_user, publish_page) +from cms.admin.forms import save_permissions +from cms.constants import PUBLISHER_STATE_PENDING +from cms.management.commands.subcommands.moderator import log +from cms.menu import get_visible_pages +from cms.models import Page, CMSPlugin, Title, ACCESS_PAGE +from cms.models.permissionmodels import (ACCESS_DESCENDANTS, + ACCESS_PAGE_AND_DESCENDANTS, + PagePermission, + GlobalPagePermission) +from cms.plugin_pool import plugin_pool +from cms.test_utils.testcases import (URL_CMS_PAGE_ADD, URL_CMS_PLUGIN_REMOVE, + URL_CMS_PLUGIN_ADD, CMSTestCase) +from cms.test_utils.util.context_managers import disable_logger +from cms.test_utils.util.fuzzy_int import FuzzyInt +from cms.utils.i18n import force_language +from cms.utils.page_resolver import get_page_from_path +from cms.utils.permissions import (has_page_add_permission, + has_page_change_permission, + has_generic_permission) + + +def fake_tree_attrs(page): + page.depth = 1 + page.path = '0001' + page.numchild = 0 + + +@override_settings(CMS_PERMISSION=True) +class PermissionModeratorTests(CMSTestCase): + """Permissions and moderator together + + Fixtures contains 3 users and 1 published page and some other stuff + + Users: + 1. `super`: superuser + 2. `master`: user with permissions to all applications + 3. `slave`: user assigned to page `slave-home` + + Pages: + 1. `home`: + - published page + - master can do anything on its subpages, but not on home! + + 2. `master`: + - published page + - created by super + - `master` can do anything on it and its descendants + - subpages: + + 3. `slave-home`: + - not published + - assigned slave user which can add/change/delete/ + move/publish this page and its descendants + - `master` user want to moderate this page and all descendants + + 4. `pageA`: + - created by super + - master can add/change/delete on it and descendants + """ + #TODO: Split this test case into one that tests publish functionality, and + #TODO: one that tests permission inheritance. This is too complex. + + def setUp(self): + # create super user + self.user_super = self._create_user("super", is_staff=True, + is_superuser=True) + self.user_staff = self._create_user("staff", is_staff=True, + add_default_permissions=True) + self.user_master = self._create_user("master", is_staff=True, + add_default_permissions=True) + self.user_slave = self._create_user("slave", is_staff=True, + add_default_permissions=True) + self.user_normal = self._create_user("normal", is_staff=False) + self.user_normal.user_permissions.add( + Permission.objects.get(codename='publish_page')) + + with self.login_user_context(self.user_super): + self.home_page = create_page("home", "nav_playground.html", "en", + created_by=self.user_super) + + # master page & master user + + self.master_page = create_page("master", "nav_playground.html", "en") + + # create non global, non staff user + self.user_non_global = self._create_user("nonglobal") + + # assign master user under home page + assign_user_to_page(self.home_page, self.user_master, + grant_on=ACCESS_DESCENDANTS, grant_all=True) + + # and to master page + assign_user_to_page(self.master_page, self.user_master, + grant_on=ACCESS_PAGE_AND_DESCENDANTS, grant_all=True) + + # slave page & slave user + + self.slave_page = create_page("slave-home", "col_two.html", "en", + parent=self.master_page, created_by=self.user_super) + + assign_user_to_page(self.slave_page, self.user_slave, grant_all=True) + + # create page_b + page_b = create_page("pageB", "nav_playground.html", "en", created_by=self.user_super) + # Normal user + + # it's allowed for the normal user to view the page + assign_user_to_page(page_b, self.user_normal, can_view=True) + + # create page_a - sample page from master + + page_a = create_page("pageA", "nav_playground.html", "en", + created_by=self.user_super) + assign_user_to_page(page_a, self.user_master, + can_add=True, can_change=True, can_delete=True, can_publish=True, + can_move_page=True) + + # publish after creating all drafts + publish_page(self.home_page, self.user_super, 'en') + + publish_page(self.master_page, self.user_super, 'en') + + self.page_b = publish_page(page_b, self.user_super, 'en') + + def _add_plugin(self, user, page): + """ + Add a plugin using the test client to check for permissions. + """ + with self.login_user_context(user): + placeholder = page.placeholders.all()[0] + post_data = { + 'plugin_language': 'en', + 'plugin_parent': '', + 'placeholder_id': placeholder.pk, + 'plugin_type': 'TextPlugin' + } + url = URL_CMS_PLUGIN_ADD + response = self.client.post(url, post_data) + self.assertEqual(response.status_code, 200) + return response.content.decode('utf8') + + def test_super_can_add_page_to_root(self): + with self.login_user_context(self.user_super): + response = self.client.get(URL_CMS_PAGE_ADD) + self.assertEqual(response.status_code, 200) + + def test_master_cannot_add_page_to_root(self): + with self.login_user_context(self.user_master): + response = self.client.get(URL_CMS_PAGE_ADD) + self.assertEqual(response.status_code, 403) + + def test_slave_cannot_add_page_to_root(self): + with self.login_user_context(self.user_slave): + response = self.client.get(URL_CMS_PAGE_ADD) + self.assertEqual(response.status_code, 403) + + def test_slave_can_add_page_under_slave_home(self): + with self.login_user_context(self.user_slave): + # move to admin.py? + # url = URL_CMS_PAGE_ADD + "?target=%d&position=last-child" % slave_page.pk + + # can he even access it over get? + # response = self.client.get(url) + # self.assertEqual(response.status_code, 200) + + # add page + page = create_page("page", "nav_playground.html", "en", + parent=self.slave_page, created_by=self.user_slave) + # adds user_slave as page moderator for this page + # public model shouldn't be available yet, because of the moderation + # moderators and approval ok? + + # must not have public object yet + self.assertFalse(page.publisher_public) + + self.assertObjectExist(Title.objects, slug="page") + self.assertObjectDoesNotExist(Title.objects.public(), slug="page") + + self.assertTrue(has_generic_permission(page.pk, self.user_slave, "publish", 1)) + + # publish as slave, published as user_master before + publish_page(page, self.user_slave, 'en') + # user_slave is moderator for this page + # approve / publish as user_slave + # user master should be able to approve as well + + @override_settings( + CMS_PLACEHOLDER_CONF={ + 'col_left': { + 'default_plugins': [ + { + 'plugin_type': 'TextPlugin', + 'values': { + 'body': 'Lorem ipsum dolor sit amet, consectetur adipisicing elit. Culpa, repellendus, delectus, quo quasi ullam inventore quod quam aut voluptatum aliquam voluptatibus harum officiis officia nihil minus unde accusamus dolorem repudiandae.' + }, + }, + ] + }, + }, + ) + def test_default_plugins(self): + with self.login_user_context(self.user_slave): + self.assertEqual(CMSPlugin.objects.count(), 0) + response = self.client.get(self.slave_page.get_absolute_url(), {'edit': 1}) + self.assertEqual(response.status_code, 200) + self.assertEqual(CMSPlugin.objects.count(), 1) + + def test_page_added_by_slave_can_be_published_by_user_master(self): + # add page + page = create_page("page", "nav_playground.html", "en", + parent=self.slave_page, created_by=self.user_slave) + # same as test_slave_can_add_page_under_slave_home + + # must not have public object yet + self.assertFalse(page.publisher_public) + + self.assertTrue(has_generic_permission(page.pk, self.user_master, "publish", page.site.pk)) + # should be True user_master should have publish permissions for children as well + publish_page(self.slave_page, self.user_master, 'en') + page = publish_page(page, self.user_master, 'en') + self.assertTrue(page.publisher_public_id) + # user_master is moderator for top level page / but can't approve descendants? + # approve / publish as user_master + # user master should be able to approve descendants + + def test_super_can_add_plugin(self): + self._add_plugin(self.user_super, page=self.slave_page) + + def test_master_can_add_plugin(self): + self._add_plugin(self.user_master, page=self.slave_page) + + def test_slave_can_add_plugin(self): + self._add_plugin(self.user_slave, page=self.slave_page) + + def test_same_order(self): + # create 4 pages + slugs = [] + for i in range(0, 4): + page = create_page("page", "nav_playground.html", "en", + parent=self.home_page) + slug = page.title_set.drafts()[0].slug + slugs.append(slug) + + # approve last 2 pages in reverse order + for slug in reversed(slugs[2:]): + page = self.assertObjectExist(Page.objects.drafts(), title_set__slug=slug) + page = publish_page(page, self.user_master, 'en') + self.check_published_page_attributes(page) + + def test_create_copy_publish(self): + # create new page to copy + page = create_page("page", "nav_playground.html", "en", + parent=self.slave_page) + + # copy it under home page... + # TODO: Use page.copy_page here + with self.login_user_context(self.user_master): + copied_page = self.copy_page(page, self.home_page) + + page = publish_page(copied_page, self.user_master, 'en') + self.check_published_page_attributes(page) + + def test_create_publish_copy(self): + # create new page to copy + page = create_page("page", "nav_playground.html", "en", + parent=self.home_page) + + page = publish_page(page, self.user_master, 'en') + + # copy it under master page... + # TODO: Use page.copy_page here + with self.login_user_context(self.user_master): + copied_page = self.copy_page(page, self.master_page) + + self.check_published_page_attributes(page) + copied_page = publish_page(copied_page, self.user_master, 'en') + self.check_published_page_attributes(copied_page) + + def test_subtree_needs_approval(self): + # create page under slave_page + page = create_page("parent", "nav_playground.html", "en", + parent=self.home_page) + self.assertFalse(page.publisher_public) + + # create subpage under page + subpage = create_page("subpage", "nav_playground.html", "en", parent=page) + self.assertFalse(subpage.publisher_public) + + # publish both of them in reverse order + subpage = publish_page(subpage, self.user_master, 'en') + + # subpage should not be published, because parent is not published + # yet, should be marked as `publish when parent` + self.assertFalse(subpage.publisher_public) + + # publish page (parent of subage), so subpage must be published also + page = publish_page(page, self.user_master, 'en') + self.assertNotEqual(page.publisher_public, None) + + # reload subpage, it was probably changed + subpage = self.reload(subpage) + + # parent was published, so subpage must be also published.. + self.assertNotEqual(subpage.publisher_public, None) + + #check attributes + self.check_published_page_attributes(page) + self.check_published_page_attributes(subpage) + + def test_subtree_with_super(self): + # create page under root + page = create_page("page", "nav_playground.html", "en") + self.assertFalse(page.publisher_public) + + # create subpage under page + subpage = create_page("subpage", "nav_playground.html", "en", + parent=page) + self.assertFalse(subpage.publisher_public) + + # tree id must be the same + self.assertEqual(page.path[0:4], subpage.path[0:4]) + + # publish both of them + page = self.reload(page) + page = publish_page(page, self.user_super, 'en') + # reload subpage, there were an path change + subpage = self.reload(subpage) + self.assertEqual(page.path[0:4], subpage.path[0:4]) + + subpage = publish_page(subpage, self.user_super, 'en') + # tree id must stay the same + self.assertEqual(page.path[0:4], subpage.path[0:4]) + + # published pages must also have the same root-path + self.assertEqual(page.publisher_public.path[0:4], subpage.publisher_public.path[0:4]) + + #check attributes + self.check_published_page_attributes(page) + self.check_published_page_attributes(subpage) + + def test_super_add_page_to_root(self): + """Create page which is not under moderation in root, and check if + some properties are correct. + """ + # create page under root + page = create_page("page", "nav_playground.html", "en") + + # public must not exist + self.assertFalse(page.publisher_public) + + def test_moderator_flags(self): + """Add page under slave_home and check its flag + """ + page = create_page("page", "nav_playground.html", "en", + parent=self.slave_page) + + # No public version + self.assertIsNone(page.publisher_public) + self.assertFalse(page.publisher_public_id) + + # check publish box + page = publish_page(page, self.user_slave, 'en') + + # public page must not exist because of parent + self.assertFalse(page.publisher_public) + + # waiting for parents + self.assertEqual(page.get_publisher_state('en'), PUBLISHER_STATE_PENDING) + + # publish slave page + self.slave_page = self.slave_page.reload() + slave_page = publish_page(self.slave_page, self.user_master, 'en') + + self.assertFalse(page.publisher_public) + self.assertTrue(slave_page.publisher_public) + + def test_plugins_get_published(self): + # create page under root + page = create_page("page", "nav_playground.html", "en") + placeholder = page.placeholders.all()[0] + add_plugin(placeholder, "TextPlugin", "en", body="test") + # public must not exist + self.assertEqual(CMSPlugin.objects.all().count(), 1) + publish_page(page, self.user_super, 'en') + self.assertEqual(CMSPlugin.objects.all().count(), 2) + + def test_remove_plugin_page_under_moderation(self): + # login as slave and create page + page = create_page("page", "nav_playground.html", "en", parent=self.slave_page) + + # add plugin + placeholder = page.placeholders.all()[0] + plugin = add_plugin(placeholder, "TextPlugin", "en", body="test") + + # publish page + page = self.reload(page) + page = publish_page(page, self.user_slave, 'en') + + # only the draft plugin should exist + self.assertEqual(CMSPlugin.objects.all().count(), 1) + + # page should require approval + self.assertEqual(page.get_publisher_state('en'), PUBLISHER_STATE_PENDING) + + # master approves and publishes the page + # first approve slave-home + slave_page = self.reload(self.slave_page) + publish_page(slave_page, self.user_master, 'en') + page = self.reload(page) + page = publish_page(page, self.user_master, 'en') + + # draft and public plugins should now exist + self.assertEqual(CMSPlugin.objects.all().count(), 2) + + # login as slave and delete the plugin - should require moderation + with self.login_user_context(self.user_slave): + plugin_data = { + 'plugin_id': plugin.pk + } + remove_url = URL_CMS_PLUGIN_REMOVE + "%s/" % plugin.pk + response = self.client.post(remove_url, plugin_data) + self.assertEqual(response.status_code, 302) + + # there should only be a public plugin - since the draft has been deleted + self.assertEqual(CMSPlugin.objects.all().count(), 1) + + page = self.reload(page) + + # login as super user and approve/publish the page + publish_page(page, self.user_super, 'en') + + # there should now be 0 plugins + self.assertEqual(CMSPlugin.objects.all().count(), 0) + + def test_superuser_can_view(self): + url = self.page_b.get_absolute_url(language='en') + with self.login_user_context(self.user_super): + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + + def test_staff_can_view(self): + url = self.page_b.get_absolute_url(language='en') + all_view_perms = PagePermission.objects.filter(can_view=True) + # verifiy that the user_staff has access to this page + has_perm = False + for perm in all_view_perms: + if perm.page == self.page_b: + if perm.user == self.user_staff: + has_perm = True + self.assertEqual(has_perm, False) + login_ok = self.client.login(username=getattr(self.user_staff, get_user_model().USERNAME_FIELD), + password=getattr(self.user_staff, get_user_model().USERNAME_FIELD)) + self.assertTrue(login_ok) + + # really logged in + self.assertTrue('_auth_user_id' in self.client.session) + login_user_id = self.client.session.get('_auth_user_id') + user = get_user_model().objects.get(pk=self.user_staff.pk) + self.assertEqual(login_user_id, user.id) + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + + def test_user_normal_can_view(self): + url = self.page_b.get_absolute_url(language='en') + all_view_perms = PagePermission.objects.filter(can_view=True) + # verifiy that the normal_user has access to this page + normal_has_perm = False + for perm in all_view_perms: + if perm.page == self.page_b: + if perm.user == self.user_normal: + normal_has_perm = True + self.assertTrue(normal_has_perm) + with self.login_user_context(self.user_normal): + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + + # verifiy that the user_non_global has not access to this page + non_global_has_perm = False + for perm in all_view_perms: + if perm.page == self.page_b: + if perm.user == self.user_non_global: + non_global_has_perm = True + self.assertFalse(non_global_has_perm) + with self.login_user_context(self.user_non_global): + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + # non logged in user + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + + def test_user_globalpermission(self): + # Global user + user_global = self._create_user("global") + + with self.login_user_context(self.user_super): + user_global = create_page_user(user_global, user_global) + user_global.is_staff = False + user_global.save() # Prevent is_staff permission + global_page = create_page("global", "nav_playground.html", "en", + published=True) + # Removed call since global page user doesn't have publish permission + #global_page = publish_page(global_page, user_global) + # it's allowed for the normal user to view the page + assign_user_to_page(global_page, user_global, + global_permission=True, can_view=True) + + url = global_page.get_absolute_url('en') + all_view_perms = PagePermission.objects.filter(can_view=True) + has_perm = False + for perm in all_view_perms: + if perm.page == self.page_b and perm.user == user_global: + has_perm = True + self.assertEqual(has_perm, False) + + global_page_perm_q = Q(user=user_global) & Q(can_view=True) + global_view_perms = GlobalPagePermission.objects.filter(global_page_perm_q).exists() + self.assertEqual(global_view_perms, True) + + # user_global + with self.login_user_context(user_global): + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + # self.non_user_global + has_perm = False + for perm in all_view_perms: + if perm.page == self.page_b and perm.user == self.user_non_global: + has_perm = True + self.assertEqual(has_perm, False) + + global_page_perm_q = Q(user=self.user_non_global) & Q(can_view=True) + global_view_perms = GlobalPagePermission.objects.filter(global_page_perm_q).exists() + self.assertEqual(global_view_perms, False) + + with self.login_user_context(self.user_non_global): + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + + def test_anonymous_user_public_for_all(self): + url = self.page_b.get_absolute_url('en') + with self.settings(CMS_PUBLIC_FOR='all'): + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + + def test_anonymous_user_public_for_none(self): + # default of when to show pages to anonymous user doesn't take + # global permissions into account + url = self.page_b.get_absolute_url('en') + with self.settings(CMS_PUBLIC_FOR=None): + response = self.client.get(url) + self.assertEqual(response.status_code, 404) + + +@override_settings(CMS_PERMISSION=True) +class PatricksMoveTest(CMSTestCase): + """ + Fixtures contains 3 users and 1 published page and some other stuff + + Users: + 1. `super`: superuser + 2. `master`: user with permissions to all applications + 3. `slave`: user assigned to page `slave-home` + + Pages: + 1. `home`: + - published page + - master can do anything on its subpages, but not on home! + + 2. `master`: + - published page + - crated by super + - `master` can do anything on it and its descendants + - subpages: + + 3. `slave-home`: + - not published + - assigned slave user which can add/change/delete/ + move/publish/moderate this page and its descendants + - `master` user want to moderate this page and all descendants + + 4. `pageA`: + - created by super + - master can add/change/delete on it and descendants + """ + + def setUp(self): + # create super user + self.user_super = self._create_user("super", True, True) + + with self.login_user_context(self.user_super): + self.home_page = create_page("home", "nav_playground.html", "en", + created_by=self.user_super) + + # master page & master user + + self.master_page = create_page("master", "nav_playground.html", "en") + + # create master user + self.user_master = self._create_user("master", True) + self.user_master.user_permissions.add(Permission.objects.get(codename='publish_page')) + #self.user_master = create_page_user(self.user_super, master, grant_all=True) + + # assign master user under home page + assign_user_to_page(self.home_page, self.user_master, + grant_on=ACCESS_DESCENDANTS, grant_all=True) + + # and to master page + assign_user_to_page(self.master_page, self.user_master, grant_all=True) + + # slave page & slave user + + self.slave_page = create_page("slave-home", "nav_playground.html", "en", + parent=self.master_page, created_by=self.user_super) + slave = self._create_user("slave", True) + self.user_slave = create_page_user(self.user_super, slave, can_add_page=True, + can_change_page=True, can_delete_page=True) + + assign_user_to_page(self.slave_page, self.user_slave, grant_all=True) + + # create page_a - sample page from master + + page_a = create_page("pageA", "nav_playground.html", "en", + created_by=self.user_super) + assign_user_to_page(page_a, self.user_master, + can_add=True, can_change=True, can_delete=True, can_publish=True, + can_move_page=True) + + # publish after creating all drafts + publish_page(self.home_page, self.user_super, 'en') + publish_page(self.master_page, self.user_super, 'en') + + with self.login_user_context(self.user_slave): + # all of them are under moderation... + self.pa = create_page("pa", "nav_playground.html", "en", parent=self.slave_page) + self.pb = create_page("pb", "nav_playground.html", "en", parent=self.pa, position="right") + self.pc = create_page("pc", "nav_playground.html", "en", parent=self.pb, position="right") + + self.pd = create_page("pd", "nav_playground.html", "en", parent=self.pb) + self.pe = create_page("pe", "nav_playground.html", "en", parent=self.pd, position="right") + + self.pf = create_page("pf", "nav_playground.html", "en", parent=self.pe) + self.pg = create_page("pg", "nav_playground.html", "en", parent=self.pf, position="right") + self.ph = create_page("ph", "nav_playground.html", "en", parent=self.pf, position="right") + + self.assertFalse(self.pg.publisher_public) + + # login as master for approval + self.slave_page = self.slave_page.reload() + + publish_page(self.slave_page, self.user_master, 'en') + + # publish and approve them all + publish_page(self.pa, self.user_master, 'en') + publish_page(self.pb, self.user_master, 'en') + publish_page(self.pc, self.user_master, 'en') + publish_page(self.pd, self.user_master, 'en') + publish_page(self.pe, self.user_master, 'en') + publish_page(self.pf, self.user_master, 'en') + publish_page(self.pg, self.user_master, 'en') + publish_page(self.ph, self.user_master, 'en') + self.reload_pages() + + def reload_pages(self): + self.pa = self.pa.reload() + self.pb = self.pb.reload() + self.pc = self.pc.reload() + self.pd = self.pd.reload() + self.pe = self.pe.reload() + self.pf = self.pf.reload() + self.pg = self.pg.reload() + self.ph = self.ph.reload() + + + def test_patricks_move(self): + """ + + Tests permmod when moving trees of pages. + + 1. build following tree (master node is approved and published) + + slave-home + / | \ + A B C + / \ + D E + / | \ + F G H + + 2. perform move operations: + 1. move G under C + 2. move E under G + + slave-home + / | \ + A B C + / \ + D G + \ + E + / \ + F H + + 3. approve nodes in following order: + 1. approve H + 2. approve G + 3. approve E + 4. approve F + """ + # TODO: this takes 5 seconds to run on my MBP. That's TOO LONG! + self.assertEqual(self.pg.parent_id, self.pe.pk) + self.assertEqual(self.pg.publisher_public.parent_id, self.pe.publisher_public_id) + # perform moves under slave... + self.move_page(self.pg, self.pc) + self.reload_pages() + # Draft page is now under PC + self.assertEqual(self.pg.parent_id, self.pc.pk) + # Public page is under PC + self.assertEqual(self.pg.publisher_public.parent_id, self.pc.publisher_public_id) + self.assertEqual(self.pg.publisher_public.parent.get_absolute_url(), + self.pc.publisher_public.get_absolute_url()) + self.assertEqual(self.pg.get_absolute_url(), self.pg.publisher_public.get_absolute_url()) + self.move_page(self.pe, self.pg) + self.reload_pages() + self.assertEqual(self.pe.parent_id, self.pg.pk) + self.assertEqual(self.pe.publisher_public.parent_id, self.pg.publisher_public_id) + self.ph = self.ph.reload() + # check urls - they should stay be the same now after the move + self.assertEqual( + self.pg.publisher_public.get_absolute_url(), + self.pg.get_absolute_url() + ) + self.assertEqual( + self.ph.publisher_public.get_absolute_url(), + self.ph.get_absolute_url() + ) + + # public parent check after move + self.assertEqual(self.pg.publisher_public.parent.pk, self.pc.publisher_public_id) + self.assertEqual(self.pe.publisher_public.parent.pk, self.pg.publisher_public_id) + self.assertEqual(self.ph.publisher_public.parent.pk, self.pe.publisher_public_id) + + # check if urls are correct after move + self.assertEqual( + self.pg.publisher_public.get_absolute_url(), + u'%smaster/slave-home/pc/pg/' % self.get_pages_root() + ) + self.assertEqual( + self.ph.publisher_public.get_absolute_url(), + u'%smaster/slave-home/pc/pg/pe/ph/' % self.get_pages_root() + ) + + +class ModeratorSwitchCommandTest(CMSTestCase): + def test_switch_moderator_on(self): + with force_language("en"): + pages_root = unquote(reverse("pages-root")) + page1 = create_page('page', 'nav_playground.html', 'en', published=True) + with disable_logger(log): + call_command('cms', 'moderator', 'on') + with force_language("en"): + path = page1.get_absolute_url()[len(pages_root):].strip('/') + page2 = get_page_from_path(path) + self.assertEqual(page1.get_absolute_url(), page2.get_absolute_url()) + + def test_table_name_patching(self): + """ + This tests the plugin models patching when publishing from the command line + """ + self.get_superuser() + create_page("The page!", "nav_playground.html", "en", published=True) + draft = Page.objects.drafts()[0] + draft.reverse_id = 'a_test' # we have to change *something* + draft.save() + add_plugin(draft.placeholders.get(slot=u"body"), + u"TextPlugin", u"en", body="Test content") + draft.publish('en') + add_plugin(draft.placeholders.get(slot=u"body"), + u"TextPlugin", u"en", body="Test content") + + # Manually undoing table name patching + Text._meta.db_table = 'djangocms_text_ckeditor_text' + plugin_pool.patched = False + + with disable_logger(log): + call_command('cms', 'moderator', 'on') + # Sanity check the database (we should have one draft and one public) + not_drafts = len(Page.objects.filter(publisher_is_draft=False)) + drafts = len(Page.objects.filter(publisher_is_draft=True)) + self.assertEqual(not_drafts, 1) + self.assertEqual(drafts, 1) + + def test_switch_moderator_off(self): + with force_language("en"): + pages_root = unquote(reverse("pages-root")) + page1 = create_page('page', 'nav_playground.html', 'en', published=True) + path = page1.get_absolute_url()[len(pages_root):].strip('/') + page2 = get_page_from_path(path) + self.assertIsNotNone(page2) + self.assertEqual(page1.get_absolute_url(), page2.get_absolute_url()) + + def tearDown(self): + plugin_pool.patched = False + plugin_pool.set_plugin_meta() + + +class ViewPermissionBaseTests(CMSTestCase): + + def setUp(self): + self.page = create_page('testpage', 'nav_playground.html', 'en') + + def get_request(self, user=None): + attrs = { + 'user': user or AnonymousUser(), + 'REQUEST': {}, + 'session': {}, + } + return type('Request', (object,), attrs) + + +@override_settings( + CMS_PERMISSION=False, + CMS_PUBLIC_FOR='staff', +) +class BasicViewPermissionTests(ViewPermissionBaseTests): + """ + Test functionality with CMS_PERMISSION set to false, as this is the + normal use case + """ + + @override_settings(CMS_PUBLIC_FOR="all") + def test_unauth_public(self): + request = self.get_request() + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + def test_unauth_non_access(self): + request = self.get_request() + with self.assertNumQueries(0): + self.assertFalse(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + []) + + @override_settings(CMS_PUBLIC_FOR="all") + def test_staff_public_all(self): + request = self.get_request(self.get_staff_user_with_no_permissions()) + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + def test_staff_public_staff(self): + request = self.get_request(self.get_staff_user_with_no_permissions()) + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + @override_settings(CMS_PUBLIC_FOR="none") + def test_staff_basic_auth(self): + request = self.get_request(self.get_staff_user_with_no_permissions()) + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + @override_settings(CMS_PUBLIC_FOR="none") + def test_normal_basic_auth(self): + request = self.get_request(self.get_standard_user()) + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + +@override_settings( + CMS_PERMISSION=True, + CMS_PUBLIC_FOR='none' +) +class UnrestrictedViewPermissionTests(ViewPermissionBaseTests): + """ + Test functionality with CMS_PERMISSION set to True but no restrictions + apply to this specific page + """ + + def test_unauth_non_access(self): + request = self.get_request() + with self.assertNumQueries(1): + """ + The query is: + PagePermission query for the affected page (is the page restricted?) + """ + self.assertFalse(self.page.has_view_permission(request)) + with self.assertNumQueries(0): + self.assertFalse(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + []) + + def test_global_access(self): + user = self.get_standard_user() + GlobalPagePermission.objects.create(can_view=True, user=user) + request = self.get_request(user) + with self.assertNumQueries(2): + """The queries are: + PagePermission query for the affected page (is the page restricted?) + GlobalPagePermission query for the page site + """ + self.assertTrue(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + [self.page.pk]) + + def test_normal_denied(self): + request = self.get_request(self.get_standard_user()) + with self.assertNumQueries(4): + """ + The queries are: + PagePermission query for the affected page (is the page restricted?) + GlobalPagePermission query for the page site + User permissions query + Content type query + """ + self.assertFalse(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertFalse(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, [self.page], self.page.site), + []) + + +@override_settings( + CMS_PERMISSION=True, + CMS_PUBLIC_FOR='all' +) +class RestrictedViewPermissionTests(ViewPermissionBaseTests): + """ + Test functionality with CMS_PERMISSION set to True and view restrictions + apply to this specific page + """ + def setUp(self): + super(RestrictedViewPermissionTests, self).setUp() + self.group = Group.objects.create(name='testgroup') + self.pages = [self.page] + self.expected = [self.page.pk] + PagePermission.objects.create(page=self.page, group=self.group, can_view=True, grant_on=ACCESS_PAGE) + + def test_unauthed(self): + request = self.get_request() + with self.assertNumQueries(1): + """The queries are: + PagePermission query for the affected page (is the page restricted?) + """ + self.assertFalse(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertFalse(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + []) + + def test_page_permissions(self): + user = self.get_standard_user() + request = self.get_request(user) + PagePermission.objects.create(can_view=True, user=user, page=self.page, grant_on=ACCESS_PAGE) + with self.assertNumQueries(3): + """ + The queries are: + PagePermission query (is this page restricted) + GlobalpagePermission query for user + PagePermission query for this user + """ + self.assertTrue(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + self.expected) + + def test_page_group_permissions(self): + user = self.get_standard_user() + user.groups.add(self.group) + request = self.get_request(user) + with self.assertNumQueries(3): + self.assertTrue(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + self.expected) + + def test_global_permission(self): + user = self.get_standard_user() + GlobalPagePermission.objects.create(can_view=True, user=user) + request = self.get_request(user) + with self.assertNumQueries(2): + """ + The queries are: + PagePermission query (is this page restricted) + GlobalpagePermission query for user + """ + self.assertTrue(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + self.expected) + + def test_basic_perm_denied(self): + request = self.get_request(self.get_staff_user_with_no_permissions()) + with self.assertNumQueries(5): + """ + The queries are: + PagePermission query (is this page restricted) + GlobalpagePermission query for user + PagePermission query for this user + Generic django permission lookup + content type lookup by permission lookup + """ + self.assertFalse(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertFalse(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + []) + + def test_basic_perm(self): + user = self.get_standard_user() + user.user_permissions.add(Permission.objects.get(codename='view_page')) + request = self.get_request(user) + with self.assertNumQueries(5): + """ + The queries are: + PagePermission query (is this page restricted) + GlobalpagePermission query for user + PagePermission query for this user + Generic django permission lookup + content type lookup by permission lookup + """ + self.assertTrue(self.page.has_view_permission(request)) + + with self.assertNumQueries(0): + self.assertTrue(self.page.has_view_permission(request)) # test cache + + self.assertEqual(get_visible_pages(request, self.pages, self.page.site), + self.expected) + + +class PublicViewPermissionTests(RestrictedViewPermissionTests): + """ Run the same tests as before, but on the public page instead. """ + + def setUp(self): + super(PublicViewPermissionTests, self).setUp() + self.page.publish('en') + self.pages = [self.page.publisher_public] + self.expected = [self.page.publisher_public_id] + + +class GlobalPermissionTests(CMSTestCase): + + def test_sanity_check(self): + """ Because we have a new manager, we'll do some basic checks.""" + # manager is still named the same. + self.assertTrue(hasattr(GlobalPagePermission, 'objects')) + self.assertEqual(0, GlobalPagePermission.objects.all().count()) + + # we are correctly inheriting from BasicPagePermissionManager + self.assertTrue(hasattr(GlobalPagePermission.objects, 'with_user')) + + # If we're using the new manager, we have extra methods which ensure + # This site access OR all site access. + self.assertTrue(hasattr(GlobalPagePermission.objects, 'user_has_permission')) + # these are just convienence methods for the above. + self.assertTrue(hasattr(GlobalPagePermission.objects, 'user_has_add_permission')) + self.assertTrue(hasattr(GlobalPagePermission.objects, 'user_has_change_permission')) + self.assertTrue(hasattr(GlobalPagePermission.objects, 'user_has_view_permission')) + + def test_emulate_admin_index(self): + """ Call methods that emulate the adminsite instance's index. + This test was basically the reason for the new manager, in light of the + problem highlighted in ticket #1120, which asserts that giving a user + no site-specific rights when creating a GlobalPagePermission should + allow access to all sites. + """ + # create and then ignore this user. + superuser = self._create_user("super", is_staff=True, is_active=True, + is_superuser=True) + superuser.set_password("super") + superuser.save() + # create 2 staff users + SITES = [ + Site.objects.get(pk=1), + Site.objects.create(domain='example2.com', name='example2.com'), + ] + USERS = [ + self._create_user("staff", is_staff=True, is_active=True), + self._create_user("staff_2", is_staff=True, is_active=True), + ] + for user in USERS: + user.set_password('staff') + # re-use the same methods the UserPage form does. + # Note that it internally calls .save(), as we've not done so. + save_permissions({ + 'can_add_page': True, + 'can_change_page': True, + 'can_delete_page': False + }, user) + + GlobalPagePermission.objects.create(can_add=True, can_change=True, + can_delete=False, user=USERS[0]) + # we're querying here to ensure that even though we've created two users + # above, we should have successfully filtered to just one perm. + self.assertEqual(1, GlobalPagePermission.objects.with_user(USERS[0]).count()) + + # this will confirm explicit permissions still work, by adding the first + # site instance to the many2many relationship 'sites' + GlobalPagePermission.objects.create(can_add=True, can_change=True, + can_delete=False, + user=USERS[1]).sites.add(SITES[0]) + self.assertEqual(1, GlobalPagePermission.objects.with_user(USERS[1]).count()) + + homepage = create_page(title="master", template="nav_playground.html", + language="en", in_navigation=True, slug='/') + publish_page(page=homepage, user=superuser, language='en') + + with self.settings(CMS_PERMISSION=True): + # for all users, they should have access to site 1 + request = RequestFactory().get(path='/', data={'site__exact': 1}) + # we need a session attribute for current_site(request), which is + # used by has_page_add_permission and has_page_change_permission + request.session = {} + for user in USERS: + # has_page_add_permission and has_page_change_permission both test + # for this explicitly, to see if it's a superuser. + request.user = user + # Note, the query count is inflated by doing additional lookups + # because there's a site param in the request. + with self.assertNumQueries(FuzzyInt(6,7)): + # PageAdmin swaps out the methods called for permissions + # if the setting is true, it makes use of cms.utils.permissions + self.assertTrue(has_page_add_permission(request)) + self.assertTrue(has_page_change_permission(request)) + # internally this calls PageAdmin.has_[add|change|delete]_permission() + self.assertEqual({'add': True, 'change': True, 'delete': False}, + site._registry[Page].get_model_perms(request)) + + # can't use the above loop for this test, as we're testing that + # user 1 has access, but user 2 does not, as they are only assigned + # to site 1 + request = RequestFactory().get('/', data={'site__exact': 2}) + request.session = {} + # As before, the query count is inflated by doing additional lookups + # because there's a site param in the request + with self.assertNumQueries(FuzzyInt(11, 20)): + # this user shouldn't have access to site 2 + request.user = USERS[1] + self.assertTrue(not has_page_add_permission(request)) + self.assertTrue(not has_page_change_permission(request)) + self.assertEqual({'add': False, 'change': False, 'delete': False}, + site._registry[Page].get_model_perms(request)) + # but, going back to the first user, they should. + request = RequestFactory().get('/', data={'site__exact': 2}) + request.user = USERS[0] + self.assertTrue(has_page_add_permission(request)) + self.assertTrue(has_page_change_permission(request)) + self.assertEqual({'add': True, 'change': True, 'delete': False}, + site._registry[Page].get_model_perms(request)) + + def test_has_page_add_permission_with_target(self): + page = create_page('Test', 'nav_playground.html', 'en') + user = self._create_user('user') + request = RequestFactory().get('/', data={'target': page.pk}) + request.session = {} + request.user = user + has_perm = has_page_add_permission(request) + self.assertFalse(has_perm) + +# Transformer/IO/_VASP.py + + +# ------- +# Imports +# ------- + +import warnings; + +from Transformer import Constants; +from Transformer import Structure; +from Transformer.Utilities import StructureTools; + + +# --------- +# Functions +# --------- + +def ReadPOSCARFile(inputReader, atomicSymbolLookupTable = None): + # Variables to collect. + + systemName = None; + scaleFactor = None; + latticeVectors = None; + atomTypes, atomCounts = None, None; + coordinateType, atomPositions = None, None; + + # Read the system name. + + systemName = next(inputReader).strip(); + + # Read the scale factor. + + scaleFactor = float(next(inputReader).strip()); + + # Read the lattice vectors. + + latticeVectors = []; + + for i in range(0, 3): + latticeVectors.append( + [float(element) for element in next(inputReader).strip().split()][:3] + ); + + # Although we sliced the list returned from split(), this does not guarentee that there were at least three elements. + + for latticeVector in latticeVectors: + if len(latticeVector) != 3: + raise Exception("Error: The lattice vector specification in the supplied VASP POSCAR file is invalid."); + + # Read the atom types and/or atom counts. + + atomTypes = [element for element in next(inputReader).strip().split()]; + + atomCounts = None; + + if atomTypes[0].isdigit(): + atomCounts = [int(item) for item in atomTypes]; + atomTypes = None; + else: + atomCounts = [int(element) for element in next(inputReader).strip().split()]; + + # If atom types were given in the file, check the number of atom types listed is consistent with the number of atom counts. + + if atomTypes != None and len(atomTypes) != len(atomCounts): + raise Exception("Error: The atom-type and atom-count lines in the supplied VASP POSCAR file contain different numbers of entries."); + + # Read the coordinate type. + + coordinateType = None; + + keyword = next(inputReader).strip().lower(); + + # Check for and skip the "selective dynamics" keyword. + + if keyword[0] == "s": + keyword = next(inputReader).strip().lower(); + + if keyword[0] == 'd': + coordinateType = 'd'; + elif keyword[0] == 'c' or keyword[0] == 'k': + coordinateType = 'c'; + else: + raise Exception("Error: The coordinate-type line in the supplied VASP POSCAR file contains an unexpected keyword."); + + # Read the atom positions. + + totalAtomCount = 0; + + for atomCount in atomCounts: + totalAtomCount = totalAtomCount + atomCount; + + atomPositions = []; + + for i in range(0, totalAtomCount): + elements = next(inputReader).strip().split(); + + atomPositions.append( + [float(element) for element in elements[:3]] + ); + + for atomPosition in atomPositions: + if len(atomPosition) != 3: + raise Exception("Error: One or more atomic position specifications in the supplied VASP POSCAR file is invalid."); + + # If a scale factor other than 1 has been set, adjust the lattice vectors. + + if scaleFactor != 1.0: + for i, vector in enumerate(latticeVectors): + latticeVectors[i] = [scaleFactor * x for x in vector]; + + # Build a list of atom-type numbers. + + atomTypeNumbers = None; + + if atomTypes != None: + # If atom types were read from the POSCAR file, convert these to atomic numbers. + + atomicSymbols = []; + + for atomType, atomCount in zip(atomTypes, atomCounts): + atomicSymbols = atomicSymbols + [atomType] * atomCount; + + # Convert the atomic symbols to atom-type numbers. + + atomTypeNumbers = [ + Structure.AtomTypeToAtomTypeNumber(symbol, atomicSymbolLookupTable = atomicSymbolLookupTable) + for symbol in atomicSymbols + ]; + else: + # If not, issue a warning and assign negative type numbers from -1. + + warnings.warn("Structure objects returned by reading VASP 4-format POSCAR files numbers will be initialised with negative atomic numbers from -1.", UserWarning); + + atomTypeNumbers = []; + + for i, atomCount in enumerate(atomCounts): + atomTypeNumbers = atomTypeNumbers + [-1 * (i + 1)] * atomCount; + + # If the atom positions are given in Cartesian coordinates, convert them to fractional coordinates. + + if coordinateType == 'c': + atomPositions = StructureTools.CartesianToFractionalCoordinates(latticeVectors, atomPositions); + + # Return a Structure object. + + return Structure.Structure(latticeVectors, atomPositions, atomTypeNumbers, name = systemName); + +def WritePOSCARFile(structure, outputWriter, atomicSymbolLookupTable = None): + # Write the system name; Structure.GetName() returns a sensible default value if a name is not set. + + outputWriter.write("{0}\n".format(structure.GetName())); + + # Write the scale factor. + + outputWriter.write(" {0: >19.16f}\n".format(1.0)); + + # Write the lattice vectors. + + for ax, ay, az in structure.GetLatticeVectors(): + outputWriter.write(" {0: >21.16f} {1: >21.16f} {2: >21.16f}\n".format(ax, ay, az)); + + # Write the atom types and counts. + + atomicSymbols, atomCounts = structure.GetAtomicSymbolsCounts(atomicSymbolLookupTable = atomicSymbolLookupTable); + + for atomicSymbol in atomicSymbols: + outputWriter.write(" {0: >3}".format(atomicSymbol)); + + outputWriter.write("\n"); + + for atomCount in atomCounts: + outputWriter.write(" {0: >3}".format(atomCount)); + + outputWriter.write("\n"); + + # Write the coordinate type. + + outputWriter.write("Direct\n"); + + # Write the atom positions. + + for x, y, z in structure.GetAtomPositions(): + outputWriter.write(" {0: >21.16f} {1: >21.16f} {2: >21.16f}\n".format(x, y, z)); + +"""SCons.Tool.gas + +Tool-specific initialization for as, the Gnu assembler. + +There normally shouldn't be any need to import this module directly. +It will usually be imported through the generic SCons.Tool.Tool() +selection method. + +""" + +# +# Copyright (c) 2001 - 2014 The SCons Foundation +# +# Permission is hereby granted, free of charge, to any person obtaining +# a copy of this software and associated documentation files (the +# "Software"), to deal in the Software without restriction, including +# without limitation the rights to use, copy, modify, merge, publish, +# distribute, sublicense, and/or sell copies of the Software, and to +# permit persons to whom the Software is furnished to do so, subject to +# the following conditions: +# +# The above copyright notice and this permission notice shall be included +# in all copies or substantial portions of the Software. +# +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY +# KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE +# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +# LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +# OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +# WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. +# + +__revision__ = "src/engine/SCons/Tool/gas.py 2014/09/27 12:51:43 garyo" + +as_module = __import__('as', globals(), locals(), []) + +assemblers = ['as', 'gas'] + +def generate(env): + """Add Builders and construction variables for as to an Environment.""" + as_module.generate(env) + + env['AS'] = env.Detect(assemblers) or 'as' + +def exists(env): + return env.Detect(assemblers) + +# Local Variables: +# tab-width:4 +# indent-tabs-mode:nil +# End: +# vim: set expandtab tabstop=4 shiftwidth=4: + +#!/usr/bin/env python +# +# Copyright 2014 The LibYuv Project Authors. All rights reserved. +# +# Use of this source code is governed by a BSD-style license +# that can be found in the LICENSE file in the root of the source +# tree. An additional intellectual property rights grant can be found +# in the file PATENTS. All contributing project authors may +# be found in the AUTHORS file in the root of the source tree. + + +# This script is a modified copy of the src/build/gyp_chromium.py file. +# It is needed for parallel processing. + +# This file is (possibly, depending on python version) imported by +# gyp_libyuv when GYP_PARALLEL=1 and it creates sub-processes +# through the multiprocessing library. + +# Importing in Python 2.6 (fixed in 2.7) on Windows doesn't search for +# imports that don't end in .py (and aren't directories with an +# __init__.py). This wrapper makes "import gyp_libyuv" work with +# those old versions and makes it possible to execute gyp_libyuv.py +# directly on Windows where the extension is useful. + +import os + +path = os.path.abspath(os.path.split(__file__)[0]) +execfile(os.path.join(path, 'gyp_libyuv')) + +import warnings +warnings.warn("eventlet.processes is deprecated in favor of " + "eventlet.green.subprocess, which is API-compatible with the standard " + " library subprocess module.", + DeprecationWarning, stacklevel=2) + + +import errno +import os +import popen2 +import signal + +from eventlet import api +from eventlet import pools +from eventlet import greenio + + +class DeadProcess(RuntimeError): + pass + +def cooperative_wait(pobj, check_interval=0.01): + """ Waits for a child process to exit, returning the status + code. + + Unlike ``os.wait``, :func:`cooperative_wait` does not block the entire + process, only the calling coroutine. If the child process does not die, + :func:`cooperative_wait` could wait forever. + + The argument *check_interval* is the amount of time, in seconds, that + :func:`cooperative_wait` will sleep between calls to ``os.waitpid``. + """ + try: + while True: + status = pobj.poll() + if status >= 0: + return status + api.sleep(check_interval) + except OSError, e: + if e.errno == errno.ECHILD: + # no child process, this happens if the child process + # already died and has been cleaned up, or if you just + # called with a random pid value + return -1 + else: + raise + + +class Process(object): + """Construct Process objects, then call read, and write on them.""" + process_number = 0 + def __init__(self, command, args, dead_callback=lambda:None): + self.process_number = self.process_number + 1 + Process.process_number = self.process_number + self.command = command + self.args = args + self._dead_callback = dead_callback + self.run() + + def run(self): + self.dead = False + self.started = False + self.popen4 = None + + ## We use popen4 so that read() will read from either stdout or stderr + self.popen4 = popen2.Popen4([self.command] + self.args) + child_stdout_stderr = self.popen4.fromchild + child_stdin = self.popen4.tochild + self.child_stdout_stderr = greenio.GreenPipe(child_stdout_stderr, child_stdout_stderr.mode, 0) + self.child_stdin = greenio.GreenPipe(child_stdin, child_stdin.mode, 0) + + self.sendall = self.child_stdin.write + self.send = self.child_stdin.write + self.recv = self.child_stdout_stderr.read + self.readline = self.child_stdout_stderr.readline + self._read_first_result = False + + def wait(self): + return cooperative_wait(self.popen4) + + def dead_callback(self): + self.wait() + self.dead = True + if self._dead_callback: + self._dead_callback() + + def makefile(self, mode, *arg): + if mode.startswith('r'): + return self.child_stdout_stderr + if mode.startswith('w'): + return self.child_stdin + raise RuntimeError("Unknown mode", mode) + + def read(self, amount=None): + """Reads from the stdout and stderr of the child process. + The first call to read() will return a string; subsequent + calls may raise a DeadProcess when EOF occurs on the pipe. + """ + result = self.child_stdout_stderr.read(amount) + if result == '' and self._read_first_result: + # This process is dead. + self.dead_callback() + raise DeadProcess + else: + self._read_first_result = True + return result + + def write(self, stuff): + written = 0 + try: + written = self.child_stdin.write(stuff) + self.child_stdin.flush() + except ValueError, e: + ## File was closed + assert str(e) == 'I/O operation on closed file' + if written == 0: + self.dead_callback() + raise DeadProcess + + def flush(self): + self.child_stdin.flush() + + def close(self): + self.child_stdout_stderr.close() + self.child_stdin.close() + self.dead_callback() + + def close_stdin(self): + self.child_stdin.close() + + def kill(self, sig=None): + if sig == None: + sig = signal.SIGTERM + pid = self.getpid() + os.kill(pid, sig) + + def getpid(self): + return self.popen4.pid + + +class ProcessPool(pools.Pool): + def __init__(self, command, args=None, min_size=0, max_size=4): + """*command* + the command to run + """ + self.command = command + if args is None: + args = [] + self.args = args + pools.Pool.__init__(self, min_size, max_size) + + def create(self): + """Generate a process + """ + def dead_callback(): + self.current_size -= 1 + return Process(self.command, self.args, dead_callback) + + def put(self, item): + if not item.dead: + if item.popen4.poll() != -1: + item.dead_callback() + else: + pools.Pool.put(self, item) + +# Permission is hereby granted, free of charge, to any person obtaining a +# copy of this software and associated documentation files (the +# "Software"), to deal in the Software without restriction, including +# without limitation the rights to use, copy, modify, merge, publish, dis- +# tribute, sublicense, and/or sell copies of the Software, and to permit +# persons to whom the Software is furnished to do so, subject to the fol- +# lowing conditions: +# +# The above copyright notice and this permission notice shall be included +# in all copies or substantial portions of the Software. +# +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS +# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL- +# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT +# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, +# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS +# IN THE SOFTWARE. +# +# Created by Chris Huegle for TellApart, Inc. + +class ConnectionSettingAttribute(object): + """ + Represents the ConnectionSetting segment of ELB Attributes. + """ + def __init__(self, connection=None): + self.idle_timeout = None + + def __repr__(self): + return 'ConnectionSettingAttribute(%s)' % ( + self.idle_timeout) + + def startElement(self, name, attrs, connection): + pass + + def endElement(self, name, value, connection): + if name == 'IdleTimeout': + self.idle_timeout = int(value) + +class CrossZoneLoadBalancingAttribute(object): + """ + Represents the CrossZoneLoadBalancing segement of ELB Attributes. + """ + def __init__(self, connection=None): + self.enabled = None + + def __repr__(self): + return 'CrossZoneLoadBalancingAttribute(%s)' % ( + self.enabled) + + def startElement(self, name, attrs, connection): + pass + + def endElement(self, name, value, connection): + if name == 'Enabled': + if value.lower() == 'true': + self.enabled = True + else: + self.enabled = False + + +class AccessLogAttribute(object): + """ + Represents the AccessLog segment of ELB attributes. + """ + def __init__(self, connection=None): + self.enabled = None + self.s3_bucket_name = None + self.s3_bucket_prefix = None + self.emit_interval = None + + def __repr__(self): + return 'AccessLog(%s, %s, %s, %s)' % ( + self.enabled, + self.s3_bucket_name, + self.s3_bucket_prefix, + self.emit_interval + ) + + def startElement(self, name, attrs, connection): + pass + + def endElement(self, name, value, connection): + if name == 'Enabled': + if value.lower() == 'true': + self.enabled = True + else: + self.enabled = False + elif name == 'S3BucketName': + self.s3_bucket_name = value + elif name == 'S3BucketPrefix': + self.s3_bucket_prefix = value + elif name == 'EmitInterval': + self.emit_interval = int(value) + + +class ConnectionDrainingAttribute(object): + """ + Represents the ConnectionDraining segment of ELB attributes. + """ + def __init__(self, connection=None): + self.enabled = None + self.timeout = None + + def __repr__(self): + return 'ConnectionDraining(%s, %s)' % ( + self.enabled, + self.timeout + ) + + def startElement(self, name, attrs, connection): + pass + + def endElement(self, name, value, connection): + if name == 'Enabled': + if value.lower() == 'true': + self.enabled = True + else: + self.enabled = False + elif name == 'Timeout': + self.timeout = int(value) + + +class LbAttributes(object): + """ + Represents the Attributes of an Elastic Load Balancer. + """ + def __init__(self, connection=None): + self.connection = connection + self.cross_zone_load_balancing = CrossZoneLoadBalancingAttribute( + self.connection) + self.access_log = AccessLogAttribute(self.connection) + self.connection_draining = ConnectionDrainingAttribute(self.connection) + self.connecting_settings = ConnectionSettingAttribute(self.connection) + + def __repr__(self): + return 'LbAttributes(%s, %s, %s, %s)' % ( + repr(self.cross_zone_load_balancing), + repr(self.access_log), + repr(self.connection_draining), + repr(self.connecting_settings)) + + def startElement(self, name, attrs, connection): + if name == 'CrossZoneLoadBalancing': + return self.cross_zone_load_balancing + if name == 'AccessLog': + return self.access_log + if name == 'ConnectionDraining': + return self.connection_draining + if name == 'ConnectionSettings': + return self.connecting_settings + + def endElement(self, name, value, connection): + pass + +#======================================================================= +# +# Python Lexical Analyser +# +# Actions for use in token specifications +# +#======================================================================= + +class Action(object): + + def perform(self, token_stream, text): + pass # abstract + + def same_as(self, other): + return self is other + + +class Return(Action): + """ + Internal Plex action which causes |value| to + be returned as the value of the associated token + """ + + def __init__(self, value): + self.value = value + + def perform(self, token_stream, text): + return self.value + + def same_as(self, other): + return isinstance(other, Return) and self.value == other.value + + def __repr__(self): + return "Return(%s)" % repr(self.value) + + +class Call(Action): + """ + Internal Plex action which causes a function to be called. + """ + + def __init__(self, function): + self.function = function + + def perform(self, token_stream, text): + return self.function(token_stream, text) + + def __repr__(self): + return "Call(%s)" % self.function.__name__ + + def same_as(self, other): + return isinstance(other, Call) and self.function is other.function + + +class Begin(Action): + """ + Begin(state_name) is a Plex action which causes the Scanner to + enter the state |state_name|. See the docstring of Plex.Lexicon + for more information. + """ + + def __init__(self, state_name): + self.state_name = state_name + + def perform(self, token_stream, text): + token_stream.begin(self.state_name) + + def __repr__(self): + return "Begin(%s)" % self.state_name + + def same_as(self, other): + return isinstance(other, Begin) and self.state_name == other.state_name + + +class Ignore(Action): + """ + IGNORE is a Plex action which causes its associated token + to be ignored. See the docstring of Plex.Lexicon for more + information. + """ + def perform(self, token_stream, text): + return None + + def __repr__(self): + return "IGNORE" + +IGNORE = Ignore() +#IGNORE.__doc__ = Ignore.__doc__ + +class Text(Action): + """ + TEXT is a Plex action which causes the text of a token to + be returned as the value of the token. See the docstring of + Plex.Lexicon for more information. + """ + + def perform(self, token_stream, text): + return text + + def __repr__(self): + return "TEXT" + +TEXT = Text() +#TEXT.__doc__ = Text.__doc__ + + + +try: + from urllib.parse import urljoin +except ImportError: + # Python 2 + from urlparse import urljoin + +from django.conf import settings + +# See http://docs.cksource.com/ckeditor_api/symbols/CKEDITOR.config.html +# for all settings + +CKEDITOR_SETTINGS = getattr(settings, 'CKEDITOR_SETTINGS', { + 'language': '{{ language }}', + 'toolbar': 'CMS', + 'skin': 'moono', + 'toolbarCanCollapse': False, +}) + +INSTALLED_APPS = getattr(settings, 'INSTALLED_APPS', []) +if 'cms.plugins.picture' in INSTALLED_APPS or 'djangocms_picture' in INSTALLED_APPS: + save_function_default = 'djangocms_text_ckeditor.picture_save.create_picture_plugin' +else: + save_function_default = None + +TEXT_SAVE_IMAGE_FUNCTION = getattr(settings, 'TEXT_SAVE_IMAGE_FUNCTION', save_function_default) +TEXT_ADDITIONAL_TAGS = getattr(settings, 'TEXT_ADDITIONAL_TAGS', ()) +TEXT_ADDITIONAL_ATTRIBUTES = getattr(settings, 'TEXT_ADDITIONAL_ATTRIBUTES', ()) +TEXT_ADDITIONAL_PROTOCOLS = getattr(settings, 'TEXT_ADDITIONAL_PROTOCOLS', ()) +TEXT_CKEDITOR_CONFIGURATION = getattr(settings, 'TEXT_CKEDITOR_CONFIGURATION', None) +TEXT_HTML_SANITIZE = getattr(settings, 'TEXT_HTML_SANITIZE', True) +TEXT_CKEDITOR_BASE_PATH = getattr(settings, 'TEXT_CKEDITOR_BASE_PATH', urljoin(settings.STATIC_URL, 'djangocms_text_ckeditor/ckeditor/')) +TEXT_AUTO_HYPHENATE = getattr(settings, 'TEXT_AUTO_HYPHENATE', True) + +ALLOW_TOKEN_PARSERS = ( + 'djangocms_text_ckeditor.attribute_parsers.DataAttributeParser', +) + +# -*- coding: utf-8 -*- +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. + +import unittest +from mock import Mock + +from airflow.models import TaskInstance +from airflow.ti_deps.deps.dag_ti_slots_available_dep import DagTISlotsAvailableDep + + +class DagTISlotsAvailableDepTest(unittest.TestCase): + + def test_concurrency_reached(self): + """ + Test concurrency reached should fail dep + """ + dag = Mock(concurrency=1, concurrency_reached=True) + task = Mock(dag=dag) + ti = TaskInstance(task, execution_date=None) + + self.assertFalse(DagTISlotsAvailableDep().is_met(ti=ti)) + + def test_all_conditions_met(self): + """ + Test all conditions met should pass dep + """ + dag = Mock(concurrency=1, concurrency_reached=False) + task = Mock(dag=dag) + ti = TaskInstance(task, execution_date=None) + + self.assertTrue(DagTISlotsAvailableDep().is_met(ti=ti)) + +# This file is part of Buildbot. Buildbot is free software: you can +# redistribute it and/or modify it under the terms of the GNU General Public +# License as published by the Free Software Foundation, version 2. +# +# This program is distributed in the hope that it will be useful, but WITHOUT +# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS +# FOR A PARTICULAR PURPOSE. See the GNU General Public License for more +# details. +# +# You should have received a copy of the GNU General Public License along with +# this program; if not, write to the Free Software Foundation, Inc., 51 +# Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. +# +# Copyright Buildbot Team Members + +from __future__ import absolute_import +from __future__ import print_function + +import sqlalchemy as sa + +from twisted.trial import unittest + +from buildbot.test.util import migration +from buildbot.util import sautils + + +class Migration(migration.MigrateTestMixin, unittest.TestCase): + + def setUp(self): + return self.setUpMigrateTest() + + def tearDown(self): + return self.tearDownMigrateTest() + + def create_tables_thd(self, conn): + metadata = sa.MetaData() + metadata.bind = conn + + patches = sautils.Table( + 'patches', metadata, + sa.Column('id', sa.Integer, primary_key=True), + sa.Column('patchlevel', sa.Integer, nullable=False), + sa.Column('patch_base64', sa.Text, nullable=False), + sa.Column('patch_author', sa.Text, nullable=False), + sa.Column('patch_comment', sa.Text, nullable=False), + sa.Column('subdir', sa.Text), + ) + + sourcestamps = sautils.Table( + 'sourcestamps', metadata, + sa.Column('id', sa.Integer, primary_key=True), + sa.Column('ss_hash', sa.String(40), nullable=False), + sa.Column('branch', sa.String(256)), + sa.Column('revision', sa.String(256)), + sa.Column('patchid', sa.Integer, sa.ForeignKey('patches.id')), + sa.Column('repository', sa.String(length=512), nullable=False, + server_default=''), + sa.Column('codebase', sa.String(256), nullable=False, + server_default=sa.DefaultClause("")), + sa.Column('project', sa.String(length=512), nullable=False, + server_default=''), + sa.Column('created_at', sa.Integer, nullable=False), + ) + + changes = sautils.Table( + 'changes', metadata, + sa.Column('changeid', sa.Integer, primary_key=True), + sa.Column('author', sa.String(256), nullable=False), + sa.Column('comments', sa.Text, nullable=False), + sa.Column('branch', sa.String(256)), + sa.Column('revision', sa.String(256)), + sa.Column('revlink', sa.String(256)), + sa.Column('when_timestamp', sa.Integer, nullable=False), + sa.Column('category', sa.String(256)), + sa.Column('repository', sa.String(length=512), nullable=False, + server_default=''), + sa.Column('codebase', sa.String(256), nullable=False, + server_default=sa.DefaultClause("")), + sa.Column('project', sa.String(length=512), nullable=False, + server_default=''), + sa.Column('sourcestampid', sa.Integer, + sa.ForeignKey('sourcestamps.id')), + ) + patches.create() + sourcestamps.create() + changes.create() + + def test_update(self): + def setup_thd(conn): + self.create_tables_thd(conn) + + def verify_thd(conn): + metadata = sa.MetaData() + metadata.bind = conn + + changes = sautils.Table('changes', metadata, autoload=True) + self.assertIsInstance(changes.c.parent_changeids.type, sa.Integer) + + return self.do_test_migration(42, 43, setup_thd, verify_thd) + +#!/usr/bin/env python + +import ConfigParser +import socket +import os +import json + +def main(): + cf = ConfigParser.ConfigParser() + cf.read("conf/oj.conf") + + judge_host = cf.get('sandbox', 'judgeHost') + userid = int(cf.get('sandbox', 'userid')) + + server = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM) + if os.path.exists("/tmp/judge_root.sock"): + os.unlink("/tmp/judge_root.sock") + server.bind("/tmp/judge_root.sock") + os.system('chmod 777 /tmp/judge_root.sock') + server.listen(0) + while True: + connection, address = server.accept() + infor = json.loads(connection.recv(1024).decode()) + work_dir, bin, usrout, errout, input_dir, stdin, time_limit, mem_limit = infor + cmd = "%s %s %s %s %s %s %s %s %s %s"%(judge_host, work_dir, bin, usrout, errout, input_dir, stdin, time_limit, mem_limit, userid) + print cmd + tmp = os.popen(cmd).read() + result, time_used, mem_used = [int(s) for s in tmp.split()] + success = result == 0 + time_exceeded = result == 2 + mem_exceeded = result == 3 + connection.send(json.dumps([success, time_exceeded, mem_exceeded, time_used, mem_used]).encode()) + + +if __name__ == '__main__': + main() + +"""Tests for quotient rings.""" + +from sympy import QQ, ZZ +from sympy.abc import x, y + +from sympy.polys.polyerrors import NotReversible + +from sympy.utilities.pytest import raises +from sympy.core.compatibility import range + + +def test_QuotientRingElement(): + R = QQ.old_poly_ring(x)/[x**10] + X = R.convert(x) + + assert X*(X + 1) == R.convert(x**2 + x) + assert X*x == R.convert(x**2) + assert x*X == R.convert(x**2) + assert X + x == R.convert(2*x) + assert x + X == 2*X + assert X**2 == R.convert(x**2) + assert 1/(1 - X) == R.convert(sum(x**i for i in range(10))) + assert X**10 == R.zero + assert X != x + + raises(NotReversible, lambda: 1/X) + + +def test_QuotientRing(): + I = QQ.old_poly_ring(x).ideal(x**2 + 1) + R = QQ.old_poly_ring(x)/I + + assert R == QQ.old_poly_ring(x)/[x**2 + 1] + assert R == QQ.old_poly_ring(x)/QQ.old_poly_ring(x).ideal(x**2 + 1) + assert R != QQ.old_poly_ring(x) + + assert R.convert(1)/x == -x + I + assert -1 + I == x**2 + I + assert R.convert(ZZ(1), ZZ) == 1 + I + assert R.convert(R.convert(x), R) == R.convert(x) + + X = R.convert(x) + Y = QQ.old_poly_ring(x).convert(x) + assert -1 + I == X**2 + I + assert -1 + I == Y**2 + I + assert R.to_sympy(X) == x + + raises(ValueError, lambda: QQ.old_poly_ring(x)/QQ.old_poly_ring(x, y).ideal(x)) + + R = QQ.old_poly_ring(x, order="ilex") + I = R.ideal(x) + assert R.convert(1) + I == (R/I).convert(1) + +# automatically generated by the FlatBuffers compiler, do not modify + +# namespace: NamespaceA + +import flatbuffers + +class SecondTableInA(object): + __slots__ = ['_tab'] + + @classmethod + def GetRootAsSecondTableInA(cls, buf, offset): + n = flatbuffers.encode.Get(flatbuffers.packer.uoffset, buf, offset) + x = SecondTableInA() + x.Init(buf, n + offset) + return x + + # SecondTableInA + def Init(self, buf, pos): + self._tab = flatbuffers.table.Table(buf, pos) + + # SecondTableInA + def ReferToC(self): + o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(4)) + if o != 0: + x = self._tab.Indirect(o + self._tab.Pos) + from .TableInC import TableInC + obj = TableInC() + obj.Init(self._tab.Bytes, x) + return obj + return None + +def SecondTableInAStart(builder): builder.StartObject(1) +def SecondTableInAAddReferToC(builder, referToC): builder.PrependUOffsetTRelativeSlot(0, flatbuffers.number_types.UOffsetTFlags.py_type(referToC), 0) +def SecondTableInAEnd(builder): return builder.EndObject() + +# Copyright 2015, Google Inc. +# All rights reserved. +# +# Redistribution and use in source and binary forms, with or without +# modification, are permitted provided that the following conditions are +# met: +# +# * Redistributions of source code must retain the above copyright +# notice, this list of conditions and the following disclaimer. +# * Redistributions in binary form must reproduce the above +# copyright notice, this list of conditions and the following disclaimer +# in the documentation and/or other materials provided with the +# distribution. +# * Neither the name of Google Inc. nor the names of its +# contributors may be used to endorse or promote products derived from +# this software without specific prior written permission. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + + +""" +Connectors wrap the details of communicating with different Bitcoin clients and implementations. +""" +import io +import logging +import time + +import bitcoin.rpc +import requests +from bitcoin.core import CTransaction +from cert_schema import Chain +from pycoin.serialize import b2h +from pycoin.services import providers +from pycoin.services.blockr_io import BlockrioProvider +from pycoin.services.insight import InsightProvider +from pycoin.services.providers import service_provider_methods +from pycoin.tx import Spendable + +from cert_issuer.errors import ConnectorError, BroadcastError +from cert_issuer.helpers import hexlify +from cert_issuer.helpers import unhexlify + +BROADCAST_RETRY_INTERVAL = 30 + +try: + from urllib2 import urlopen, HTTPError + from urllib import urlencode +except ImportError: + from urllib.request import urlopen, HTTPError + from urllib.parse import urlencode + +MAX_BROADCAST_ATTEMPTS = 3 + + +def try_get(url): + """throw error if call fails""" + response = requests.get(url) + if int(response.status_code) != 200: + error_message = 'Error! status_code={}, error={}'.format( + response.status_code, response.json()['error']) + logging.error(error_message) + raise ConnectorError(error_message) + return response + + +def to_hex(transaction): + s = io.BytesIO() + transaction.stream(s) + tx_as_hex = b2h(s.getvalue()) + return tx_as_hex + + +class BlockExplorerBroadcaster(object): + def __init__(self, base_url): + self.base_url = base_url + + def broadcast_tx(self, tx): + hextx = to_hex(tx) + broadcast_url = self.base_url + '/tx/send' + response = requests.post(broadcast_url, json={'rawtx': hextx}) + if int(response.status_code) == 200: + tx_id = response.json().get('txid', None) + return tx_id + logging.error('Error broadcasting the transaction through the BlockExplorer API. Error msg: %s', response.text) + raise BroadcastError(response.text) + + +class BlockcypherBroadcaster(object): + """ + Note that this needs an API token + """ + + def __init__(self, base_url, api_token): + self.base_url = base_url + self.api_token = api_token + + def broadcast_tx(self, tx): + hextx = to_hex(tx) + broadcast_url = self.base_url + '/txs/push?token=' + self.api_token + response = requests.post(broadcast_url, json={'tx': hextx}) + if int(response.status_code) == 200: + tx_id = response.json().get('txid', None) + return tx_id + logging.error('Error broadcasting the transaction through the Blockcypher API. Error msg: %s', response.text) + raise BroadcastError(response.text) + + +class BlockrIOBroadcaster(object): + def __init__(self, base_url): + self.base_url = base_url + + def broadcast_tx(self, tx): + hextx = to_hex(tx) + url = self.base_url + '/tx/push' + response = requests.post(url, json={'hex': hextx}) + if int(response.status_code) == 200: + tx_id = response.json().get('data', None) + return tx_id + logging.error('Error broadcasting the transaction through the Blockr.IO API. Error msg: %s', response.text) + raise BroadcastError(response.text) + + +class BitcoindConnector(object): + def __init__(self, netcode): + self.netcode = netcode + + def broadcast_tx(self, transaction): + as_hex = transaction.as_hex() + transaction = CTransaction.deserialize(unhexlify(as_hex)) + tx_id = bitcoin.rpc.Proxy().sendrawtransaction(transaction) + # reverse endianness for bitcoind + return hexlify(bytearray(tx_id)[::-1]) + + def spendables_for_address(self, address): + """ + Converts to pycoin Spendable type + :param address: + :return: list of Spendables + """ + unspent_outputs = bitcoin.rpc.Proxy().listunspent(addrs=[address]) + logging.debug('spendables_for_address %s', address) + + spendables = [] + for unspent in unspent_outputs: + coin_value = unspent.get('amount', 0) + outpoint = unspent.get('outpoint') + script = unspent.get('scriptPubKey') + previous_hash = outpoint.hash + previous_index = outpoint.n + spendables.append(Spendable(coin_value, script, previous_hash, previous_index)) + return spendables + + +class ServiceProviderConnector(object): + def __init__(self, bitcoin_chain): + self.bitcoin_chain = bitcoin_chain + + def spendables_for_address(self, bitcoin_address): + for m in service_provider_methods('spendables_for_address', get_providers_for_chain(self.bitcoin_chain)): + try: + logging.debug('m=%s', m) + spendables = m(bitcoin_address) + return spendables + except Exception as e: + logging.warning(e) + pass + return [] + + def get_unspent_outputs(self, address): + """ + Get unspent outputs at the address + :param address: + :return: + """ + logging.debug('get_unspent_outputs for address=%s', address) + spendables = self.spendables_for_address(bitcoin_address=address) + if spendables: + return sorted(spendables, key=lambda x: hash(x.coin_value)) + return None + + def get_balance(self, address): + """ + Get balance available to spend at the address + :param address: + :return: + """ + spendables = self.get_unspent_outputs(address) + if not spendables: + logging.warning('address %s has a balance of 0', address) + return 0 + + balance = sum(s.coin_value for s in spendables) + return balance + + def broadcast_tx(self, tx): + """ + Broadcast the transaction through the configured set of providers + + :param tx: + :return: + """ + + return ServiceProviderConnector.broadcast_tx_with_chain(tx, self.bitcoin_chain) + + @staticmethod + def broadcast_tx_with_chain(tx, bitcoin_chain): + """ + Broadcast the transaction through the configured set of providers + + :param tx: + :param bitcoin_chain: + :return: + """ + last_exception = None + final_tx_id = None + + # Unlike other providers, we want to broadcast to all available apis + for attempt_number in range(0, MAX_BROADCAST_ATTEMPTS): + for method_provider in service_provider_methods('broadcast_tx', + get_providers_for_chain(bitcoin_chain)): + try: + tx_id = method_provider(tx) + if tx_id: + logging.info('Broadcasting succeeded with method_provider=%s, txid=%s', str(method_provider), + tx_id) + if final_tx_id and final_tx_id != tx_id: + logging.error( + 'This should never happen; fail and investigate if it does. Got conflicting tx_ids=%s and %s. Hextx=%s', + final_tx_id, tx_id, tx.as_hex()) + raise Exception('Got conflicting tx_ids.') + final_tx_id = tx_id + except Exception as e: + logging.warning('Caught exception trying provider %s. Trying another. Exception=%s', + str(method_provider), e) + last_exception = e + # At least 1 provider succeeded, so return + if final_tx_id: + return final_tx_id + else: + logging.warning('Broadcasting failed. Waiting before retrying. This is attempt number %d', + attempt_number) + time.sleep(BROADCAST_RETRY_INTERVAL) + logging.error('Failed broadcasting through all providers') + logging.error(last_exception, exc_info=True) + raise BroadcastError(last_exception) + + +PYCOIN_BTC_PROVIDERS = "blockchain.info blockexplorer.com blockcypher.com chain.so" +PYCOIN_XTN_PROVIDERS = "blockexplorer.com" # chain.so + +# initialize connectors +connectors = {} + +# configure mainnet providers +provider_list = providers.providers_for_config_string(PYCOIN_BTC_PROVIDERS, Chain.mainnet.netcode) +provider_list.append(BlockrIOBroadcaster('https://btc.blockr.io/api/v1')) +provider_list.append(BlockExplorerBroadcaster('https://blockexplorer.com/api')) +provider_list.append(BlockrioProvider(Chain.mainnet.netcode)) +provider_list.append(InsightProvider(netcode=Chain.mainnet.netcode)) +connectors[Chain.mainnet.netcode] = provider_list + +# configure testnet providers +xtn_provider_list = providers.providers_for_config_string(PYCOIN_XTN_PROVIDERS, Chain.testnet.netcode) +xtn_provider_list.append(InsightProvider(base_url='https://test-insight.bitpay.com', netcode=Chain.testnet.netcode)) +xtn_provider_list.append(BlockrIOBroadcaster('https://tbtc.blockr.io/api/v1')) +xtn_provider_list.append(BlockExplorerBroadcaster('https://testnet.blockexplorer.com/api')) +xtn_provider_list.append(BlockrioProvider(Chain.testnet.netcode)) +connectors[Chain.testnet.netcode] = xtn_provider_list + +# workaround for regtest +connectors['REG'] = [BitcoindConnector(Chain.testnet.netcode)] + + +def get_providers_for_chain(bitcoin_chain): + if bitcoin_chain == Chain.regtest: + return connectors['REG'] + return connectors[bitcoin_chain.netcode] + +__author__ = 'firatlepirate' + + +import socket +import threading +import Queue +import time +import errno + + +class WriteThread (threading.Thread): + def __init__(self, name, cSocket, address,fihrist,threadQueue, logQueue ): + threading.Thread.__init__(self) + self.name = name + self.cSocket = cSocket + self.address = address + self.lQueue = logQueue + self.tQueue = threadQueue + + self.fihrist = fihrist + self.nickname="" + self.flag=False + def run(self): + self.lQueue.put("Starting " + self.name) + + while True: + # burasi kuyrukta sirasi gelen mesajlari + # gondermek icin kullanilacak + + if self.tQueue.qsize()>0 or self.flag: + if self.nickname=="": + self.nickname=self.tQueue.get() + self.flag=True + if self.fihrist[self.nickname].qsize()>0: + queue_message = self.fihrist[self.nickname].get() + + # gonderilen ozel mesajsa + if not queue_message[0]=="SAY" and len(queue_message)==3 and not queue_message=="QUI" : + message_to_send = "MSG "+str(queue_message[0])+":"+str(queue_message[1])+";"+str(queue_message[2]) + self.cSocket.send(str(message_to_send)) + # genel mesajsa + elif queue_message[0]=="SAY": + message_to_send = "SAY "+str(queue_message[1])+":"+str(queue_message[2]) + self.cSocket.send(str(message_to_send)) + print(message_to_send) + elif queue_message=="QUI": + # fihristten sil + del self.fihrist[self.nickname] + break + # hicbiri degilse sistem mesajidir + else: + message_to_send = "SYS "+str(queue_message[1]) + self.cSocket.send(str(message_to_send)) + + self.lQueue.put("Exiting " + self.name) + + + +class ReadThread (threading.Thread): + def __init__(self, name, cSocket, address,fihrist,threadQueue,logQueue): + threading.Thread.__init__(self) + self.name = name + self.cSocket = cSocket + self.address = address + self.lQueue = logQueue + self.fihrist = fihrist + self.tQueue = threadQueue + + self.nickname="" + def parser(self, data): + #data = data.strip() + + # henuz login olmadiysa + if not self.nickname and not data[0:3] == "USR": + response="ERL" + self.cSocket.send(response) + else: + # data sekli bozuksa + if len(data)<3: + response = "ERR" + self.cSocket.send(response) + return 0 + if data[0:3] == "USR": + if len(data)>4 and data[3]==" " and not data[3:len(data)]==" ": + self.nickname = data[4:] + if not self.nickname in self.fihrist: + # kullanici yoksa + + + response = "HEL " + self.nickname + self.cSocket.send(response) + self.fihrist[self.nickname]=Queue.Queue(10) + + # fihristi guncelle + #self.fihrist.update(...) + self.lQueue.put(self.nickname + " has joined.") + self.tQueue.put(self.nickname) + queue_message = ("SYS", self.nickname) + for items in self.fihrist.keys(): + self.fihrist[items].put(queue_message) + return 0 + else: + # kullanici reddedilecek + response = "REJ " + self.nickname + self.cSocket.send(response) + # baglantiyi kapat + # self.cSocket.close() + return 1 + else: + response = "ERR" + self.cSocket.send(response) + + elif data[0:3] == "QUI": + response = "BYE " + self.nickname + self.cSocket.send(response) + queue_message="QUI" + self.fihrist[self.nickname].put(queue_message) + + # log gonder + self.lQueue.put(self.nickname + " has left.") + # baglantiyi sil + self.cSocket.close() + return queue_message + + + elif data[0:3] == "LSQ": + a=" " + for i in self.fihrist.keys(): + a=a+i+":" + response="LSA"+a[:-1] + self.cSocket.send(response) + elif data[0:3] == "TIC": + response="TOC" + self.cSocket.send(response) + elif data[0:3] == "SAY": + if len(data)>4 and data[3]==" " and not data[4:]==" ": + message=data[4:] + queue_message = ("SAY", self.nickname, message) + for items in self.fihrist.keys(): + self.fihrist[items].put(queue_message) + + response="SOK" + self.cSocket.send(response) + + elif data[0:3] == "MSG": + + c=":" + if not data[4:]==" " and c in data[4:]: + to_nickname=data[4:data.index(":")] + message=data[data.index(":")+1:] + if not to_nickname in self.fihrist.keys(): + response = "MNO" + else: + + queue_message = (to_nickname, self.nickname, message) + # gonderilecek threadQueueyu fihristten alip icine yaz + + self.fihrist[to_nickname].put(queue_message) + + + response = "MOK" + self.cSocket.send(response) + + + + else: + # bir seye uymadiysa protokol hatasi verilecek + response = "ERR" + self.cSocket.send(response) + + + + + def run(self): + self.lQueue.put("Starting " + self.name) + + while True: + + try: + incoming_data=self.cSocket.recv(1024) + except socket.error ,e: + err=e.args[0] + if err == errno.EAGAIN or err == errno.EWOULDBLOCK: + time.sleep(1) + print 'No data available' + continue + else: + print("ERROR"+str(e)) + + queue_message = self.parser(incoming_data) + + + if(queue_message)=="QUI": + break + + + + + self.lQueue.put("Exiting " + self.name) + print(threading.activeCount()) + +class LoggerThread (threading.Thread): + def __init__(self, name, logQueue, logFileName): + threading.Thread.__init__(self) + self.name = name + self.lQueue = logQueue + self.fileName=logFileName + # dosyayi appendable olarak ac + self.fid = open(self.fileName, "a") + def log(self,message): + # gelen mesaji zamanla beraber bastir + t = time.ctime() + self.fid.write(t+":"+" "+ message+"\n") + self.fid.flush() + def run(self): + self.log("Starting " + self.name) + while True: + if self.lQueue.qsize() > 0: + # lQueue'da yeni mesaj varsa + # self.log() metodunu cagir + to_be_logged = self.lQueue.get() + self.log(to_be_logged) + self.log("Exiting" + self.name) + self.fid.close() + +userList={} + + +loggerQueue= Queue.Queue() + + + +thread3=LoggerThread("LoggerThread",loggerQueue,"log.txt") +thread3.start() +s = socket.socket() +#host = socket.gethostname() +host="127.0.0.1" +print("host"+host) +port = 12345 +s.bind((host, port)) +s.listen(5) +threadCounter=0 +threadCounter2=0 +while True: + loggerQueue.put("Waiting for connection") + print "Waiting for connection" + c, addr = s.accept() + workQueue = Queue.Queue() + loggerQueue.put("Got a connection from " + str(addr)) + print "Got a connection from ", addr + threadCounter += 1 + thread = ReadThread("ReadThread"+str(threadCounter), c, addr,userList,workQueue,loggerQueue) + threadCounter2 += 1 + thread2 = WriteThread("WriteThread"+str(threadCounter2), c, addr,userList,workQueue,loggerQueue) + + thread.start() + thread2.start() + +# -*- coding: utf-8 -*- + +"PDF Template Helper for FPDF.py" + +__author__ = "Mariano Reingart " +__copyright__ = "Copyright (C) 2010 Mariano Reingart" +__license__ = "LGPL 3.0" + +#PyFPDF-cover-test:format=PDF +#PyFPDF-cover-test:fn=invoice.pdf +#PyFPDF-cover-test:hash=5844bbebe3e33b0ac9cc15ac39327a81 +#PyFPDF-cover-test:res=invoice.csv + +import common +from fpdf import Template + +import os + +class randomfake: + RINT1_10 = [8, 5, 7, 9, 10, 8, 1, 9, 1, 7, 6, 2, 3, 7, 8, 4, 6, 5, 7, 2, \ + 5, 8, 6, 5, 5, 8, 7, 7, 6] + RINT65_90 = [67, 67, 87, 78, 84, 67, 86, 75, 86, 89, 81, 69, 72, 71, 84, \ + 80, 71 , 86 , 82 , 70 , 84 , 69 , 70] + RFLT = [0.820710198665, 0.342854771472, 0.0238515965298, 0.177658111957, \ + 0.422301628067, 0.701867781693, 0.168650983171, 0.329723498664, \ + 0.490481106182, 0.892634029991, 0.994758791625, 0.998243714035, \ + 0.596244312914 ,0.318601111178 ,0.321593673214 ,0.203486335469] + def __init__(self): + self.icnt1_10 = 0 + self.icnt65_90 = 0 + self.fcnt = 0 + + def randint(self, beg, end): + if beg == 1 and end == 10: + self.icnt1_10 += 1 + if self.icnt1_10 > len(self.RINT1_10): + self.icnt1_10 = 1 + return self.RINT1_10[self.icnt1_10 - 1] + if beg == 65 and end == 90: + self.icnt65_90 += 1 + if self.icnt65_90 > len(self.RINT65_90): + self.icnt65_90 = 1 + return self.RINT65_90[self.icnt65_90 - 1] + raise Exception("Not implemented") + + def random(self): + self.fcnt += 1 + if self.fcnt > len(self.RFLT): + self.fcnt = 1 + return self.RFLT[self.fcnt - 1] + +@common.add_unittest +def dotest(outputname, nostamp): + + # generate sample invoice (according Argentina's regulations) + from decimal import Decimal + + f = Template(format="A4", + title="Sample Invoice", author="Sample Company", + subject="Sample Customer", keywords="Electronic TAX Invoice") + if nostamp: + f.pdf._putinfo = lambda: common.test_putinfo(f.pdf) + random = randomfake() + else: + import random + + csvpath = os.path.join(common.basepath, "invoice.csv") + f.parse_csv(infile=csvpath, delimiter=";", decimal_sep=",") + + detail = "Lorem ipsum dolor sit amet, consectetur. " * 30 + items = [] + for i in range(1, 30): + ds = "Sample product %s" % i + qty = random.randint(1,10) + price = round(random.random()*100,3) + code = "%s%s%02d" % (chr(random.randint(65,90)), chr(random.randint(65,90)),i) + items.append(dict(code=code, unit='u', + qty=qty, price=price, + amount=qty*price, + ds="%s: %s" % (i,ds))) + + # divide and count lines + lines = 0 + li_items = [] + for it in items: + qty = it['qty'] + code = it['code'] + unit = it['unit'] + for ds in f.split_multicell(it['ds'], 'item_description01'): + # add item description line (without price nor amount) + li_items.append(dict(code=code, ds=ds, qty=qty, unit=unit, price=None, amount=None)) + # clean qty and code (show only at first) + unit = qty = code = None + # set last item line price and amount + li_items[-1].update(amount = it['amount'], + price = it['price']) + + obs="\nDetail:\n\n" + detail + for ds in f.split_multicell(obs, 'item_description01'): + li_items.append(dict(code=code, ds=ds, qty=qty, unit=unit, price=None, amount=None)) + + # calculate pages: + lines = len(li_items) + max_lines_per_page = 24 + pages = int(lines / (max_lines_per_page - 1)) + if lines % (max_lines_per_page - 1): pages = pages + 1 + + # completo campos y hojas + for page in range(1, int(pages)+1): + f.add_page() + f['page'] = 'Page %s of %s' % (page, pages) + if pages>1 and page page * (max_lines_per_page - 1): + break + if it['amount']: + total += Decimal("%.6f" % it['amount']) + if k > (page - 1) * (max_lines_per_page - 1): + li += 1 + if it['qty'] is not None: + f['item_quantity%02d' % li] = it['qty'] + if it['code'] is not None: + f['item_code%02d' % li] = it['code'] + if it['unit'] is not None: + f['item_unit%02d' % li] = it['unit'] + f['item_description%02d' % li] = it['ds'] + if it['price'] is not None: + f['item_price%02d' % li] = "%0.3f" % it['price'] + if it['amount'] is not None: + f['item_amount%02d' % li] = "%0.2f" % it['amount'] + + if pages == page: + f['net'] = "%0.2f" % (total/Decimal("1.21")) + f['vat'] = "%0.2f" % (total*(1-1/Decimal("1.21"))) + f['total_label'] = 'Total:' + else: + f['total_label'] = 'SubTotal:' + f['total'] = "%0.2f" % total + + f.render(outputname) + +if __name__ == "__main__": + common.testmain(__file__, dotest) + + +#!/usr/bin/env python +# -*- coding: utf-8 -*- +# ----------------------------------------------------------------------------- +# +# FreeType high-level python API - Copyright 2011 Nicolas P. Rougier +# Distributed under the terms of the new BSD license. +# +# ----------------------------------------------------------------------------- +''' +Show how to access glyph outline description. +''' +from freetype import * + +if __name__ == '__main__': + import numpy + import matplotlib.pyplot as plt + from matplotlib.path import Path + import matplotlib.patches as patches + + face = Face(b'./Vera.ttf') + face.set_char_size( 32*64 ) + face.load_char('g') + slot = face.glyph + + bitmap = face.glyph.bitmap + width = face.glyph.bitmap.width + rows = face.glyph.bitmap.rows + pitch = face.glyph.bitmap.pitch + + data = [] + for i in range(rows): + data.extend(bitmap.buffer[i*pitch:i*pitch+width]) + Z = numpy.array(data,dtype=numpy.ubyte).reshape(rows, width) + + outline = slot.outline + points = numpy.array(outline.points, dtype=[('x',float), ('y',float)]) + x, y = points['x'], points['y'] + + figure = plt.figure(figsize=(8,10)) + axis = figure.add_subplot(111) + #axis.scatter(points['x'], points['y'], alpha=.25) + start, end = 0, 0 + + VERTS, CODES = [], [] + # Iterate over each contour + for i in range(len(outline.contours)): + end = outline.contours[i] + points = outline.points[start:end+1] + points.append(points[0]) + tags = outline.tags[start:end+1] + tags.append(tags[0]) + + segments = [ [points[0],], ] + for j in range(1, len(points) ): + segments[-1].append(points[j]) + if tags[j] & (1 << 0) and j < (len(points)-1): + segments.append( [points[j],] ) + verts = [points[0], ] + codes = [Path.MOVETO,] + for segment in segments: + if len(segment) == 2: + verts.extend(segment[1:]) + codes.extend([Path.LINETO]) + elif len(segment) == 3: + verts.extend(segment[1:]) + codes.extend([Path.CURVE3, Path.CURVE3]) + else: + verts.append(segment[1]) + codes.append(Path.CURVE3) + for i in range(1,len(segment)-2): + A,B = segment[i], segment[i+1] + C = ((A[0]+B[0])/2.0, (A[1]+B[1])/2.0) + verts.extend([ C, B ]) + codes.extend([ Path.CURVE3, Path.CURVE3]) + verts.append(segment[-1]) + codes.append(Path.CURVE3) + VERTS.extend(verts) + CODES.extend(codes) + start = end+1 + + + # Draw glyph + path = Path(VERTS, CODES) + glyph = patches.PathPatch(path, fill = True, facecolor=(0.8,0.5,0.8), alpha=.25, lw=0) + glyph_outline = patches.PathPatch(path, fill = False, edgecolor='black', lw=3) + + plt.imshow(Z, extent=[x.min(), x.max(),y.min(), y.max()], + interpolation='nearest', cmap = plt.cm.gray_r, vmin=0, vmax=400) + plt.xticks(numpy.linspace(x.min(), x.max(), Z.shape[1]+1), ()) + plt.yticks(numpy.linspace(y.min(), y.max(), Z.shape[0]+1), ()) + plt.grid(color='k', linewidth=1, linestyle='-') + axis.add_patch(glyph) + axis.add_patch(glyph_outline) + axis.set_xlim(x.min(), x.max()) + axis.set_ylim(y.min(), y.max()) + + plt.savefig('test.pdf') + plt.show() + + + + +#!/usr/bin/env python2.7 +from __future__ import absolute_import, unicode_literals, print_function, division + +from sys import argv +from os import environ, stat, remove as _delete_file +from os.path import isfile, dirname, basename, abspath +from hashlib import sha256 +from subprocess import check_call as run + +from boto.s3.connection import S3Connection +from boto.s3.key import Key +from boto.exception import S3ResponseError + + +NEED_TO_UPLOAD_MARKER = '.need-to-upload' +BYTES_PER_MB = 1024 * 1024 +try: + BUCKET_NAME = environ['TWBS_S3_BUCKET'] +except KeyError: + raise SystemExit("TWBS_S3_BUCKET environment variable not set!") + + +def _sha256_of_file(filename): + hasher = sha256() + with open(filename, 'rb') as input_file: + hasher.update(input_file.read()) + file_hash = hasher.hexdigest() + print('sha256({}) = {}'.format(filename, file_hash)) + return file_hash + + +def _delete_file_quietly(filename): + try: + _delete_file(filename) + except (OSError, IOError): + pass + + +def _tarball_size(directory): + kib = stat(_tarball_filename_for(directory)).st_size // BYTES_PER_MB + return "{} MiB".format(kib) + + +def _tarball_filename_for(directory): + return abspath('./{}.tar.gz'.format(basename(directory))) + + +def _create_tarball(directory): + print("Creating tarball of {}...".format(directory)) + run(['tar', '-czf', _tarball_filename_for(directory), '-C', dirname(directory), basename(directory)]) + + +def _extract_tarball(directory): + print("Extracting tarball of {}...".format(directory)) + run(['tar', '-xzf', _tarball_filename_for(directory), '-C', dirname(directory)]) + + +def download(directory): + _delete_file_quietly(NEED_TO_UPLOAD_MARKER) + try: + print("Downloading {} tarball from S3...".format(friendly_name)) + key.get_contents_to_filename(_tarball_filename_for(directory)) + except S3ResponseError as err: + open(NEED_TO_UPLOAD_MARKER, 'a').close() + print(err) + raise SystemExit("Cached {} download failed!".format(friendly_name)) + print("Downloaded {}.".format(_tarball_size(directory))) + _extract_tarball(directory) + print("{} successfully installed from cache.".format(friendly_name)) + + +def upload(directory): + _create_tarball(directory) + print("Uploading {} tarball to S3... ({})".format(friendly_name, _tarball_size(directory))) + key.set_contents_from_filename(_tarball_filename_for(directory)) + print("{} cache successfully updated.".format(friendly_name)) + _delete_file_quietly(NEED_TO_UPLOAD_MARKER) + + +if __name__ == '__main__': + # Uses environment variables: + # AWS_ACCESS_KEY_ID -- AWS Access Key ID + # AWS_SECRET_ACCESS_KEY -- AWS Secret Access Key + argv.pop(0) + if len(argv) != 4: + raise SystemExit("USAGE: s3_cache.py ") + mode, friendly_name, dependencies_file, directory = argv + + conn = S3Connection() + bucket = conn.lookup(BUCKET_NAME, validate=False) + if bucket is None: + raise SystemExit("Could not access bucket!") + + dependencies_file_hash = _sha256_of_file(dependencies_file) + + key = Key(bucket, dependencies_file_hash) + key.storage_class = 'REDUCED_REDUNDANCY' + + if mode == 'download': + download(directory) + elif mode == 'upload': + if isfile(NEED_TO_UPLOAD_MARKER): # FIXME + upload(directory) + else: + print("No need to upload anything.") + else: + raise SystemExit("Unrecognized mode {!r}".format(mode)) + +# Copyright 2012 New Dream Network, LLC (DreamHost) +# +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +import hashlib +import hmac + +import httplib2 +from neutronclient.v2_0 import client +from oslo_config import cfg +from oslo_log import log as logging +import oslo_messaging +from oslo_service import loopingcall +import six +import six.moves.urllib.parse as urlparse +import webob + +from neutron.agent.linux import utils as agent_utils +from neutron.agent.metadata import config +from neutron.agent import rpc as agent_rpc +from neutron.common import constants as n_const +from neutron.common import rpc as n_rpc +from neutron.common import topics +from neutron.common import utils +from neutron import context +from neutron.i18n import _LE, _LW +from neutron.openstack.common.cache import cache + +LOG = logging.getLogger(__name__) + +MODE_MAP = { + config.USER_MODE: 0o644, + config.GROUP_MODE: 0o664, + config.ALL_MODE: 0o666, +} + + +class MetadataPluginAPI(object): + """Agent-side RPC for metadata agent-to-plugin interaction. + + This class implements the client side of an rpc interface used by the + metadata service to make calls back into the Neutron plugin. The server + side is defined in + neutron.api.rpc.handlers.metadata_rpc.MetadataRpcCallback. For more + information about changing rpc interfaces, see + doc/source/devref/rpc_api.rst. + + API version history: + 1.0 - Initial version. + """ + + def __init__(self, topic): + target = oslo_messaging.Target( + topic=topic, + namespace=n_const.RPC_NAMESPACE_METADATA, + version='1.0') + self.client = n_rpc.get_client(target) + + def get_ports(self, context, filters): + cctxt = self.client.prepare() + return cctxt.call(context, 'get_ports', filters=filters) + + +class MetadataProxyHandler(object): + + def __init__(self, conf): + self.conf = conf + self.auth_info = {} + if self.conf.cache_url: + self._cache = cache.get_cache(self.conf.cache_url) + else: + self._cache = False + + self.plugin_rpc = MetadataPluginAPI(topics.PLUGIN) + self.context = context.get_admin_context_without_session() + # Use RPC by default + self.use_rpc = True + + def _get_neutron_client(self): + params = { + 'username': self.conf.admin_user, + 'password': self.conf.admin_password, + 'tenant_name': self.conf.admin_tenant_name, + 'auth_url': self.conf.auth_url, + 'auth_strategy': self.conf.auth_strategy, + 'region_name': self.conf.auth_region, + 'token': self.auth_info.get('auth_token'), + 'insecure': self.conf.auth_insecure, + 'ca_cert': self.conf.auth_ca_cert, + } + if self.conf.endpoint_url: + params['endpoint_url'] = self.conf.endpoint_url + else: + params['endpoint_url'] = self.auth_info.get('endpoint_url') + params['endpoint_type'] = self.conf.endpoint_type + + return client.Client(**params) + + @webob.dec.wsgify(RequestClass=webob.Request) + def __call__(self, req): + try: + LOG.debug("Request: %s", req) + + instance_id, tenant_id = self._get_instance_and_tenant_id(req) + if instance_id: + return self._proxy_request(instance_id, tenant_id, req) + else: + return webob.exc.HTTPNotFound() + + except Exception: + LOG.exception(_LE("Unexpected error.")) + msg = _('An unknown error has occurred. ' + 'Please try your request again.') + explanation = six.text_type(msg) + return webob.exc.HTTPInternalServerError(explanation=explanation) + + def _get_ports_from_server(self, router_id=None, ip_address=None, + networks=None): + """Either get ports from server by RPC or fallback to neutron client""" + filters = self._get_port_filters(router_id, ip_address, networks) + if self.use_rpc: + try: + return self.plugin_rpc.get_ports(self.context, filters) + except (oslo_messaging.MessagingException, AttributeError): + # TODO(obondarev): remove fallback once RPC is proven + # to work fine with metadata agent (K or L release at most) + LOG.warning(_LW('Server does not support metadata RPC, ' + 'fallback to using neutron client')) + self.use_rpc = False + + return self._get_ports_using_client(filters) + + def _get_port_filters(self, router_id=None, ip_address=None, + networks=None): + filters = {} + if router_id: + filters['device_id'] = [router_id] + filters['device_owner'] = n_const.ROUTER_INTERFACE_OWNERS + if ip_address: + filters['fixed_ips'] = {'ip_address': [ip_address]} + if networks: + filters['network_id'] = networks + + return filters + + @utils.cache_method_results + def _get_router_networks(self, router_id): + """Find all networks connected to given router.""" + internal_ports = self._get_ports_from_server(router_id=router_id) + return tuple(p['network_id'] for p in internal_ports) + + @utils.cache_method_results + def _get_ports_for_remote_address(self, remote_address, networks): + """Get list of ports that has given ip address and are part of + given networks. + + :param networks: list of networks in which the ip address will be + searched for + + """ + return self._get_ports_from_server(networks=networks, + ip_address=remote_address) + + def _get_ports_using_client(self, filters): + # reformat filters for neutron client + if 'device_id' in filters: + filters['device_id'] = filters['device_id'][0] + if 'fixed_ips' in filters: + filters['fixed_ips'] = [ + 'ip_address=%s' % filters['fixed_ips']['ip_address'][0]] + + client = self._get_neutron_client() + ports = client.list_ports(**filters) + self.auth_info = client.get_auth_info() + return ports['ports'] + + def _get_ports(self, remote_address, network_id=None, router_id=None): + """Search for all ports that contain passed ip address and belongs to + given network. + + If no network is passed ports are searched on all networks connected to + given router. Either one of network_id or router_id must be passed. + + """ + if network_id: + networks = (network_id,) + elif router_id: + networks = self._get_router_networks(router_id) + else: + raise TypeError(_("Either one of parameter network_id or router_id" + " must be passed to _get_ports method.")) + + return self._get_ports_for_remote_address(remote_address, networks) + + def _get_instance_and_tenant_id(self, req): + remote_address = req.headers.get('X-Forwarded-For') + network_id = req.headers.get('X-Neutron-Network-ID') + router_id = req.headers.get('X-Neutron-Router-ID') + + ports = self._get_ports(remote_address, network_id, router_id) + + if len(ports) == 1: + return ports[0]['device_id'], ports[0]['tenant_id'] + return None, None + + def _proxy_request(self, instance_id, tenant_id, req): + headers = { + 'X-Forwarded-For': req.headers.get('X-Forwarded-For'), + 'X-Instance-ID': instance_id, + 'X-Tenant-ID': tenant_id, + 'X-Instance-ID-Signature': self._sign_instance_id(instance_id) + } + + nova_ip_port = '%s:%s' % (self.conf.nova_metadata_ip, + self.conf.nova_metadata_port) + url = urlparse.urlunsplit(( + self.conf.nova_metadata_protocol, + nova_ip_port, + req.path_info, + req.query_string, + '')) + + h = httplib2.Http( + ca_certs=self.conf.auth_ca_cert, + disable_ssl_certificate_validation=self.conf.nova_metadata_insecure + ) + if self.conf.nova_client_cert and self.conf.nova_client_priv_key: + h.add_certificate(self.conf.nova_client_priv_key, + self.conf.nova_client_cert, + nova_ip_port) + resp, content = h.request(url, method=req.method, headers=headers, + body=req.body) + + if resp.status == 200: + LOG.debug(str(resp)) + req.response.content_type = resp['content-type'] + req.response.body = content + return req.response + elif resp.status == 403: + LOG.warn(_LW( + 'The remote metadata server responded with Forbidden. This ' + 'response usually occurs when shared secrets do not match.' + )) + return webob.exc.HTTPForbidden() + elif resp.status == 400: + return webob.exc.HTTPBadRequest() + elif resp.status == 404: + return webob.exc.HTTPNotFound() + elif resp.status == 409: + return webob.exc.HTTPConflict() + elif resp.status == 500: + msg = _( + 'Remote metadata server experienced an internal server error.' + ) + LOG.warn(msg) + explanation = six.text_type(msg) + return webob.exc.HTTPInternalServerError(explanation=explanation) + else: + raise Exception(_('Unexpected response code: %s') % resp.status) + + def _sign_instance_id(self, instance_id): + secret = self.conf.metadata_proxy_shared_secret + if isinstance(secret, six.text_type): + secret = secret.encode('utf-8') + if isinstance(instance_id, six.text_type): + instance_id = instance_id.encode('utf-8') + return hmac.new(secret, instance_id, hashlib.sha256).hexdigest() + + +class UnixDomainMetadataProxy(object): + + def __init__(self, conf): + self.conf = conf + agent_utils.ensure_directory_exists_without_file( + cfg.CONF.metadata_proxy_socket) + self._init_state_reporting() + + def _init_state_reporting(self): + self.context = context.get_admin_context_without_session() + self.state_rpc = agent_rpc.PluginReportStateAPI(topics.PLUGIN) + self.agent_state = { + 'binary': 'neutron-metadata-agent', + 'host': cfg.CONF.host, + 'topic': 'N/A', + 'configurations': { + 'metadata_proxy_socket': cfg.CONF.metadata_proxy_socket, + 'nova_metadata_ip': cfg.CONF.nova_metadata_ip, + 'nova_metadata_port': cfg.CONF.nova_metadata_port, + 'log_agent_heartbeats': cfg.CONF.AGENT.log_agent_heartbeats, + }, + 'start_flag': True, + 'agent_type': n_const.AGENT_TYPE_METADATA} + report_interval = cfg.CONF.AGENT.report_interval + if report_interval: + self.heartbeat = loopingcall.FixedIntervalLoopingCall( + self._report_state) + self.heartbeat.start(interval=report_interval) + + def _report_state(self): + try: + self.state_rpc.report_state( + self.context, + self.agent_state, + use_call=self.agent_state.get('start_flag')) + except AttributeError: + # This means the server does not support report_state + LOG.warn(_LW('Neutron server does not support state report.' + ' State report for this agent will be disabled.')) + self.heartbeat.stop() + return + except Exception: + LOG.exception(_LE("Failed reporting state!")) + return + self.agent_state.pop('start_flag', None) + + def _get_socket_mode(self): + mode = self.conf.metadata_proxy_socket_mode + if mode == config.DEDUCE_MODE: + user = self.conf.metadata_proxy_user + if (not user or user == '0' or user == 'root' + or agent_utils.is_effective_user(user)): + # user is agent effective user or root => USER_MODE + mode = config.USER_MODE + else: + group = self.conf.metadata_proxy_group + if not group or agent_utils.is_effective_group(group): + # group is agent effective group => GROUP_MODE + mode = config.GROUP_MODE + else: + # otherwise => ALL_MODE + mode = config.ALL_MODE + return MODE_MAP[mode] + + def run(self): + server = agent_utils.UnixDomainWSGIServer('neutron-metadata-agent') + server.start(MetadataProxyHandler(self.conf), + self.conf.metadata_proxy_socket, + workers=self.conf.metadata_workers, + backlog=self.conf.metadata_backlog, + mode=self._get_socket_mode()) + server.wait() + +# -*- coding: utf-8 -*- +"""QGIS Unit tests for QgsFieldModel + +.. note:: This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; either version 2 of the License, or +(at your option) any later version. +""" +__author__ = 'Nyall Dawson' +__date__ = '14/11/2016' +__copyright__ = 'Copyright 2016, The QGIS Project' + +import qgis # NOQA + +from qgis.core import (QgsField, + QgsFields, + QgsVectorLayer, + QgsFieldModel, + QgsFieldProxyModel, + QgsEditorWidgetSetup, + QgsProject, + QgsVectorLayerJoinInfo, + QgsFieldConstraints) +from qgis.PyQt.QtCore import QVariant, Qt, QModelIndex + +from qgis.testing import start_app, unittest + +start_app() + + +def create_layer(): + layer = QgsVectorLayer("Point?field=fldtxt:string&field=fldint:integer", + "addfeat", "memory") + layer.setEditorWidgetSetup(0, QgsEditorWidgetSetup('Hidden', {})) + layer.setEditorWidgetSetup(1, QgsEditorWidgetSetup('ValueMap', {})) + assert layer.isValid() + return layer + + +def create_model(): + l = create_layer() + m = QgsFieldModel() + m.setLayer(l) + return l, m + + +class TestQgsFieldModel(unittest.TestCase): + + def testGettersSetters(self): + """ test model getters/setters """ + l = create_layer() + m = QgsFieldModel() + + self.assertFalse(m.layer()) + m.setLayer(l) + self.assertEqual(m.layer(), l) + + m.setAllowExpression(True) + self.assertTrue(m.allowExpression()) + m.setAllowExpression(False) + self.assertFalse(m.allowExpression()) + + m.setAllowEmptyFieldName(True) + self.assertTrue(m.allowEmptyFieldName()) + m.setAllowEmptyFieldName(False) + self.assertFalse(m.allowEmptyFieldName()) + + fields = QgsFields() + fields.append(QgsField('test1', QVariant.String)) + fields.append(QgsField('test2', QVariant.String)) + m.setFields(fields) + self.assertIsNone(m.layer()) + self.assertEqual(m.fields(), fields) + + def testIndexFromName(self): + l, m = create_model() + i = m.indexFromName('fldtxt') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 0) + i = m.indexFromName('fldint') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 1) + i = m.indexFromName('not a field') + self.assertFalse(i.isValid()) + + # test with alias + i = m.indexFromName('text field') + self.assertFalse(i.isValid()) + l.setFieldAlias(0, 'text field') + i = m.indexFromName('text field') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 0) + i = m.indexFromName('int field') + self.assertFalse(i.isValid()) + l.setFieldAlias(1, 'int field') + i = m.indexFromName('int field') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 1) + + # should be case insensitive + i = m.indexFromName('FLDTXT') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 0) + i = m.indexFromName('FLDINT') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 1) + + # try with expression + m.setAllowExpression(True) + i = m.indexFromName('not a field') + # still not valid - needs expression set first + self.assertFalse(i.isValid()) + m.setExpression('not a field') + i = m.indexFromName('not a field') + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 2) + + # try with null + i = m.indexFromName(None) + self.assertFalse(i.isValid()) + m.setAllowEmptyFieldName(True) + i = m.indexFromName(None) + self.assertTrue(i.isValid()) + self.assertEqual(i.row(), 0) + # when null is shown, all other rows should be offset + self.assertEqual(m.indexFromName('fldtxt').row(), 1) + self.assertEqual(m.indexFromName('fldint').row(), 2) + self.assertEqual(m.indexFromName('not a field').row(), 3) + self.assertEqual(m.indexFromName('FLDTXT').row(), 1) + self.assertEqual(m.indexFromName('FLDINT').row(), 2) + + def testIsField(self): + l, m = create_model() + self.assertTrue(m.isField('fldtxt')) + self.assertTrue(m.isField('fldint')) + self.assertFalse(m.isField(None)) + self.assertFalse(m.isField('an expression')) + + def testRowCount(self): + l, m = create_model() + self.assertEqual(m.rowCount(), 2) + m.setAllowEmptyFieldName(True) + self.assertEqual(m.rowCount(), 3) + m.setAllowExpression(True) + m.setExpression('not a field') + self.assertEqual(m.rowCount(), 4) + m.setExpression('not a field') + self.assertEqual(m.rowCount(), 4) + m.setExpression('not a field 2') + self.assertEqual(m.rowCount(), 4) + m.removeExpression() + self.assertEqual(m.rowCount(), 3) + + def testFieldNameRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldNameRole), 'fldtxt') + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.FieldNameRole), 'fldint') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldNameRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldNameRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldNameRole)) + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldNameRole)) + + def testExpressionRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.ExpressionRole), 'fldtxt') + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.ExpressionRole), 'fldint') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.ExpressionRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.ExpressionRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertEqual(m.data(m.indexFromName('an expression'), QgsFieldModel.ExpressionRole), 'an expression') + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.ExpressionRole)) + + def testFieldIndexRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldIndexRole), 0) + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.FieldIndexRole), 1) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldIndexRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldIndexRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldIndexRole)) + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldIndexRole)) + + def testIsExpressionRole(self): + l, m = create_model() + self.assertFalse(m.data(m.indexFromName('fldtxt'), QgsFieldModel.IsExpressionRole)) + self.assertFalse(m.data(m.indexFromName('fldint'), QgsFieldModel.IsExpressionRole)) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.IsExpressionRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.IsExpressionRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertTrue(m.data(m.indexFromName('an expression'), QgsFieldModel.IsExpressionRole)) + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.IsExpressionRole)) + + def testExpressionValidityRole(self): + l, m = create_model() + self.assertTrue(m.data(m.indexFromName('fldtxt'), QgsFieldModel.ExpressionValidityRole)) + self.assertTrue(m.data(m.indexFromName('fldint'), QgsFieldModel.ExpressionValidityRole)) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.ExpressionValidityRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.ExpressionValidityRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.ExpressionValidityRole)) + m.setAllowEmptyFieldName(True) + self.assertTrue(m.data(m.indexFromName(None), QgsFieldModel.ExpressionValidityRole)) + + def testFieldTypeRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldTypeRole), QVariant.String) + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.FieldTypeRole), QVariant.Int) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldTypeRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldTypeRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldTypeRole)) + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldTypeRole)) + + def testFieldOriginRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldOriginRole), QgsFields.OriginProvider) + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.FieldOriginRole), QgsFields.OriginProvider) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldOriginRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldOriginRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldOriginRole)) + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldOriginRole)) + + def testIsEmptyRole(self): + l, m = create_model() + self.assertFalse(m.data(m.indexFromName('fldtxt'), QgsFieldModel.IsEmptyRole), QgsFields.OriginProvider) + self.assertFalse(m.data(m.indexFromName('fldint'), QgsFieldModel.IsEmptyRole), QgsFields.OriginProvider) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.IsEmptyRole)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.IsEmptyRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.IsEmptyRole)) + m.setAllowEmptyFieldName(True) + self.assertTrue(m.data(m.indexFromName(None), QgsFieldModel.IsEmptyRole)) + + def testDisplayRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), Qt.DisplayRole), 'fldtxt') + self.assertEqual(m.data(m.indexFromName('fldint'), Qt.DisplayRole), 'fldint') + self.assertFalse(m.data(m.indexFromName('an expression'), Qt.DisplayRole)) + self.assertFalse(m.data(m.indexFromName(None), Qt.DisplayRole)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertEqual(m.data(m.indexFromName('an expression'), Qt.DisplayRole), 'an expression') + m.setAllowEmptyFieldName(True) + self.assertFalse(m.data(m.indexFromName(None), Qt.DisplayRole)) + + def testManualFields(self): + _, m = create_model() + fields = QgsFields() + fields.append(QgsField('f1', QVariant.String)) + fields.append(QgsField('f2', QVariant.String)) + m.setFields(fields) + self.assertEqual(m.rowCount(), 2) + self.assertEqual(m.data(m.index(0, 0, QModelIndex()), Qt.DisplayRole), 'f1') + self.assertEqual(m.data(m.index(1, 0, QModelIndex()), Qt.DisplayRole), 'f2') + + def testEditorWidgetTypeRole(self): + l, m = create_model() + self.assertEqual(m.data(m.indexFromName('fldtxt'), QgsFieldModel.EditorWidgetType), 'Hidden') + self.assertEqual(m.data(m.indexFromName('fldint'), QgsFieldModel.EditorWidgetType), 'ValueMap') + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.EditorWidgetType)) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.EditorWidgetType)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.EditorWidgetType)) + m.setAllowEmptyFieldName(True) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.EditorWidgetType)) + + def testJoinedFieldIsEditableRole(self): + layer = QgsVectorLayer("Point?field=id_a:integer", + "addfeat", "memory") + layer2 = QgsVectorLayer("Point?field=id_b:integer&field=value_b", + "addfeat", "memory") + QgsProject.instance().addMapLayers([layer, layer2]) + + # editable join + join_info = QgsVectorLayerJoinInfo() + join_info.setTargetFieldName("id_a") + join_info.setJoinLayer(layer2) + join_info.setJoinFieldName("id_b") + join_info.setPrefix("B_") + join_info.setEditable(True) + join_info.setUpsertOnEdit(True) + layer.addJoin(join_info) + + m = QgsFieldModel() + m.setLayer(layer) + + self.assertIsNone(m.data(m.indexFromName('id_a'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertTrue(m.data(m.indexFromName('B_value_b'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.JoinedFieldIsEditable)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.JoinedFieldIsEditable)) + m.setAllowEmptyFieldName(True) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.JoinedFieldIsEditable)) + + proxy_m = QgsFieldProxyModel() + proxy_m.setFilters(QgsFieldProxyModel.AllTypes | QgsFieldProxyModel.HideReadOnly) + proxy_m.sourceFieldModel().setLayer(layer) + self.assertEqual(proxy_m.rowCount(), 2) + self.assertEqual(proxy_m.data(proxy_m.index(0, 0)), 'id_a') + self.assertEqual(proxy_m.data(proxy_m.index(1, 0)), 'B_value_b') + + # not editable join + layer3 = QgsVectorLayer("Point?field=id_a:integer", + "addfeat", "memory") + QgsProject.instance().addMapLayers([layer3]) + join_info = QgsVectorLayerJoinInfo() + join_info.setTargetFieldName("id_a") + join_info.setJoinLayer(layer2) + join_info.setJoinFieldName("id_b") + join_info.setPrefix("B_") + join_info.setEditable(False) + + layer3.addJoin(join_info) + m = QgsFieldModel() + m.setLayer(layer3) + + self.assertIsNone(m.data(m.indexFromName('id_a'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertFalse(m.data(m.indexFromName('B_value_b'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.JoinedFieldIsEditable)) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.JoinedFieldIsEditable)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertIsNone(m.data(m.indexFromName('an expression'), QgsFieldModel.JoinedFieldIsEditable)) + m.setAllowEmptyFieldName(True) + self.assertIsNone(m.data(m.indexFromName(None), QgsFieldModel.JoinedFieldIsEditable)) + + proxy_m = QgsFieldProxyModel() + proxy_m.sourceFieldModel().setLayer(layer3) + proxy_m.setFilters(QgsFieldProxyModel.AllTypes | QgsFieldProxyModel.HideReadOnly) + self.assertEqual(proxy_m.rowCount(), 1) + self.assertEqual(proxy_m.data(proxy_m.index(0, 0)), 'id_a') + + def testFieldIsWidgetEditableRole(self): + l, m = create_model() + self.assertTrue(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldIsWidgetEditable)) + self.assertTrue(m.data(m.indexFromName('fldint'), QgsFieldModel.FieldIsWidgetEditable)) + self.assertFalse(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldIsWidgetEditable)) + self.assertFalse(m.data(m.indexFromName(None), QgsFieldModel.FieldIsWidgetEditable)) + m.setAllowExpression(True) + m.setExpression('an expression') + self.assertTrue(m.data(m.indexFromName('an expression'), QgsFieldModel.FieldIsWidgetEditable)) + m.setAllowEmptyFieldName(True) + self.assertTrue(m.data(m.indexFromName(None), QgsFieldModel.FieldIsWidgetEditable)) + + editFormConfig = l.editFormConfig() + idx = l.fields().indexOf('fldtxt') + # Make fldtxt readOnly + editFormConfig.setReadOnly(idx, True) + l.setEditFormConfig(editFormConfig) + # It's read only, so the widget is NOT editable + self.assertFalse(m.data(m.indexFromName('fldtxt'), QgsFieldModel.FieldIsWidgetEditable)) + + def testFieldTooltip(self): + f = QgsField('my_string', QVariant.String, 'string') + self.assertEqual(QgsFieldModel.fieldToolTip(f), "my_string
string NULL") + f.setAlias('my alias') + self.assertEqual(QgsFieldModel.fieldToolTip(f), "my alias (my_string)
string NULL") + f.setLength(20) + self.assertEqual(QgsFieldModel.fieldToolTip(f), "my alias (my_string)
string(20) NULL") + f = QgsField('my_real', QVariant.Double, 'real', 8, 3) + self.assertEqual(QgsFieldModel.fieldToolTip(f), "my_real
real(8, 3) NULL") + f.setComment('Comment text') + self.assertEqual(QgsFieldModel.fieldToolTip(f), "my_real
real(8, 3) NULL
Comment text") + + def testFieldTooltipExtended(self): + layer = QgsVectorLayer("Point?", "tooltip", "memory") + f = QgsField('my_real', QVariant.Double, 'real', 8, 3, 'Comment text') + layer.addExpressionField('1+1', f) + layer.updateFields() + self.assertEqual(QgsFieldModel.fieldToolTipExtended(QgsField('my_string', QVariant.String, 'string'), layer), '') + self.assertEqual(QgsFieldModel.fieldToolTipExtended(f, layer), "my_real
real(8, 3) NULL
Comment text
1+1") + f.setAlias('my alias') + constraints = f.constraints() + constraints.setConstraint(QgsFieldConstraints.ConstraintUnique) + f.setConstraints(constraints) + self.assertEqual(QgsFieldModel.fieldToolTipExtended(f, layer), "my alias (my_real)
real(8, 3) NULL UNIQUE
Comment text
1+1") + + +if __name__ == '__main__': + unittest.main() + +# -*- test-case-name: twisted.internet.test.test_sigchld -*- +# Copyright (c) Twisted Matrix Laboratories. +# See LICENSE for details. + +""" +This module is used to integrate child process termination into a +reactor event loop. This is a challenging feature to provide because +most platforms indicate process termination via SIGCHLD and do not +provide a way to wait for that signal and arbitrary I/O events at the +same time. The naive implementation involves installing a Python +SIGCHLD handler; unfortunately this leads to other syscalls being +interrupted (whenever SIGCHLD is received) and failing with EINTR +(which almost no one is prepared to handle). This interruption can be +disabled via siginterrupt(2) (or one of the equivalent mechanisms); +however, if the SIGCHLD is delivered by the platform to a non-main +thread (not a common occurrence, but difficult to prove impossible), +the main thread (waiting on select() or another event notification +API) may not wake up leading to an arbitrary delay before the child +termination is noticed. + +The basic solution to all these issues involves enabling SA_RESTART +(ie, disabling system call interruption) and registering a C signal +handler which writes a byte to a pipe. The other end of the pipe is +registered with the event loop, allowing it to wake up shortly after +SIGCHLD is received. See L{twisted.internet.posixbase._SIGCHLDWaker} +for the implementation of the event loop side of this solution. The +use of a pipe this way is known as the U{self-pipe +trick}. + +From Python version 2.6, C{signal.siginterrupt} and C{signal.set_wakeup_fd} +provide the necessary C signal handler which writes to the pipe to be +registered with C{SA_RESTART}. +""" + +from __future__ import division, absolute_import + +import signal + + +def installHandler(fd): + """ + Install a signal handler which will write a byte to C{fd} when + I{SIGCHLD} is received. + + This is implemented by installing a SIGCHLD handler that does nothing, + setting the I{SIGCHLD} handler as not allowed to interrupt system calls, + and using L{signal.set_wakeup_fd} to do the actual writing. + + @param fd: The file descriptor to which to write when I{SIGCHLD} is + received. + @type fd: C{int} + """ + if fd == -1: + signal.signal(signal.SIGCHLD, signal.SIG_DFL) + else: + def noopSignalHandler(*args): + pass + signal.signal(signal.SIGCHLD, noopSignalHandler) + signal.siginterrupt(signal.SIGCHLD, False) + return signal.set_wakeup_fd(fd) + + + +def isDefaultHandler(): + """ + Determine whether the I{SIGCHLD} handler is the default or not. + """ + return signal.getsignal(signal.SIGCHLD) == signal.SIG_DFL + +import os +import platform + +from twisted.internet import defer + +from .. import data, helper +from p2pool.util import pack + + +P2P_PREFIX = 'fcc1b7dc'.decode('hex') +P2P_PORT = 19333 +ADDRESS_VERSION = 111 +RPC_PORT = 19332 +RPC_CHECK = defer.inlineCallbacks(lambda bitcoind: defer.returnValue( + 'litecoinaddress' in (yield bitcoind.rpc_help()) and + (yield bitcoind.rpc_getinfo())['testnet'] + )) +SUBSIDY_FUNC = lambda height: 50*100000000 >> (height + 1)//840000 +POW_FUNC = lambda data: pack.IntType(256).unpack(__import__('ltc_scrypt').getPoWHash(data)) +BLOCK_PERIOD = 150 # s +SYMBOL = 'tLTC' +CONF_FILE_FUNC = lambda: os.path.join(os.path.join(os.environ['APPDATA'], 'Litecoin') if platform.system() == 'Windows' else os.path.expanduser('~/Library/Application Support/Litecoin/') if platform.system() == 'Darwin' else os.path.expanduser('~/.litecoin'), 'litecoin.conf') +BLOCK_EXPLORER_URL_PREFIX = 'http://nonexistent-litecoin-testnet-explorer/block/' +ADDRESS_EXPLORER_URL_PREFIX = 'http://nonexistent-litecoin-testnet-explorer/address/' +TX_EXPLORER_URL_PREFIX = 'http://nonexistent-litecoin-testnet-explorer/tx/' +SANE_TARGET_RANGE = (2**256//1000000000 - 1, 2**256 - 1) +DUMB_SCRYPT_DIFF = 2**16 +DUST_THRESHOLD = 1e8 + +# Copyright 2014 The Chromium Authors. All rights reserved. +# Use of this source code is governed by a BSD-style license that can be +# found in the LICENSE file. + +# This script computs the number of concurrent links we want to run in the build +# as a function of machine spec. It's based on GetDefaultConcurrentLinks in GYP. + +import optparse +import os +import re +import subprocess +import sys + +def _GetTotalMemoryInBytes(): + if sys.platform in ('win32', 'cygwin'): + import ctypes + + class MEMORYSTATUSEX(ctypes.Structure): + _fields_ = [ + ("dwLength", ctypes.c_ulong), + ("dwMemoryLoad", ctypes.c_ulong), + ("ullTotalPhys", ctypes.c_ulonglong), + ("ullAvailPhys", ctypes.c_ulonglong), + ("ullTotalPageFile", ctypes.c_ulonglong), + ("ullAvailPageFile", ctypes.c_ulonglong), + ("ullTotalVirtual", ctypes.c_ulonglong), + ("ullAvailVirtual", ctypes.c_ulonglong), + ("sullAvailExtendedVirtual", ctypes.c_ulonglong), + ] + + stat = MEMORYSTATUSEX(dwLength=ctypes.sizeof(MEMORYSTATUSEX)) + ctypes.windll.kernel32.GlobalMemoryStatusEx(ctypes.byref(stat)) + return stat.ullTotalPhys + elif sys.platform.startswith('linux'): + if os.path.exists("/proc/meminfo"): + with open("/proc/meminfo") as meminfo: + memtotal_re = re.compile(r'^MemTotal:\s*(\d*)\s*kB') + for line in meminfo: + match = memtotal_re.match(line) + if not match: + continue + return float(match.group(1)) * 2**10 + elif sys.platform == 'darwin': + try: + return int(subprocess.check_output(['sysctl', '-n', 'hw.memsize'])) + except Exception: + return 0 + # TODO(scottmg): Implement this for other platforms. + return 0 + + +def _GetDefaultConcurrentLinks(mem_per_link_gb, reserve_mem_gb): + # Inherit the legacy environment variable for people that have set it in GYP. + pool_size = int(os.getenv('GYP_LINK_CONCURRENCY', 0)) + if pool_size: + return pool_size + + mem_total_bytes = _GetTotalMemoryInBytes() + mem_total_bytes = max(0, mem_total_bytes - reserve_mem_gb * 2**30) + num_concurrent_links = int(max(1, mem_total_bytes / mem_per_link_gb / 2**30)) + hard_cap = max(1, int(os.getenv('GYP_LINK_CONCURRENCY_MAX', 2**32))) + return min(num_concurrent_links, hard_cap) + + +def main(): + parser = optparse.OptionParser() + parser.add_option('--mem_per_link_gb', action="store", type="int", default=8) + parser.add_option('--reserve_mem_gb', action="store", type="int", default=0) + parser.disable_interspersed_args() + options, _ = parser.parse_args() + + print _GetDefaultConcurrentLinks(options.mem_per_link_gb, + options.reserve_mem_gb) + return 0 + +if __name__ == '__main__': + sys.exit(main()) + +from __future__ import print_function, division, absolute_import + +try: + from .core import (DataFrame, Series, Index, _Frame, map_partitions, + repartition, to_datetime, to_timedelta) + from .groupby import Aggregation + from .io import (from_array, from_pandas, from_bcolz, + from_dask_array, read_hdf, read_sql_table, + from_delayed, read_csv, to_csv, read_table, + demo, to_hdf, to_records, to_bag, read_json, to_json) + from .optimize import optimize + from .multi import merge, concat + from . import rolling + from ..base import compute + from .reshape import get_dummies, pivot_table, melt + from .utils import assert_eq + from .io.orc import read_orc + try: + from .io import read_parquet, to_parquet + except ImportError: + pass + try: + from .core import isna + except ImportError: + pass +except ImportError as e: + msg = ("Dask dataframe requirements are not installed.\n\n" + "Please either conda or pip install as follows:\n\n" + " conda install dask # either conda install\n" + " pip install dask[dataframe] --upgrade # or pip install") + raise ImportError(str(e) + '\n\n' + msg) + +# Copyright (c) Victor van den Elzen +# Released under the Expat license, see LICENSE file for details + +from struct import pack, unpack, calcsize +from collections import OrderedDict + +def getbytes(s, n): + b = s.read(n) + assert len(b) == n, "Unexpected EOF" + return b + +def getbyte(s): + return getbytes(s, 1) + +class Seek(object): + def __init__(self, s, *args, **kwargs): + self.old_pos = None + self.s = s + self.args = args + self.kwargs = kwargs + + def __enter__(self): + self.old_pos = self.s.tell() + self.s.seek(*self.args, **self.kwargs) + + def __exit__(self, exc_type, exc_value, traceback): + self.s.seek(self.old_pos) + +class FakeWriteStream(object): + def __init__(self, offset=0): + self.offset = offset + + def seek(self, offset): + self.offset = offset + + def tell(self): + return self.offset + + def write(self, data): + self.offset += len(data) + return len(data) + +class BaseField(object): + def unpack(self, s): + self.data = self.unpack_data(s) + + def unpack_data(self, s): + raise notImplementedError + + def pack(self, s): + self.pack_data(s, self.data) + + def pack_data(self, s, data): + raise NotImplementedError(self) + + def full_pack(self, s): + new_data = self.data + while True: + old_data = new_data + self.pack(FakeWriteStream(s.tell())) + new_data = self.data + if old_data == new_data: + break + self.pack(s) + + def serialize(self): + return self.data + +class ContainerField(BaseField): + def __getitem__(self, key): + return self.field[key] + + def __setitem__(self, key, value): + self.field[key] = value + + def __delitem__(self, key): + del self.field[key] + + def __len__(self): + return len(self.field) + + def __iter__(self): + return iter(self.field) + + def __contains__(self, key): + return key in self.field + + def serialize(self): + return self.field.serialize() + +class Struct(ContainerField): + def __init__(self, *args, **kwargs): + self.args = args + self.kwargs = kwargs + + def add_field(self, name, f): + assert name not in self, name + self[name] = f + input_type, v = self.input + if input_type == "data": + f.data = v.get(name, None) + elif input_type == "stream": + f.unpack(v) + else: + assert False, input_type + return f + + def F(self, name, f): + return self.add_field(name, f) + + def unpack(self, s): + self.field = OrderedDict() + self.input = ("stream", s) + self.fields(*self.args, **self.kwargs) + del self.input + + def pack(self, s): + for name, f in self.field.items(): + f.pack(s) + + @property + def data(self): + data = OrderedDict() + for k, v in self.field.items(): + data[k] = v.data + return data + + @data.setter + def data(self, v): + self.field = OrderedDict() + self.input = ("data", v) + self.fields(*self.args, **self.kwargs) + del self.input + + def serialize(self): + data = OrderedDict() + for k, v in self.field.items(): + if self.should_serialize(k, v): + data[k] = v.serialize() + return data + + def should_serialize(self, k, v): + return True + + def fields(self): + raise NotImplementedError(self) + +class Magic(BaseField): + def __init__(self, magic): + if isinstance(magic, str): + magic = magic.encode() + self.magic = magic + + def unpack(self, s): + data = getbytes(s, len(self.magic)) + assert data == self.magic + + def pack(self, s): + s.write(self.magic) + + @property + def data(self): + return self.magic.decode() + + @data.setter + def data(self, v): + assert v == self.magic or v is None, v + +class Format(BaseField): + def __init__(self, fmt): + if fmt[0] in "@=<>!": + bosa = fmt[0] + fmt = fmt[1:] + else: + bosa = "<" + self.bosa = bosa + self.fmt = fmt + self.single = len(fmt) == 1 + + def unpack_data(self, s): + fmt = self.bosa + self.fmt + size = calcsize(fmt) + b = getbytes(s, size) + data = unpack(fmt, b) + if self.single: + assert len(data) == 1 + data = data[0] + return data + + def pack_data(self, s, data): + if self.single: + data = (data,) + s.write(pack(self.fmt, *data)) + +class BaseArray(ContainerField): + def __init__(self, field_maker=None, field_function=None): + if field_function is None: + field_function = lambda i, f: field_maker() + self.field_fun = field_function + self._dict = None + + def unpack(self, s): + self.field = [self.field_fun(i, self) for i in range(self.size)] + for f in self: + f.unpack(s) + + def pack(self, s): + for f in self: + f.pack(s) + + @property + def data(self): + return [f.data for f in self] + + def index(self, field): + if self._dict is None: + self._dict = {} + for i in range(len(self.field)): + self._dict[self.field[i]] = i + return self._dict[field] + + @data.setter + def data(self, v): + self.field = [self.field_fun(i, self) for i in range(len(v))] + for f, fv in zip(self.field, v): + f.data = fv + self._dict = None + + def serialize(self): + return [f.serialize() for f in self] + + def append_data(self, v): + idx = len(self.field) + f = self.field_fun(idx, self) + self.field.append(f) + f.data = v + if self._dict is not None: + self._dict[f] = idx + +class Array(BaseArray): + def __init__(self, size, *args, **kwargs): + self.size = size + BaseArray.__init__(self, *args, **kwargs) + +class PrefixedArray(BaseArray): + def __init__(self, prefix_field, *args, **kwargs): + self.prefix_field = prefix_field + BaseArray.__init__(self, *args, **kwargs) + + @property + def size(self): + return self.prefix_field.data + + def unpack(self, s): + self.prefix_field.unpack(s) + BaseArray.unpack(self, s) + + def pack(self, s): + self.prefix_field.data = len(self) + self.prefix_field.pack(s) + BaseArray.pack(self, s) + +class BaseBlob(BaseField): + def unpack_data(self, s): + return getbytes(s, self.size) + + def pack_data(self, s, data): + s.write(data) + +class Blob(BaseBlob): + def __init__(self, size): + self.size = size + + def serialize(self): + return None + +class PrefixedBlob(BaseBlob): + def __init__(self, prefix_field, *args, **kwargs): + self.prefix_field = prefix_field + BaseBlob.__init__(self, *args, **kwargs) + + @property + def size(self): + return self.prefix_field.data + + def unpack(self, s): + self.prefix_field.unpack(s) + BaseBlob.unpack(self, s) + + def pack(self, s): + self.prefix_field.data = len(self) + self.prefix_field.pack(s) + BaseBlob.pack(self, s) + +class String(BaseField): + def unpack_data(self, s): + lc = [] + c = getbyte(s) + while c != b"\0": + lc.append(c) + c = getbyte(s) + return b"".join(lc).decode() + + def pack_data(self, s, data): + s.write(data.encode()) + s.write(b"\0") + +class FixedString(BaseField): + def __init__(self, size): + self.size = size + + def unpack_data(self, s): + data = getbytes(s, self.size) + data = data.rstrip(b"\0").decode() + return data + + def pack_data(self, s, data): + data = data.encode().ljust(self.size, b"\0") + s.write(data) + +class Index(BaseField): + def __init__(self, array, index_field): + self.array = array + self.index_field = index_field + + def unpack_data(self, s): + self.index_field.unpack(s) + return self.array[self.index_field.data].data + + def pack_data(self, s, data): + try: + index = self.array.data.index(data) + except ValueError: + index = len(self.array) + self.array.append_data(data) + self.index_field.data = index + self.index_field.pack(s) + +class Offset(BaseField): + def unpack_data(self, s): + return s.tell() + + def pack_data(self, s, data): + self.data = s.tell() + +class Pointer(ContainerField): + def __init__(self, offset, field): + self.offset = offset + self.field = field + + def unpack(self, s): + with Seek(s, self.offset): + self.field.unpack(s) + + @property + def data(self): + return self.field.data + + @data.setter + def data(self, v): + self.field.data = v + + def pack_data(self, s, data): + pass + +class DataPointer(ContainerField): + def __init__(self, offset_field, field): + self.offset_field = offset_field + self.field = field + + def unpack(self, s): + self.offset_field.unpack(s) + with Seek(s, self.offset_field.data): + self.field.unpack(s) + + @property + def data(self): + return self.field.data + + @data.setter + def data(self, v): + self.field.data = v + +class Mapping(BaseField): + def __init__(self, field, mapping): + self.field = field + self.mapping = mapping + + def unpack_data(self, s): + data = self.field.unpack_data(s) + return self.mapping[data] + +class Flags(BaseField): + def __init__(self, field, flags): + self.field = field + self.flags = flags + + def unpack_data(self, s): + data = self.field.unpack_data(s) + flag_data = [] + for mask, name in self.flags: + if mask & data: + flag_data.append(name) + return flag_data + +# (c) 2015, Marius Gedminas +# +# This file is part of Ansible +# +# Ansible is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Ansible is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# alongwith Ansible. If not, see . + +# Make coding more python3-ish +from __future__ import (absolute_import, division, print_function) +__metaclass__ = type + +import shlex +from ansible.module_utils.six import PY3 +from ansible.module_utils._text import to_bytes, to_text + + +if PY3: + # shlex.split() wants Unicode (i.e. ``str``) input on Python 3 + shlex_split = shlex.split +else: + # shlex.split() wants bytes (i.e. ``str``) input on Python 2 + def shlex_split(s, comments=False, posix=True): + return map(to_text, shlex.split(to_bytes(s), comments, posix)) + shlex_split.__doc__ = shlex.split.__doc__ + +#!/usr/bin/python +# -*- coding: utf-8 -*- +# +# (c) 2015, René Moser +# +# This file is part of Ansible +# +# Ansible is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Ansible is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with Ansible. If not, see . + +ANSIBLE_METADATA = {'metadata_version': '1.0', + 'status': ['stableinterface'], + 'supported_by': 'community'} + + +DOCUMENTATION = ''' +--- +module: cs_sshkeypair +short_description: Manages SSH keys on Apache CloudStack based clouds. +description: + - Create, register and remove SSH keys. + - If no key was found and no public key was provided and a new SSH + private/public key pair will be created and the private key will be returned. +version_added: '2.0' +author: "René Moser (@resmo)" +options: + name: + description: + - Name of public key. + required: true + domain: + description: + - Domain the public key is related to. + required: false + default: null + account: + description: + - Account the public key is related to. + required: false + default: null + project: + description: + - Name of the project the public key to be registered in. + required: false + default: null + state: + description: + - State of the public key. + required: false + default: 'present' + choices: [ 'present', 'absent' ] + public_key: + description: + - String of the public key. + required: false + default: null +extends_documentation_fragment: cloudstack +''' + +EXAMPLES = ''' +# create a new private / public key pair: +- cs_sshkeypair: + name: linus@example.com + delegate_to: localhost + register: key +- debug: + msg: 'Private key is {{ key.private_key }}' + +# remove a public key by its name: +- cs_sshkeypair: + name: linus@example.com + state: absent + delegate_to: localhost + +# register your existing local public key: +- cs_sshkeypair: + name: linus@example.com + public_key: "{{ lookup('file', '~/.ssh/id_rsa.pub') }}" + delegate_to: localhost +''' + +RETURN = ''' +--- +id: + description: UUID of the SSH public key. + returned: success + type: string + sample: a6f7a5fc-43f8-11e5-a151-feff819cdc9f +name: + description: Name of the SSH public key. + returned: success + type: string + sample: linus@example.com +fingerprint: + description: Fingerprint of the SSH public key. + returned: success + type: string + sample: "86:5e:a3:e8:bd:95:7b:07:7c:c2:5c:f7:ad:8b:09:28" +private_key: + description: Private key of generated SSH keypair. + returned: changed + type: string + sample: "-----BEGIN RSA PRIVATE KEY-----\nMII...8tO\n-----END RSA PRIVATE KEY-----\n" +''' + +try: + import sshpubkeys + HAS_LIB_SSHPUBKEYS = True +except ImportError: + HAS_LIB_SSHPUBKEYS = False + +from ansible.module_utils.basic import AnsibleModule +from ansible.module_utils._text import to_native +from ansible.module_utils.cloudstack import ( + AnsibleCloudStack, + CloudStackException, + cs_required_together, + cs_argument_spec +) + + +class AnsibleCloudStackSshKey(AnsibleCloudStack): + + def __init__(self, module): + super(AnsibleCloudStackSshKey, self).__init__(module) + self.returns = { + 'privatekey': 'private_key', + 'fingerprint': 'fingerprint', + } + self.ssh_key = None + + def register_ssh_key(self, public_key): + ssh_key = self.get_ssh_key() + args = self._get_common_args() + name = self.module.params.get('name') + + res = None + if not ssh_key: + self.result['changed'] = True + args['publickey'] = public_key + if not self.module.check_mode: + args['name'] = name + res = self.cs.registerSSHKeyPair(**args) + else: + fingerprint = self._get_ssh_fingerprint(public_key) + if ssh_key['fingerprint'] != fingerprint: + self.result['changed'] = True + if not self.module.check_mode: + # delete the ssh key with matching name but wrong fingerprint + args['name'] = name + self.cs.deleteSSHKeyPair(**args) + + elif ssh_key['name'].lower() != name.lower(): + self.result['changed'] = True + if not self.module.check_mode: + # delete the ssh key with matching fingerprint but wrong name + args['name'] = ssh_key['name'] + self.cs.deleteSSHKeyPair(**args) + # First match for key retrievment will be the fingerprint. + # We need to make another lookup if there is a key with identical name. + self.ssh_key = None + ssh_key = self.get_ssh_key() + if ssh_key['fingerprint'] != fingerprint: + args['name'] = name + self.cs.deleteSSHKeyPair(**args) + + if not self.module.check_mode and self.result['changed']: + args['publickey'] = public_key + args['name'] = name + res = self.cs.registerSSHKeyPair(**args) + + if res and 'keypair' in res: + ssh_key = res['keypair'] + + return ssh_key + + def create_ssh_key(self): + ssh_key = self.get_ssh_key() + if not ssh_key: + self.result['changed'] = True + args = self._get_common_args() + args['name'] = self.module.params.get('name') + if not self.module.check_mode: + res = self.cs.createSSHKeyPair(**args) + ssh_key = res['keypair'] + return ssh_key + + def remove_ssh_key(self, name=None): + ssh_key = self.get_ssh_key() + if ssh_key: + self.result['changed'] = True + args = self._get_common_args() + args['name'] = name or self.module.params.get('name') + if not self.module.check_mode: + self.cs.deleteSSHKeyPair(**args) + return ssh_key + + def _get_common_args(self): + return { + 'domainid': self.get_domain('id'), + 'account': self.get_account('name'), + 'projectid': self.get_project('id') + } + + def get_ssh_key(self): + if not self.ssh_key: + public_key = self.module.params.get('public_key') + if public_key: + # Query by fingerprint of the public key + args_fingerprint = self._get_common_args() + args_fingerprint['fingerprint'] = self._get_ssh_fingerprint(public_key) + ssh_keys = self.cs.listSSHKeyPairs(**args_fingerprint) + if ssh_keys and 'sshkeypair' in ssh_keys: + self.ssh_key = ssh_keys['sshkeypair'][0] + # When key has not been found by fingerprint, use the name + if not self.ssh_key: + args_name = self._get_common_args() + args_name['name'] = self.module.params.get('name') + ssh_keys = self.cs.listSSHKeyPairs(**args_name) + if ssh_keys and 'sshkeypair' in ssh_keys: + self.ssh_key = ssh_keys['sshkeypair'][0] + return self.ssh_key + + def _get_ssh_fingerprint(self, public_key): + key = sshpubkeys.SSHKey(public_key) + if hasattr(key, 'hash_md5'): + return key.hash_md5().replace(to_native('MD5:'), to_native('')) + return key.hash() + + +def main(): + argument_spec = cs_argument_spec() + argument_spec.update(dict( + name=dict(required=True), + public_key=dict(), + domain=dict(), + account=dict(), + project=dict(), + state=dict(choices=['present', 'absent'], default='present'), + )) + + module = AnsibleModule( + argument_spec=argument_spec, + required_together=cs_required_together(), + supports_check_mode=True + ) + + if not HAS_LIB_SSHPUBKEYS: + module.fail_json(msg="python library sshpubkeys required: pip install sshpubkeys") + + try: + acs_sshkey = AnsibleCloudStackSshKey(module) + state = module.params.get('state') + if state in ['absent']: + ssh_key = acs_sshkey.remove_ssh_key() + else: + public_key = module.params.get('public_key') + if public_key: + ssh_key = acs_sshkey.register_ssh_key(public_key) + else: + ssh_key = acs_sshkey.create_ssh_key() + + result = acs_sshkey.get_result(ssh_key) + + except CloudStackException as e: + module.fail_json(msg='CloudStackException: %s' % str(e)) + + module.exit_json(**result) + + +if __name__ == '__main__': + main() + +from django.views.generic import TemplateView +from django.utils.decorators import method_decorator +from django.contrib.auth.decorators import login_required +from django.shortcuts import get_object_or_404 +from django.http import JsonResponse + +from Hindlebook.models import Author + + +class FriendsView(TemplateView): + template_name = 'friends/friends.html' + + @method_decorator(login_required) + def dispatch(self, *args, **kwargs): + uuid = kwargs.get('authorUUID', self.request.user.author.uuid) + self.author = get_object_or_404(Author, uuid=uuid) + return super(FriendsView, self).dispatch(*args, **kwargs) + + def get_context_data(self, **kwargs): + context = super(FriendsView, self).get_context_data(**kwargs) + context['author'] = self.author + return context + + +class NotificationsView(TemplateView): + template_name = 'friends/notifications.html' + + @method_decorator(login_required) + def dispatch(self, *args, **kwargs): + self.author = self.request.user.author + return super(NotificationsView, self).dispatch(*args, **kwargs) + + def get_context_data(self, **kwargs): + context = super(NotificationsView, self).get_context_data(**kwargs) + context['author'] = self.author + return context + +""" +Russian-specific forms helpers +""" +from __future__ import absolute_import + +import re + +from django.contrib.localflavor.ru.ru_regions import RU_COUNTY_CHOICES, RU_REGIONS_CHOICES +from django.forms.fields import RegexField, Select +from django.utils.translation import ugettext_lazy as _ + + +phone_digits_re = re.compile(r'^(?:[78]-?)?(\d{3})[-\.]?(\d{3})[-\.]?(\d{4})$') + +class RUCountySelect(Select): + """ + A Select widget that uses a list of Russian Counties as its choices. + """ + def __init__(self, attrs=None): + super(RUCountySelect, self).__init__(attrs, choices=RU_COUNTY_CHOICES) + + +class RURegionSelect(Select): + """ + A Select widget that uses a list of Russian Regions as its choices. + """ + def __init__(self, attrs=None): + super(RURegionSelect, self).__init__(attrs, choices=RU_REGIONS_CHOICES) + + +class RUPostalCodeField(RegexField): + """ + Russian Postal code field. + Format: XXXXXX, where X is any digit, and first digit is not zero. + """ + default_error_messages = { + 'invalid': _(u'Enter a postal code in the format XXXXXX.'), + } + def __init__(self, max_length=None, min_length=None, *args, **kwargs): + super(RUPostalCodeField, self).__init__(r'^\d{6}$', + max_length, min_length, *args, **kwargs) + + +class RUPassportNumberField(RegexField): + """ + Russian internal passport number format: + XXXX XXXXXX where X - any digit. + """ + default_error_messages = { + 'invalid': _(u'Enter a passport number in the format XXXX XXXXXX.'), + } + def __init__(self, max_length=None, min_length=None, *args, **kwargs): + super(RUPassportNumberField, self).__init__(r'^\d{4} \d{6}$', + max_length, min_length, *args, **kwargs) + + +class RUAlienPassportNumberField(RegexField): + """ + Russian alien's passport number format: + XX XXXXXXX where X - any digit. + """ + default_error_messages = { + 'invalid': _(u'Enter a passport number in the format XX XXXXXXX.'), + } + def __init__(self, max_length=None, min_length=None, *args, **kwargs): + super(RUAlienPassportNumberField, self).__init__(r'^\d{2} \d{7}$', + max_length, min_length, *args, **kwargs) + +#!/usr/bin/env python + +# Copyright 2008-2013 Nokia Siemens Networks Oyj +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""Module implementing the command line entry point for the `Libdoc` tool. + +This module can be executed from the command line using the following +approaches:: + + python -m robot.libdoc + python path/to/robot/libdoc.py + +Instead of ``python`` it is possible to use also other Python interpreters. + +This module also provides :func:`libdoc` and :func:`libdoc_cli` functions +that can be used programmatically. Other code is for internal usage. + +Libdoc itself is implemented in the :mod:`~robot.libdocpkg` package. +""" + +USAGE = """robot.libdoc -- Robot Framework library documentation generator + +Version: + +Usage: python -m robot.libdoc [options] library output_file + or: python -m robot.libdoc [options] library list|show|version [names] + +Libdoc tool can generate keyword documentation in HTML and XML formats both +for test libraries and resource files. HTML format is suitable for humans and +XML specs for RIDE and other tools. Libdoc also has few special commands to +show library or resource information on the console. + +Libdoc supports all library and resource types and also earlier generated XML +specs can be used as input. If a library needs arguments, they must be given +as part of the library name and separated by two colons, for example, like +`LibraryName::arg1::arg2`. + +Options +======= + + -f --format HTML|XML Specifies whether to generate HTML or XML output. + If this options is not used, the format is got + from the extension of the output file. + -F --docformat ROBOT|HTML|TEXT|REST + Specifies the source documentation format. Possible + values are Robot Framework's documentation format, + HTML, plain text, and reStructuredText. The default + value can be specified in test library source code + and the initial default value is `ROBOT`. + New in Robot Framework 2.7.5. + -n --name newname Sets the name of the documented library or resource. + -v --version newversion Sets the version of the documented library or + resource. + -P --pythonpath path * Additional locations where to search for libraries + and resources. + -E --escape what:with * Escapes characters which are problematic in console. + 'what' is the name of the character to escape and + 'with' is the string to escape it with. + <-------------------ESCAPES------------------------> + -h -? --help Print this help. + +Creating documentation +====================== + +When creating documentation in HTML or XML format, the output file must +be specified as a second argument after the library/resource name or path. +Output format is got automatically from the extension but can also be set +with `--format` option. + +Examples: + + python -m robot.libdoc src/MyLib.py doc/MyLib.html + jython -m robot.libdoc MyJavaLibrary.java MyJavaLibrary.html + python -m robot.libdoc --name MyLib Remote::10.0.0.42:8270 MyLib.xml + +Viewing information on console +============================== + +Libdoc has three special commands to show information on the console. These +commands are used instead of the name of the output file, and they can also +take additional arguments. + +list: List names of the keywords the library/resource contains. Can be + limited to show only certain keywords by passing optional patterns as + arguments. Keyword is listed if its name contains any given pattern. +show: Show library/resource documentation. Can be limited to show only + certain keywords by passing names as arguments. Keyword is shown if + its name matches any given name. Special argument `intro` will show + the library introduction and importing sections. +version: Show library version + +Optional patterns given to `list` and `show` are case and space insensitive. +Both also accept `*` and `?` as wildcards. + +Examples: + + python -m robot.libdoc Dialogs list + python -m robot.libdoc Selenium2Library list browser + python -m robot.libdoc Remote::10.0.0.42:8270 show + python -m robot.libdoc Dialogs show PauseExecution execute* + python -m robot.libdoc Selenium2Library show intro + python -m robot.libdoc Selenium2Library version + +Alternative execution +===================== + +Libdoc works with all interpreters supported by Robot Framework (Python, +Jython and IronPython). In the examples above libdoc is executed as an +installed module, but it can also be executed as a script like +`python path/robot/libdoc.py`. + +For more information see libdoc section in Robot Framework User Guide at +http://code.google.com/p/robotframework/wiki/UserGuide +""" + +import sys +import os + +# Allows running as a script. __name__ check needed with multiprocessing: +# http://code.google.com/p/robotframework/issues/detail?id=1137 +if 'robot' not in sys.modules and __name__ == '__main__': + import pythonpathsetter + +from robot.utils import Application, seq2str +from robot.errors import DataError +from robot.libdocpkg import LibraryDocumentation, ConsoleViewer + + +class LibDoc(Application): + + def __init__(self): + Application.__init__(self, USAGE, arg_limits=(2,), auto_version=False) + + def validate(self, options, arguments): + if ConsoleViewer.handles(arguments[1]): + ConsoleViewer.validate_command(arguments[1], arguments[2:]) + elif len(arguments) > 2: + raise DataError('Only two arguments allowed when writing output.') + return options, arguments + + def main(self, args, name='', version='', format=None, docformat=None): + lib_or_res, output = args[:2] + libdoc = LibraryDocumentation(lib_or_res, name, version, + self._get_doc_format(docformat)) + if ConsoleViewer.handles(output): + ConsoleViewer(libdoc).view(output, *args[2:]) + else: + libdoc.save(output, self._get_output_format(format, output)) + self.console(os.path.abspath(output)) + + def _get_doc_format(self, format): + if not format: + return None + return self._verify_format('Doc format', format, + ['ROBOT', 'TEXT', 'HTML', 'REST']) + + def _get_output_format(self, format, output): + default = os.path.splitext(output)[1][1:] + return self._verify_format('Format', format or default, ['HTML', 'XML']) + + def _verify_format(self, type, format, valid): + format = format.upper() + if format not in valid: + raise DataError("%s must be %s, got '%s'." + % (type, seq2str(valid, lastsep=' or '), format)) + return format + + +def libdoc_cli(arguments): + """Executes Libdoc similarly as from the command line. + + :param arguments: Command line arguments as a list of strings. + + For programmatic usage the :func:`libdoc` function is typically better. It + has a better API for that usage and does not call :func:`sys.exit` like + this function. + + Example:: + + from robot.libdoc import libdoc_cli + + libdoc_cli(['--version', '1.0', 'MyLibrary.py', 'MyLibraryDoc.html']) + """ + LibDoc().execute_cli(arguments) +