label
stringlengths 1
61
| code
stringlengths 4k
8k
| code_compressed
null |
---|---|---|
test spectral model | # Licensed under a 3-clause BSD style license - see LICENSE.rst
import pytest
import numpy as np
from numpy.testing import assert_allclose
import astropy.units as u
from gammapy.catalog import SourceCatalog1LHAASO
from gammapy.modeling.models import (
GaussianSpatialModel,
PointSpatialModel,
PowerLawNormSpectralModel,
PowerLawSpectralModel,
TemplateSpatialModel,
)
from gammapy.utils.testing import requires_data
@pytest.fixture(scope="session")
def lhaaso1():
return SourceCatalog1LHAASO()
@requires_data()
class TestSourceCatalog1LHAASO:
@staticmethod
def test_source_table(lhaaso1):
assert lhaaso1.tag == "1LHAASO"
assert len(lhaaso1.table) == 90
@staticmethod
def test_positions(lhaaso1):
assert len(lhaaso1.positions) == 90
@staticmethod
def test_to_models(lhaaso1):
models = lhaaso1.to_models(which="both")
assert len(models) == 90
models = lhaaso1.to_models(which="KM2A")
assert np.all(
[m.spectral_model.reference.quantity == 50 * u.TeV for m in models]
)
assert len(models) == 75
models = lhaaso1.to_models(which="WCDA")
assert np.all(
[m.spectral_model.reference.quantity == 3 * u.TeV for m in models]
)
assert len(models) == 69
@requires_data()
class TestSourceCatalogObject1LHAASO:
@staticmethod
def test_data(lhaaso1):
assert lhaaso1[0].data["Source_Name"] == "1LHAASO J0007+5659u"
assert "KM2A" in lhaaso1[0].data["Model_a"]
assert_allclose(lhaaso1[0].data["r39_ul"].value, 0.18)
assert lhaaso1[0].data["r39_ul"].unit == u.deg
assert_allclose(lhaaso1[0].data["N0"].value, 0.33e-16)
assert lhaaso1[0].data["N0"].unit == u.Unit("cm−2 s−1 TeV−1")
assert_allclose(lhaaso1[0].data["N0_err"].value, 0.05e-16)
assert lhaaso1[0].data["N0_err"].unit == u.Unit("cm−2 s−1 TeV−1")
assert_allclose(lhaaso1[0].data["N0_ul_b"].value, 0.27e-13)
assert lhaaso1[0].data["N0_ul_b"].unit == u.Unit("cm−2 s−1 TeV−1")
assert lhaaso1[1].data["ASSO_Name"] == "CTA 1"
assert_allclose(lhaaso1[1].data["ASSO_Sep"].value, 0.12)
assert lhaaso1[0].data["ASSO_Sep"].unit == u.deg
assert lhaaso1[10].data["Source_Name"] == "1LHAASO J0428+5531*"
assert "WCDA" in lhaaso1[10].data["Model_a"]
assert_allclose(lhaaso1[10].data["RAJ2000"].value, 67.23)
assert_allclose(lhaaso1[10].data["DECJ2000"].value, 55.53)
assert_allclose(lhaaso1[10].data["pos_err"].value, 0.36)
assert lhaaso1[10].data["RAJ2000"].unit == u.deg
assert lhaaso1[10].data["DECJ2000"].unit == u.deg
assert lhaaso1[10].data["pos_err"].unit == u.deg
assert_allclose(lhaaso1[10].data["r39"].value, 1.18)
assert_allclose(lhaaso1[10].data["r39_b"].value, 0.32)
assert lhaaso1[10].data["r39_b"].unit == u.deg
assert_allclose(lhaaso1[10].data["r39_err"].value, 0.12)
assert_allclose(lhaaso1[10].data["r39_err_b"].value, 0.06)
assert lhaaso1[10].data["r39_err_b"].unit == u.deg
@staticmethod
def test_position(lhaaso1):
position = lhaaso1[0].position
assert_allclose(position.ra.deg, 1.86, atol=1e-3)
assert_allclose(position.dec.deg, 57.00, atol=1e-3)
@staticmethod
def test_sky_model(lhaaso1):
model = lhaaso1[0].sky_model("both")
assert model.name == "1LHAASO J0007+5659u"
assert isinstance(model.spectral_model, PowerLawSpectralModel)
assert isinstance(model.spatial_model, PointSpatialModel)
assert lhaaso1[0].sky_model("WCDA") is None
model = lhaaso1[1].sky_model("both")
assert model.name == "1LHAASO J0007+7303u"
assert isinstance(model.spectral_model, PowerLawNormSpectralModel)
assert isinstance(model.spatial_model, TemplateSpatialModel)
model = lhaaso1[1].sky_model("KM2A")
assert model.name == "1LHAASO J0007+7303u"
assert isinstance(model.spectral_model, PowerLawSpectralModel)
assert isinstance(model.spatial_model, GaussianSpatialModel)
model = lhaaso1[1].sky_model("WCDA")
assert model.name == "1LHAASO J0007+7303u"
assert isinstance(model.spectral_model, PowerLawSpectralModel)
assert isinstance(model.spatial_model, PointSpatialModel)
model = lhaaso1[11].sky_model("both")
assert model.name == "1LHAASO J0500+4454"
assert isinstance(model.spectral_model, PowerLawSpectralModel)
assert isinstance(model.spatial_model, GaussianSpatialModel)
@staticmethod
def METHOD_NAME(lhaaso1):
m = lhaaso1[0].spectral_model("KM2A")
dnde, dnde_err = m.evaluate_error(50 * u.TeV)
assert dnde.unit == "cm-2 s-1 TeV-1"
assert_allclose(dnde.value, 0.33e-16, rtol=1e-3)
assert_allclose(dnde_err.value, 0.05e-16, rtol=1e-3)
m = lhaaso1[11].spectral_model("WCDA")
dnde, dnde_err = m.evaluate_error(3 * u.TeV)
assert dnde.unit == "cm-2 s-1 TeV-1"
assert_allclose(dnde.value, 0.69e-13, rtol=1e-3)
assert_allclose(dnde_err.value, 0.16e-13, rtol=1e-3)
@staticmethod
def test_spatial_model(lhaaso1):
m = lhaaso1[0].spatial_model("KM2A")
assert isinstance(m, PointSpatialModel)
assert m.lon_0.unit == "deg"
assert_allclose(m.lon_0.value, 1.86, atol=1e-2)
assert_allclose(m.lon_0.error, 0.09, atol=1e-2)
assert m.lat_0.unit == "deg"
assert_allclose(m.lat_0.value, 57.00, atol=1e-2)
assert_allclose(m.lat_0.error, 0.049, atol=1e-2)
assert m.frame == "fk5"
m = lhaaso1[11].spatial_model("WCDA")
assert isinstance(m, GaussianSpatialModel)
assert m.lon_0.unit == "deg"
assert_allclose(m.lon_0.value, 75.01, atol=1e-10)
assert m.lat_0.unit == "deg"
assert_allclose(m.lat_0.value, 44.92, atol=1e-10)
assert m.frame == "fk5"
assert m.sigma.unit == "deg"
assert_allclose(m.sigma.value, 0.41, atol=1e-3)
model = lhaaso1["1LHAASO J0007+5659u"].spatial_model("KM2A")
pos_err = model.position_error
assert_allclose(pos_err.height.value, 2 * 0.12, rtol=1e-4)
assert_allclose(pos_err.width.value, 2 * 0.12, rtol=1e-4)
assert_allclose(model.position.ra.value, pos_err.center.ra.value)
assert_allclose(model.position.dec.value, pos_err.center.dec.value) | null |
update dynamic services | """
Copyright 2020 The Magma Authors.
This source code is licensed under the BSD-style license found in the
LICENSE file in the root directory of this source tree.
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
import logging
import time
from typing import List, Optional
import grpc
from magma.common.job import Job
from magma.common.rpc_utils import (
grpc_async_wrapper,
indicates_connection_error,
)
from magma.common.sentry import EXCLUDE_FROM_ERROR_MONITORING
from magma.common.service_registry import ServiceRegistry
from magma.magmad.metrics import UNEXPECTED_SERVICE_RESTARTS
from orc8r.protos.common_pb2 import Void
from orc8r.protos.service303_pb2_grpc import Service303Stub
class ServiceInfo(object):
"""
Stores info about the individual services
"""
# Time buffer for services to restart, in seconds
SERVICE_RESTART_BUFFER_TIME = 30
def __init__(self, service_name):
self.continuous_timeouts = 0
self._service_name = service_name
self._expected_start_time = time.time()
self._status = None
self._linked_services = []
# Initialize the counter for each service
UNEXPECTED_SERVICE_RESTARTS.labels(
service_name=self._service_name,
).inc(0)
@property
def status(self):
return self._status
@property
def linked_services(self):
return self._linked_services
def add_linked_services(self, service_list):
for service in service_list:
if service != self._service_name and \
service not in self._linked_services:
self._linked_services.append(service)
def update(self, start_time, status):
self._status = status
self.continuous_timeouts = 0
if start_time <= self._expected_start_time:
# Probably a race in service starts, or magmad restarted
return
if (
start_time - self._expected_start_time
> self.SERVICE_RESTART_BUFFER_TIME
):
UNEXPECTED_SERVICE_RESTARTS.labels(
service_name=self._service_name,
).inc()
self._expected_start_time = start_time
def process_service_restart(self):
self._expected_start_time = time.time()
class ServicePoller(Job):
"""
Periodically query the services' Service303 interface
"""
# Periodicity for getting status from other services, in seconds
GET_STATUS_INTERVAL = 10
# Timeout when getting status from other local services, in seconds
GET_STATUS_TIMEOUT = 8
def __init__(self, loop, config, dynamic_services: Optional[List[str]] = None):
"""
Initialize the ServicePooler
Args:
loop: loop
config: configuration
dynamic_services: list of dynamic services
"""
super().__init__(
interval=self.GET_STATUS_INTERVAL,
loop=loop,
)
self._config = config
# Holds a map of service name -> ServiceInfo
self._service_info = {}
for service in config['magma_services']:
self._service_info[service] = ServiceInfo(service)
if dynamic_services is not None:
for service in dynamic_services:
self._service_info[service] = ServiceInfo(service)
for service_list in config.get('linked_services', []):
for service in service_list:
self._service_info[service].add_linked_services(service_list)
def METHOD_NAME(
self,
new_services: List[str],
stopped_services: List[str],
):
"""
Update the service poller when dynamic services are enabled or disabled
Args:
new_services: New services which were enabled
stopped_services: Old services which were disabled
"""
for service in new_services:
self._service_info[service] = ServiceInfo(service)
for service in stopped_services:
self._service_info.pop(service)
def get_service_timeouts(self):
ret = {}
for service_name, service in self._service_info.items():
ret[service_name] = service.continuous_timeouts
return ret
def reset_timeout_counter(self, service):
self._service_info[service].continuous_timeouts = 0
@property
def service_info(self):
return self._service_info
def process_service_restart(self, service_name):
self._service_info[service_name].process_service_restart()
for linked_service in self._service_info[service_name].linked_services:
self._service_info[linked_service].process_service_restart()
async def _run(self):
await self._get_service_info()
async def _get_service_info(self):
"""
Make RPC calls to 'GetServiceInfo' functions of other services, to
get current status.
"""
for service in list(self._service_info):
# Check whether service provides service303 interface
if service in self._config['non_service303_services']:
continue
try:
chan = ServiceRegistry.get_rpc_channel(
service, ServiceRegistry.LOCAL,
)
except ValueError:
# Service can't be contacted
logging.error('Cant get RPC channel to %s', service)
continue
client = Service303Stub(chan)
try:
future = client.GetServiceInfo.future(
Void(),
self.GET_STATUS_TIMEOUT,
)
info = await grpc_async_wrapper(future, self._loop)
self._service_info[service].update(
info.start_time_secs,
info.status,
)
self._service_info[service].continuous_timeouts = 0
except grpc.RpcError as err:
logging.error(
"GetServiceInfo Error for %s! [%s] %s",
service,
err.code(),
err.details(),
extra=EXCLUDE_FROM_ERROR_MONITORING if indicates_connection_error(err) else None,
)
self._service_info[service].continuous_timeouts += 1 | null |
libs | # Copyright 2013-2023 Lawrence Livermore National Security, LLC and other
# Spack Project Developers. See the top-level COPYRIGHT file for details.
#
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
import sys
from spack.package import *
class Mesa18(AutotoolsPackage):
"""Mesa is an open-source implementation of the OpenGL specification
- a system for rendering interactive 3D graphics."""
homepage = "https://www.mesa3d.org"
maintainers("v-dobrev", "ChristianTackeGSI")
# Note that we always want to build from the git repo instead of a
# tarball since the tarball has pre-generated files for certain versions
# of LLVM while the git repo doesn't so it can adapt at build time to
# whatever version of LLVM you're using.
git = "https://gitlab.freedesktop.org/mesa/mesa.git"
version("18.3.6", tag="mesa-18.3.6", commit="11049bcff86da8013a4f63bd68daf637e3af22f3")
depends_on("autoconf", type="build")
depends_on("automake", type="build")
depends_on("libtool", type="build")
depends_on("m4", type="build")
depends_on("pkgconfig", type="build")
depends_on("binutils+plugins", when=(sys.platform != "darwin"), type="build")
depends_on("bison", type="build")
depends_on("flex", type="build")
depends_on("gettext", type="build")
depends_on("pkgconfig", type="build")
depends_on("python@:3.8", type="build") # https://github.com/spack/spack/issues/28219
depends_on("[email protected]:", type="build")
depends_on("libxml2")
depends_on("zlib-api")
depends_on("expat")
depends_on("ncurses+termlib")
# Internal options
variant("llvm", default=True, description="Enable LLVM.")
variant(
"swr",
values=spack.variant.DisjointSetsOfValues(
("none",), ("auto",), ("avx", "avx2", "knl", "skx")
)
.with_non_feature_values("auto")
.with_non_feature_values("none")
.with_default("auto"),
when="+llvm",
description="Enable the SWR driver.",
)
# Front ends
variant("osmesa", default=True, description="Enable the OSMesa frontend.")
is_linux = sys.platform.startswith("linux")
variant("glx", default=is_linux, description="Enable the GLX frontend.")
# Additional backends
variant("opengles", default=False, description="Enable OpenGL ES support.")
# Provides
provides("libglx", when="+glx")
provides("libosmesa", when="+osmesa")
# Variant dependencies
depends_on("libllvm@6:10", when="+llvm")
depends_on("libx11", when="+glx")
depends_on("libxcb", when="+glx")
depends_on("libxext", when="+glx")
depends_on("[email protected]:", when="+glx")
# Require at least 1 front-end
conflicts("~osmesa ~glx")
# Prevent an unnecessary xcb-dri dependency
patch("autotools-x11-nodri.patch")
# Backport Mesa MR#6053 to prevent multiply-defined symbols
patch("multiple-symbols_hash.patch", when="@:20.1.4%gcc@10:")
def setup_build_environment(self, env):
env.set("PYTHON", self.spec["python"].command.path)
def autoreconf(self, spec, prefix):
which("autoreconf")("--force", "--verbose", "--install")
def configure_args(self):
spec = self.spec
args = [
"LDFLAGS={0}".format(self.spec["ncurses"].METHOD_NAME.search_flags),
"--enable-shared",
"--disable-static",
"--disable-libglvnd",
"--disable-nine",
"--disable-omx-bellagio",
"--disable-omx-tizonia",
"--disable-opencl",
"--disable-opencl-icd",
"--disable-va",
"--disable-vdpau",
"--disable-xa",
"--disable-xvmc",
"--disable-osmesa",
"--with-vulkan-drivers=",
"--disable-egl",
"--disable-gbm",
"--disable-dri",
"--enable-opengl",
]
args_platforms = []
args_gallium_drivers = ["swrast"]
args_dri_drivers = []
if spec.target.family == "arm" or spec.target.family == "aarch64":
args.append("--disable-libunwind")
num_frontends = 0
if "+osmesa" in spec:
num_frontends += 1
args.append("--enable-gallium-osmesa")
else:
args.append("--disable-gallium-osmesa")
if "+glx" in spec:
num_frontends += 1
args.append("--enable-glx=gallium-xlib")
args_platforms.append("x11")
else:
args.append("--disable-glx")
if "+opengles" in spec:
args.extend(["--enable-gles1", "--enable-gles2"])
else:
args.extend(["--disable-gles1", "--disable-gles2"])
if num_frontends > 1:
args.append("--enable-shared-glapi")
else:
args.append("--disable-shared-glapi")
if "+llvm" in spec:
args.append("--enable-llvm")
args.append("--with-llvm-prefix=%s" % spec["libllvm"].prefix)
if "+llvm_dylib" in spec["libllvm"]:
args.append("--enable-llvm-shared-libs")
else:
args.append("--disable-llvm-shared-libs")
else:
args.append("--disable-llvm")
args_swr_arches = []
if "swr=auto" in spec:
if "avx" in spec.target:
args_swr_arches.append("avx")
if "avx2" in spec.target:
args_swr_arches.append("avx2")
if "avx512f" in spec.target:
if "avx512er" in spec.target:
args_swr_arches.append("knl")
if "avx512bw" in spec.target:
args_swr_arches.append("skx")
else:
if "swr=avx" in spec:
args_swr_arches.append("avx")
if "swr=avx2" in spec:
args_swr_arches.append("avx2")
if "swr=knl" in spec:
args_swr_arches.append("knl")
if "swr=skx" in spec:
args_swr_arches.append("skx")
if args_swr_arches:
args_gallium_drivers.append("swr")
args.append("--with-swr-archs=" + ",".join(args_swr_arches))
# Add the remaining list args
args.append("--with-platforms=" + ",".join(args_platforms))
args.append("--with-gallium-drivers=" + ",".join(args_gallium_drivers))
args.append("--with-dri-drivers=" + ",".join(args_dri_drivers))
return args
@property
def METHOD_NAME(self):
spec = self.spec
libs_to_seek = set()
if "platform=windows" in spec:
libs_to_seek.add("opengl32")
if "+osmesa" in spec:
libs_to_seek.add("osmesa")
else:
libs_to_seek.add("libGL")
if "+osmesa" in spec:
libs_to_seek.add("libOSMesa")
if "+glx" in spec:
libs_to_seek.add("libGL")
if "+opengles" in spec:
libs_to_seek.add("libGLESv1_CM")
libs_to_seek.add("libGLESv2")
return find_libraries(
list(libs_to_seek), root=self.spec.prefix, shared=True, recursive=True
)
@property
def libglx_headers(self):
return find_headers("GL/glx", root=self.spec.prefix.include, recursive=False)
@property
def libglx_libs(self):
return find_libraries("libGL", root=self.spec.prefix, recursive=True)
@property
def libosmesa_headers(self):
return find_headers("GL/osmesa", root=self.spec.prefix.include, recursive=False)
@property
def libosmesa_libs(self):
if "platform=windows" in self.spec:
lib_name = "osmesa"
else:
lib_name = "libOSMesa"
return find_libraries(lib_name, root=self.spec.prefix, recursive=True) | null |
all interfaces | #!/usr/bin/env python3
# Copyright (c) 2014-2022 The Bitcoin Core developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
"""Linux network utilities.
Roughly based on http://voorloopnul.com/blog/a-python-netstat-in-less-than-100-lines-of-code/ by Ricardo Pascal
"""
import sys
import socket
import struct
import array
import os
# STATE_ESTABLISHED = '01'
# STATE_SYN_SENT = '02'
# STATE_SYN_RECV = '03'
# STATE_FIN_WAIT1 = '04'
# STATE_FIN_WAIT2 = '05'
# STATE_TIME_WAIT = '06'
# STATE_CLOSE = '07'
# STATE_CLOSE_WAIT = '08'
# STATE_LAST_ACK = '09'
STATE_LISTEN = '0A'
# STATE_CLOSING = '0B'
def get_socket_inodes(pid):
'''
Get list of socket inodes for process pid.
'''
base = '/proc/%i/fd' % pid
inodes = []
for item in os.listdir(base):
target = os.readlink(os.path.join(base, item))
if target.startswith('socket:'):
inodes.append(int(target[8:-1]))
return inodes
def _remove_empty(array):
return [x for x in array if x !='']
def _convert_ip_port(array):
host,port = array.split(':')
# convert host from mangled-per-four-bytes form as used by kernel
host = bytes.fromhex(host)
host_out = ''
for x in range(0, len(host) // 4):
(val,) = struct.unpack('=I', host[x*4:(x+1)*4])
host_out += '%08x' % val
return host_out,int(port,16)
def netstat(typ='tcp'):
'''
Function to return a list with status of tcp connections at linux systems
To get pid of all network process running on system, you must run this script
as superuser
'''
with open('/proc/net/'+typ,'r',encoding='utf8') as f:
content = f.readlines()
content.pop(0)
result = []
for line in content:
line_array = _remove_empty(line.split(' ')) # Split lines and remove empty spaces.
tcp_id = line_array[0]
l_addr = _convert_ip_port(line_array[1])
r_addr = _convert_ip_port(line_array[2])
state = line_array[3]
inode = int(line_array[9]) # Need the inode to match with process pid.
nline = [tcp_id, l_addr, r_addr, state, inode]
result.append(nline)
return result
def get_bind_addrs(pid):
'''
Get bind addresses as (host,port) tuples for process pid.
'''
inodes = get_socket_inodes(pid)
bind_addrs = []
for conn in netstat('tcp') + netstat('tcp6'):
if conn[3] == STATE_LISTEN and conn[4] in inodes:
bind_addrs.append(conn[1])
return bind_addrs
# from: https://code.activestate.com/recipes/439093/
def METHOD_NAME():
'''
Return all interfaces that are up
'''
import fcntl # Linux only, so only import when required
is_64bits = sys.maxsize > 2**32
struct_size = 40 if is_64bits else 32
s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
max_possible = 8 # initial value
while True:
bytes = max_possible * struct_size
names = array.array('B', b'\0' * bytes)
outbytes = struct.unpack('iL', fcntl.ioctl(
s.fileno(),
0x8912, # SIOCGIFCONF
struct.pack('iL', bytes, names.buffer_info()[0])
))[0]
if outbytes == bytes:
max_possible *= 2
else:
break
namestr = names.tobytes()
return [(namestr[i:i+16].split(b'\0', 1)[0],
socket.inet_ntoa(namestr[i+20:i+24]))
for i in range(0, outbytes, struct_size)]
def addr_to_hex(addr):
'''
Convert string IPv4 or IPv6 address to binary address as returned by
get_bind_addrs.
Very naive implementation that certainly doesn't work for all IPv6 variants.
'''
if '.' in addr: # IPv4
addr = [int(x) for x in addr.split('.')]
elif ':' in addr: # IPv6
sub = [[], []] # prefix, suffix
x = 0
addr = addr.split(':')
for i,comp in enumerate(addr):
if comp == '':
if i == 0 or i == (len(addr)-1): # skip empty component at beginning or end
continue
x += 1 # :: skips to suffix
assert x < 2
else: # two bytes per component
val = int(comp, 16)
sub[x].append(val >> 8)
sub[x].append(val & 0xff)
nullbytes = 16 - len(sub[0]) - len(sub[1])
assert (x == 0 and nullbytes == 0) or (x == 1 and nullbytes > 0)
addr = sub[0] + ([0] * nullbytes) + sub[1]
else:
raise ValueError('Could not parse address %s' % addr)
return bytearray(addr).hex()
def test_ipv6_local():
'''
Check for (local) IPv6 support.
'''
# By using SOCK_DGRAM this will not actually make a connection, but it will
# fail if there is no route to IPv6 localhost.
have_ipv6 = True
try:
s = socket.socket(socket.AF_INET6, socket.SOCK_DGRAM)
s.connect(('::1', 1))
except socket.error:
have_ipv6 = False
return have_ipv6 | null |
test args | import contextlib
import sys
import os
import unittest
from test import support
import time
resource = support.import_module('resource')
# This test is checking a few specific problem spots with the resource module.
class ResourceTest(unittest.TestCase):
def METHOD_NAME(self):
self.assertRaises(TypeError, resource.getrlimit)
self.assertRaises(TypeError, resource.getrlimit, 42, 42)
self.assertRaises(TypeError, resource.setrlimit)
self.assertRaises(TypeError, resource.setrlimit, 42, 42, 42)
def test_fsize_ismax(self):
try:
(cur, max) = resource.getrlimit(resource.RLIMIT_FSIZE)
except AttributeError:
pass
else:
# RLIMIT_FSIZE should be RLIM_INFINITY, which will be a really big
# number on a platform with large file support. On these platforms,
# we need to test that the get/setrlimit functions properly convert
# the number to a C long long and that the conversion doesn't raise
# an error.
self.assertEqual(resource.RLIM_INFINITY, max)
resource.setrlimit(resource.RLIMIT_FSIZE, (cur, max))
def test_fsize_enforced(self):
try:
(cur, max) = resource.getrlimit(resource.RLIMIT_FSIZE)
except AttributeError:
pass
else:
# Check to see what happens when the RLIMIT_FSIZE is small. Some
# versions of Python were terminated by an uncaught SIGXFSZ, but
# pythonrun.c has been fixed to ignore that exception. If so, the
# write() should return EFBIG when the limit is exceeded.
# At least one platform has an unlimited RLIMIT_FSIZE and attempts
# to change it raise ValueError instead.
try:
try:
resource.setrlimit(resource.RLIMIT_FSIZE, (1024, max))
limit_set = True
except ValueError:
limit_set = False
f = open(support.TESTFN, "wb")
try:
f.write(b"X" * 1024)
try:
f.write(b"Y")
f.flush()
# On some systems (e.g., Ubuntu on hppa) the flush()
# doesn't always cause the exception, but the close()
# does eventually. Try flushing several times in
# an attempt to ensure the file is really synced and
# the exception raised.
for i in range(5):
time.sleep(.1)
f.flush()
except OSError:
if not limit_set:
raise
if limit_set:
# Close will attempt to flush the byte we wrote
# Restore limit first to avoid getting a spurious error
resource.setrlimit(resource.RLIMIT_FSIZE, (cur, max))
finally:
f.close()
finally:
if limit_set:
resource.setrlimit(resource.RLIMIT_FSIZE, (cur, max))
support.unlink(support.TESTFN)
def test_fsize_toobig(self):
# Be sure that setrlimit is checking for really large values
too_big = 10**50
try:
(cur, max) = resource.getrlimit(resource.RLIMIT_FSIZE)
except AttributeError:
pass
else:
try:
resource.setrlimit(resource.RLIMIT_FSIZE, (too_big, max))
except (OverflowError, ValueError):
pass
try:
resource.setrlimit(resource.RLIMIT_FSIZE, (max, too_big))
except (OverflowError, ValueError):
pass
def test_getrusage(self):
self.assertRaises(TypeError, resource.getrusage)
self.assertRaises(TypeError, resource.getrusage, 42, 42)
usageself = resource.getrusage(resource.RUSAGE_SELF)
usagechildren = resource.getrusage(resource.RUSAGE_CHILDREN)
# May not be available on all systems.
try:
usageboth = resource.getrusage(resource.RUSAGE_BOTH)
except (ValueError, AttributeError):
pass
try:
usage_thread = resource.getrusage(resource.RUSAGE_THREAD)
except (ValueError, AttributeError):
pass
# Issue 6083: Reference counting bug
def test_setrusage_refcount(self):
try:
limits = resource.getrlimit(resource.RLIMIT_CPU)
except AttributeError:
pass
else:
class BadSequence:
def __len__(self):
return 2
def __getitem__(self, key):
if key in (0, 1):
return len(tuple(range(1000000)))
raise IndexError
resource.setrlimit(resource.RLIMIT_CPU, BadSequence())
def test_pagesize(self):
pagesize = resource.getpagesize()
self.assertIsInstance(pagesize, int)
self.assertGreaterEqual(pagesize, 0)
@unittest.skipUnless(sys.platform == 'linux', 'test requires Linux')
def test_linux_constants(self):
for attr in ['MSGQUEUE', 'NICE', 'RTPRIO', 'RTTIME', 'SIGPENDING']:
with contextlib.suppress(AttributeError):
self.assertIsInstance(getattr(resource, 'RLIMIT_' + attr), int)
@support.requires_freebsd_version(9)
def test_freebsd_contants(self):
for attr in ['SWAP', 'SBSIZE', 'NPTS']:
with contextlib.suppress(AttributeError):
self.assertIsInstance(getattr(resource, 'RLIMIT_' + attr), int)
@unittest.skipUnless(hasattr(resource, 'prlimit'), 'no prlimit')
@support.requires_linux_version(2, 6, 36)
def test_prlimit(self):
self.assertRaises(TypeError, resource.prlimit)
if os.geteuid() != 0:
self.assertRaises(PermissionError, resource.prlimit,
1, resource.RLIMIT_AS)
self.assertRaises(ProcessLookupError, resource.prlimit,
-1, resource.RLIMIT_AS)
limit = resource.getrlimit(resource.RLIMIT_AS)
self.assertEqual(resource.prlimit(0, resource.RLIMIT_AS), limit)
self.assertEqual(resource.prlimit(0, resource.RLIMIT_AS, limit),
limit)
def test_main(verbose=None):
support.run_unittest(ResourceTest)
if __name__ == "__main__":
test_main() | null |
process classlist | from docutils.parsers.rst import Directive
from docutils import nodes
import pdb
from collections import defaultdict
from sphinx import addnodes
from pathlib import Path
class classlist(nodes.General, nodes.Element):
pass
class ClasslistDirective(Directive):
def run(self):
return [classlist('')]
def generate_classlist(app, fromdocname, subtree, class_list, prefix, level=2):
for e in class_list:
ref = nodes.reference('', '')
ref['refdocname'] = e[3]
ref['refuri'] = app.builder.get_relative_uri(fromdocname, e[3])
ref['refuri'] += '#' + e[4]
ref.append(nodes.Text(prefix + e[0].split(".")[-1]))
class_item = nodes.list_item('', addnodes.compact_paragraph('', '', ref), classes=["toctree-l" + str(level)])
#print(fromdocname, e[3])
if fromdocname.startswith(e[3].replace(".api.", ".")):
#print("current")
class_item['classes'].append('current')
subtree += class_item
def generate_collapsible_classlist(app, fromdocname, classes, container, caption, module_index, label_prefix):
toc = nodes.bullet_list()
toc += nodes.caption(caption, '', *[nodes.Text(caption)])
if module_index is not None:
entries = defaultdict(list)
#print("test", classes, fromdocname, caption)
prefix = ".".join(classes[0][0].split(".")[:module_index]) + "."
for e in classes:
module = e[0].split(".")[module_index]
entries[module].append(e)
#print("t", fromdocname)
for module, class_list in entries.items():
#print("t2", "src." + prefix + module)
ref = nodes.reference('', '')
ref['refuri'] = app.builder.get_relative_uri(fromdocname, prefix + module)
ref.append(nodes.Text((label_prefix + module) if label_prefix != "" else module.capitalize()))
module_item = nodes.list_item('', addnodes.compact_paragraph('', '', ref), classes=["toctree-l1"])
if fromdocname.startswith(prefix + module):
module_item["classes"].append('current')
toc += module_item
subtree = nodes.bullet_list()
module_item += subtree
generate_classlist(app, fromdocname, subtree, class_list, "")
else:
generate_classlist(app, fromdocname, toc, classes, label_prefix, level=1)
container += toc
def generate_tutorials_sidebar(app, fromdocname, container):
tutorials_dir = Path(__file__).absolute().parent.parent / "docs" / "tutorials"
tutorials = [
("Loading and manipulating objects", "loader"),
("Configuring the camera", "camera"),
("Rendering the scene", "renderer"),
("Writing the results to file", "writer"),
("How key frames work", "key_frames"),
("Positioning objects via the physics simulator", "physics"),
]
container += nodes.caption("Tutorials", '', *[nodes.Text("Tutorials")])
for tutorial in tutorials:
toc = nodes.bullet_list()
ref = nodes.reference('', '')
ref['refuri'] = app.builder.get_relative_uri(fromdocname, "docs/tutorials/" + tutorial[1])
ref.append(nodes.Text(tutorial[0]))
module_item = nodes.list_item('', addnodes.compact_paragraph('', '', ref), classes=["toctree-l1"])
if fromdocname.startswith("docs/tutorials/" + tutorial[1]):
module_item["classes"].append('current')
toc += module_item
container += toc
def generate_examples_sidebar(app, fromdocname, container):
examples = Path(__file__).absolute().parent.parent / "examples"
container += nodes.caption("Examples", '', *[nodes.Text("Examples")])
for example_groups in [examples / group for group in ["basics", "advanced", "datasets"]]:
if example_groups.is_dir():
toc = nodes.bullet_list()
ref = nodes.reference('', '')
ref['refuri'] = app.builder.get_relative_uri(fromdocname, "examples/" + example_groups.name + "/README")
ref.append(nodes.Text(example_groups.name.capitalize()))
module_item = nodes.list_item('', addnodes.compact_paragraph('', '', ref), classes=["toctree-l1"])
if fromdocname.startswith("examples/" + example_groups.name):
module_item["classes"].append('current')
toc += module_item
subtree = nodes.bullet_list()
module_item += subtree
for example in sorted(example_groups.rglob("*/README.md"), key=lambda x: x.parent.name):
ref = nodes.reference('', '')
ref['refuri'] = app.builder.get_relative_uri(fromdocname, str(example).replace(str(examples), "examples").replace("README.md", "README"))
ref.append(nodes.Text(example.parent.name))
class_item = nodes.list_item('', addnodes.compact_paragraph('', '', ref), classes=["toctree-l2"])
if fromdocname == ref['refuri'].replace(".html", ""):
class_item['classes'].append('current')
subtree += class_item
container += toc
def generate_sidebar(app, fromdocname):
env = app.builder.env
container = nodes.compound(classes=['toctree-wrapper'])#addnodes.compact_paragraph('', '', classes=['toctree-wrapper'])
py = env.get_domain('py')
classes = py.get_objects()
#print("classes", classes, [_[2] for _ in py.get_objects()])
classes_per_group = {"api": ([], None, "bproc."), "internal": ([], 2, "bproc.python."), "modules (deprecated)": ([], 3, "")}#"modules": ([], 1), "provider": ([], 2),
for e in classes:
if e[2] == 'module' and e[3].startswith("blenderproc.api.") or e[2] == 'class' and not e[3].startswith("blenderproc.api."):
#print(e)
if e[3].startswith("blenderproc.api."):
group = "api"
elif e[0].startswith("blenderproc.python.modules."):
group = "modules (deprecated)"
else:
group = "internal"
#print(group, e)
classes_per_group[group][0].append(e)
generate_tutorials_sidebar(app, fromdocname, container)
generate_examples_sidebar(app, fromdocname, container)
for key, items in classes_per_group.items():
generate_collapsible_classlist(app, fromdocname, items[0], container, key.capitalize(), items[1], items[2])
return container
def METHOD_NAME(app, doctree, fromdocname):
container = generate_sidebar(app, fromdocname)
ctx = app.env.config['html_context']
ctx['classlist'] = container
for node in doctree.traverse(classlist):
node.replace_self([container])
continue
def add_classlist_handler(app):
def _print_classlist(**kwargs):
ctx = app.env.config['html_context']
return app.builder.render_partial(ctx['classlist'])['fragment']
ctx = app.env.config['html_context']
if 'print_classlist' not in ctx:
ctx['print_classlist'] = _print_classlist
def html_page_context(app, pagename, templatename, context, doctree):
def make_toctree(collapse=True, maxdepth=-1, includehidden=True, titles_only=False):
if "page_source_suffix" in context and context["page_source_suffix"] == ".md":
fromdocname = context["current_page_name"]
else:
fromdocname = "" if "title" not in context else context["title"]
fulltoc = generate_sidebar(app, fromdocname)
rendered_toc = app.builder.render_partial(fulltoc)['fragment']
return rendered_toc
context['toctree'] = make_toctree
def setup(app):
app.add_node(classlist)
app.add_directive('classlist', ClasslistDirective)
app.connect('doctree-resolved', METHOD_NAME)
app.connect('builder-inited', add_classlist_handler)
app.connect('html-page-context', html_page_context) | null |
fake authz key file |
from Utils.Utilities import encodeUnicodeToBytes
from future.utils import viewitems, viewvalues, listitems
import os, hmac, hashlib, cherrypy
from tempfile import NamedTemporaryFile
from Utils.PythonVersion import PY3
from WMCore.REST.Main import RESTMain
from WMCore.REST.Auth import authz_canonical
from WMCore.Configuration import Configuration
def fake_authz_headers(hmac_key, method = 'HNLogin',
login='testuser', name='Test User',
dn="/test/dn", roles={}, format="list"):
"""Create fake authentication and authorisation headers compatible
with the CMSWEB front-ends. Assumes you have the HMAC signing key
the back-end will use to validate the headers.
:arg str hmac_key: binary key data for signing headers.
:arg str method: authentication method, one of X509Cert, X509Proxy,
HNLogin, HostIP, AUCookie or None.
:arg str login: account login name.
:arg str name: account user name.
:arg str dn: account X509 subject.
:arg dict roles: role dictionary, each role with 'site' and 'group' lists.
:returns: list of header name, value tuples to add to a HTTP request."""
headers = { 'cms-auth-status': 'OK', 'cms-authn-method': method }
if login:
headers['cms-authn-login'] = login
if name:
headers['cms-authn-name'] = name
if dn:
headers['cms-authn-dn'] = dn
for name, role in viewitems(roles):
name = 'cms-authz-' + authz_canonical(name)
headers[name] = []
for r in 'site', 'group':
if r in role:
headers[name].extend(["%s:%s" % (r, authz_canonical(v)) for v in role[r]])
headers[name] = " ".join(headers[name])
prefix = suffix = ""
hkeys = list(headers)
for hk in sorted(hkeys):
if hk != 'cms-auth-status':
prefix += "h%xv%x" % (len(hk), len(headers[hk]))
suffix += "%s%s" % (hk, headers[hk])
msg = prefix + "#" + suffix
if PY3:
hmac_key = encodeUnicodeToBytes(hmac_key)
msg = encodeUnicodeToBytes(msg)
cksum = hmac.new(hmac_key, msg, hashlib.sha1).hexdigest()
headers['cms-authn-hmac'] = cksum
if format == "list":
return listitems(headers)
else:
return headers
def METHOD_NAME(delete=True):
"""Create temporary file for fake authorisation hmac signing key.
:returns: Instance of :class:`~.NamedTemporaryFile`, whose *data*
attribute contains the HMAC signing binary key."""
t = NamedTemporaryFile(delete=delete)
with open("/dev/urandom", "rb") as fd:
t.data = fd.read(20)
t.write(t.data)
t.seek(0)
return t
def setup_dummy_server(module_name, class_name, app_name = None, authz_key_file=None, port=8888):
"""Helper function to set up a :class:`~.RESTMain` server from given
module and class. Creates a fake server configuration and instantiates
the server application from it.
:arg str module_name: module from which to import test class.
:arg str class_type: name of the server test class.
:arg str app_name: optional test application name, 'test' by default.
:returns: tuple with the server object and authz hmac signing key."""
if authz_key_file:
test_authz_key = authz_key_file
else:
test_authz_key = METHOD_NAME()
cfg = Configuration()
main = cfg.section_('main')
main.application = app_name or 'test'
main.silent = True
main.index = 'top'
main.authz_defaults = { 'role': None, 'group': None, 'site': None }
main.section_('tools').section_('cms_auth').key_file = test_authz_key.name
app = cfg.section_(app_name or 'test')
app.admin = '[email protected]'
app.description = app.title = 'Test'
views = cfg.section_('views')
top = views.section_('top')
top.object = module_name + "." + class_name
server = RESTMain(cfg, os.getcwd())
server.validate_config()
server.setup_server()
server.install_application()
cherrypy.config.update({'server.socket_port': port})
cherrypy.config.update({'server.socket_host': '127.0.0.1'})
cherrypy.config.update({'request.show_tracebacks': True})
cherrypy.config.update({'environment': 'test_suite'})
for app in viewvalues(cherrypy.tree.apps):
if '/' in app.config:
app.config["/"]["request.show_tracebacks"] = True
return server, test_authz_key | null |
run | import os
import shutil
import attr
import plistlib
from packaging import version
import tempfile
from drawBot.drawBotSettings import __version__
from drawBot.scriptTools import ScriptRunner
from drawBot.drawBotDrawingTools import _drawBotDrawingTool
"""
DrawBot support for .drawbot packages.
* Read and build packages.
* Run packages
MyTool.drawbot (package)
info.plist
lib/
main.py
anotherScript.py
anImage.jpg
info.plist
name: optional, default to <name>.drawBot
version: optional, default to 0.0, but advised to set
developer: optional, default to None
developerURL: optional, default to None
requiresVersion: optional, default to all versions
mainScript: optional, default to 'lib/main.py' if this isn't specified?)
"""
drawBotVersion = version.Version(__version__)
@attr.s(slots=True)
class DrawBotPackageInfo(object):
name = attr.ib(default="", init=False, type=str, validator=attr.validators.instance_of(str))
version = attr.ib(default="0.0", init=False, type=str, validator=attr.validators.instance_of(str))
developer = attr.ib(default="", init=False, type=str, validator=attr.validators.instance_of(str))
developerURL = attr.ib(default="", init=False, type=str, validator=attr.validators.instance_of(str))
requiresVersion = attr.ib(default="0.0", init=False, type=str, validator=attr.validators.instance_of(str))
mainScript = attr.ib(default="main.py", init=False, type=str, validator=attr.validators.instance_of(str))
def asDict(self):
"""
Return only keys which are different from the defaults.
"""
defaults = attr.asdict(self.__class__())
data = attr.asdict(self)
for key, value in defaults.items():
if data[key] == value:
del data[key]
return data
def fromDict(self, data):
"""
Set data from dict.
"""
for key, value in data.items():
setattr(self, key, value)
class DrawBotPackage(object):
def __init__(self, path=None):
self.path = path
self.info = DrawBotPackageInfo()
self._readInfo()
def _readInfo(self):
"""
Read the info.plist if the file is availble.
"""
# get the info.plist path
infoPath = self.infoPath()
if infoPath and os.path.exists(infoPath):
with open(infoPath, "rb") as f:
info = plistlib.load(f)
self.info.fromDict(info)
# validate incoming info
attr.validate(self.info)
def infoPath(self):
"""
Return the info.plist path.
Return None if there is no root path.
"""
if self.path:
return os.path.join(self.path, "info.plist")
return None
def mainScriptPath(self):
"""
Return the main scripting python file path.
Return None if ther is no root path.
"""
if self.path:
return os.path.join(self.path, "lib", self.info.mainScript)
return None
def METHOD_NAME(self):
"""
Execute the .drawBot package.
Return if executing was succesfull with a report on failure.
"""
# check if the package can run in this version of DrawBot
if version.Version(self.info.requiresVersion) > drawBotVersion:
return False, "Requires a newer version of DrawBot (%s)." % self.info.requiresVersion
# get the main scriptin path
path = self.mainScriptPath()
if path is None:
return False, "Cannot execute an empty package."
if not os.path.exists(path):
return False, "Cannot find '%s'." % path
# create a namespace
namespace = {}
# add the tool callbacks in the name space
_drawBotDrawingTool._addToNamespace(namespace)
# run the script
ScriptRunner(path=path, namespace=namespace)
return True, ""
def buildPackage(self, destinationPath, scriptRoot):
"""
Build a .drawbot package
"""
if not destinationPath.endswith(".drawbot"):
return False, "The path to save the package must have a .drawbot file extension."
# check if the script root exists
if not os.path.exists(scriptRoot):
return False, "Cannot find the script root '%s'" % scriptRoot
# check if the main script path exists
mainScriptPath = os.path.join(scriptRoot, self.info.mainScript)
if not os.path.exists(mainScriptPath):
return False, "Main script path '%s' does not exists in '%s'." % (self.info.mainScript, mainScriptPath)
# build packages in temp folder
tempDir = tempfile.mkdtemp()
# set the temp folder
self.path = tempDir
# validate info
attr.validate(self.info)
# write the plist
infoData = self.info.asDict()
# only write info that is different from
if infoData:
with open(self.infoPath(), "wb") as f:
plistlib.dump(infoData, f)
# build lib root path
libRoot = os.path.join(self.path, "lib")
# copy the script root
shutil.copytree(scriptRoot, libRoot)
# remove existing destination paths
if os.path.exists(destinationPath):
shutil.rmtree(destinationPath)
# copy the temp to the destination path
shutil.copytree(tempDir, destinationPath)
# remove the temp
shutil.rmtree(tempDir)
# set the destination path
self.path = destinationPath
return True, ""
# ####
# # an empty package
# p = DrawBotPackage()
# print(p.path)
# print(p.info.name)
# print(p.infoPath())
# print(p.info)
# # create on the deskop such a file
# tempScriptDir = tempfile.mkdtemp()
# with open(os.path.join(tempScriptDir, "main.py"), "w") as f:
# f.write("print('hello world! Im running from a .drawbot package!!')")
# packagePath = os.path.expanduser(os.path.join("~", "Desktop", "simple.drawbot"))
# package = DrawBotPackage()
# package.buildPackage(packagePath, tempScriptDir)
# shutil.rmtree(tempScriptDir)
# p = DrawBotPackage(packagePath)
# print(p.path)
# print(p.info.name)
# print(p.infoPath())
# print(p.info)
# print(p.run()) | null |
get info | """This is the main component of the game, the core of the game that squishes
everything together and make the game work. Usually user can just import this
module and use the beat() function and everything will be fine, but will there
be any certain things that can't be accomplished that way, the user may also
freely import another modules.
"""
import re
from zulip_bots.game_handler import BadMoveException
from . import database, mechanics
COMMAND_PATTERN = re.compile("^(\\w*).*(\\d,\\d).*(\\d,\\d)|^(\\w+).*(\\d,\\d)")
def METHOD_NAME():
"""Gets the info on starting the game
:return: Info on how to start the game
"""
return "To start a game, mention me and add `create`. A game will start " "in that topic. "
def getHelp():
"""Gets the help message
:return: Help message
"""
return """Commands:
put (v,h): Put a man into the grid in phase 1
move (v,h) -> (v,h): Moves a man from one point to -> another point
take (v,h): Take an opponent's man from the grid in phase 2/3
v: vertical position of grid
h: horizontal position of grid"""
def unknown_command():
"""Returns an unknown command info
:return: A string containing info about available commands
"""
message = "Unknown command. Available commands: put (v,h), take (v,h), move (v,h) -> (v,h)"
raise BadMoveException(message)
def beat(message, topic_name, merels_storage):
"""This gets triggered every time a user send a message in any topic
:param message: User's message
:param topic_name: User's current topic
:param merels_storage: Merels' storage
:return: a tuple of response string and message, non-empty string
we want to keep the turn of the same played,
an empty string otherwise.
"""
database.MerelsStorage(topic_name, merels_storage)
match = COMMAND_PATTERN.match(message)
same_player_move = "" # message indicating move of the same player
if match is None:
return unknown_command()
if match.group(1) is not None and match.group(2) is not None and match.group(3) is not None:
responses = ""
command = match.group(1)
if command.lower() == "move":
p1 = [int(x) for x in match.group(2).split(",")]
p2 = [int(x) for x in match.group(3).split(",")]
if mechanics.get_take_status(topic_name, merels_storage) == 1:
raise BadMoveException("Take is required to proceed." " Please try again.\n")
responses += mechanics.move_man(topic_name, p1, p2, merels_storage) + "\n"
no_moves = after_event_checkup(responses, topic_name, merels_storage)
mechanics.update_hill_uid(topic_name, merels_storage)
responses += mechanics.display_game(topic_name, merels_storage) + "\n"
if no_moves != "":
same_player_move = no_moves
else:
return unknown_command()
if mechanics.get_take_status(topic_name, merels_storage) == 1:
same_player_move = "Take is required to proceed.\n"
return responses, same_player_move
elif match.group(4) is not None and match.group(5) is not None:
command = match.group(4)
p1 = [int(x) for x in match.group(5).split(",")]
# put 1,2
if command == "put":
responses = ""
if mechanics.get_take_status(topic_name, merels_storage) == 1:
raise BadMoveException("Take is required to proceed." " Please try again.\n")
responses += mechanics.put_man(topic_name, p1[0], p1[1], merels_storage) + "\n"
no_moves = after_event_checkup(responses, topic_name, merels_storage)
mechanics.update_hill_uid(topic_name, merels_storage)
responses += mechanics.display_game(topic_name, merels_storage) + "\n"
if no_moves != "":
same_player_move = no_moves
if mechanics.get_take_status(topic_name, merels_storage) == 1:
same_player_move = "Take is required to proceed.\n"
return responses, same_player_move
# take 5,3
elif command == "take":
responses = ""
if mechanics.get_take_status(topic_name, merels_storage) == 1:
responses += mechanics.take_man(topic_name, p1[0], p1[1], merels_storage) + "\n"
if "Failed" in responses:
raise BadMoveException(responses)
mechanics.update_toggle_take_mode(topic_name, merels_storage)
no_moves = after_event_checkup(responses, topic_name, merels_storage)
mechanics.update_hill_uid(topic_name, merels_storage)
responses += mechanics.display_game(topic_name, merels_storage) + "\n"
responses += check_win(topic_name, merels_storage)
if no_moves != "":
same_player_move = no_moves
return responses, same_player_move
else:
raise BadMoveException("Taking is not possible.")
else:
return unknown_command()
def check_take_mode(response, topic_name, merels_storage):
"""This checks whether the previous action can result in a take mode for
current player. This assumes that the previous action is successful and not
failed.
:param response: A response string
:param topic_name: Topic name
:param merels_storage: Merels' storage
:return: None
"""
if not ("Failed" in response):
if mechanics.can_take_mode(topic_name, merels_storage):
mechanics.update_toggle_take_mode(topic_name, merels_storage)
else:
mechanics.update_change_turn(topic_name, merels_storage)
def check_any_moves(topic_name, merels_storage):
"""Check whether the player can make any moves, if can't switch to another
player
:param topic_name: Topic name
:param merels_storage: MerelsDatabase object
:return: A response string
"""
if not mechanics.can_make_any_move(topic_name, merels_storage):
mechanics.update_change_turn(topic_name, merels_storage)
return "Cannot make any move on the grid. Switching to " "previous player.\n"
return ""
def after_event_checkup(response, topic_name, merels_storage):
"""After doing certain moves in the game, it will check for take mode
availability and check for any possible moves
:param response: Current response string. This is useful for checking
any failed previous commands
:param topic_name: Topic name
:param merels_storage: Merels' storage
:return: A response string
"""
check_take_mode(response, topic_name, merels_storage)
return check_any_moves(topic_name, merels_storage)
def check_win(topic_name, merels_storage):
"""Checks whether the current grid has a winner, if it does, finish the
game and remove it from the database
:param topic_name: Topic name
:param merels_storage: Merels' storage
:return:
"""
merels = database.MerelsStorage(topic_name, merels_storage)
win = mechanics.who_won(topic_name, merels_storage)
if win != "None":
merels.remove_game(topic_name)
return f"{win} wins the game!"
return "" | null |
run | import logging
import os
from avocado.utils import process
from virttest import libvirt_version
from virttest import virt_vm, utils_misc
from virttest.libvirt_xml import vm_xml, xcepts
from virttest.utils_test import libvirt
from virttest.utils_libvirt import libvirt_disk
LOG = logging.getLogger('avocado.' + __name__)
cleanup_files = []
def create_customized_disk(params, device_target, source_file_path):
"""
Create one customized disk with related attributes
:param params: dict wrapped with params
:param device_target: device target
:param source_file_path: source file path
"""
type_name = params.get("type_name")
disk_device = params.get("device_type")
device_bus = params.get("target_bus")
device_format = params.get("target_format")
source_dict = {}
if source_file_path:
if 'block' in type_name:
source_dict.update({"dev": source_file_path})
else:
source_dict.update({"file": source_file_path})
disk_src_dict = {"attrs": source_dict}
addr_str = params.get("addr_attrs")
customized_disk = libvirt_disk.create_primitive_disk_xml(
type_name, disk_device,
device_target, device_bus,
device_format, disk_src_dict, None)
if addr_str:
addr_dict = eval(addr_str)
customized_disk.address = customized_disk.new_disk_address(
**{"attrs": addr_dict})
LOG.debug("create customized xml: %s", customized_disk)
return customized_disk
def create_multiple_file_source_disks(params):
"""
Create multiple file source disks
:param params: dict wrapped with params
"""
device_format = params.get("target_format")
source_file_path = params.get("virt_disk_device_source")
source_file_list = ["%s.1" % source_file_path, "%s.2" % source_file_path, "%s.3" % source_file_path]
device_target_list = ['vdb', 'vdc', 'vdd']
created_file_source_disks = []
for device_target, source_file in zip(device_target_list, source_file_list):
libvirt.create_local_disk("file", source_file, 1, device_format)
cleanup_files.append(source_file)
source_disk = create_customized_disk(params, device_target, source_file)
created_file_source_disks.append(source_disk)
return created_file_source_disks
def check_multiple_file_source_disks(params, log_config_path, test):
"""
Check related information in libvirtd log
:param params: wrapped parameters in dictionary
:param log_config_path: log config path
:param test: test assert object
"""
msg1 = params.get('message_1', 'Setting up disks')
msg2 = params.get('message_2', 'Setup all disks')
for message in [msg1, msg2]:
result = utils_misc.wait_for(lambda: libvirt.check_logfile(message, log_config_path), timeout=20)
if not result:
test.fail("Failed to get expected messages: %s from log file: %s."
% (message, log_config_path))
def METHOD_NAME(test, params, env):
"""
Test start Vm with device without device control file.
1.Prepare test environment,destroy or suspend a VM.
2.Prepare test xml.
3.Perform test operation.
4.Recover test environment.
5.Confirm the test result.
"""
vm_name = params.get("main_vm")
vm = env.get_vm(vm_name)
virsh_dargs = {'debug': True, 'ignore_status': True}
# Disk specific attributes.
backend_device = params.get("backend_device", "disk")
hotplug = "yes" == params.get("virt_device_hotplug")
status_error = "yes" == params.get("status_error")
define_error = "yes" == params.get("define_error", "no")
log_config_path = os.path.join(test.debugdir, "libvirtd.log")
control_path = '/dev/mapper/control'
disk_objects = None
# Back up xml file.
if vm.is_alive():
vm.destroy(gracefully=False)
vmxml_backup = vm_xml.VMXML.new_from_inactive_dumpxml(vm_name)
# Skip this case if libvirt version doesn't support this feature
libvirt_version.is_libvirt_feature_supported(params)
try:
vmxml = vm_xml.VMXML.new_from_dumpxml(vm_name)
if backend_device == "multiple_disks":
disk_objects = create_multiple_file_source_disks(params)
if os.path.exists(control_path):
process.METHOD_NAME('rm -rf /dev/mapper/control', ignore_status=True, shell=True)
if not hotplug:
# Sync VM xml.
for disk_xml in disk_objects:
vmxml.add_device(disk_xml)
vmxml.sync()
vm.start()
vm.wait_for_login().close()
except virt_vm.VMStartError as e:
if status_error:
if hotplug:
test.fail("In hotplug scenario, VM should "
"start successfully but not."
"Error: %s" % str(e))
else:
LOG.debug("VM failed to start as expected."
"Error: %s", str(e))
else:
test.fail("VM failed to start."
"Error: %s" % str(e))
except xcepts.LibvirtXMLError as xml_error:
if not define_error:
test.fail("Failed to define VM:\n%s" % xml_error)
else:
LOG.info("As expected, failed to define VM")
except Exception as ex:
test.error("unexpected exception happen: %s" % str(ex))
else:
if backend_device == "multiple_disks":
check_multiple_file_source_disks(params, log_config_path, test)
finally:
# Recover VM.
if vm.is_alive():
vm.destroy(gracefully=False)
LOG.info("Restoring vm...")
vmxml_backup.sync()
# Clean up images
for file_path in cleanup_files:
if os.path.exists(file_path):
os.remove(file_path)
if not os.path.exists(control_path):
process.METHOD_NAME('mknod /dev/mapper/control c 10 236', ignore_status=True, shell=True) | null |
get selected filter options | import platform
from os.path import expanduser
from PySide6 import QtWidgets, QtGui
from src.Controller.PathHandler import resource_path
import logging
class SelectSubgroupOptions(QtWidgets.QWidget):
"""
Select subgroup options for batch processing.
"""
def __init__(self):
"""
Initialise the class.
"""
QtWidgets.QWidget.__init__(self)
# Create the main layout
self.main_layout = QtWidgets.QVBoxLayout()
# Get the stylesheet
if platform.system() == 'Darwin':
self.stylesheet_path = "res/stylesheet.qss"
else:
self.stylesheet_path = "res/stylesheet-win-linux.qss"
self.stylesheet = open(resource_path(self.stylesheet_path)).read()
# Info messages
self.message = QtWidgets.QLabel(
"No Clinical-data-SR files located in current selected directory"
)
# Filter options table
self.filter_table = QtWidgets.QTableWidget(0, 0)
self.filter_table.setStyleSheet(self.stylesheet)
# Set up layout
self.main_layout.addWidget(self.message)
self.main_layout.addWidget(self.filter_table)
self.setLayout(self.main_layout)
# storing the currently selected filter options
self._selected_filters = {}
self.filter_table.cellClicked.connect(self.select_filter_option_cell)
logging.debug("SelectSubgroupOptions successfully initialised.")
def METHOD_NAME(self):
"""
Getter for the selected filters
"""
return self._selected_filters
def display_no_data(self):
"""
Utility method to display an empty table and appropriate message
"""
self.message.setText(
"No Clinical-data-SR files located in current selected directory"
)
self.filter_table.setRowCount(0)
self.filter_table.setColumnCount(0)
self.filter_table.setHorizontalHeaderLabels([])
def show_filtering_options_in_table(self, options_data_dict):
"""
Displays the data in the table
:param options_data_dict: dictionary of clinical
data attributes and values
"""
if not options_data_dict:
self.display_no_data()
return
self.message.setText("Select values to filter by:")
self.filter_table.setRowCount(0)
self.filter_table.setColumnCount(0)
# removes the Patient Identifier (assumed to be first column
# in the dataset)
# As the column name may be changing we cannot hard code the
# column name
# not a necessary filter option as specified in requirements
options_data_dict.pop(list(options_data_dict)[0])
columns_to_remove = []
# remove the keys with an empty list of options
for title, values_list in options_data_dict.items():
filtered_values = [x for x in values_list if x != ""]
if len(filtered_values) == 0:
columns_to_remove.append(title)
for title in columns_to_remove:
options_data_dict.pop(title)
for title in options_data_dict.keys():
col = self.filter_table.columnCount()
self.filter_table.insertColumn(col)
for row in range(0, len(options_data_dict[title])):
str_value = str(options_data_dict[title][row])
# filters out blank options
if str_value == "":
continue
filter_value = QtWidgets.QTableWidgetItem(str_value)
if row >= self.filter_table.rowCount():
self.filter_table.insertRow(row)
self.filter_table.setItem(row, col, filter_value)
self.filter_table.setHorizontalHeaderLabels(options_data_dict.keys())
def select_filter_option_cell(self, row, column):
"""
Toggles the selected options green and stores value
:param row: row index that was clicked
:param column: column index that was clicked
"""
item = self.filter_table.item(row, column)
# in the case they select empty cell
if not item:
return
header = self.filter_table.horizontalHeaderItem(column).text()
text_filter = item.text()
logging.debug(f"Cell selected: ({row}, {column}). " \
"Column header: '{header}'. String-value: '{text_filter}'")
if header in self._selected_filters.keys():
current_filters = self._selected_filters[header]
if text_filter in current_filters:
item.setBackground(QtGui.QColor(255, 255, 255))
# and remove item
self.remove_selected_filters(header, text_filter)
logging.debug(f"Value: '{text_filter}' deselected")
return
item.setBackground(QtGui.QColor(144, 238, 144))
# store in model for reference at output
self.set_selected_filters(header, text_filter)
logging.debug(f"Value: '{text_filter}' selected")
def set_selected_filters(self, filter_type, value):
"""
Setter for the selected filters
:param filter_type: the column name
:param value: the actual value within that column
"""
if filter_type not in self._selected_filters.keys():
self._selected_filters[filter_type] = [value]
else:
self._selected_filters[filter_type].append(value)
def remove_selected_filters(self, filter_type, value):
"""
Remove filter from the selected_filters store
:param filter_type: the column name
:param value: the actual value within that column
"""
if filter_type not in self._selected_filters.keys():
return
else:
self._selected_filters[filter_type].remove(value) | null |
format plotly | import pandas as pd
import streamlit as st
import yfinance
from openbb_terminal.core.plots.plotly_helper import OpenBBFigure
from openbb_terminal.core.session.current_system import set_system_variable
from openbb_terminal.dashboards.stream import streamlit_helpers as st_helpers
pd.options.plotting.backend = "plotly"
# Suppressing sdk logs
set_system_variable("LOGGING_SUPPRESS", True)
st.set_page_config(
layout="wide",
page_title="Chains",
initial_sidebar_state="expanded",
)
st_helpers.set_current_page("Chains")
st_helpers.set_css()
st.sidebar.markdown(
"<h2 style='text-align: center;'>Option Chain Dashboard</h2>",
unsafe_allow_html=True,
)
st.sidebar.write("Select a ticker and expiry to view the option chain")
TICKER_WIDGET = st.sidebar.empty()
EXPIRY_WIDGET = st.sidebar.empty()
INST_WIDGET = st.sidebar.empty()
X_WIDGET = st.sidebar.empty()
Y_WIDGET = st.sidebar.empty()
OPTS = [
"lastTradeDate",
"strike",
"lastPrice",
"bid",
"ask",
"change",
"percentChange",
"volume",
"openInterest",
"impliedVolatility",
]
def clean_str(string):
new_str = ""
for letter in string:
if letter.isupper():
new_str += " "
new_str += letter
return new_str.title()
def METHOD_NAME(fig: OpenBBFigure, x, y, ticker, expiry, inst):
fig.update_yaxes(title=clean_str(y))
fig.update_xaxes(title=clean_str(x))
expires = ", ".join(expiry)
title = (
f"{clean_str(y)} vs. {clean_str(x)} for {ticker.upper()} {inst}s on {expires}"
)
fig.update_layout(
margin=dict(t=40),
autosize=False,
width=1000,
height=500,
title=dict(
text=title,
y=0.98,
x=0.5,
xanchor="center",
yanchor="top",
),
)
class Chart:
def __init__(self):
self.stock = st_helpers.load_state("stock", {})
self.last_ticker = st_helpers.load_state("last_ticker", "")
self.expiry = st_helpers.load_state("expiry", [])
self.dfs = st_helpers.load_state("dfs", {})
self.options = st_helpers.load_state("options", [])
def update(self, ticker, inst, x, y):
if self.expiry:
fig = OpenBBFigure()
for expire in self.expiry:
if expire not in self.dfs:
self.dfs[expire] = self.stock.option_chain(expire)
group = self.dfs[expire]
df = group.calls if inst == "Call" else group.puts
fig.add_scatter(
x=df[x],
y=df[y],
mode="lines",
connectgaps=True,
name=expire,
)
METHOD_NAME(fig, x, y, ticker, self.expiry, inst)
fig.update_layout(margin=dict(l=40))
st.plotly_chart(fig, use_container_width=True)
def on_change(self):
ticker = st_helpers.load_state("ticker", "")
if ticker and ticker != self.last_ticker:
stock = yfinance.Ticker(ticker)
self.options = list([*stock.options])
st_helpers.save_state("options", self.options)
self.last_ticker = ticker
st_helpers.save_state("last_ticker", self.last_ticker)
self.stock = stock
st_helpers.save_state("stock", self.stock)
def run(self):
with TICKER_WIDGET.container():
ticker = TICKER_WIDGET.text_input(
"Ticker", key="ticker", on_change=self.on_change
)
if self.options:
with EXPIRY_WIDGET.container():
exp = EXPIRY_WIDGET.multiselect(
"Expiry",
st_helpers.load_state("options", []),
key="exp",
default=self.options[0] if self.options else None,
)
self.expiry = exp
with INST_WIDGET.container():
inst = INST_WIDGET.selectbox("Type", ["Put", "Call"])
with X_WIDGET.container():
x = X_WIDGET.selectbox("X", OPTS, index=1)
with Y_WIDGET.container():
y = Y_WIDGET.selectbox("Y", OPTS, index=3)
self.update(ticker, inst, x, y)
if __name__ == "__main__":
Chart().run() | null |
resolve available attributes | from typing import List
import graphene
from ...attribute import models as attribute_models
from ...page import models
from ...permission.enums import PagePermissions, PageTypePermissions
from ..attribute.filters import AttributeFilterInput, AttributeWhereInput
from ..attribute.types import Attribute, AttributeCountableConnection, SelectedAttribute
from ..core import ResolveInfo
from ..core.connection import (
CountableConnection,
create_connection_slice,
filter_connection_queryset,
)
from ..core.descriptions import ADDED_IN_33, DEPRECATED_IN_3X_FIELD, RICH_CONTENT
from ..core.doc_category import DOC_CATEGORY_PAGES
from ..core.federation import federated_entity, resolve_federation_references
from ..core.fields import FilterConnectionField, JSONString, PermissionsField
from ..core.scalars import Date
from ..core.types import ModelObjectType, NonNullList
from ..meta.types import ObjectWithMetadata
from ..translations.fields import TranslationField
from ..translations.types import PageTranslation
from .dataloaders import (
PageAttributesByPageTypeIdLoader,
PagesByPageTypeIdLoader,
PageTypeByIdLoader,
SelectedAttributesByPageIdLoader,
)
@federated_entity("id")
class PageType(ModelObjectType[models.PageType]):
id = graphene.GlobalID(required=True, description="ID of the page type.")
name = graphene.String(required=True, description="Name of the page type.")
slug = graphene.String(required=True, description="Slug of the page type.")
attributes = NonNullList(
Attribute, description="Page attributes of that page type."
)
available_attributes = FilterConnectionField(
AttributeCountableConnection,
filter=AttributeFilterInput(),
where=AttributeWhereInput(),
description="Attributes that can be assigned to the page type.",
permissions=[
PagePermissions.MANAGE_PAGES,
PageTypePermissions.MANAGE_PAGE_TYPES_AND_ATTRIBUTES,
],
)
has_pages = PermissionsField(
graphene.Boolean,
description="Whether page type has pages assigned.",
permissions=[
PagePermissions.MANAGE_PAGES,
PageTypePermissions.MANAGE_PAGE_TYPES_AND_ATTRIBUTES,
],
)
class Meta:
description = (
"Represents a type of page. It defines what attributes are available to "
"pages of this type."
)
interfaces = [graphene.relay.Node, ObjectWithMetadata]
model = models.PageType
@staticmethod
def get_model():
return models.PageType
@staticmethod
def resolve_attributes(root: models.PageType, info: ResolveInfo):
return PageAttributesByPageTypeIdLoader(info.context).load(root.pk)
@staticmethod
def METHOD_NAME(
root: models.PageType, info: ResolveInfo, **kwargs
):
qs = attribute_models.Attribute.objects.get_unassigned_page_type_attributes(
root.pk
)
qs = filter_connection_queryset(qs, kwargs, info.context)
return create_connection_slice(qs, info, kwargs, AttributeCountableConnection)
@staticmethod
def resolve_has_pages(root: models.PageType, info: ResolveInfo):
return (
PagesByPageTypeIdLoader(info.context)
.load(root.pk)
.then(lambda pages: bool(pages))
)
@staticmethod
def __resolve_references(roots: List["PageType"], _info: ResolveInfo):
return resolve_federation_references(PageType, roots, models.PageType.objects)
class PageTypeCountableConnection(CountableConnection):
class Meta:
doc_category = DOC_CATEGORY_PAGES
node = PageType
class Page(ModelObjectType[models.Page]):
id = graphene.GlobalID(required=True, description="ID of the page.")
seo_title = graphene.String(description="Title of the page for SEO.")
seo_description = graphene.String(description="Description of the page for SEO.")
title = graphene.String(required=True, description="Title of the page.")
content = JSONString(description="Content of the page." + RICH_CONTENT)
publication_date = Date(
deprecation_reason=(
f"{DEPRECATED_IN_3X_FIELD} "
"Use the `publishedAt` field to fetch the publication date."
),
)
published_at = graphene.DateTime(
description="The page publication date." + ADDED_IN_33
)
is_published = graphene.Boolean(
required=True, description="Determines if the page is published."
)
slug = graphene.String(required=True, description="Slug of the page.")
page_type = graphene.Field(
PageType, required=True, description="Determines the type of page"
)
created = graphene.DateTime(
required=True, description="Date and time at which page was created."
)
content_json = JSONString(
description="Content of the page." + RICH_CONTENT,
deprecation_reason=f"{DEPRECATED_IN_3X_FIELD} Use the `content` field instead.",
required=True,
)
translation = TranslationField(PageTranslation, type_name="page")
attributes = NonNullList(
SelectedAttribute,
required=True,
description="List of attributes assigned to this product.",
)
class Meta:
description = (
"A static page that can be manually added by a shop operator through the "
"dashboard."
)
interfaces = [graphene.relay.Node, ObjectWithMetadata]
model = models.Page
@staticmethod
def resolve_publication_date(root: models.Page, _info: ResolveInfo):
return root.published_at
@staticmethod
def resolve_created(root: models.Page, _info: ResolveInfo):
return root.created_at
@staticmethod
def resolve_page_type(root: models.Page, info: ResolveInfo):
return PageTypeByIdLoader(info.context).load(root.page_type_id)
@staticmethod
def resolve_content_json(root: models.Page, _info: ResolveInfo):
content = root.content
return content if content is not None else {}
@staticmethod
def resolve_attributes(root: models.Page, info: ResolveInfo):
return SelectedAttributesByPageIdLoader(info.context).load(root.id)
class PageCountableConnection(CountableConnection):
class Meta:
doc_category = DOC_CATEGORY_PAGES
node = Page | null |
profile gen | import os, sys
import time
import random
import string
from .astf_general_test import CASTFGeneral_Test, CTRexScenario
from nose.tools import assert_raises
from trex.astf.api import *
from trex.stl.trex_stl_packet_builder_scapy import ip2int, int2ip
class ASTFResilience_Test(CASTFGeneral_Test):
"""Checking stability of ASTF in non-usual conditions """
def setUp(self):
CASTFGeneral_Test.setUp(self)
self.weak = self.is_VM
setup = CTRexScenario.setup_name
if 'no_resilience' in CTRexScenario.modes:
self.skip('not enough memory for this test')
if setup in ['trex12']:
self.weak = True
self.low_memory = self.weak
if setup in ['trex41']:
self.low_memory = True # trex-41 uses the memory for the driver and crash
def ip_gen(self, client_base, server_base, client_ips, server_ips):
assert client_ips>0
assert server_ips>0
ip_gen_c = ASTFIPGenDist(ip_range = [client_base, int2ip(ip2int(client_base) + client_ips - 1)])
ip_gen_s = ASTFIPGenDist(ip_range = [server_base, int2ip(ip2int(server_base) + server_ips - 1)])
return ASTFIPGen(dist_client = ip_gen_c,
dist_server = ip_gen_s)
def progs_gen(self, msg_len = 16):
msg = 'x' * msg_len
prog_c = ASTFProgram(side = 'c')
prog_c.send(msg)
prog_c.recv(len(msg))
prog_s = ASTFProgram(side = 's')
prog_s.recv(len(msg))
#prog_s.delay(15000000)
prog_s.send(msg)
return prog_c, prog_s
def METHOD_NAME(self, client_ips, server_ips, templates):
ip_gen = self.ip_gen('16.0.0.1', '48.0.0.1', client_ips, server_ips)
prog_c, prog_s = self.progs_gen()
templates_arr = []
for i in range(templates):
temp_c = ASTFTCPClientTemplate(program = prog_c, ip_gen = ip_gen, cps = i + 1)
temp_s = ASTFTCPServerTemplate(program = prog_s, assoc = ASTFAssociationRule(port = 80 + i))
template = ASTFTemplate(client_template = temp_c, server_template = temp_s)
templates_arr.append(template)
return ASTFProfile(default_ip_gen = ip_gen, templates = templates_arr)
def test_astf_params(self):
print('')
for client_ips in (1<<8, 1<<16):
for server_ips in (1<<8, 1<<16):
for templates in (1, 1<<8, 1<<12):
if self.weak and templates > 1<<8:
continue
if self.weak:
if (client_ips > (1<<8)) and (server_ips >(1<<8)) :
continue;
params = {
'client_ips': client_ips,
'server_ips': server_ips,
'templates': templates,
}
print('Creating profile with params: %s' % params)
profile = self.METHOD_NAME(**params)
profile_str = profile.to_json_str()
print('Profile size: %s' % len(profile_str))
start_time = time.time()
self.astf_trex.load_profile(profile)
print('Load took: %g' % round(time.time() - start_time, 3))
start_time = time.time()
self.astf_trex.start(duration = 1, nc = True)
print('Start took: %g' % round(time.time() - start_time, 3))
self.astf_trex.stop()
def randomString(self, stringLength=10):
"""Generate a random string of fixed length """
letters = string.ascii_lowercase
return ''.join(random.choice(letters) for i in range(stringLength))
def test_astf_params_dynamic_profile(self):
print('')
for client_ips in (1<<8, 1<<16):
for server_ips in (1<<8, 1<<16):
for templates in (1, 1<<8, 1<<12):
if self.weak:
if ( (templates > 1<<8) or
(server_ips > 1<<8 ) or
(client_ips > 1<<8 ) ):
continue
params = {
'client_ips': client_ips,
'server_ips': server_ips,
'templates': templates,
}
print('Creating profile with params: %s' % params)
profile = self.METHOD_NAME(**params)
profile_str = profile.to_json_str()
print('Profile size: %s' % len(profile_str))
start_time = time.time()
print('Creating random name for the dynamic profile')
random_profile = self.randomString()
print('Dynamic profile name : %s' % str(random_profile))
self.astf_trex.load_profile(profile, pid_input=str(random_profile))
print('Load took: %g' % round(time.time() - start_time, 3))
start_time = time.time()
self.astf_trex.start(duration = 1, nc = True, pid_input=str(random_profile))
print('Start took: %g' % round(time.time() - start_time, 3))
self.astf_trex.stop(pid_input=str(random_profile))
self.astf_trex.reset()
def test_double_start_stop(self):
print('')
c = self.astf_trex
c.load_profile(os.path.join(CTRexScenario.scripts_path, 'astf', 'udp1.py'))
c.start(duration = 20)
with assert_raises(TRexError):
c.start()
c.stop()
c.stop()
def test_double_start_stop_dynamic_profile(self):
print('')
c = self.astf_trex
random_profile_1 = self.randomString()
print('Dynamic profile(1) : %s' % str(random_profile_1))
c.load_profile(os.path.join(CTRexScenario.scripts_path, 'astf', 'udp1.py'), pid_input=str(random_profile_1))
c.start(duration = 20, pid_input=str(random_profile_1))
with assert_raises(TRexError):
random_profile_2 = self.randomString()
print('Dynamic profile(2) : %s' % str(random_profile_2))
c.load_profile(os.path.join(CTRexScenario.scripts_path, 'astf', 'udp1.py'), pid_input=str(random_profile_2))
c.start(pid_input=str(random_profile_2))
c.stop(pid_input=str(random_profile_1))
c.stop(pid_input=str(random_profile_2))
def test_stress_start_stop(self):
print('')
c = self.astf_trex
c.load_profile(os.path.join(CTRexScenario.scripts_path, 'astf', 'udp1.py'))
for _ in range(99):
c.start()
c.stop()
def test_stress_start_stop_dynamic_profile(self):
print('')
c = self.astf_trex
profiles =1000
if self.low_memory:
profiles = 100
for n in range(profiles):
profile_n = self.randomString()
port_n = 9000 + n
tunables = {'port': port_n}
print('Dynamic profile added : %s' % str(profile_n))
print('Port added : %s' % str(port_n))
c.load_profile(os.path.join(CTRexScenario.scripts_path, 'astf', 'http_simple_port_tunable.py'),
tunables=tunables,
pid_input=str(profile_n))
c.start(duration = 20, pid_input=str(profile_n))
c.wait_on_traffic()
c.stop(pid_input = "*")
c.reset()
| null |
test dict | # Copyright (c) MONAI Consortium
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from __future__ import annotations
import unittest
import numpy as np
import torch
from monai.data import MetaTensor
from monai.transforms import EnsureTyped
from tests.utils import assert_allclose
class TestEnsureTyped(unittest.TestCase):
def test_array_input(self):
test_datas = [np.array([[1, 2], [3, 4]]), torch.as_tensor([[1, 2], [3, 4]])]
if torch.cuda.is_available():
test_datas.append(test_datas[-1].cuda())
for test_data in test_datas:
for dtype in ("tensor", "NUMPY"):
result = EnsureTyped(
keys="data", data_type=dtype, dtype=np.float32 if dtype == "NUMPY" else None, device="cpu"
)({"data": test_data})["data"]
if dtype == "NUMPY":
self.assertTrue(result.dtype == np.float32)
self.assertTrue(isinstance(result, torch.Tensor if dtype == "tensor" else np.ndarray))
assert_allclose(result, test_data, type_test=False)
self.assertTupleEqual(result.shape, (2, 2))
def test_single_input(self):
test_datas = [5, 5.0, False, np.asarray(5), torch.tensor(5)]
if torch.cuda.is_available():
test_datas.append(test_datas[-1].cuda())
for test_data in test_datas:
for dtype in ("tensor", "numpy"):
result = EnsureTyped(keys="data", data_type=dtype)({"data": test_data})["data"]
self.assertTrue(isinstance(result, torch.Tensor if dtype == "tensor" else np.ndarray))
if isinstance(test_data, bool):
self.assertFalse(result)
else:
assert_allclose(result, test_data, type_test=False)
self.assertEqual(result.ndim, 0)
def test_string(self):
for dtype in ("tensor", "numpy"):
# string input
result = EnsureTyped(keys="data", data_type=dtype)({"data": "test_string"})["data"]
self.assertTrue(isinstance(result, str))
self.assertEqual(result, "test_string")
# numpy array of string
result = EnsureTyped(keys="data", data_type=dtype)({"data": np.array(["test_string"])})["data"]
self.assertTrue(isinstance(result, np.ndarray))
self.assertEqual(result[0], "test_string")
def test_list_tuple(self):
for dtype in ("tensor", "numpy"):
result = EnsureTyped(keys="data", data_type=dtype, wrap_sequence=False, track_meta=True)(
{"data": [[1, 2], [3, 4]]}
)["data"]
self.assertTrue(isinstance(result, list))
self.assertTrue(isinstance(result[0][1], MetaTensor if dtype == "tensor" else np.ndarray))
assert_allclose(result[1][0], torch.as_tensor(3), type_test=False)
# tuple of numpy arrays
result = EnsureTyped(keys="data", data_type=dtype, wrap_sequence=False)(
{"data": (np.array([1, 2]), np.array([3, 4]))}
)["data"]
self.assertTrue(isinstance(result, tuple))
self.assertTrue(isinstance(result[0], torch.Tensor if dtype == "tensor" else np.ndarray))
assert_allclose(result[1], torch.as_tensor([3, 4]), type_test=False)
def METHOD_NAME(self):
# simulate complicated input data
test_data = {
"img": np.array([1.0, 2.0], dtype=np.float32),
"meta": {"dims": 3, "size": np.array([1, 2, 3]), "path": "temp/test"},
"extra": None,
}
for dtype in ("tensor", "numpy"):
result = EnsureTyped(keys="data", data_type=dtype, device="cpu")({"data": test_data})["data"]
self.assertTrue(isinstance(result, dict))
self.assertTrue(isinstance(result["img"], torch.Tensor if dtype == "tensor" else np.ndarray))
assert_allclose(result["img"], torch.as_tensor([1.0, 2.0]), type_test=False)
self.assertTrue(isinstance(result["meta"]["size"], torch.Tensor if dtype == "tensor" else np.ndarray))
assert_allclose(result["meta"]["size"], torch.as_tensor([1, 2, 3]), type_test=False)
self.assertEqual(result["meta"]["path"], "temp/test")
self.assertEqual(result["extra"], None)
if __name__ == "__main__":
unittest.main() | null |
write file | from __future__ import annotations
import logging
import os
import pathlib
import shutil
import struct
import tempfile
import zipfile
from typing import cast
from comicapi.archivers import Archiver
logger = logging.getLogger(__name__)
class ZipArchiver(Archiver):
"""ZIP implementation"""
def __init__(self) -> None:
super().__init__()
def get_comment(self) -> str:
with zipfile.ZipFile(self.path, "r") as zf:
comment = zf.comment.decode("utf-8")
return comment
def set_comment(self, comment: str) -> bool:
with zipfile.ZipFile(self.path, mode="a") as zf:
zf.comment = bytes(comment, "utf-8")
return True
def read_file(self, archive_file: str) -> bytes:
with zipfile.ZipFile(self.path, mode="r") as zf:
try:
data = zf.read(archive_file)
except (zipfile.BadZipfile, OSError) as e:
logger.error("Error reading zip archive [%s]: %s :: %s", e, self.path, archive_file)
raise
return data
def remove_file(self, archive_file: str) -> bool:
return self.rebuild([archive_file])
def METHOD_NAME(self, archive_file: str, data: bytes) -> bool:
# At the moment, no other option but to rebuild the whole
# zip archive w/o the indicated file. Very sucky, but maybe
# another solution can be found
files = self.get_filename_list()
if archive_file in files:
if not self.rebuild([archive_file]):
return False
try:
# now just add the archive file as a new one
with zipfile.ZipFile(self.path, mode="a", allowZip64=True, compression=zipfile.ZIP_DEFLATED) as zf:
zf.writestr(archive_file, data)
return True
except (zipfile.BadZipfile, OSError) as e:
logger.error("Error writing zip archive [%s]: %s :: %s", e, self.path, archive_file)
return False
def get_filename_list(self) -> list[str]:
try:
with zipfile.ZipFile(self.path, mode="r") as zf:
namelist = [file.filename for file in zf.infolist() if not file.is_dir()]
return namelist
except (zipfile.BadZipfile, OSError) as e:
logger.error("Error listing files in zip archive [%s]: %s", e, self.path)
return []
def rebuild(self, exclude_list: list[str]) -> bool:
"""Zip helper func
This recompresses the zip archive, without the files in the exclude_list
"""
try:
with zipfile.ZipFile(
tempfile.NamedTemporaryFile(dir=os.path.dirname(self.path), delete=False), "w", allowZip64=True
) as zout:
with zipfile.ZipFile(self.path, mode="r") as zin:
for item in zin.infolist():
buffer = zin.read(item.filename)
if item.filename not in exclude_list:
zout.writestr(item, buffer)
# preserve the old comment
zout.comment = zin.comment
# replace with the new file
self.path.unlink(missing_ok=True)
zout.close() # Required on windows
shutil.move(cast(str, zout.filename), self.path)
except (zipfile.BadZipfile, OSError) as e:
logger.error("Error rebuilding zip file [%s]: %s", e, self.path)
return False
return True
def copy_from_archive(self, other_archive: Archiver) -> bool:
"""Replace the current zip with one copied from another archive"""
try:
with zipfile.ZipFile(self.path, mode="w", allowZip64=True) as zout:
for filename in other_archive.get_filename_list():
data = other_archive.read_file(filename)
if data is not None:
zout.writestr(filename, data)
# preserve the old comment
comment = other_archive.get_comment()
if comment is not None:
if not self.write_zip_comment(self.path, comment):
return False
except Exception as e:
logger.error("Error while copying to zip archive [%s]: from %s to %s", e, other_archive.path, self.path)
return False
else:
return True
def is_writable(self) -> bool:
return True
def extension(self) -> str:
return ".cbz"
def name(self) -> str:
return "ZIP"
@classmethod
def is_valid(cls, path: pathlib.Path) -> bool:
return zipfile.is_zipfile(path)
def write_zip_comment(self, filename: pathlib.Path | str, comment: str) -> bool:
"""
This is a custom function for writing a comment to a zip file,
since the built-in one doesn't seem to work on Windows and Mac OS/X
Fortunately, the zip comment is at the end of the file, and it's
easy to manipulate. See this website for more info:
see: http://en.wikipedia.org/wiki/Zip_(file_format)#Structure
"""
# get file size
statinfo = os.stat(filename)
file_length = statinfo.st_size
try:
with open(filename, mode="r+b") as file:
# the starting position, relative to EOF
pos = -4
found = False
# walk backwards to find the "End of Central Directory" record
while (not found) and (-pos != file_length):
# seek, relative to EOF
file.seek(pos, 2)
value = file.read(4)
# look for the end of central directory signature
if bytearray(value) == bytearray([0x50, 0x4B, 0x05, 0x06]):
found = True
else:
# not found, step back another byte
pos = pos - 1
if found:
# now skip forward 20 bytes to the comment length word
pos += 20
file.seek(pos, 2)
# Pack the length of the comment string
fmt = "H" # one 2-byte integer
comment_length = struct.pack(fmt, len(comment)) # pack integer in a binary string
# write out the length
file.write(comment_length)
file.seek(pos + 2, 2)
# write out the comment itself
file.write(comment.encode("utf-8"))
file.truncate()
else:
raise Exception("Could not find the End of Central Directory record!")
except Exception as e:
logger.error("Error writing comment to zip archive [%s]: %s", e, self.path)
return False
else:
return True | null |
test get usage for trial services | from datetime import date, datetime
import pytest
from freezegun import freeze_time
from app.errors import InvalidRequest
from app.models import EMAIL_TYPE, SMS_TYPE
from app.platform_stats.rest import validate_date_range_is_within_a_financial_year
from tests.app.db import (
create_ft_notification_status,
create_notification,
create_service,
create_template,
save_notification,
set_up_usage_data,
)
@freeze_time("2018-06-01")
def test_get_platform_stats_uses_todays_date_if_no_start_or_end_date_is_provided(admin_request, mocker):
today = datetime.now().date()
dao_mock = mocker.patch("app.platform_stats.rest.fetch_notification_status_totals_for_all_services")
mocker.patch("app.service.rest.statistics.format_statistics")
admin_request.get("platform_stats.get_platform_stats")
dao_mock.assert_called_once_with(start_date=today, end_date=today)
def test_get_platform_stats_can_filter_by_date(admin_request, mocker):
start_date = date(2017, 1, 1)
end_date = date(2018, 1, 1)
dao_mock = mocker.patch("app.platform_stats.rest.fetch_notification_status_totals_for_all_services")
mocker.patch("app.service.rest.statistics.format_statistics")
admin_request.get("platform_stats.get_platform_stats", start_date=start_date, end_date=end_date)
dao_mock.assert_called_once_with(start_date=start_date, end_date=end_date)
def test_get_platform_stats_validates_the_date(admin_request):
start_date = "1234-56-78"
response = admin_request.get("platform_stats.get_platform_stats", start_date=start_date, _expected_status=400)
assert response["errors"][0]["message"] == "start_date time data {} does not match format %Y-%m-%d".format(start_date)
@freeze_time("2018-10-31 14:00")
def test_get_platform_stats_with_real_query(admin_request, notify_db_session):
service_1 = create_service(service_name="service_1")
sms_template = create_template(service=service_1, template_type=SMS_TYPE)
email_template = create_template(service=service_1, template_type=EMAIL_TYPE)
create_ft_notification_status(date(2018, 10, 29), "sms", service_1, count=10)
create_ft_notification_status(date(2018, 10, 29), "email", service_1, count=3)
save_notification(create_notification(sms_template, created_at=datetime(2018, 10, 31, 11, 0, 0), key_type="test"))
save_notification(create_notification(sms_template, created_at=datetime(2018, 10, 31, 12, 0, 0), status="delivered"))
save_notification(create_notification(email_template, created_at=datetime(2018, 10, 31, 13, 0, 0), status="delivered"))
response = admin_request.get(
"platform_stats.get_platform_stats",
start_date=date(2018, 10, 29),
)
assert response == {
"email": {
"failures": {
"virus-scan-failed": 0,
"temporary-failure": 0,
"permanent-failure": 0,
"technical-failure": 0,
},
"total": 4,
"test-key": 0,
},
"letter": {
"failures": {
"virus-scan-failed": 0,
"temporary-failure": 0,
"permanent-failure": 0,
"technical-failure": 0,
},
"total": 0,
"test-key": 0,
},
"sms": {
"failures": {
"virus-scan-failed": 0,
"temporary-failure": 0,
"permanent-failure": 0,
"technical-failure": 0,
},
"total": 11,
"test-key": 1,
},
}
@pytest.mark.parametrize(
"start_date, end_date",
[
("2019-04-01", "2019-06-30"),
("2019-08-01", "2019-09-30"),
("2019-01-01", "2019-03-31"),
("2019-12-01", "2020-02-28"),
],
)
def test_validate_date_range_is_within_a_financial_year(start_date, end_date):
validate_date_range_is_within_a_financial_year(start_date, end_date)
@pytest.mark.parametrize(
"start_date, end_date",
[
("2019-04-01", "2020-06-30"),
("2019-01-01", "2019-04-30"),
("2019-12-01", "2020-04-30"),
("2019-03-31", "2019-04-01"),
],
)
def test_validate_date_range_is_within_a_financial_year_raises(start_date, end_date):
with pytest.raises(expected_exception=InvalidRequest) as e:
validate_date_range_is_within_a_financial_year(start_date, end_date)
assert e.value.message == "Date must be in a single financial year."
assert e.value.status_code == 400
def test_validate_date_is_within_a_financial_year_raises_validation_error():
start_date = "2019-08-01"
end_date = "2019-06-01"
with pytest.raises(expected_exception=InvalidRequest) as e:
validate_date_range_is_within_a_financial_year(start_date, end_date)
assert e.value.message == "Start date must be before end date"
assert e.value.status_code == 400
@pytest.mark.parametrize("start_date, end_date", [("22-01-2019", "2019-08-01"), ("2019-07-01", "not-date")])
def test_validate_date_is_within_a_financial_year_when_input_is_not_a_date(start_date, end_date):
with pytest.raises(expected_exception=InvalidRequest) as e:
validate_date_range_is_within_a_financial_year(start_date, end_date)
assert e.value.message == "Input must be a date in the format: YYYY-MM-DD"
assert e.value.status_code == 400
def test_get_usage_for_all_services(notify_db_session, admin_request):
org, org_2, service, service_2, service_3, service_sms_only = set_up_usage_data(datetime(2019, 5, 1))
response = admin_request.get(
"platform_stats.get_usage_for_all_services",
start_date="2019-05-01",
end_date="2019-06-30",
)
assert len(response) == 4
assert response[0]["organisation_id"] == str(org.id)
assert response[0]["service_id"] == str(service.id)
assert response[0]["sms_cost"] == 0
assert response[0]["sms_fragments"] == 0
assert response[0]["letter_cost"] == 3.40
assert response[0]["letter_breakdown"] == "6 second class letters at 45p\n2 first class letters at 35p\n"
assert response[1]["organisation_id"] == str(org_2.id)
assert response[1]["service_id"] == str(service_2.id)
assert response[1]["sms_cost"] == 0
assert response[1]["sms_fragments"] == 0
assert response[1]["letter_cost"] == 14
assert response[1]["letter_breakdown"] == "20 second class letters at 65p\n2 first class letters at 50p\n"
assert response[2]["organisation_id"] == ""
assert response[2]["service_id"] == str(service_sms_only.id)
assert response[2]["sms_cost"] == 0.33
assert response[2]["sms_fragments"] == 3
assert response[2]["letter_cost"] == 0
assert response[2]["letter_breakdown"] == ""
assert response[3]["organisation_id"] == ""
assert response[3]["service_id"] == str(service_3.id)
assert response[3]["sms_cost"] == 0
assert response[3]["sms_fragments"] == 0
assert response[3]["letter_cost"] == 8.25
assert response[3]["letter_breakdown"] == "15 second class letters at 55p\n"
def METHOD_NAME(mocker, admin_request):
# The DAO method is already covered by tests
mock = mocker.patch(
"app.platform_stats.rest.fetch_notification_stats_for_trial_services",
return_value=[],
)
response = admin_request.get("platform_stats.get_usage_for_trial_services")
assert len(response) == 0
mock.assert_called_once()
def test_get_send_methods_stats_by_service(mocker, admin_request):
# The DAO method is already covered by tests
mock = mocker.patch("app.platform_stats.rest.send_method_stats_by_service", return_value=[])
response = admin_request.get(
"platform_stats.get_send_methods_stats_by_service",
start_date="2020-12-01",
end_date="2020-12-07",
)
assert len(response) == 0
mock.assert_called_once_with(
date(2020, 12, 1),
date(2020, 12, 7),
) | null |
ok | # Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"). You
# may not use this file except in compliance with the License. A copy of
# the License is located at
#
# http://aws.amazon.com/apache2.0/
#
# or in the "license" file accompanying this file. This file is
# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
# ANY KIND, either express or implied. See the License for the specific
# language governing permissions and limitations under the License.
import os
import shutil
import sys
from tempfile import mkdtemp, mkstemp
import mock
import unittest
from cement.core import backend, foundation
from cement.utils.misc import rando
from ebcli.core import ebcore, ebrun, fileoperations
from ebcli.lib import aws
from tests.unit.integration import mockservice
ebrun.fix_path()
class TestApp(foundation.CementApp):
"""
Basic CementApp for generic testing.
"""
class Meta:
label = "app-%s" % rando()[:12]
config_files = []
argv = []
base_controller = None
arguments = []
exit_on_close = False
class CementTestCase(unittest.TestCase):
"""
A copy of cement.utils.test.CementTestCase inorder to avoid external, unnecessary dependency
of `nose`
"""
app_class = TestApp
def __init__(self, *args, **kw):
super(CementTestCase, self).__init__(*args, **kw)
def setUp(self):
self.app = self.make_app()
_, self.tmp_file = mkstemp()
self.tmp_dir = mkdtemp()
def make_app(self, *args, **kw):
self.reset_backend()
return self.app_class(*args, **kw)
def reset_backend(self):
for _handler in backend.__handlers__.copy():
del backend.__handlers__[_handler]
for _hook in backend.__hooks__.copy():
del backend.__hooks__[_hook]
def METHOD_NAME(self, expr, msg=None):
if not expr:
raise AssertionError(msg)
def eq(self, a, b, msg=None):
if not a == b:
raise AssertionError(msg or "%r != %r" % (a, b))
class BaseIntegrationTest(CementTestCase):
app_class = ebcore.EB
def setUp(self):
super(BaseIntegrationTest, self).setUp()
aws._flush()
aws.set_region('us-east-1')
self.reset_backend()
self.patcher_input = mock.patch('ebcli.core.io.get_input')
self.patcher_output = mock.patch('ebcli.core.io.echo')
self.patcher_warning = mock.patch('ebcli.core.io.log_warning')
self.mock_input = self.patcher_input.start()
self.mock_output = self.patcher_output.start()
self.mock_warning = self.patcher_warning.start()
self.patcher_endpoint = mock.patch('botocore.endpoint.Endpoint')
self.mock_endpoint = self.patcher_endpoint.start()
instance = self.mock_endpoint.return_value
instance.make_request = mockservice.handle_response
instance.host = 'http://someurl.test/something'
if not os.path.exists('testDir/'):
os.makedirs('testDir/')
os.chdir('testDir')
if not os.path.exists(fileoperations.beanstalk_directory):
os.makedirs(fileoperations.beanstalk_directory)
fileoperations.default_section = 'ebcli_test_default'
if not os.path.exists('home'):
os.makedirs('home')
fileoperations.aws_config_folder = 'home' + os.path.sep
fileoperations.aws_config_location \
= fileoperations.aws_config_folder + 'config'
def run_command(self, *args):
self.app = ebcore.EB(argv=list(args))
self.app.setup()
self.app.run()
self.app.close()
def tearDown(self):
self.patcher_input.stop()
self.patcher_output.stop()
self.patcher_endpoint.stop()
self.patcher_warning.stop()
os.chdir(os.path.pardir)
if os.path.exists('testDir'):
if sys.platform.startswith('win'):
os.system('rmdir /S /Q testDir')
else:
shutil.rmtree('testDir')
mockservice.reset()
class BaseOperationsTest(CementTestCase):
module_name = 'base'
def setUp(self):
super(BaseOperationsTest, self).setUp()
self.reset_backend()
self.patcher_input = mock.patch('ebcli.core.io.get_input')
self.patcher_eb = mock.patch('ebcli.operations.' + self.module_name + '.elasticbeanstalk')
self.patcher_output = mock.patch('ebcli.core.io.echo')
self.patcher_file = mock.patch('ebcli.operations.' + self.module_name + '.fileoperations')
self.mock_input = self.patcher_input.start()
self.mock_elasticbeanstalk = self.patcher_eb.start()
self.mock_output = self.patcher_output.start()
def tearDown(self):
self.patcher_eb.stop()
self.patcher_input.stop()
self.patcher_output.stop() | null |
test image | from recipe_scrapers.saveur import Saveur
from tests import ScraperTest
class TestSaveurScraper(ScraperTest):
scraper_class = Saveur
def test_host(self):
self.assertEqual("saveur.com", self.harvester_class.host())
def test_author(self):
self.assertEqual(None, self.harvester_class.author())
def test_title(self):
self.assertEqual(
"Baked Saffron Yogurt Rice with Chicken (Tahcheen-e Morgh)",
self.harvester_class.title(),
)
def test_total_time(self):
self.assertEqual(285, self.harvester_class.total_time())
def test_yields(self):
self.assertEqual("8 servings", self.harvester_class.yields())
def METHOD_NAME(self):
self.assertEqual(
"https://www.saveur.com/uploads/2019/03/07/httpsrecipes.saveur.comuploadsimagesabb46018d2ab45231975b33416325e96ca3afeab.jpeg?auto=webp",
self.harvester_class.image(),
)
def test_ingredients(self):
self.assertEqual(
[
"3 cups extra-long white basmati rice (1 lb. 4 oz.)",
"1⁄4 cup plus 2 Tbsp. kosher salt, divided, plus more as needed",
"1⁄4 tsp. saffron, lightly ground between your fingers, plus more for sprinkling",
"3 tbsp. extra-virgin olive oil",
"1 large yellow onion, halved and sliced ¼-inch-thick",
"4 medium garlic cloves, finely chopped (1 Tbsp.)",
"10 boneless, skinless chicken thighs (2 lb. 3 oz.), halved",
"1⁄4 tsp. freshly ground black pepper",
"1⁄4 cup fresh lemon juice",
"1 cup plain whole-milk Greek yogurt",
"1 large egg",
"5 tbsp. unsalted butter, divided",
"1 tbsp. unsalted butter",
"1⁄2 cup dried barberries, picked, soaked in cold water for 15 minutes, then drained (or substitute finely chopped dried cranberries)",
"1 tsp. sugar (omit if using cranberries)",
"1⁄3 cup raw pistachios, coarsely chopped",
],
self.harvester_class.ingredients(),
)
def test_instructions(self):
self.assertEqual(
"""instruction 0\nParboil the rice by following steps 1 and 2 of the Steamed Saffron Rice with Tahdig recipe (p. 36). Set the rice aside to drain as you prepare the dish’s other components.\ninstruction 1\nIn a small bowl, add the saffron and 2 tablespoons hot water. (Be sure to pre-soak your barberries now, too. See toppings list.)\ninstruction 2\nIn a large, deep skillet, add the olive oil and set over medium heat. Once hot, add the onion and ¼ teaspoon salt, and cook, stirring occasionally, until soft and golden, about 20 minutes. Add the garlic and cook until fragrant, 2 minutes. Add the chicken thighs, 2¾ teaspoons salt, and the pepper. Cook, turning the chicken as needed, until the pieces begin to brown slightly, about 6 minutes total. Add the lemon juice and saffron water, and turn the chicken to coat well. Lower the heat to medium-low, partially cover the pot, and let simmer, stirring occasionally, until the chicken is tender and just cooked through, 20–25 minutes. Remove from the heat and let cool slightly. Cut the chicken into ½-inch-long strips, return them to the pan, and toss to coat in the onions and pan juices. Remove the pan from the heat and set aside while you finish the rice.\ninstruction 3\nPreheat the oven to 400°F and set a rack in the lowest position.\ninstruction 4\nIn a large bowl, combine the yogurt, egg, and a tiny pinch of ground saffron. Add half of the rice and stir well to combine. Set the mixture aside.\ninstruction 5\nIn a 9-by-13-inch glass baking dish (preferably clear), add 3 tablespoons of butter; set the dish in the oven to melt the butter, about 3 minutes. Remove the dish, swirling the melted butter all over the bottom and sides. Spread the yogurt-rice mixture evenly along the bottom of the dish, packing it down firmly. Add the chicken pieces evenly over the rice, then sprinkle the remaining plain rice over the chicken in an even layer. Drizzle with 2 tablespoons of chicken juices and dot with the remaining 2 tablespoons of butter. Cover tightly with aluminum foil, transfer to the oven, and bake until the bottom of the rice is crispy and golden, 80–90 minutes. (If using a clear glass dish, carefully lift it to check the bottom.) Remove the baking dish and let the rice rest at room temperature for 5 minutes.\ninstruction 6\nWhile the tahcheen rests, prepare the topping: In a small pot, melt the butter over medium heat. Stir in the drained barberries, pistachios, and sugar (if using), and cook until the berries are just plumped, about 2 minutes. Remove from the heat.\ninstruction 7\nRun a knife along the sides of the baking dish to help release the rice. Place a large rectangular serving platter, baking tray, or cutting board over the tahcheen, take a deep breath, and quickly and confidently flip the baking dish over to unmold. You should hear a swish when the tahdig releases. Remove the baking dish and sprinkle the crispy surface of the rice with the barberry topping. Cut into pieces and serve.""",
self.harvester_class.instructions(),
) | null |
get metadata | import json
import sys
if sys.version_info >= (3,0):
import urllib.parse
dku_quote_fn = urllib.parse.quote
else:
import urllib
dku_quote_fn = urllib.quote
class DSSObjectDiscussions(object):
"""
A handle to manage discussions on a DSS object.
.. important::
Do not create this class directly, instead use :meth:`dataikuapi.DSSClient.get_object_discussions` on any commentable DSS object.
"""
def __init__(self, client, project_key, object_type, object_id):
self.client = client
self.project_key = project_key
self.object_type = object_type
self.object_id = object_id
# encode in UTF-8 if its python2 and unicode
if sys.version_info < (3,0) and isinstance(self.object_id, unicode):
self.object_id = self.object_id.encode('utf-8')
def list_discussions(self):
"""
Gets the list of discussions on the object.
:returns: list of discussions on the object
:rtype: list of :class:`dataikuapi.dss.discussion.DSSDiscussion`
"""
data = self.client._perform_json("GET", "/projects/%s/discussions/%s/%s/" % (self.project_key, self.object_type, self.object_id))
return [DSSDiscussion(self.client, self.project_key, self.object_type, self.object_id, discu_data['id'], discu_data, False) for discu_data in data]
def create_discussion(self, topic, message):
"""
Creates a new discussion with one message.
:param str topic: the discussion topic
:param str message: the markdown formatted first message
:returns: the newly created discussion
:rtype: :class:`dataikuapi.dss.discussion.DSSDiscussion`
"""
creation_data = {
"topic" : topic,
"reply" : message
}
discu_data = self.client._perform_json("POST", "/projects/%s/discussions/%s/%s/" % (self.project_key, self.object_type, self.object_id), body=creation_data)
return DSSDiscussion(self.client, self.project_key, self.object_type, self.object_id, discu_data['id'], discu_data, True)
def get_discussion(self, discussion_id):
"""
Gets a specific discussion.
:param str discussion_id: the discussion ID
:returns: the discussion
:rtype: :class:`dataikuapi.dss.discussion.DSSDiscussion`
"""
discu_data = self.client._perform_json("GET", "/projects/%s/discussions/%s/%s/%s" % (self.project_key, self.object_type, self.object_id, discussion_id))
return DSSDiscussion(self.client, self.project_key, self.object_type, self.object_id, discussion_id, discu_data, True)
class DSSDiscussion(object):
"""
A handle to interact with a discussion.
.. important::
Do not call directly, use :meth:`dataikuapi.dss.discussion.DSSObjectDiscussions.get_discussion`,
:meth:`dataikuapi.dss.discussion.DSSObjectDiscussions.create_discussion`
or :meth:`dataikuapi.dss.discussion.DSSObjectDiscussions.list_discussions`.
"""
def __init__(self, client, project_key, object_type, object_id, discussion_id, discussion_data, discussion_data_has_replies):
"""
:param :class:`dataikuapi.dssclient.DSSClient` client: an api client to connect to the DSS backend
:param str project_key: identifier of the project to access
:param str object_type: DSS object type
:param str object_id: DSS object ID
:param str discussion_id: identifier of the discussion
:param dict discussion_data: the discussion data
:param bool discussion_data_has_replies: a flag that indicates if the replies has been loaded
"""
self.client = client
self.project_key = project_key
self.object_type = object_type
self.object_id = object_id
self.discussion_id = discussion_id
self.discussion_data = discussion_data
self.discussion_data_has_replies = discussion_data_has_replies
def _get_with_replies(self):
"""
Reloads the discussion data from the backend including the replies.
"""
self.discussion_data = self.client._perform_json("GET", "/projects/%s/discussions/%s/%s/%s" % (self.project_key, self.object_type, self.object_id, self.discussion_id))
self.discussion_data_has_replies = True
def METHOD_NAME(self):
"""
Gets the discussion metadata.
:returns: the discussion metadata
:rtype: dict
"""
metadata = dict(self.discussion_data)
if "replies" in metadata:
del metadata["replies"]
return metadata
def set_metadata(self, discussion_metadata):
"""
Updates the discussion metadata.
:param dict discussion_metadata: the discussion metadata
"""
if not self.discussion_data_has_replies:
self._get_with_replies()
edited_metadata = dict(discussion_metadata)
edited_metadata["replies"] = self.discussion_data["replies"]
self.discussion_data = self.client._perform_json("PUT", "/projects/%s/discussions/%s/%s/%s" % (self.project_key, self.object_type, self.object_id, self.discussion_id), body=edited_metadata)
self.discussion_data_has_replies = True
def get_replies(self):
"""
Gets the list of replies in this discussion.
:returns: a list of replies
:rtype: list of :class:`dataikuapi.dss.discussion.DSSDiscussionReply`
"""
if not self.discussion_data_has_replies:
self._get_with_replies()
return [DSSDiscussionReply(reply_data) for reply_data in self.discussion_data["replies"]]
def add_reply(self, text):
"""
Adds a reply to a discussion.
:param str text: the markdown formatted text to reply
"""
reply_data = {
"reply": text
}
self.discussion_data = self.client._perform_json("POST", "/projects/%s/discussions/%s/%s/%s/replies/" % (self.project_key, self.object_type, self.object_id, self.discussion_id), body=reply_data)
self.discussion_data_has_replies = True
class DSSDiscussionReply(object):
"""
A read-only handle to access a discussion reply.
.. important::
Do not create this class directly, use :meth:`dataikuapi.dss.discussion.DSSDiscussion.get_replies`
"""
def __init__(self, reply_data):
self.reply_data = reply_data
def get_raw_data(self):
"""
Gets the reply raw data.
:returns: the reply data
:rtype: dict
"""
return self.reply_data
def get_text(self):
"""
Gets the reply text.
:returns: the reply text
:rtype: str
"""
return self.reply_data["text"]
def get_author(self):
"""
Gets the reply author.
:returns: the author ID
:rtype: str
"""
return self.reply_data["author"]
def get_timestamp(self):
"""
Gets the reply timestamp.
:returns: the reply timestamp
:rtype: long
"""
return self.reply_data["time"]
def get_edited_timestamp(self):
"""
Gets the last edition timestamp.
:returns: the last edition timestamp
:rtype: long
"""
return self.reply_data["editedOn"] | null |
test turbomole num hessian | import numpy as np
import pytest
import qcelemental
from qcelemental.testing import compare_values
import qcengine as qcng
from qcengine.programs.turbomole.harvester import parse_hessian
from qcengine.testing import using
@pytest.fixture
def h2o():
mol = qcelemental.models.Molecule.from_data(
"""
O 0.000000000000 0.000000000000 -0.068516245955
H 0.000000000000 -0.790689888800 0.543701278274
H 0.000000000000 0.790689888800 0.543701278274
"""
)
return mol
@pytest.fixture
def h2o_ricc2_def2svp():
"""NumForce calls only make sense for stationary points. So this
geometry was optimized at the ricc2/def2-svp level of theory and
can be used to run NumForce with ricc2."""
mol = qcelemental.models.Molecule.from_data(
"""
O 0.0000000 0.0000000 -0.0835835
H 0.7501772 0.0000000 0.5210589
H -0.7501772 0.0000000 0.5210589
"""
)
return mol
@pytest.mark.parametrize(
"method, keywords, ref_energy",
[
pytest.param("hf", {}, -75.95536954370, marks=using("turbomole")),
pytest.param("pbe0", {"grid": "m5"}, -76.27371135900, marks=using("turbomole")),
pytest.param("ricc2", {}, -76.1603807755, marks=using("turbomole")),
pytest.param("rimp2", {}, -76.1593614075, marks=using("turbomole")),
pytest.param(
"hf",
{"scf_conv": 4, "scf_iters": 1},
-75.95536954370,
marks=[using("turbomole"), pytest.mark.xfail(raises=AssertionError, strict=True)],
),
],
)
def test_turbomole_energy(method, keywords, ref_energy, h2o):
resi = {"molecule": h2o, "driver": "energy", "model": {"method": method, "basis": "def2-SVP"}, "keywords": keywords}
res = qcng.compute(resi, "turbomole", raise_error=True, return_dict=True)
assert res["driver"] == "energy"
assert res["success"] is True
assert compare_values(ref_energy, res["return_result"])
@pytest.mark.parametrize(
"method, keywords, ref_norm",
[
pytest.param("hf", {}, 0.099340, marks=using("turbomole")),
pytest.param("pbe0", {"grid": "m5"}, 0.0606266, marks=using("turbomole")),
pytest.param("ricc2", {}, 0.059378, marks=using("turbomole")),
pytest.param("rimp2", {}, 0.061576, marks=using("turbomole")),
],
)
def test_turbomole_gradient(method, keywords, ref_norm, h2o):
resi = {
"molecule": h2o,
"driver": "gradient",
"model": {"method": method, "basis": "def2-SVP"},
"keywords": keywords,
}
res = qcng.compute(resi, "turbomole", raise_error=True)
assert res.driver == "gradient"
assert res.success is True
assert res.properties.return_energy
grad = res.return_result
grad_norm = np.linalg.norm(grad)
assert compare_values(ref_norm, grad_norm)
@using("turbomole")
def test_turbomole_ri_dsp(h2o):
resi = {
"molecule": h2o,
"driver": "energy",
"model": {"method": "b-p", "basis": "def2-SVP"},
"keywords": {"ri": True, "d3bj": True},
}
res = qcng.compute(resi, "turbomole", raise_error=True)
assert res.driver == "energy"
assert res.success is True
energy = res.return_result
ref_energy = -76.36275642866
assert compare_values(ref_energy, energy)
def assert_hessian(H, ref_eigvals, ref_size):
w, v = np.linalg.eigh(H)
last_eigvals = w[-3:]
# Hessian must be symmetric
np.testing.assert_allclose(H, H.T)
# Check eigenvalues
np.testing.assert_allclose(last_eigvals, ref_eigvals)
# Hessian must be of shape (3N x 3N)
assert H.shape == (ref_size, ref_size)
@using("turbomole")
@pytest.mark.parametrize(
"method, keywords, ref_eigvals",
[
("hf", {}, (2.00771683e-01, 7.77977644e-01, 9.91091318e-01)),
("pbe0", {"grid": "m5"}, (1.72092719e-01, 7.38603449e-01, 9.73783598e-01)),
("b-p", {"grid": "m5", "ri": True}, (1.59729409e-01, 7.21364827e-01, 9.63399519e-01)),
],
)
def test_turbomole_hessian(method, keywords, ref_eigvals, h2o):
resi = {
"molecule": h2o,
"driver": "hessian",
"model": {
"method": method,
"basis": "def2-SVP",
},
"keywords": keywords,
}
res = qcng.compute(resi, "turbomole", raise_error=True)
H = res.return_result
size = h2o.geometry.size
assert res.driver == "hessian"
assert res.success is True
assert res.properties.return_energy
assert_hessian(H, ref_eigvals, size)
@using("turbomole")
@pytest.mark.parametrize(
"method, keywords, ref_eigvals",
[
("ricc2", {}, (1.65405531e-01, 9.63690706e-01, 1.24676634e00)),
],
)
def METHOD_NAME(method, keywords, ref_eigvals, h2o_ricc2_def2svp):
resi = {
"molecule": h2o_ricc2_def2svp,
"driver": "hessian",
"model": {
"method": method,
"basis": "def2-SVP",
},
"keywords": keywords,
}
res = qcng.compute(resi, "turbomole", raise_error=True)
H = res.return_result
size = h2o_ricc2_def2svp.geometry.size
assert res.driver == "hessian"
assert res.success is True
assert res.properties.return_energy
assert_hessian(H, ref_eigvals, size)
@pytest.fixture
def h2o_nprhessian():
return """$nprhessian
1 1 0.6142699252 -0.0000000000 0.0000000000 -0.3071349626 -0.2479448514
1 2 -0.0000000000 -0.3071349626 0.2479448514 -0.0000000000
2 1 -0.0000000000 0.4365036678 0.0000000000 -0.1885017686 -0.2182518339
2 2 -0.0000000000 0.1885017686 -0.2182518339 0.0000000000
3 1 0.0000000000 0.0000000000 -0.0000524175 -0.0000000000 -0.0000000000
3 2 0.0000262088 -0.0000000000 0.0000000000 0.0000262088
4 1 -0.3071349626 -0.1885017686 -0.0000000000 0.3389423895 0.2182233100
4 2 0.0000000000 -0.0318074269 -0.0297215414 -0.0000000000
5 1 -0.2479448514 -0.2182518339 -0.0000000000 0.2182233100 0.2092172237
5 2 0.0000000000 0.0297215414 0.0090346102 0.0000000000
6 1 -0.0000000000 -0.0000000000 0.0000262088 0.0000000000 0.0000000000
6 2 -0.0000125560 -0.0000000000 0.0000000000 -0.0000136528
7 1 -0.3071349626 0.1885017686 -0.0000000000 -0.0318074269 0.0297215414
7 2 -0.0000000000 0.3389423895 -0.2182233100 0.0000000000
8 1 0.2479448514 -0.2182518339 0.0000000000 -0.0297215414 0.0090346102
8 2 0.0000000000 -0.2182233100 0.2092172237 -0.0000000000
9 1 -0.0000000000 0.0000000000 0.0000262088 -0.0000000000 0.0000000000
9 2 -0.0000136528 0.0000000000 -0.0000000000 -0.0000125560
$end"""
def test_turbomole_parse_hessian(h2o_nprhessian):
"""Test parsing of unproject Turbomole Hessian for water."""
hessian = parse_hessian(h2o_nprhessian)
assert hessian.shape == (9, 9)
eigvals, _ = np.linalg.eigh(hessian)
assert eigvals[-1] == pytest.approx(1.12157030e00) | null |
on 200 | # --------------------------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
#
# Code generated by aaz-dev-tools
# --------------------------------------------------------------------------------------------
# pylint: skip-file
# flake8: noqa
from azure.cli.core.aaz import *
@register_command(
"network vhub routing-intent list",
is_preview=True,
)
class List(AAZCommand):
"""Retrieve the details of all routing intent resources of the virtual hub.
:example: Retrieve the details of all routing intent resources of the virtual hub.
az network vhub routing-intent list -g MyResourceGroup --vhub MyVirtualHub
"""
_aaz_info = {
"version": "2021-05-01",
"resources": [
["mgmt-plane", "/subscriptions/{}/resourcegroups/{}/providers/microsoft.network/virtualhubs/{}/routingintent", "2021-05-01"],
]
}
def _handler(self, command_args):
super()._handler(command_args)
return self.build_paging(self._execute_operations, self._output)
_args_schema = None
@classmethod
def _build_arguments_schema(cls, *args, **kwargs):
if cls._args_schema is not None:
return cls._args_schema
cls._args_schema = super()._build_arguments_schema(*args, **kwargs)
# define Arg Group ""
_args_schema = cls._args_schema
_args_schema.resource_group = AAZResourceGroupNameArg(
required=True,
)
_args_schema.vhub = AAZStrArg(
options=["--vhub"],
help="Name of the virtual hub.",
required=True,
)
return cls._args_schema
def _execute_operations(self):
self.RoutingIntentList(ctx=self.ctx)()
def _output(self, *args, **kwargs):
result = self.deserialize_output(self.ctx.vars.instance.value, client_flatten=True)
next_link = self.deserialize_output(self.ctx.vars.instance.next_link)
return result, next_link
class RoutingIntentList(AAZHttpOperation):
CLIENT_TYPE = "MgmtClient"
def __call__(self, *args, **kwargs):
request = self.make_request()
session = self.client.send_request(request=request, stream=False, **kwargs)
if session.http_response.status_code in [200]:
return self.METHOD_NAME(session)
return self.on_error(session.http_response)
@property
def url(self):
return self.client.format_url(
"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Network/virtualHubs/{virtualHubName}/routingIntent",
**self.url_parameters
)
@property
def method(self):
return "GET"
@property
def error_format(self):
return "ODataV4Format"
@property
def url_parameters(self):
parameters = {
**self.serialize_url_param(
"resourceGroupName", self.ctx.args.resource_group,
required=True,
),
**self.serialize_url_param(
"subscriptionId", self.ctx.subscription_id,
required=True,
),
**self.serialize_url_param(
"virtualHubName", self.ctx.args.vhub,
required=True,
),
}
return parameters
@property
def query_parameters(self):
parameters = {
**self.serialize_query_param(
"api-version", "2021-05-01",
required=True,
),
}
return parameters
@property
def header_parameters(self):
parameters = {
**self.serialize_header_param(
"Accept", "application/json",
),
}
return parameters
def METHOD_NAME(self, session):
data = self.deserialize_http_content(session)
self.ctx.set_var(
"instance",
data,
schema_builder=self._build_schema_on_200
)
_schema_on_200 = None
@classmethod
def _build_schema_on_200(cls):
if cls._schema_on_200 is not None:
return cls._schema_on_200
cls._schema_on_200 = AAZObjectType()
_schema_on_200 = cls._schema_on_200
_schema_on_200.next_link = AAZStrType(
serialized_name="nextLink",
)
_schema_on_200.value = AAZListType()
value = cls._schema_on_200.value
value.Element = AAZObjectType()
_element = cls._schema_on_200.value.Element
_element.etag = AAZStrType(
flags={"read_only": True},
)
_element.id = AAZStrType()
_element.name = AAZStrType()
_element.properties = AAZObjectType(
flags={"client_flatten": True},
)
_element.type = AAZStrType(
flags={"read_only": True},
)
properties = cls._schema_on_200.value.Element.properties
properties.provisioning_state = AAZStrType(
serialized_name="provisioningState",
flags={"read_only": True},
)
properties.routing_policies = AAZListType(
serialized_name="routingPolicies",
)
routing_policies = cls._schema_on_200.value.Element.properties.routing_policies
routing_policies.Element = AAZObjectType()
_element = cls._schema_on_200.value.Element.properties.routing_policies.Element
_element.destinations = AAZListType(
flags={"required": True},
)
_element.name = AAZStrType(
flags={"required": True},
)
_element.next_hop = AAZStrType(
serialized_name="nextHop",
flags={"required": True},
)
destinations = cls._schema_on_200.value.Element.properties.routing_policies.Element.destinations
destinations.Element = AAZStrType()
return cls._schema_on_200
__all__ = ["List"] | null |
target selectors | import datetime
import random
import uuid
import factory
from sqlalchemy import orm
from h import models
from h.db.types import URLSafeUUID
from h.models.document import update_document_metadata
from .base import FAKER, ModelFactory
from .document import Document, DocumentMeta, DocumentURI
class Annotation(ModelFactory):
class Meta:
model = models.Annotation
sqlalchemy_session_persistence = (
"flush" # Always flush the db to generate annotation.id.
)
tags = factory.LazyFunction(
lambda: list(FAKER.words(nb=random.randint(0, 5))) # pylint:disable=no-member
)
target_uri = factory.Faker("uri")
text = factory.Faker("paragraph")
userid = factory.LazyFunction(
lambda: f"acct:{FAKER.user_name()}@localhost" # pylint:disable=no-member
)
document = factory.SubFactory(Document)
groupid = "__world__"
@factory.lazy_attribute
def METHOD_NAME(self):
return [
{
"endContainer": "/div[1]/article[1]/section[1]/div[1]/div[2]/div[1]",
"endOffset": 76,
"startContainer": "/div[1]/article[1]/section[1]/div[1]/div[2]/div[1]",
"startOffset": 0,
"type": "RangeSelector",
},
{"end": 362, "start": 286, "type": "TextPositionSelector"},
{
# pylint: disable=line-too-long
"exact": "If you wish to install Hypothesis on your own site then head over to GitHub.",
"prefix": " browser extension.\n ",
"suffix": "\n \n \n \n ",
"type": "TextQuoteSelector",
},
]
@factory.post_generation
def make_metadata( # pylint:disable=unused-argument
self, create, extracted, **kwargs
):
"""Create associated document metadata for the annotation."""
# The metadata objects are going to be added to the db, so if we're not
# using the create strategy then simply don't make any.
if not create:
return
def document_uri_dict():
"""
Return a randomly generated DocumentURI dict for this annotation.
This doesn't add anything to the database session yet.
"""
document_uri = DocumentURI.build(
document=None, claimant=self.target_uri, uri=self.target_uri
)
return {
"claimant": document_uri.claimant,
"uri": document_uri.uri,
"type": document_uri.type,
"content_type": document_uri.content_type,
}
document_uri_dicts = [document_uri_dict() for _ in range(random.randint(1, 3))]
def document_meta_dict(type_=None):
"""
Return a randomly generated DocumentMeta dict for this annotation.
This doesn't add anything to the database session yet.
"""
kwargs = {"document": None, "claimant": self.target_uri}
if type_ is not None:
kwargs["type"] = type_
document_meta = DocumentMeta.build(**kwargs)
return {
"claimant": document_meta.claimant,
"type": document_meta.type,
"value": document_meta.value,
}
document_meta_dicts = [
document_meta_dict() for _ in range(random.randint(1, 3))
]
# Make sure that there's always at least one DocumentMeta with
# type='title', so that we never get annotation.document.title is None:
if "title" not in [m["type"] for m in document_meta_dicts]:
document_meta_dicts.append(document_meta_dict(type_="title"))
self.document = update_document_metadata(
orm.object_session(self),
self.target_uri,
document_meta_dicts=document_meta_dicts,
document_uri_dicts=document_uri_dicts,
created=self.created,
updated=self.updated,
)
@factory.post_generation
def make_id(self, create, extracted, **kwargs): # pylint:disable=unused-argument
"""Add a randomly ID if the annotation doesn't have one yet."""
# If using the create strategy don't generate an id.
# models.Annotation.id's server_default function will generate one
# when the annotation is saved to the DB.
if create:
return
# Don't generate an id if the user passed in one of their own.
if getattr(self, "id", None):
return
# Ids in the DB are in hex, but in the code they should be URL safe
self.id = URLSafeUUID().process_result_value( # pylint:disable=attribute-defined-outside-init,invalid-name
uuid.uuid4().hex, None
)
@factory.post_generation
def timestamps(self, create, extracted, **kwargs): # pylint:disable=unused-argument
# If using the create strategy let sqlalchemy set the created and
# updated times when saving to the DB.
if create:
return
# When using the build or stub strategy sqlalchemy won't set created or updated
# times for us, so do it ourselves instead.
#
# We're generating created and updated separately (calling now() twice
# instead of just once) so created and updated won't be exactly the
# same. This is consistent with how models.Annotation does it when
# saving to the DB.
# pylint:disable=attribute-defined-outside-init
self.created = self.created or datetime.datetime.now()
self.updated = self.updated or datetime.datetime.now() | null |
test set log level error | # Copyright (c) Streamlit Inc. (2018-2022) Snowflake Inc. (2022)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Logger Unittest."""
import logging
import unittest
from collections import OrderedDict
from unittest.mock import patch
import pytest
from parameterized import parameterized
from streamlit import config, logger
DUMMY_CONFIG_OPTIONS = OrderedDict()
class LoggerTest(unittest.TestCase):
"""Logger Unittest class."""
# Need to fix this test:
# https://trello.com/c/ZwNR7fWI
# def test_set_log_level_by_name(self):
# """Test streamlit.logger.set_log_level."""
# data = {
# 'critical': logging.CRITICAL,
# 'error': logging.ERROR,
# 'warning': logging.WARNING,
# 'info': logging.INFO,
# 'debug': logging.DEBUG,
# }
# for k, v in data.items():
# streamlit.logger.set_log_level(k)
# self.assertEqual(v, logging.getLogger().getEffectiveLevel())
def test_set_log_level_by_constant(self):
"""Test streamlit.logger.set_log_level."""
data = [
logging.CRITICAL,
logging.ERROR,
logging.WARNING,
logging.INFO,
logging.DEBUG,
]
for k in data:
logger.set_log_level(k)
self.assertEqual(k, logging.getLogger("streamlit").getEffectiveLevel())
def METHOD_NAME(self):
"""Test streamlit.logger.set_log_level."""
with pytest.raises(SystemExit) as e:
logger.set_log_level(90)
self.assertEqual(e.type, SystemExit)
self.assertEqual(e.value.code, 1)
# Need to fix this test:
# https://trello.com/c/ZwNR7fWI
# def test_set_log_level_resets(self):
# """Test streamlit.logger.set_log_level."""
# streamlit.logger.set_log_level('debug')
# test1 = streamlit.logger.get_logger('test1')
# self.assertEqual(logging.DEBUG, test1.getEffectiveLevel())
#
# streamlit.logger.set_log_level('warning')
# self.assertEqual(logging.WARNING, test1.getEffectiveLevel())
#
# streamlit.logger.set_log_level('critical')
# test2 = streamlit.logger.get_logger('test2')
# self.assertEqual(logging.CRITICAL, test2.getEffectiveLevel())
@parameterized.expand(
[
("%(asctime)s.%(msecs)03d %(name)s: %(message)s", None),
("%(asctime)s.%(msecs)03d %(name)s: %(message)s", DUMMY_CONFIG_OPTIONS),
(None, None),
(None, DUMMY_CONFIG_OPTIONS),
]
)
def test_setup_log_formatter(self, messageFormat, config_options):
"""Test streamlit.logger.setup_log_formatter."""
LOGGER = logger.get_logger("test")
config._set_option("logger.messageFormat", messageFormat, "test")
config._set_option("logger.level", logging.DEBUG, "test")
with patch.object(config, "_config_options", new=config_options):
logger.setup_formatter(LOGGER)
self.assertEqual(len(LOGGER.handlers), 1)
if config_options:
self.assertEqual(
LOGGER.handlers[0].formatter._fmt, messageFormat or "%(message)s"
)
else:
self.assertEqual(
LOGGER.handlers[0].formatter._fmt, logger.DEFAULT_LOG_MESSAGE
)
def test_init_tornado_logs(self):
"""Test streamlit.logger.init_tornado_logs."""
logger.init_tornado_logs()
loggers = [x for x in logger._loggers.keys() if "tornado." in x]
truth = ["tornado.access", "tornado.application", "tornado.general"]
self.assertEqual(sorted(truth), sorted(loggers))
# Need to fix this test:
# https://trello.com/c/ZwNR7fWI
# def test_get_logger(self):
# """Test streamlit.logger.get_logger."""
# # Test that get_logger with no args, figures out its caller
# logger = streamlit.logger.get_logger()
# self.assertTrue('.logger_test' in streamlit.logger.LOGGERS.keys()) | null |
get mro | """
Patch recently added ABCs into the standard lib module
``collections.abc`` (Py3) or ``collections`` (Py2).
Usage::
import backports_abc
backports_abc.patch()
or::
try:
from collections.abc import Generator
except ImportError:
from backports_abc import Generator
"""
try:
import collections.abc as _collections_abc
except ImportError:
import collections as _collections_abc
def METHOD_NAME(cls):
try:
return cls.__mro__
except AttributeError:
return old_style_mro(cls)
def old_style_mro(cls):
yield cls
for base in cls.__bases__:
for c in old_style_mro(base):
yield c
def mk_gen():
from abc import abstractmethod
required_methods = (
'__iter__', '__next__' if hasattr(iter(()), '__next__') else 'next',
'send', 'throw', 'close')
class Generator(_collections_abc.Iterator):
__slots__ = ()
if '__next__' in required_methods:
def __next__(self):
return self.send(None)
else:
def next(self):
return self.send(None)
@abstractmethod
def send(self, value):
raise StopIteration
@abstractmethod
def throw(self, typ, val=None, tb=None):
if val is None:
if tb is None:
raise typ
val = typ()
if tb is not None:
val = val.with_traceback(tb)
raise val
def close(self):
try:
self.throw(GeneratorExit)
except (GeneratorExit, StopIteration):
pass
else:
raise RuntimeError('generator ignored GeneratorExit')
@classmethod
def __subclasshook__(cls, C):
if cls is Generator:
mro = METHOD_NAME(C)
for method in required_methods:
for base in mro:
if method in base.__dict__:
break
else:
return NotImplemented
return True
return NotImplemented
generator = type((lambda: (yield))())
Generator.register(generator)
return Generator
def mk_awaitable():
from abc import abstractmethod, ABCMeta
@abstractmethod
def __await__(self):
yield
@classmethod
def __subclasshook__(cls, C):
if cls is Awaitable:
for B in METHOD_NAME(C):
if '__await__' in B.__dict__:
if B.__dict__['__await__']:
return True
break
return NotImplemented
# calling metaclass directly as syntax differs in Py2/Py3
Awaitable = ABCMeta('Awaitable', (), {
'__slots__': (),
'__await__': __await__,
'__subclasshook__': __subclasshook__,
})
return Awaitable
def mk_coroutine():
from abc import abstractmethod
class Coroutine(Awaitable):
__slots__ = ()
@abstractmethod
def send(self, value):
"""Send a value into the coroutine.
Return next yielded value or raise StopIteration.
"""
raise StopIteration
@abstractmethod
def throw(self, typ, val=None, tb=None):
"""Raise an exception in the coroutine.
Return next yielded value or raise StopIteration.
"""
if val is None:
if tb is None:
raise typ
val = typ()
if tb is not None:
val = val.with_traceback(tb)
raise val
def close(self):
"""Raise GeneratorExit inside coroutine.
"""
try:
self.throw(GeneratorExit)
except (GeneratorExit, StopIteration):
pass
else:
raise RuntimeError('coroutine ignored GeneratorExit')
@classmethod
def __subclasshook__(cls, C):
if cls is Coroutine:
mro = METHOD_NAME(C)
for method in ('__await__', 'send', 'throw', 'close'):
for base in mro:
if method in base.__dict__:
break
else:
return NotImplemented
return True
return NotImplemented
return Coroutine
###
# make all ABCs available in this module
try:
Generator = _collections_abc.Generator
except AttributeError:
Generator = mk_gen()
try:
Awaitable = _collections_abc.Awaitable
except AttributeError:
Awaitable = mk_awaitable()
try:
Coroutine = _collections_abc.Coroutine
except AttributeError:
Coroutine = mk_coroutine()
try:
from inspect import isawaitable
except ImportError:
def isawaitable(obj):
return isinstance(obj, Awaitable)
###
# allow patching the stdlib
PATCHED = {}
def patch(patch_inspect=True):
"""
Main entry point for patching the ``collections.abc`` and ``inspect``
standard library modules.
"""
PATCHED['collections.abc.Generator'] = _collections_abc.Generator = Generator
PATCHED['collections.abc.Coroutine'] = _collections_abc.Coroutine = Coroutine
PATCHED['collections.abc.Awaitable'] = _collections_abc.Awaitable = Awaitable
if patch_inspect:
import inspect
PATCHED['inspect.isawaitable'] = inspect.isawaitable = isawaitable | null |
operation group two | # coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for
# license information.
#
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is
# regenerated.
# --------------------------------------------------------------------------
from typing import Any, Optional, TYPE_CHECKING
from azure.mgmt.core import ARMPipelineClient
from azure.profiles import KnownProfiles, ProfileDefinition
from azure.profiles.multiapiclient import MultiApiClientMixin
from ._configuration import MultiapiServiceClientConfiguration
from ._operations_mixin import MultiapiServiceClientOperationsMixin
from ._serialization import Deserializer, Serializer
if TYPE_CHECKING:
# pylint: disable=unused-import,ungrouped-imports
from azure.core.credentials import TokenCredential
class _SDKClient(object):
def __init__(self, *args, **kwargs):
"""This is a fake class to support current implemetation of MultiApiClientMixin."
Will be removed in final version of multiapi azure-core based client
"""
pass
class MultiapiServiceClient(MultiapiServiceClientOperationsMixin, MultiApiClientMixin, _SDKClient):
"""Service client for multiapi client testing.
This ready contains multiple API versions, to help you deal with all of the Azure clouds
(Azure Stack, Azure Government, Azure China, etc.).
By default, it uses the latest API version available on public Azure.
For production, you should stick to a particular api-version and/or profile.
The profile sets a mapping between an operation group and its API version.
The api-version parameter sets the default API version if the operation
group is not described in the profile.
:param credential: Credential needed for the client to connect to Azure. Required.
:type credential: ~azure.core.credentials.TokenCredential
:param api_version: API version to use if no profile is provided, or if missing in profile.
:type api_version: str
:param base_url: Service URL
:type base_url: str
:param profile: A profile definition, from KnownProfiles to dict.
:type profile: azure.profiles.KnownProfiles
:keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present.
"""
DEFAULT_API_VERSION = '3.0.0'
_PROFILE_TAG = "multiapinoasync.MultiapiServiceClient"
LATEST_PROFILE = ProfileDefinition({
_PROFILE_TAG: {
None: DEFAULT_API_VERSION,
'begin_test_lro': '1.0.0',
'begin_test_lro_and_paging': '1.0.0',
'test_one': '2.0.0',
}},
_PROFILE_TAG + " latest"
)
def __init__(
self,
credential: "TokenCredential",
api_version: Optional[str]=None,
base_url: str = "http://localhost:3000",
profile: KnownProfiles=KnownProfiles.default,
**kwargs: Any
):
if api_version:
kwargs.setdefault('api_version', api_version)
self._config = MultiapiServiceClientConfiguration(credential, **kwargs)
self._client = ARMPipelineClient(base_url=base_url, config=self._config, **kwargs)
super(MultiapiServiceClient, self).__init__(
api_version=api_version,
profile=profile
)
@classmethod
def _models_dict(cls, api_version):
return {k: v for k, v in cls.models(api_version).__dict__.items() if isinstance(v, type)}
@classmethod
def models(cls, api_version=DEFAULT_API_VERSION):
"""Module depends on the API version:
* 1.0.0: :mod:`v1.models<multiapinoasync.v1.models>`
* 2.0.0: :mod:`v2.models<multiapinoasync.v2.models>`
* 3.0.0: :mod:`v3.models<multiapinoasync.v3.models>`
"""
if api_version == '1.0.0':
from .v1 import models
return models
elif api_version == '2.0.0':
from .v2 import models
return models
elif api_version == '3.0.0':
from .v3 import models
return models
raise ValueError("API version {} is not available".format(api_version))
@property
def operation_group_one(self):
"""Instance depends on the API version:
* 1.0.0: :class:`OperationGroupOneOperations<multiapinoasync.v1.operations.OperationGroupOneOperations>`
* 2.0.0: :class:`OperationGroupOneOperations<multiapinoasync.v2.operations.OperationGroupOneOperations>`
* 3.0.0: :class:`OperationGroupOneOperations<multiapinoasync.v3.operations.OperationGroupOneOperations>`
"""
api_version = self._get_api_version('operation_group_one')
if api_version == '1.0.0':
from .v1.operations import OperationGroupOneOperations as OperationClass
elif api_version == '2.0.0':
from .v2.operations import OperationGroupOneOperations as OperationClass
elif api_version == '3.0.0':
from .v3.operations import OperationGroupOneOperations as OperationClass
else:
raise ValueError("API version {} does not have operation group 'operation_group_one'".format(api_version))
self._config.api_version = api_version
return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version)), api_version)
@property
def METHOD_NAME(self):
"""Instance depends on the API version:
* 2.0.0: :class:`OperationGroupTwoOperations<multiapinoasync.v2.operations.OperationGroupTwoOperations>`
* 3.0.0: :class:`OperationGroupTwoOperations<multiapinoasync.v3.operations.OperationGroupTwoOperations>`
"""
api_version = self._get_api_version('operation_group_two')
if api_version == '2.0.0':
from .v2.operations import OperationGroupTwoOperations as OperationClass
elif api_version == '3.0.0':
from .v3.operations import OperationGroupTwoOperations as OperationClass
else:
raise ValueError("API version {} does not have operation group 'operation_group_two'".format(api_version))
self._config.api_version = api_version
return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version)), api_version)
def close(self):
self._client.close()
def __enter__(self):
self._client.__enter__()
return self
def __exit__(self, *exc_details):
self._client.__exit__(*exc_details) | null |
flush | # Copyright 2017 PerfKitBenchmarker Authors. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import json
import logging
from perfkitbenchmarker import background_tasks
from perfkitbenchmarker import errors
from perfkitbenchmarker import provider_info
from perfkitbenchmarker import vm_util
from perfkitbenchmarker.linux_packages import memcached_server
from perfkitbenchmarker.memcache_service import MemcacheService
from perfkitbenchmarker.providers.aws import aws_network
from perfkitbenchmarker.providers.aws import util
ELASTICACHE_PORT = 11211
class ElastiCacheMemcacheService(MemcacheService):
"""Class for AWS elasticache memcache service."""
CLOUD = provider_info.AWS
def __init__(self, network, cluster_id, region, node_type, num_servers=1):
self.cluster_id = cluster_id
self.region = region
self.node_type = node_type
self.num_servers = num_servers
self.hosts = [] # [(ip, port)]
self.vpc_id = network.subnet.vpc_id
self.security_group_id = (
network.regional_network.vpc.default_security_group_id)
self.subnet_id = network.subnet.id
self.subnet_group_name = '%ssubnet' % cluster_id
def Create(self):
# Open the port memcached needs
aws_network.AwsFirewall.GetFirewall().AllowPortInSecurityGroup(
self.region, self.security_group_id, ELASTICACHE_PORT)
# Create a cache subnet group
cmd = ['aws', 'elasticache', 'create-cache-subnet-group',
'--region=%s' % self.region,
'--cache-subnet-group-name=%s' % self.subnet_group_name,
'--cache-subnet-group-description="PKB memcached_ycsb benchmark"',
'--subnet-ids=%s' % self.subnet_id]
vm_util.IssueCommand(cmd)
# Create the cluster
cmd = ['aws', 'elasticache', 'create-cache-cluster',
'--engine=memcached',
'--cache-subnet-group-name=%s' % self.subnet_group_name,
'--cache-cluster-id=%s' % self.cluster_id,
'--num-cache-nodes=%s' % self.num_servers,
'--region=%s' % self.region,
'--cache-node-type=%s' % self.node_type,
'--tags'] + util.MakeFormattedDefaultTags()
vm_util.IssueCommand(cmd)
# Wait for the cluster to come up
cluster_info = self._WaitForClusterUp()
# Parse out the hosts
self.hosts = [(node['Endpoint']['Address'], node['Endpoint']['Port'])
for node in cluster_info['CacheNodes']]
assert len(self.hosts) == self.num_servers
def Destroy(self):
# Delete the ElastiCache cluster
cmd = ['aws', 'elasticache', 'delete-cache-cluster',
'--cache-cluster-id=%s' % self.cluster_id,
'--region=%s' % self.region]
vm_util.IssueCommand(cmd, raise_on_failure=False)
# Don't have to delete the subnet group. It will be deleted with the subnet.
def METHOD_NAME(self):
background_tasks.RunThreaded(
memcached_server.FlushMemcachedServer, self.hosts
)
def GetHosts(self):
return ['%s:%s' % (ip, port) for ip, port in self.hosts]
def GetMetadata(self):
return {'num_servers': self.num_servers,
'elasticache_region': self.region,
'elasticache_node_type': self.node_type}
def _GetClusterInfo(self):
cmd = ['aws', 'elasticache', 'describe-cache-clusters']
cmd += ['--cache-cluster-id=%s' % self.cluster_id]
cmd += ['--region=%s' % self.region]
cmd += ['--show-cache-node-info']
out, _, _ = vm_util.IssueCommand(cmd)
return json.loads(out)['CacheClusters'][0]
@vm_util.Retry(poll_interval=15, timeout=300,
retryable_exceptions=(errors.Resource.RetryableCreationError))
def _WaitForClusterUp(self):
"""Block until the ElastiCache memcached cluster is up.
Will timeout after 5 minutes, and raise an exception. Before the timeout
expires any exceptions are caught and the status check is retried.
We check the status of the cluster using the AWS CLI.
Returns:
The cluster info json as a dict
Raises:
errors.Resource.RetryableCreationError when response is not as expected or
if there is an error connecting to the port or otherwise running the
remote check command.
"""
logging.info('Trying to get ElastiCache cluster info for %s',
self.cluster_id)
cluster_status = None
try:
cluster_info = self._GetClusterInfo()
cluster_status = cluster_info['CacheClusterStatus']
if cluster_status == 'available':
logging.info('ElastiCache memcached cluster is up and running.')
return cluster_info
except errors.VirtualMachine.RemoteCommandError as e:
raise errors.Resource.RetryableCreationError(
'ElastiCache memcached cluster not up yet: %s.' % str(e))
else:
raise errors.Resource.RetryableCreationError(
'ElastiCache memcached cluster not up yet. Status: %s' %
cluster_status) | null |
get class | import numpy as np
import pytest
from aspire.image import Image
from aspire.utils import gaussian_2d, utest_tolerance
from aspire.utils.coor_trans import grid_2d
from aspire.utils.random import randn
from aspire.volume import Volume
# Parameter combinations for testing 2D bases
# Each tuple represents (resolution in pixels, datatype of basis)
basis_params_2d = [
(8, np.float32),
(8, np.float64),
(16, np.float32),
(16, np.float64),
(32, np.float32),
(32, np.float64),
]
basis_params_3d = [
(8, np.float32),
(8, np.float64),
]
pswf_params_2d = [
(8, np.float64),
]
def show_basis_params(basis):
# print descriptive test name for parametrized test
# run pytest with option -rA to see explicitly
return f"{basis.nres}-{basis.dtype}"
class Steerable2DMixin:
"""
Inheriting Test class will expect all Steerable2DMixin functions to take a Basis object
as a parameter.
"""
def testIndices(self, basis):
ell_max = basis.ell_max
k_max = basis.k_max
indices = basis.indices()
i = 0
for ell in range(ell_max + 1):
if ell == 0:
sgns = [1]
else:
sgns = [1, -1]
for sgn in sgns:
for k in range(k_max[ell]):
assert indices["ells"][i] == ell
assert indices["sgns"][i] == sgn
assert indices["ks"][i] == k
i += 1
def testGaussianExpand(self, basis):
# Offset slightly
x0 = 0.50
y0 = 0.75
L = basis.nres
# Want sigma to be as large as possible without the Gaussian
# spilling too much outside the central disk.
sigma = L / 8
im1 = gaussian_2d(L, mu=(x0, y0), sigma=sigma, dtype=basis.dtype)
coef = basis.expand(im1)
im2 = basis.evaluate(coef)
if isinstance(im2, Image):
im2 = im2.asnumpy()
im2 = im2[0]
# For small L there's too much clipping at high freqs to get 1e-3
# accuracy.
if L < 32:
atol = 1e-2
else:
atol = 1e-3
assert im1.shape == im2.shape
assert np.allclose(im1, im2, atol=atol)
def testIsotropic(self, basis):
L = basis.nres
sigma = L / 8
im = gaussian_2d(L, sigma=sigma, dtype=basis.dtype)
coef = basis.expand(im)
ells = basis.indices()["ells"]
energy_outside = np.sum(np.abs(coef[ells != 0]) ** 2)
energy_total = np.sum(np.abs(coef) ** 2)
energy_ratio = energy_outside / energy_total
assert energy_ratio < 0.01
def testModulated(self, basis):
L = basis.nres
if L < 32:
pytest.skip()
ell = 1
sigma = L / 8
im = gaussian_2d(L, sigma=sigma, dtype=basis.dtype)
g2d = grid_2d(L)
for trig_fun in (np.sin, np.cos):
im1 = im * trig_fun(ell * g2d["phi"])
coef = basis.expand(im1)
ells = basis.indices()["ells"]
energy_outside = np.sum(np.abs(coef[ells != ell]) ** 2)
energy_total = np.sum(np.abs(coef) ** 2)
energy_ratio = energy_outside / energy_total
assert energy_ratio < 0.10
def testEvaluateExpand(self, basis):
coef1 = randn(basis.count, seed=self.seed)
coef1 = coef1.astype(basis.dtype)
im = basis.evaluate(coef1)
if isinstance(im, Image):
im = im.asnumpy()
coef2 = basis.expand(im)[0]
assert coef1.shape == coef2.shape
assert np.allclose(coef1, coef2, atol=utest_tolerance(basis.dtype))
def testAdjoint(self, basis):
u = randn(basis.count, seed=self.seed)
u = u.astype(basis.dtype)
Au = basis.evaluate(u)
if isinstance(Au, Image):
Au = Au.asnumpy()
x = Image(randn(*basis.sz, seed=self.seed), dtype=basis.dtype)
ATx = basis.evaluate_t(x)
Au_dot_x = np.sum(Au * x.asnumpy())
u_dot_ATx = np.sum(u * ATx)
assert Au_dot_x.shape == u_dot_ATx.shape
assert np.isclose(Au_dot_x, u_dot_ATx)
class UniversalBasisMixin:
"""
Inheriting Test class will expect all UniversalBasisMixin functions to take a Basis object
as a parameter.
"""
def METHOD_NAME(self, basis):
if basis.ndim == 2:
return Image
elif basis.ndim == 3:
return Volume
def testEvaluate(self, basis):
# evaluate should take a NumPy array of type basis.coefficient_dtype
# and return an Image/Volume
_class = self.METHOD_NAME(basis)
result = basis.evaluate(np.zeros((basis.count), dtype=basis.coefficient_dtype))
assert isinstance(result, _class)
def testEvaluate_t(self, basis):
# evaluate_t should take an Image/Volume and return a NumPy array of type
# basis.coefficient_dtype
_class = self.METHOD_NAME(basis)
result = basis.evaluate_t(
_class(np.zeros((basis.nres,) * basis.ndim, dtype=basis.dtype))
)
assert isinstance(result, np.ndarray)
assert result.dtype == basis.coefficient_dtype
def testExpand(self, basis):
_class = self.METHOD_NAME(basis)
# expand should take an Image/Volume and return a NumPy array of type
# basis.coefficient_dtype
result = basis.expand(
_class(np.zeros((basis.nres,) * basis.ndim, dtype=basis.dtype))
)
assert isinstance(result, np.ndarray)
assert result.dtype == basis.coefficient_dtype
def testInitWithIntSize(self, basis):
# make sure we can instantiate with just an int as a shortcut
assert (basis.nres,) * basis.ndim == basis.__class__(basis.nres).sz | null |
test branch change already on branch | # -*- coding: utf-8 -*-
# Copyright (c) 2020, Frappe and Contributors
# See license.txt
import unittest
from unittest.mock import patch
import frappe
from frappe.core.utils import find
from press.press.doctype.app.app import App
from press.press.doctype.app.test_app import create_test_app
from press.press.doctype.app_release.test_app_release import create_test_app_release
from press.press.doctype.app_source.app_source import AppSource
from press.press.doctype.app_source.test_app_source import create_test_app_source
from press.press.doctype.release_group.release_group import (
ReleaseGroup,
new_release_group,
)
from press.press.doctype.team.test_team import create_test_team
def create_test_release_group(
apps: list[App], user: str = None, public=False, frappe_version="Version 14"
) -> ReleaseGroup:
"""
Create Release Group doc.
Also creates app source
"""
user = user or frappe.session.user
release_group = frappe.get_doc(
{
"doctype": "Release Group",
"version": frappe_version,
"enabled": True,
"title": f"Test ReleaseGroup {frappe.mock('name')}",
"team": frappe.get_value("Team", {"user": user}, "name"),
"public": public,
}
)
for app in apps:
app_source = create_test_app_source(release_group.version, app)
release_group.append("apps", {"app": app.name, "source": app_source.name})
release_group.insert(ignore_if_duplicate=True)
release_group.reload()
return release_group
@patch.object(AppSource, "create_release", create_test_app_release)
class TestReleaseGroup(unittest.TestCase):
def setUp(self):
self.team = create_test_team().name
def tearDown(self):
frappe.db.rollback()
def test_create_release_group(self):
app = create_test_app("frappe", "Frappe Framework")
source = app.add_source(
"Version 12", "https://github.com/frappe/frappe", "version-12", team=self.team
)
group = new_release_group(
"Test Group",
"Version 12",
[{"app": source.app, "source": source.name}],
team=self.team,
)
self.assertEqual(group.title, "Test Group")
def test_create_release_group_set_app_from_source(self):
app1 = create_test_app("frappe", "Frappe Framework")
source1 = app1.add_source(
"Version 12", "https://github.com/frappe/frappe", "version-12", team=self.team
)
app2 = create_test_app("erpnext", "ERPNext")
source2 = app2.add_source(
"Version 12", "https://github.com/frappe/erpnext", "version-12", team=self.team
)
group = new_release_group(
"Test Group",
"Version 12",
[{"app": source2.app, "source": source1.name}],
team=self.team,
)
self.assertEqual(group.apps[0].app, source1.app)
def test_create_release_group_fail_when_first_app_is_not_frappe(self):
app = create_test_app("erpnext", "ERPNext")
source = app.add_source(
"Version 12", "https://github.com/frappe/erpnext", "version-12", team=self.team
)
self.assertRaises(
frappe.ValidationError,
new_release_group,
"Test Group",
"Version 12",
[{"app": source.app, "source": source.name}],
team=self.team,
)
def test_create_release_group_fail_when_duplicate_apps(self):
app = create_test_app("frappe", "Frappe Framework")
source = app.add_source(
"Version 12", "https://github.com/frappe/frappe", "version-12", team=self.team
)
self.assertRaises(
frappe.ValidationError,
new_release_group,
"Test Group",
"Version 12",
[
{"app": source.app, "source": source.name},
{"app": source.app, "source": source.name},
],
team=self.team,
)
def test_create_release_group_fail_when_version_mismatch(self):
app = create_test_app("frappe", "Frappe Framework")
source = app.add_source(
"Version 12", "https://github.com/frappe/frappe", "version-12", team=self.team
)
self.assertRaises(
frappe.ValidationError,
new_release_group,
"Test Group",
"Version 13",
[{"app": source.app, "source": source.name}],
team=self.team,
)
def test_create_release_group_fail_with_duplicate_titles(self):
app = create_test_app("frappe", "Frappe Framework")
source = app.add_source(
"Version 12", "https://github.com/frappe/frappe", "version-12", team=self.team
)
new_release_group(
"Test Group",
"Version 12",
[{"app": source.app, "source": source.name}],
team=self.team,
)
self.assertRaises(
frappe.ValidationError,
new_release_group,
"Test Group",
"Version 12",
[{"app": source.app, "source": source.name}],
team=self.team,
)
def METHOD_NAME(self):
app = create_test_app()
rg = create_test_release_group([app])
with self.assertRaises(frappe.ValidationError):
rg.change_app_branch("frappe", "master")
def test_branch_change_app_source_exists(self):
app = create_test_app()
rg = create_test_release_group([app])
current_app_source = frappe.get_doc("App Source", rg.apps[0].source)
app_source = create_test_app_source(
current_app_source.versions[0].version,
app,
current_app_source.repository_url,
"develop",
)
rg.change_app_branch(app.name, "develop")
rg.reload()
# Source must be set to the available `app_source` for `app`
self.assertEqual(rg.apps[0].source, app_source.name)
def test_branch_change_app_source_does_not_exist(self):
app = create_test_app()
rg = create_test_release_group([app])
previous_app_source = frappe.get_doc("App Source", rg.apps[0].source)
rg.change_app_branch(app.name, "develop")
rg.reload()
new_app_source = frappe.get_doc("App Source", rg.apps[0].source)
self.assertEqual(new_app_source.branch, "develop")
self.assertEqual(
new_app_source.versions[0].version, previous_app_source.versions[0].version
)
self.assertEqual(new_app_source.repository_url, previous_app_source.repository_url)
self.assertEqual(new_app_source.app, app.name)
def test_new_release_group_loaded_with_correct_dependencies(self):
app = create_test_app("frappe", "Frappe Framework")
frappe_version = frappe.get_doc("Frappe Version", "Version 14")
group = frappe.get_doc(
{
"doctype": "Release Group",
"title": "Test Group",
"version": "Version 14",
"apps": [
{"app": app.name, "source": create_test_app_source("Version 14", app).name}
],
"team": self.team,
}
).insert()
self.assertEqual(
find(group.dependencies, lambda d: d.dependency == "PYTHON_VERSION").version,
find(
frappe_version.dependencies, lambda x: x.dependency == "PYTHON_VERSION"
).version,
) | null |
test user revoke invalid certified | # Parsec Cloud (https://parsec.cloud) Copyright (c) BUSL-1.1 2016-present Scille SAS
from __future__ import annotations
import pytest
import trio
from quart.testing.connections import WebsocketDisconnectError
from parsec._parsec import (
DateTime,
UserProfile,
)
from parsec.api.data import RevokedUserCertificate
from parsec.api.protocol import (
HandshakeRevokedDevice,
UserRevokeRepAlreadyRevoked,
UserRevokeRepInvalidCertification,
UserRevokeRepNotAllowed,
UserRevokeRepNotFound,
UserRevokeRepOk,
)
from parsec.backend.user import INVITATION_VALIDITY
from tests.backend.common import authenticated_ping, user_revoke
from tests.common import freeze_time
@pytest.mark.trio
async def test_backend_close_on_user_revoke(
backend_asgi_app, alice_ws, backend_authenticated_ws_factory, bob, alice
):
now = DateTime.now()
bob_revocation = RevokedUserCertificate(
author=alice.device_id, timestamp=now, user_id=bob.user_id
).dump_and_sign(alice.signing_key)
async with backend_authenticated_ws_factory(backend_asgi_app, bob) as bob_ws:
with backend_asgi_app.backend.event_bus.listen():
rep = await user_revoke(alice_ws, revoked_user_certificate=bob_revocation)
assert isinstance(rep, UserRevokeRepOk)
# `user.revoked` event schedules connection cancellation, so wait
# for things to settle down to make sure the cancellation is done
await trio.testing.wait_all_tasks_blocked()
# Bob cannot send new command
with pytest.raises(WebsocketDisconnectError):
await authenticated_ping(bob_ws)
@pytest.mark.trio
async def test_user_revoke_ok(
backend_asgi_app, backend_authenticated_ws_factory, adam_ws, alice, adam
):
now = DateTime.now()
alice_revocation = RevokedUserCertificate(
author=adam.device_id, timestamp=now, user_id=alice.user_id
).dump_and_sign(adam.signing_key)
with backend_asgi_app.backend.event_bus.listen():
rep = await user_revoke(adam_ws, revoked_user_certificate=alice_revocation)
assert isinstance(rep, UserRevokeRepOk)
# Alice cannot connect from now on...
with pytest.raises(HandshakeRevokedDevice):
async with backend_authenticated_ws_factory(backend_asgi_app, alice):
pass
@pytest.mark.trio
async def test_user_revoke_not_admin(
backend_asgi_app, backend_authenticated_ws_factory, bob_ws, alice, bob
):
now = DateTime.now()
alice_revocation = RevokedUserCertificate(
author=bob.device_id, timestamp=now, user_id=alice.user_id
).dump_and_sign(bob.signing_key)
rep = await user_revoke(bob_ws, revoked_user_certificate=alice_revocation)
assert isinstance(rep, UserRevokeRepNotAllowed)
@pytest.mark.trio
async def test_cannot_self_revoke(
backend_asgi_app, backend_authenticated_ws_factory, alice_ws, alice
):
now = DateTime.now()
alice_revocation = RevokedUserCertificate(
author=alice.device_id, timestamp=now, user_id=alice.user_id
).dump_and_sign(alice.signing_key)
rep = await user_revoke(alice_ws, revoked_user_certificate=alice_revocation)
assert isinstance(rep, UserRevokeRepNotAllowed)
@pytest.mark.trio
async def test_user_revoke_unknown(backend_asgi_app, alice_ws, alice, mallory):
revoked_user_certificate = RevokedUserCertificate(
author=alice.device_id, timestamp=DateTime.now(), user_id=mallory.user_id
).dump_and_sign(alice.signing_key)
rep = await user_revoke(alice_ws, revoked_user_certificate=revoked_user_certificate)
assert isinstance(rep, UserRevokeRepNotFound)
@pytest.mark.trio
async def test_user_revoke_already_revoked(backend_asgi_app, alice_ws, bob, alice):
now = DateTime.now()
bob_revocation = RevokedUserCertificate(
author=alice.device_id, timestamp=now, user_id=bob.user_id
).dump_and_sign(alice.signing_key)
rep = await user_revoke(alice_ws, revoked_user_certificate=bob_revocation)
assert isinstance(rep, UserRevokeRepOk)
rep = await user_revoke(alice_ws, revoked_user_certificate=bob_revocation)
assert isinstance(rep, UserRevokeRepAlreadyRevoked)
@pytest.mark.trio
async def METHOD_NAME(backend_asgi_app, alice_ws, alice2, bob):
revoked_user_certificate = RevokedUserCertificate(
author=alice2.device_id, timestamp=DateTime.now(), user_id=bob.user_id
).dump_and_sign(alice2.signing_key)
rep = await user_revoke(alice_ws, revoked_user_certificate=revoked_user_certificate)
assert isinstance(rep, UserRevokeRepInvalidCertification)
@pytest.mark.trio
async def test_user_revoke_certify_too_old(backend_asgi_app, alice_ws, alice, bob):
now = DateTime(2000, 1, 1)
revoked_user_certificate = RevokedUserCertificate(
author=alice.device_id, timestamp=now, user_id=bob.user_id
).dump_and_sign(alice.signing_key)
with freeze_time(now.add(seconds=INVITATION_VALIDITY + 1)):
rep = await user_revoke(alice_ws, revoked_user_certificate=revoked_user_certificate)
assert isinstance(rep, UserRevokeRepInvalidCertification)
@pytest.mark.trio
async def test_user_revoke_other_organization(
ws_from_other_organization_factory,
backend_authenticated_ws_factory,
backend_asgi_app,
alice,
bob,
):
# Organizations should be isolated even for organization admins
async with ws_from_other_organization_factory(
backend_asgi_app, mimic=alice.device_id, profile=UserProfile.ADMIN
) as sock:
revocation = RevokedUserCertificate(
author=sock.device.device_id, timestamp=DateTime.now(), user_id=bob.user_id
).dump_and_sign(sock.device.signing_key)
rep = await user_revoke(sock, revoked_user_certificate=revocation)
assert isinstance(rep, UserRevokeRepNotFound)
# Make sure bob still works
async with backend_authenticated_ws_factory(backend_asgi_app, bob):
pass | null |
test boltzmann typo deprecation | # -*- Mode: python; tab-width: 4; indent-tabs-mode:nil; coding:utf-8 -*-
# vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4 fileencoding=utf-8
#
# MDAnalysis --- https://www.mdanalysis.org
# Copyright (c) 2006-2017 The MDAnalysis Development Team and contributors
# (see the file AUTHORS for the full list of names)
#
# Released under the GNU Public Licence, v2 or any higher version
#
# Please cite your use of MDAnalysis in published work:
#
# R. J. Gowers, M. Linke, J. Barnoud, T. J. E. Reddy, M. N. Melo, S. L. Seyler,
# D. L. Dotson, J. Domanski, S. Buchoux, I. M. Kenney, and O. Beckstein.
# MDAnalysis: A Python package for the rapid analysis of molecular dynamics
# simulations. In S. Benthall and S. Rostrup editors, Proceedings of the 15th
# Python in Science Conference, pages 102-109, Austin, TX, 2016. SciPy.
# doi: 10.25080/majora-629e541a-00e
#
# N. Michaud-Agrawal, E. J. Denning, T. B. Woolf, and O. Beckstein.
# MDAnalysis: A Toolkit for the Analysis of Molecular Dynamics Simulations.
# J. Comput. Chem. 32 (2011), 2319--2327, doi:10.1002/jcc.21787
#
import pytest
import numpy as np
from numpy.testing import assert_equal, assert_almost_equal
from MDAnalysis import units
class TestUnitEncoding(object):
def test_unicode(self):
try:
assert_equal(units.lengthUnit_factor[u"\u212b"], 1.0)
except KeyError:
raise AssertionError("Unicode symbol for Angtrom not supported")
def test_unicode_encoding_with_symbol(self):
try:
assert_equal(units.lengthUnit_factor[u"Å"], 1.0)
except KeyError:
raise AssertionError("UTF-8-encoded symbol for Angtrom not supported")
class TestConstants(object):
# CODATA 2010 (NIST): http://physics.nist.gov/cuu/Constants/
# (accessed 2015-02-15)
# Add a reference value to this dict for every entry in
# units.constants
constants_reference = (
('N_Avogadro', 6.02214129e+23), # mol**-1
('elementary_charge', 1.602176565e-19), # As
('calorie', 4.184), # J
('Boltzmann_constant', 8.314462159e-3), # KJ (mol K)**-1
('Boltzman_constant', 8.314462159e-3), # remove in 2.8.0
('electric_constant', 5.526350e-3), # As (Angstroms Volts)**-1
)
@pytest.mark.parametrize('name, value', constants_reference)
def test_constant(self, name, value):
assert_almost_equal(units.constants[name], value)
def METHOD_NAME(self):
wmsg = ("Please use 'Boltzmann_constant' henceforth. The key "
"'Boltzman_constant' was a typo and will be removed "
"in MDAnalysis 2.8.0.")
with pytest.warns(DeprecationWarning, match=wmsg):
units.constants['Boltzman_constant']
class TestConversion(object):
@staticmethod
def _assert_almost_equal_convert(value, u1, u2, ref):
val = units.convert(value, u1, u2)
assert_almost_equal(val, ref,
err_msg="Conversion {0} --> {1} failed".format(u1, u2))
nm = 12.34567
A = nm * 10.
@pytest.mark.parametrize('quantity, unit1, unit2, ref', (
(nm, 'nm', 'A', A),
(A, 'Angstrom', 'nm', nm),
))
def test_length(self, quantity, unit1, unit2, ref):
self._assert_almost_equal_convert(quantity, unit1, unit2, ref)
@pytest.mark.parametrize('quantity, unit1, unit2, ref', (
(1, 'ps', 'AKMA', 20.45482949774598),
(1, 'AKMA', 'ps', 0.04888821),
(1, 'ps', 'ms', 1e-9),
(1, 'ms', 'ps', 1e9),
(1, 'ps', 'us', 1e-6),
(1, 'us', 'ps', 1e6),
))
def test_time(self, quantity, unit1, unit2, ref):
self._assert_almost_equal_convert(quantity, unit1, unit2, ref)
@pytest.mark.parametrize('quantity, unit1, unit2, ref', (
(1, 'kcal/mol', 'kJ/mol', 4.184),
(1, 'kcal/mol', 'eV', 0.0433641),
))
def test_energy(self, quantity, unit1, unit2, ref):
self._assert_almost_equal_convert(quantity, unit1, unit2, ref)
@pytest.mark.parametrize('quantity, unit1, unit2, ref', (
(1, 'kJ/(mol*A)', 'J/m', 1.66053892103219e-11),
(2.5, 'kJ/(mol*nm)', 'kJ/(mol*A)', 0.25),
(1, 'kcal/(mol*Angstrom)', 'kJ/(mol*Angstrom)', 4.184),
))
def test_force(self, quantity, unit1, unit2, ref):
self._assert_almost_equal_convert(quantity, unit1, unit2, ref)
@pytest.mark.parametrize('quantity, unit1, unit2, ref', (
(1, 'A/ps', 'm/s', 1e-10/1e-12),
(1, 'A/ps', 'nm/ps', 0.1),
(1, 'A/ps', 'pm/ps', 1e2),
(1, 'A/ms', 'A/ps', 1e9),
(1, 'A/us', 'A/ps', 1e6),
(1, 'A/fs', 'A/ps', 1e-3),
(1, 'A/AKMA', 'A/ps', 1/4.888821e-2),
))
def test_speed(self, quantity, unit1, unit2, ref):
self._assert_almost_equal_convert(quantity, unit1, unit2, ref)
@pytest.mark.parametrize('quantity, unit1, unit2', ((nm, 'Stone', 'nm'),
(nm, 'nm', 'Stone')))
def test_unit_unknown(self, quantity, unit1, unit2):
with pytest.raises(ValueError):
units.convert(quantity, unit1, unit2)
def test_unit_unconvertable(self):
nm = 12.34567
A = nm * 10.
with pytest.raises(ValueError):
units.convert(A, 'A', 'ps')
class TestBaseUnits:
@staticmethod
@pytest.fixture
def ref():
# This is a copy of the dictionary we expect.
# We want to know if base units are added or altered.
ref = {"length": "A",
"time": "ps",
"energy": "kJ/mol",
"charge": "e",
"force": "kJ/(mol*A)",
"speed": "A/ps"}
return ref
def test_MDANALYSIS_BASE_UNITS_correct(self, ref):
assert ref == units.MDANALYSIS_BASE_UNITS | null |
register all arguments | import argparse
import os
import sys
from collections.abc import Callable
from typing import IO
from xdsl.backend.riscv import riscv_scf_to_asm
from xdsl.backend.riscv.lowering import (
convert_arith_to_riscv,
convert_func_to_riscv_func,
convert_memref_to_riscv,
convert_scf_to_riscv_scf,
reduce_register_pressure,
)
from xdsl.dialects.affine import Affine
from xdsl.dialects.arith import Arith
from xdsl.dialects.builtin import Builtin, ModuleOp
from xdsl.dialects.cf import Cf
from xdsl.dialects.cmath import CMath
from xdsl.dialects.experimental.dmp import DMP
from xdsl.dialects.experimental.fir import FIR
from xdsl.dialects.experimental.math import Math
from xdsl.dialects.func import Func
from xdsl.dialects.gpu import GPU
from xdsl.dialects.irdl.irdl import IRDL
from xdsl.dialects.linalg import Linalg
from xdsl.dialects.llvm import LLVM
from xdsl.dialects.memref import MemRef
from xdsl.dialects.mpi import MPI
from xdsl.dialects.pdl import PDL
from xdsl.dialects.printf import Printf
from xdsl.dialects.riscv import RISCV
from xdsl.dialects.riscv_func import RISCV_Func
from xdsl.dialects.riscv_scf import RISCV_Scf
from xdsl.dialects.scf import Scf
from xdsl.dialects.snitch import Snitch
from xdsl.dialects.snitch_runtime import SnitchRuntime
from xdsl.dialects.stencil import Stencil
from xdsl.dialects.test import Test
from xdsl.dialects.vector import Vector
from xdsl.frontend.passes.desymref import DesymrefyPass
from xdsl.frontend.symref import Symref
from xdsl.ir import Dialect, MLContext
from xdsl.parser import Parser
from xdsl.passes import ModulePass
from xdsl.transforms import (
canonicalize,
canonicalize_dmp,
dead_code_elimination,
lower_affine,
lower_mpi,
lower_riscv_func,
lower_snitch,
lower_snitch_runtime,
mlir_opt,
printf_to_llvm,
printf_to_putchar,
reconcile_unrealized_casts,
riscv_register_allocation,
)
from xdsl.transforms.experimental import (
convert_stencil_to_ll_mlir,
stencil_shape_inference,
stencil_storage_materialization,
)
from xdsl.transforms.experimental.dmp import stencil_global_to_local
from xdsl.utils.exceptions import ParseError
def get_all_dialects() -> list[Dialect]:
"""Return the list of all available dialects."""
return [
Affine,
Arith,
Builtin,
Cf,
CMath,
DMP,
FIR,
Func,
GPU,
Linalg,
IRDL,
LLVM,
Math,
MemRef,
MPI,
PDL,
Printf,
RISCV,
RISCV_Func,
RISCV_Scf,
Scf,
Snitch,
SnitchRuntime,
Stencil,
Symref,
Test,
Vector,
]
def get_all_passes() -> list[type[ModulePass]]:
"""Return the list of all available passes."""
return [
canonicalize.CanonicalizePass,
canonicalize_dmp.CanonicalizeDmpPass,
convert_stencil_to_ll_mlir.ConvertStencilToLLMLIRPass,
dead_code_elimination.DeadCodeElimination,
DesymrefyPass,
stencil_global_to_local.GlobalStencilToLocalStencil2DHorizontal,
stencil_global_to_local.LowerHaloToMPI,
lower_affine.LowerAffinePass,
lower_mpi.LowerMPIPass,
lower_riscv_func.LowerRISCVFunc,
lower_snitch.LowerSnitchPass,
lower_snitch_runtime.LowerSnitchRuntimePass,
mlir_opt.MLIROptPass,
printf_to_llvm.PrintfToLLVM,
printf_to_putchar.PrintfToPutcharPass,
reduce_register_pressure.RiscvReduceRegisterPressurePass,
riscv_register_allocation.RISCVRegisterAllocation,
convert_arith_to_riscv.ConvertArithToRiscvPass,
convert_func_to_riscv_func.ConvertFuncToRiscvFuncPass,
convert_memref_to_riscv.ConvertMemrefToRiscvPass,
convert_scf_to_riscv_scf.ConvertScfToRiscvPass,
riscv_scf_to_asm.LowerScfForToLabels,
stencil_shape_inference.StencilShapeInferencePass,
stencil_storage_materialization.StencilStorageMaterializationPass,
reconcile_unrealized_casts.ReconcileUnrealizedCastsPass,
]
class CommandLineTool:
ctx: MLContext
args: argparse.Namespace
"""
The argument parsers namespace which holds the parsed commandline
attributes.
"""
available_frontends: dict[str, Callable[[IO[str]], ModuleOp]]
"""
A mapping from file extension to a frontend that can handle this
file type.
"""
def METHOD_NAME(self, arg_parser: argparse.ArgumentParser):
arg_parser.add_argument(
"input_file", type=str, nargs="?", help="path to input file"
)
frontends = [name for name in self.available_frontends]
arg_parser.add_argument(
"-f",
"--frontend",
type=str,
required=False,
choices=frontends,
help="Frontend to be used for the input. If not set, "
"the xdsl frontend or the one for the file extension "
"is used.",
)
arg_parser.add_argument("--disable-verify", default=False, action="store_true")
arg_parser.add_argument(
"--allow-unregistered-dialect",
default=False,
action="store_true",
help="Allow the parsing of unregistered dialects.",
)
arg_parser.add_argument(
"--no-implicit-module",
default=False,
action="store_true",
help="Disable implicit addition of a top-level module op during parsing.",
)
def get_input_stream(self) -> tuple[IO[str], str]:
"""
Get the input stream to parse from, along with the file extension.
"""
if self.args.input_file is None:
f = sys.stdin
file_extension = "mlir"
else:
f = open(self.args.input_file)
_, file_extension = os.path.splitext(self.args.input_file)
file_extension = file_extension.replace(".", "")
return f, file_extension
def get_input_name(self):
return self.args.input_file or "stdin"
def register_all_dialects(self):
"""
Register all dialects that can be used.
Add other/additional dialects by overloading this function.
"""
for dialect in get_all_dialects():
self.ctx.register_dialect(dialect)
def register_all_frontends(self):
"""
Register all frontends that can be used.
Add other/additional frontends by overloading this function.
"""
def parse_mlir(io: IO[str]):
return Parser(
self.ctx,
io.read(),
self.get_input_name(),
).parse_module(not self.args.no_implicit_module)
self.available_frontends["mlir"] = parse_mlir
def parse_chunk(self, chunk: IO[str], file_extension: str) -> ModuleOp | None:
"""
Parse the input file by invoking the parser specified by the `parser`
argument. If not set, the parser registered for this file extension
is used.
"""
try:
return self.available_frontends[file_extension](chunk)
except ParseError as e:
if "parsing_diagnostics" in self.args and self.args.parsing_diagnostics:
print(e.with_context())
else:
raise Exception("Failed to parse:\n" + e.with_context()) from e
finally:
chunk.close() | null |
make app | import os
import uuid
from unittest import mock
from pyramid import testing
from kinto.core import initialization
from kinto.core.events import ACTIONS, ResourceChanged, ResourceRead
from kinto.core.listeners import ListenerBase
from kinto.core.testing import unittest
UID = str(uuid.uuid4())
class ViewSet:
def get_name(*args, **kw):
return "mushrooms"
class Service:
viewset = ViewSet()
class Match:
cornice_services = {"watev": Service()}
pattern = "watev"
class Request:
path = "/1/school/students/"
prefixed_userid = "tarek"
matchdict = {"id": UID}
registry = matched_route = Match()
current_resource_name = "school"
class ListenerSetupTest(unittest.TestCase):
def setUp(self):
demo_patch = mock.patch("tests.core.listeners.load_from_config")
self.addCleanup(demo_patch.stop)
self.demo_mocked = demo_patch.start()
def METHOD_NAME(self, extra_settings={}):
settings = {"event_listeners": "tests.core.listeners"}
settings.update(**extra_settings)
config = testing.setUp(settings=settings)
config.commit()
initialization.setup_listeners(config)
return config
def test_listener_module_is_specified_via_settings(self):
self.METHOD_NAME(
{"event_listeners": "demo", "event_listeners.demo.use": "tests.core.listeners"}
)
self.assertTrue(self.demo_mocked.called)
def test_listener_module_can_be_specified_via_listeners_list(self):
self.METHOD_NAME(
{"event_listeners": "demo", "event_listeners.demo.use": "tests.core.listeners"}
)
self.assertTrue(self.demo_mocked.called)
def test_callback_called_when_action_is_not_filtered(self):
config = self.METHOD_NAME(
{"event_listeners": "demo", "event_listeners.demo.use": "tests.core.listeners"}
)
ev = ResourceChanged({"action": ACTIONS.CREATE.value}, [], Request())
config.registry.notify(ev)
self.assertTrue(self.demo_mocked.return_value.called)
def test_callback_is_not_called_when_action_is_filtered(self):
config = self.METHOD_NAME(
{
"event_listeners": "demo",
"event_listeners.demo.use": "tests.core.listeners",
"event_listeners.demo.actions": "delete",
}
)
ev = ResourceChanged({"action": ACTIONS.CREATE.value}, [], Request())
config.registry.notify(ev)
self.assertFalse(self.demo_mocked.return_value.called)
def test_callback_called_when_resource_is_not_filtered(self):
config = self.METHOD_NAME(
{"event_listeners": "demo", "event_listeners.demo.use": "tests.core.listeners"}
)
event = ResourceChanged(
{"action": ACTIONS.CREATE.value, "resource_name": "mushroom"}, [], Request()
)
config.registry.notify(event)
self.assertTrue(self.demo_mocked.return_value.called)
def test_callback_is_not_called_when_resource_is_filtered(self):
config = self.METHOD_NAME(
{
"event_listeners": "demo",
"event_listeners.demo.use": "tests.core.listeners",
"event_listeners.demo.resources": "toad",
}
)
event = ResourceChanged(
{"action": ACTIONS.CREATE.value, "resource_name": "mushroom"}, [], Request()
)
config.registry.notify(event)
self.assertFalse(self.demo_mocked.return_value.called)
def test_callback_is_not_called_on_read_by_default(self):
config = self.METHOD_NAME(
{"event_listeners": "demo", "event_listeners.demo.use": "tests.core.listeners"}
)
event = ResourceRead({"action": ACTIONS.READ.value}, [], Request())
config.registry.notify(event)
self.assertFalse(self.demo_mocked.return_value.called)
def test_callback_is_called_on_read_if_specified(self):
config = self.METHOD_NAME(
{
"event_listeners": "demo",
"event_listeners.demo.use": "tests.core.listeners",
"event_listeners.demo.actions": "read",
}
)
event = ResourceRead({"action": ACTIONS.READ.value}, [], Request())
config.registry.notify(event)
self.assertTrue(self.demo_mocked.return_value.called)
def test_same_callback_is_called_for_read_and_write_specified(self):
config = self.METHOD_NAME(
{
"event_listeners": "demo",
"event_listeners.demo.use": "tests.core.listeners",
"event_listeners.demo.actions": "read create delete",
}
)
ev = ResourceRead({"action": ACTIONS.READ.value}, [], Request())
config.registry.notify(ev)
ev = ResourceChanged({"action": ACTIONS.CREATE.value}, [], Request())
config.registry.notify(ev)
self.assertEqual(self.demo_mocked.return_value.call_count, 2)
def test_loading_can_read_configuration_from_environment(self):
environ = {
"KINTO_EVENT_LISTENERS": "kvstore",
"KINTO_EVENT_LISTENERS_KVSTORE_USE": "tests.core.listeners",
"KINTO_EVENT_LISTENERS_KVSTORE_URL": "demo://demo:6379/0",
"KINTO_EVENT_LISTENERS_KVSTORE_POOL_SIZE": "5",
"KINTO_EVENT_LISTENERS_KVSTORE_LISTNAME": "queue",
"KINTO_EVENT_LISTENERS_KVSTORE_ACTIONS": "delete",
"KINTO_EVENT_LISTENERS_KVSTORE_RESOURCES": "toad",
}
os.environ.update(**environ)
config = self.METHOD_NAME(
{
# With real/full initialization, these should not be necessary:
"settings_prefix": "kinto",
"event_listeners": "kvstore",
}
)
# Listener is instantiated.
self.assertTrue(self.demo_mocked.called)
# Action filtering is read from ENV.
event = ResourceChanged(
{"action": ACTIONS.DELETE.value, "resource_name": "toad"}, [], Request()
)
config.registry.notify(event)
self.assertTrue(self.demo_mocked.return_value.called)
self.demo_mocked.reset_mock()
# Action filtering is read from ENV.
event = ResourceChanged({"action": ACTIONS.CREATE.value}, [], Request())
config.registry.notify(event)
self.assertFalse(self.demo_mocked.return_value.called)
# Resource filtering is read from ENV.
event = ResourceChanged(
{"action": ACTIONS.CREATE.value, "resource_name": "mushroom"}, [], Request()
)
config.registry.notify(event)
self.assertFalse(self.demo_mocked.return_value.called)
# Clean-up.
for k in environ.keys():
os.environ.pop(k)
class ListenerBaseTest(unittest.TestCase):
def test_not_implemented(self):
# make sure we can't use the base listener
listener = ListenerBase()
self.assertRaises(NotImplementedError, listener, object()) | null |
restore backup from xml file | # coding: utf-8
from datetime import datetime
import errno
import os
import shutil
import sys
import tempfile
import zipfile
from time import sleep
from onadata.apps.logger.import_tools import django_file
from onadata.apps.logger.models import Instance
from onadata.libs.utils.logger_tools import create_instance
from onadata.libs.utils.model_tools import queryset_iterator
DATE_FORMAT = "%Y-%m-%d-%H-%M-%S"
def _date_created_from_filename(filename):
base_name, ext = os.path.splitext(filename)
parts = base_name.split("-")
if len(parts) < 6:
raise ValueError(
"Inavlid filename - it must be in the form"
" 'YYYY-MM-DD-HH-MM-SS[-i].xml'")
parts_dict = dict(
zip(["year", "month", "day", "hour", "min", "sec"], parts))
return datetime.strptime(
"%(year)s-%(month)s-%(day)s-%(hour)s-%(min)s-%(sec)s" %
parts_dict, DATE_FORMAT)
def create_zip_backup(zip_output_file, user, xform=None):
# create a temp dir that we'll create our structure within and zip it
# when we are done
tmp_dir_path = tempfile.mkdtemp()
instances_path = os.path.join(tmp_dir_path, "instances")
# get the xls file from storage
# for each submission in the database - create an xml file in this
# form
# /<id_string>/YYYY/MM/DD/YYYY-MM-DD-HH-MM-SS.xml
qs = Instance.objects.filter(xform__user=user)
if xform:
qs = qs.filter(xform=xform)
num_instances = qs.count()
done = 0
sys.stdout.write("Creating XML Instances\n")
for instance in queryset_iterator(qs, 100):
# get submission time
date_time_str = instance.date_created.strftime(DATE_FORMAT)
date_parts = date_time_str.split("-")
sub_dirs = os.path.join(*date_parts[:3])
# create the directories
full_path = os.path.join(instances_path, sub_dirs)
if not os.path.exists(full_path):
try:
os.makedirs(full_path)
except OSError as e:
if e.errno != errno.EEXIST:
raise
full_xml_path = os.path.join(full_path, date_time_str + ".xml")
# check for duplicate file names
file_index = 1
while os.path.exists(full_xml_path):
full_xml_path = os.path.join(
full_path, "%s-%d.xml" % (date_time_str, file_index))
file_index += 1
# create the instance xml
with open(full_xml_path, 'wb') as f:
f.write(instance.xml.encode())
done += 1
sys.stdout.write("\r%.2f %% done" % (
float(done)/float(num_instances) * 100))
sys.stdout.flush()
sleep(0)
# write zip file
sys.stdout.write("\nWriting to ZIP arhive.\n")
zf = zipfile.ZipFile(zip_output_file, "w")
done = 0
for dir_path, dir_names, file_names in os.walk(tmp_dir_path):
for file_name in file_names:
archive_path = dir_path.replace(tmp_dir_path + os.path.sep,
"", 1)
zf.write(os.path.join(dir_path, file_name),
os.path.join(archive_path, file_name))
done += 1
sys.stdout.write("\r%.2f %% done" % (
float(done)/float(num_instances) * 100))
sys.stdout.flush()
sleep(0)
zf.close()
# removed dir tree
shutil.rmtree(tmp_dir_path)
sys.stdout.write("\nBackup saved to %s\n" % zip_output_file)
def restore_backup_from_zip(zip_file_path, username):
try:
temp_directory = tempfile.mkdtemp()
zf = zipfile.ZipFile(zip_file_path)
zf.extractall(temp_directory)
except zipfile.BadZipfile:
sys.stderr.write("Bad zip arhcive.")
else:
return restore_backup_from_path(temp_directory, username, "backup")
finally:
shutil.rmtree(temp_directory)
def METHOD_NAME(xml_instance_path, username):
# check if its a valid xml instance
file_name = os.path.basename(xml_instance_path)
xml_file = django_file(
xml_instance_path,
field_name="xml_file",
content_type="text/xml")
media_files = []
try:
date_created = _date_created_from_filename(file_name)
except ValueError as e:
sys.stderr.write(
"Couldn't determine date created from filename: '%s'\n" %
file_name)
date_created = datetime.now()
sys.stdout.write("Creating instance from '%s'\n" % file_name)
try:
create_instance(
username, xml_file, media_files,
date_created_override=date_created)
return 1
except Exception as e:
sys.stderr.write(
"Could not restore %s, create instance said: %s\n" %
(file_name, e))
return 0
def restore_backup_from_path(dir_path, username, status):
"""
Only restores xml submissions, media files are assumed to still be in
storage and will be retrieved by the filename stored within the submission
"""
num_instances = 0
num_restored = 0
for dir_path, dir_names, file_names in os.walk(dir_path):
for file_name in file_names:
# check if its a valid xml instance
xml_instance_path = os.path.join(dir_path, file_name)
num_instances += 1
num_restored += METHOD_NAME(
xml_instance_path,
username)
return num_instances, num_restored | null |
always iterable | import sys, os
from glob import glob
from shutil import copyfile
import fnmatch
import yaml, pprint
import subprocess
import jinja2
from itertools import *
import concurrent.futures
def ParseAruments() -> (dict, dict):
import argparse
parser = argparse.ArgumentParser()
parser.add_argument('configPath')
parser.add_argument('-n', '--dryRun', action='store_true')
parser.add_argument('-j', '--jobs', type=int, default=1)
parser.add_argument('-o', '--outDir')
parser.add_argument('-t', '--testName', help="Run specific test from config.yaml. By default will run all tests.")
args = parser.parse_args()
with open(args.configPath, "r") as f:
environment = jinja2.Environment()
template = environment.from_string(f.read())
config, tests = yaml.safe_load_all(template.render(glob=glob))
config['dryRun'] = args.dryRun
config['jobs'] = args.jobs
if args.outDir: config['outDir'] = args.outDir
config['onlyRunTestName'] = args.testName
config.setdefault('resolution', [600, 480])
return config, tests
def mkdir(path):
os.makedirs(path, exist_ok=True)
def extract(d:dict, k, default=None):
v = d.get(k, default)
if k in d: del d[k]
return v
def alwaysList(x):
if x is None: return []
if isinstance(x, list): return x
if isinstance(x, dict): return list(x.items())
return [x]
def alwaysDict(x):
if x is None: return {}
if isinstance(x, dict): return x
if isinstance(x, list):
assert len(set(x)) == len(x)
return {k:None for k in x}
return {x: None}
def METHOD_NAME(x):
if x is None: return iter(())
if isinstance(x, str): return iter((x,))
try: return iter(x)
except TypeError: return iter((x,))
def locateRelPath(path, root):
if os.path.isabs(path):
return path
return os.path.join(root, path)
def GetConfigForCompletedRun(newConfig):
if not os.path.exists(newConfig['outputLog']): return None
with open(newConfig['outputLog'], "r") as log:
lines = log.readlines()
oldConfigYaml = ''.join(lines[lines.index("=========== CONFIG ===========\n")+1:lines.index("==============================\n")])
return yaml.safe_load(oldConfigYaml)
import signal
import psutil
gotKeyboardInterrupt = False
def signal_handler(sig, frame):
global gotKeyboardInterrupt
gotKeyboardInterrupt = True
current_process = psutil.Process()
children = current_process.children(recursive=True)
for child in children:
signal.pthread_kill(child.pid, signal.SIGKILL)
print('******* Jobs cancelled *******')
signal.signal(signal.SIGINT, signal_handler)
def RunTest(config):
if os.path.exists(config['output']) and os.path.exists(config['outputLog']):
with open(config['outputLog'], "r") as log:
if "=========== TEST COMPLETE ===========" in log.read():
if config == GetConfigForCompletedRun(config):
# print("Skipping test", config['name'], config['renderer'])
return
else:
print("Config changed for", config['name'])
print("Run test", config['name'], config['renderer'])
cmd = ['python', '-u', 'run_test.py', yaml.dump(config)]
with open(config['outputLog'], "w", 1) as log:
try:
subprocess.run(cmd, stdout=log, stderr=subprocess.STDOUT, text=True)
except Exception:
pass
if gotKeyboardInterrupt:
print("Stopped", config['name'], config['renderer'])
return
print("\n=========== TEST COMPLETE ===========\n", file=log)
if not os.path.exists(config['output']):
copyfile("missing.png", config['output'])
def TestCanBeRunConcurrently(test:dict) -> bool:
return test.get('runConcurrently', True)
def GenerateTests(config, tests):
for name, test in tests.items():
files = [locateRelPath(f, config['dataRootDir']) for f in alwaysList(extract(test, 'files'))]
files = sorted(chain.from_iterable(map(glob, files)))
dataType = extract(test, 'type')
if extract(test, 'skip'):
continue
renderers:dict = alwaysDict(extract(test, 'renderers', config['rendererSets']['default']))
for k in renderers.copy():
if k in config['rendererSets']:
del renderers[k]
for r, v in alwaysDict(config['rendererSets'][k]).items():
renderers.setdefault(r, v)
for ren, renConfig in renderers.items():
if isinstance(renConfig, str):
renConfig = {'variable': renConfig}
renConfig = alwaysDict(renConfig)
variableKeys = [k for k in renConfig if 'variable' in k.lower()]
variables = [i for i in renConfig.items() if i[0] in variableKeys]
for k,_ in variables:
del renConfig[k]
varCombos = [*product(*[[*zip(repeat(k), alwaysList(v))] for k,v in variables])]
for combo in varCombos:
fullName = f"{name}-{ren}"
if len(varCombos) > 1:
fullName += '-'.join(['', *chain.from_iterable(combo)])
fName = fullName + ".png"
fName = fName.replace('/', '-')
yield {
'name': name,
'outputDir': config['outDir'],
'outputName': fName,
'output': f"{config['outDir']}/{fName}",
'outputSession': f"{config['outDir']}/meta/{fName}.vs3",
'outputLog': f"{config['outDir']}/meta/{fName}.log",
'files': files,
'type': dataType,
'renderer': ren,
'dryRun': config['dryRun'],
'resolution': config['resolution'],
**dict(combo),
**renConfig,
**test,
}
if __name__ == "__main__":
config, testConfigs = ParseAruments()
# print("config = ", end='')
# pprint.PrettyPrinter(indent=4).pprint(config)
mkdir(config['outDir'])
mkdir(f"{config['outDir']}/meta")
tests = GenerateTests(config, testConfigs)
if config['onlyRunTestName']:
tests = filter(lambda t: fnmatch.fnmatch(t['name'], config['onlyRunTestName']), tests)
tests = list(tests)
nJobs = config.get('jobs', 1)
if nJobs > 1:
concurrentTests = filter(TestCanBeRunConcurrently, tests)
singleTests = filter(lambda t: not TestCanBeRunConcurrently(t), tests)
else:
singleTests = tests
concurrentTests = []
if concurrentTests:
print(f"Running concurrent tests with {nJobs} jobs")
with concurrent.futures.ThreadPoolExecutor(max_workers=nJobs) as executor:
executor.map(RunTest, concurrentTests)
if singleTests:
print(f"Running single-threaded tests")
for test in singleTests:
RunTest(test) | null |
eval func | import os
import time
import argparse
from datasets import load_dataset
from transformers import WhisperForConditionalGeneration, WhisperProcessor
import torch
from evaluate import load
parser = argparse.ArgumentParser()
parser.add_argument('--int8', dest='int8', action='store_true')
parser.add_argument('--tune', dest='tune', action='store_true',
help='tune best int8 model with Neural Compressor on calibration dataset')
parser.add_argument('--accuracy_only', dest='accuracy_only', action='store_true',
help='run accuracy_only')
parser.add_argument('--benchmark', dest='benchmark', action='store_true',
help='run benchmark')
parser.add_argument('--batch_size', default=1, type=int,
help='For accuracy measurement only.')
parser.add_argument('--iters', default=0, type=int,
help='For accuracy measurement only.')
parser.add_argument('--warmup_iters', default=5, type=int,
help='For benchmark measurement only.')
parser.add_argument('--output_dir', default="saved_results", type=str,
help='the folder path to save the results.')
parser.add_argument('--cache_dir', default=None, type=str,
help='the folder path to save the results.')
args = parser.parse_args()
model_name = 'openai/whisper-large'
processor = WhisperProcessor.from_pretrained(model_name)
model = WhisperForConditionalGeneration.from_pretrained(model_name)
# dataset
librispeech_test_clean = load_dataset("librispeech_asr", "clean", split="test", cache_dir=args.cache_dir)
# metric
wer = load("wer")
def METHOD_NAME(model):
predictions = []
references = []
for batch in librispeech_test_clean:
audio = batch["audio"]
input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
reference = processor.tokenizer._normalize(batch['text'])
references.append(reference)
with torch.no_grad():
predicted_ids = model.generate(input_features)[0]
transcription = processor.decode(predicted_ids)
prediction = processor.tokenizer._normalize(transcription)
predictions.append(prediction)
wer_result = wer.compute(references=references, predictions=predictions)
print(f"Result wer: {wer_result * 100}")
accuracy = 1 - wer_result
print("Accuracy: %.5f" % accuracy)
return accuracy
if args.tune:
from neural_compressor import PostTrainingQuantConfig, quantization
op_type_dict = {
"Embedding": {"weight": {"dtype": ["fp32"]}, "activation": {"dtype": ["fp32"]}}
}
conf = PostTrainingQuantConfig(approach="dynamic", op_type_dict=op_type_dict)
q_model = quantization.fit(model,
conf=conf,
METHOD_NAME=METHOD_NAME)
q_model.save(args.output_dir)
exit(0)
#benchmark
if args.int8:
from neural_compressor.utils.pytorch import load
model = load(
os.path.abspath(os.path.expanduser(args.output_dir)), model)
if args.accuracy_only:
METHOD_NAME(model)
exit(0)
if args.benchmark:
from neural_compressor.config import BenchmarkConfig
from neural_compressor import benchmark
def b_func(model):
total_time = 0
for i, batch in enumerate(librispeech_test_clean):
if i > args.iters:
break
audio = batch["audio"]
input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
tic = time.time()
with torch.no_grad():
predicted_ids = model.generate(input_features)[0]
toc = time.time()
if i >= args.warmup_iters:
total_time += (toc - tic)
latency = total_time / (args.iters - args.warmup_iters)
print('Latency: %.3f ms' % (latency * 1000))
print('Throughput: %.3f images/sec' % (args.batch_size / latency))
print('Batch size = %d' % args.batch_size)
b_conf = BenchmarkConfig(
cores_per_instance=4,
num_of_instance=1
)
benchmark.fit(model, b_conf, b_func=b_func)
exit(0) | null |
test json gather array | import csv
import json
from io import StringIO
from urllib import parse
import numpy as np
from webtest import TestApp
import tangos
import tangos.testing.simulation_generator
import tangos.web
from tangos import testing
def setup_module():
testing.init_blank_db_for_testing()
creator = tangos.testing.simulation_generator.SimulationGeneratorForTests()
halo_offset = 0
for ts in range(1,4):
creator.add_timestep()
creator.add_objects_to_timestep(4)
creator.link_last_halos()
tangos.get_default_session().commit()
tangos.get_item("sim/ts1/halo_1")['test_image'] = np.zeros((500,500,3))
for ts in tangos.get_simulation(1).timesteps:
for h in ts.halos:
h['test_value'] = 1.0
tangos.get_default_session().commit()
creator.add_timestep() # add ts 4
creator.add_objects_to_timestep(3)
tangos.get_default_session().commit()
# now make the object have halo numbers which have a different ordering to the database IDs
tangos.get_item("sim/ts4/1")['test_value'] = 1.0
tangos.get_item("sim/ts4/2")['test_value'] = 2.0
tangos.get_item("sim/ts4/3")['test_value'] = 3.0
tangos.get_item("sim/ts4/1").halo_number = 10
tangos.get_default_session().commit()
creator = tangos.testing.simulation_generator.SimulationGeneratorForTests("simname/has/slashes")
creator.add_timestep()
creator.add_objects_to_timestep(1)
creator.add_timestep()
creator.add_objects_to_timestep(1)
creator.link_last_halos()
tangos.get_simulation(2).timesteps[0].halos[0]['test_value'] = 2.0
tangos.get_default_session().commit()
global app
app = TestApp(tangos.web.main({}))
def teardown_module():
tangos.core.close_db()
def test_root_page():
response = app.get("/")
assert response.status_int==200
assert "table" in response
def test_simulation_page():
response = app.get("/sim")
assert response.status_int==200
assert "table" in response
def test_timestep_page():
response = app.get("/sim/ts1")
assert response.status_int==200
assert "table" in response
def test_halo_page():
response = app.get("/sim/ts1/halo_1")
assert response.status_int == 200
assert "table" in response
def test_plot():
response = app.get("/sim/ts1/halo_1/t()/vs/z().png")
assert response.status_int == 200
assert response.content_type == 'image/png'
def test_plot_as_csv_timeseries():
response = app.get("/sim/ts3/halo_1/test_value/vs/z().csv")
assert response.status_int == 200
assert response.content_type == 'text/csv'
assert "filename=timeseries_test_value_vs_z().csv" in response.content_disposition
csv_reader = csv.reader(StringIO(response.body.decode('utf-8')))
csv_rows = list(csv_reader)
assert csv_rows[0]==['test_value', 'z()']
assert csv_rows[1]==['1.0','6.0']
assert csv_rows[2]==['1.0', '7.0']
assert csv_rows[3]==['1.0', '8.0']
def test_plot_as_csv_timestep():
response = app.get("/sim/ts3/test_value/vs/halo_number().csv")
assert response.status_int == 200
assert response.content_type == 'text/csv'
assert "filename=timestep_test_value_vs_halo_number().csv" in response.content_disposition
csv_reader = csv.reader(StringIO(response.body.decode('utf-8')))
csv_rows = list(csv_reader)
assert csv_rows[0]==['test_value', 'halo_number()']
assert csv_rows[1]==['1.0','1.0']
assert csv_rows[2] == ['1.0', '2.0']
assert csv_rows[3] == ['1.0', '3.0']
assert csv_rows[4] == ['1.0', '4.0']
def test_image_plot():
response = app.get("/sim/ts1/halo_1/test_image.png")
assert response.status_int == 200
assert response.content_type == 'image/png'
def test_json_gather_float():
response = app.get("/sim/ts1/gather/halo/test_value.json")
assert response.content_type == 'application/json'
assert response.status_int == 200
result = json.loads(response.body.decode('utf-8'))
assert result['timestep']=='ts1'
assert result['data_formatted']==["1.00", "1.00", "1.00", "1.00"]
assert result['is_number'] is True
assert result['is_boolean'] is False
assert result['is_array'] is False
def METHOD_NAME():
response = app.get("/sim/ts1/gather/halo/test_image.json")
assert response.content_type == 'application/json'
assert response.status_int == 200
result = json.loads(response.body.decode('utf-8'))
assert result['timestep']=='ts1'
assert result['data_formatted'][0]=="Array"
assert result['is_number'] is False
assert result['is_boolean'] is False
assert result['is_array'] is True
def test_json_gather_bool():
response = app.get("/sim/ts1/gather/halo/has_property(test_image).json")
assert response.content_type == 'application/json'
assert response.status_int == 200
result = json.loads(response.body.decode('utf-8'))
assert result['timestep'] == 'ts1'
assert result['data_formatted'] == ["True", "False", "False", "False"]
assert result['is_number'] is False
assert result['is_boolean'] is True
assert result['is_array'] is False
def test_simulation_with_slash():
response = app.get("/")
assert "simname/has/slashes" in response
simpage_response = response.click("simname/has/slashes")
assert "Simulation: simname/has/slashes" in simpage_response
ts_response = simpage_response.click("Go", index=1)
assert "Timestep: ts1" in ts_response
# Unfortunately timestep page is now generated in javascript, so we can
# no longer test ts_response.click("Go")
halo_response = app.get("/simname_has_slashes/ts1/1")
assert "halo 1 of ts1" in halo_response
calculate_url = halo_response.pyquery("#calculate_url").text()
calculate_url = parse.unquote(calculate_url)
assert "simname_has_slashes" in calculate_url
halo_next_step_response = halo_response.click(r"\+1$").follow()
assert "halo 1 of ts2" in halo_next_step_response
def test_ordering_as_expected():
"""Tests for an issue where data returned to the web interface was in a different order to the initial table,
causing results to be displayed out of order"""
assert (tangos.get_item("sim/ts4").calculate_all("halo_number()") == np.array([10,2,3])).all()
assert (tangos.get_item("sim/ts4").calculate_all("halo_number()",order_by_halo_number=True) == np.array([2,3,10])).all()
response = app.get("/sim/ts4/gather/halo/test_value.json")
assert response.content_type == 'application/json'
assert response.status_int == 200
result = json.loads(response.body.decode('utf-8'))
assert result['timestep'] == 'ts4'
assert result['data_formatted'] == ["2.00", "3.00", "1.00"] | null |
test duplicated params | # Owner(s): ["module: inductor"]
import copy
import functools
import unittest
import torch
import torch._export
import torch._inductor
import torch.fx._pytree as fx_pytree
from torch._dynamo.testing import same
from torch.testing._internal.common_utils import IS_FBCODE, TEST_WITH_ROCM, TestCase
from torch.testing._internal.inductor_utils import HAS_CUDA
from torch.utils import _pytree as pytree
aten = torch.ops.aten
requires_cuda = functools.partial(unittest.skipIf, not HAS_CUDA, "requires cuda")
class AOTInductorModelRunner:
@classmethod
def load(cls, model, example_inputs, example_outputs, options=None):
# AOTInductorModel relies on the caller to pass in output_tensors,
# so we need to explicitly allocate output tensors here.
output_tensors = []
example_outputs, output_spec = pytree.tree_flatten(example_outputs)
for output in example_outputs:
output_tensors.append(torch.empty_like(output))
# The exact API is subject to change
so_path, exported = torch._export.aot_compile(
model,
example_inputs,
options=options,
)
# Use a utility function for easier testing
source = """
#include <torch/csrc/inductor/aot_runtime/model.h>
torch::aot_inductor::AOTInductorModel model;
void run(
const std::vector<at::Tensor>& input_tensors,
std::vector<at::Tensor>& output_tensors) {
model.run(input_tensors, output_tensors, at::cuda::getCurrentCUDAStream());
}
"""
optimized = torch.utils.cpp_extension.load_inline(
name="aot_inductor",
cpp_sources=[source],
functions=["run"],
extra_ldflags=[so_path],
with_cuda=True,
).run
return optimized, exported, output_tensors, output_spec
@classmethod
def run(cls, model, example_inputs, example_outputs, options=None):
example_outputs = copy.deepcopy(example_outputs)
optimized, exported, output_tensors, output_spec = AOTInductorModelRunner.load(
model, example_inputs, example_outputs, options
)
param_buffer_values = list(exported.state_dict.values())
flat_example_inputs = fx_pytree.tree_flatten_spec(
example_inputs, exported.call_spec.in_spec
)
all_args = (*param_buffer_values, *flat_example_inputs)
optimized(all_args, output_tensors)
return pytree.tree_unflatten(output_tensors, output_spec)
class AotInductorTests(TestCase):
def test_simple(self):
class Repro(torch.nn.Module):
def __init__(self):
super().__init__()
self.weight = torch.randn(10, 10, device="cuda")
def forward(self, x, y):
return x + torch.nn.functional.linear(y, self.weight)
model = Repro()
example_inputs = (
torch.randn(10, 10, device="cuda"),
torch.randn(10, 10, device="cuda"),
)
expected = model(*example_inputs)
actual = AOTInductorModelRunner.run(model, example_inputs, expected)
self.assertTrue(same(actual, expected))
def test_missing_output(self):
class Repro(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, x, y):
a = torch.sin(x)
b = torch.mm(a, y)
c = torch.cos(b)
return c
model = Repro()
example_inputs = (
torch.randn(10, 10, device="cuda"),
torch.randn(10, 10, device="cuda"),
)
expected = model(*example_inputs)
actual = AOTInductorModelRunner.run(model, example_inputs, expected)
self.assertTrue(same(actual, expected))
def test_output_misaligned(self):
class Repro(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, x, y):
x_unsqueeze = torch.unsqueeze(x, dim=0)
y_unsqueeze = torch.unsqueeze(y, dim=0)
cat = torch.cat([x_unsqueeze, y_unsqueeze], dim=0)
x_getitem = cat[0]
y_getitem = cat[1]
x_sigmoid = torch.sigmoid(x_getitem)
return x_sigmoid, y_getitem
model = Repro()
example_inputs = (
torch.randn(10, 10, device="cuda"),
torch.randn(10, 10, device="cuda"),
)
expected = model(*example_inputs)
actual = AOTInductorModelRunner.run(model, example_inputs, expected)
self.assertTrue(same(actual, expected))
def test_dynamic_smem_above_default_limit(self):
class Repro(torch.nn.Module):
def forward(self, x, y):
return x @ y
model = Repro()
# on A100, the generated Triton kernel for this MM
# requires 55296 bytes of dynamic SMEM which is above
# the A100's default dynamic SMEM limit of 49152 bytes.
example_inputs = (
torch.randn(10285, 96, device="cuda"),
torch.randn(96, 1, device="cuda"),
)
expected = model(*example_inputs)
actual = AOTInductorModelRunner.run(
model,
example_inputs,
expected,
options={
"max_autotune": True,
"max_autotune_gemm_backends": "TRITON",
},
)
self.assertTrue(same(actual, expected))
def test_addmm(self):
class Model(torch.nn.Module):
def __init__(self, n, k):
super().__init__()
self.weight = torch.randn(n, k, device="cuda")
self.bias = torch.randn(n, device="cuda")
def forward(self, a):
return torch.nn.functional.linear(a, self.weight, self.bias)
M = 8
N = 6
K = 16
model = Model(N, K)
batch = 2
a = torch.randn(batch, M, K, device="cuda")
example_inputs = (a,)
expected = model(*example_inputs)
actual = AOTInductorModelRunner.run(model, example_inputs, expected)
self.assertTrue(same(actual, expected))
def METHOD_NAME(self):
class Model(torch.nn.Module):
def __init__(self):
super().__init__()
self.p = torch.nn.Parameter(torch.rand(6))
self.q = self.p
def forward(self, x):
return self.p * x + self.q
model = Model()
example_inputs = (torch.rand(6),)
expected = model(*example_inputs)
actual = torch._export.export(model, example_inputs)(*example_inputs)
self.assertTrue(same(actual, expected))
if __name__ == "__main__":
from torch._dynamo.test_case import run_tests
# cpp_extension N/A in fbcode
if HAS_CUDA and not TEST_WITH_ROCM and not IS_FBCODE:
run_tests(needs="filelock") | null |
test link only | import os
import pandas as pd
import pyspark.sql.functions as f
import pytest
from pyspark.sql.types import StringType, StructField, StructType
import splink.spark.comparison_level_library as cll
import splink.spark.comparison_library as cl
from splink.spark.linker import SparkLinker
from .basic_settings import get_settings_dict, name_comparison
from .decorator import mark_with_dialects_including
from .linker_utils import (
_test_write_functionality,
register_roc_data,
)
def test_full_example_spark(df_spark, tmp_path):
# Annoyingly, this needs an independent linker as csv doesn't
# accept arrays as inputs, which we are adding to df_spark below
linker = SparkLinker(df_spark, get_settings_dict())
# Test that writing to files works as expected
def spark_csv_read(x):
return linker.spark.read.csv(x, header=True).toPandas()
_test_write_functionality(linker, spark_csv_read)
# Convert a column to an array to enable testing intersection
df_spark = df_spark.withColumn("email", f.array("email"))
settings_dict = get_settings_dict()
# Only needed because the value can be overwritten by other tests
settings_dict["comparisons"][1] = cl.exact_match("surname")
settings_dict["comparisons"].append(name_comparison(cll, "surname"))
settings = {
"probability_two_random_records_match": 0.01,
"link_type": "dedupe_only",
"blocking_rules_to_generate_predictions": [
{"blocking_rule": "l.surname = r.surname", "salting_partitions": 3},
],
"comparisons": [
cl.jaro_winkler_at_thresholds("first_name", 0.9),
cl.jaro_at_thresholds("surname", 0.9),
cl.damerau_levenshtein_at_thresholds("dob", 2),
{
"comparison_levels": [
cll.array_intersect_level("email"),
cll.else_level(),
]
},
cl.jaccard_at_thresholds("city", [0.9]),
],
"retain_matching_columns": True,
"retain_intermediate_calculation_columns": True,
"additional_columns_to_retain": ["group"],
"em_convergence": 0.01,
"max_iterations": 2,
}
linker = SparkLinker(
df_spark,
settings,
break_lineage_method="checkpoint",
num_partitions_on_repartition=2,
)
linker.profile_columns(
["first_name", "surname", "first_name || surname", "concat(city, first_name)"]
)
linker.compute_tf_table("city")
linker.compute_tf_table("first_name")
linker.estimate_probability_two_random_records_match(
["l.email = r.email"], recall=0.3
)
linker.estimate_u_using_random_sampling(max_pairs=1e5, seed=1)
blocking_rule = "l.first_name = r.first_name and l.surname = r.surname"
linker.estimate_parameters_using_expectation_maximisation(blocking_rule)
blocking_rule = "l.dob = r.dob"
linker.estimate_parameters_using_expectation_maximisation(blocking_rule)
df_predict = linker.predict()
linker.comparison_viewer_dashboard(
df_predict, os.path.join(tmp_path, "test_scv_spark.html"), True, 2
)
df_clusters = linker.cluster_pairwise_predictions_at_threshold(df_predict, 0.2)
linker.cluster_studio_dashboard(
df_predict,
df_clusters,
cluster_ids=[0, 4],
cluster_names=["cluster_0", "cluster_4"],
out_path=os.path.join(tmp_path, "test_cluster_studio.html"),
)
linker.unlinkables_chart(source_dataset="Testing")
# Test that writing to files works as expected
# spark_csv_read = lambda x: linker.spark.read.csv(x, header=True).toPandas()
# _test_write_functionality(linker, spark_csv_read)
# Check spark tables are being registered correctly
StructType(
[
StructField("firstname", StringType(), True),
StructField("lastname", StringType(), True),
]
)
register_roc_data(linker)
linker.roc_chart_from_labels_table("labels")
linker.accuracy_chart_from_labels_table("labels")
linker.confusion_matrix_from_labels_table("labels")
record = {
"unique_id": 1,
"first_name": "John",
"surname": "Smith",
"dob": "1971-05-24",
"city": "London",
"email": ["[email protected]"],
"group": 10000,
}
linker.find_matches_to_new_records(
[record], blocking_rules=[], match_weight_threshold=-10000
)
# Test differing inputs are accepted
settings["link_type"] = "link_only"
linker = SparkLinker(
[df_spark, df_spark.toPandas()],
settings,
break_lineage_method="checkpoint",
num_partitions_on_repartition=2,
)
# Test saving and loading
path = os.path.join(tmp_path, "model.json")
linker.save_model_to_json(path)
linker_2 = SparkLinker(df_spark)
linker_2.load_model(path)
linker_2.load_settings(path)
linker_2.load_settings_from_json(path)
SparkLinker(df_spark, settings_dict=path)
def METHOD_NAME(df_spark):
settings = get_settings_dict()
settings["link_type"] = "link_only"
settings["source_dataset_column_name"] = "source_dataset"
df_spark_a = df_spark.withColumn("source_dataset", f.lit("my_left_ds"))
df_spark_b = df_spark.withColumn("source_dataset", f.lit("my_right_ds"))
linker = SparkLinker(
[df_spark_a, df_spark_b],
settings,
break_lineage_method="checkpoint",
num_partitions_on_repartition=2,
)
df_predict = linker.predict().as_pandas_dataframe()
assert len(df_predict) == 7257
assert set(df_predict.source_dataset_l.values) == {"my_left_ds"}
assert set(df_predict.source_dataset_r.values) == {"my_right_ds"}
@pytest.mark.parametrize(
("df"),
[
pytest.param(
pd.read_csv("./tests/datasets/fake_1000_from_splink_demos.csv"),
id="Spark load from pandas df",
)
],
)
@mark_with_dialects_including("spark")
def test_spark_load_from_file(df, spark):
settings = get_settings_dict()
linker = SparkLinker(
df,
settings,
spark=spark,
)
assert len(linker.predict().as_pandas_dataframe()) == 3167 | null |
test storage mock api | # Copyright 1999-2021 Alibaba Group Holding Ltd.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import sys
import tempfile
import numpy as np
import pandas as pd
import pyarrow as pa
import pytest
from .... import oscar as mo
from .... import tensor as mt
from ....core import tile
from ....serialization import AioDeserializer, AioSerializer
from ....storage import StorageLevel
from ....tests.core import require_ray
from ....utils import get_next_port, lazy_import
from ...cluster import MockClusterAPI
from ...meta import MockMetaAPI
from ...session import MockSessionAPI
from ...web import WebActor
from ..api import MockStorageAPI, WebStorageAPI
ray = lazy_import("ray")
vineyard = lazy_import("vineyard")
require_lib = lambda x: x
storage_configs = []
# plasma backend
plasma_storage_size = 10 * 1024 * 1024
if sys.platform == "darwin":
plasma_dir = "/tmp"
else:
plasma_dir = "/dev/shm"
plasma_setup_params = dict(
store_memory=plasma_storage_size, plasma_directory=plasma_dir, check_dir_size=False
)
if not (sys.platform.lower().startswith("win") or pa.__version__ >= "12.0.0"):
storage_configs.append({"plasma": plasma_setup_params})
# ray backend
if ray is not None:
require_lib = require_ray
storage_configs.append({"ray": dict()})
# vineyard
if vineyard is not None:
storage_configs.append({"vineyard": dict(vineyard_size="256M")})
# shared_memory
storage_configs.append({"shared_memory": dict()})
@pytest.mark.asyncio
@pytest.mark.parametrize("storage_configs", storage_configs)
@pytest.mark.parametrize(
"ray_start_regular", [{"enable": ray is not None}], indirect=True
)
@require_lib
async def METHOD_NAME(ray_start_regular, storage_configs):
start_method = "fork" if sys.platform != "win32" else None
pool = await mo.create_actor_pool(
"127.0.0.1",
2,
labels=["main", "numa-0", "io"],
subprocess_start_method=start_method,
)
async with pool:
session_id = "mock_session_id"
storage_api = await MockStorageAPI.create(
address=pool.external_address,
session_id=session_id,
storage_configs=storage_configs,
)
# test put and get
value1 = np.random.rand(10, 10)
await storage_api.put("data1", value1)
get_value1 = await storage_api.get("data1")
np.testing.assert_array_equal(value1, get_value1)
value2 = pd.DataFrame(
{
"col1": [str(i) for i in range(10)],
"col2": np.random.randint(0, 100, (10,)),
}
)
await storage_api.put("data2", value2)
get_value2 = await storage_api.get("data2")
pd.testing.assert_frame_equal(value2, get_value2)
sliced_value = await storage_api.get(
"data2", conditions=[slice(3, 5), slice(None, None)]
)
pd.testing.assert_frame_equal(value2.iloc[3:5, :], sliced_value)
infos = await storage_api.get_infos("data2")
assert infos[0].store_size > 0
await storage_api.delete("data2")
buffers = await AioSerializer(value2).run()
size = sum(getattr(buf, "nbytes", len(buf)) for buf in buffers)
# test open_reader and open_writer
writer = await storage_api.open_writer("write_key", size, StorageLevel.MEMORY)
async with writer:
for buf in buffers:
await writer.write(buf)
reader = await storage_api.open_reader("write_key")
async with reader:
read_value = await AioDeserializer(reader).run()
pd.testing.assert_frame_equal(value2, read_value)
await MockStorageAPI.cleanup(pool.external_address)
@pytest.mark.asyncio
async def test_web_storage_api():
from ..api.web import StorageWebAPIHandler
tempdir = tempfile.mkdtemp()
start_method = "fork" if sys.platform != "win32" else None
pool = await mo.create_actor_pool(
"127.0.0.1", 1, subprocess_start_method=start_method
)
async with pool:
session_id = "mock_session_id"
await MockClusterAPI.create(address=pool.external_address)
await MockSessionAPI.create(
session_id=session_id, address=pool.external_address
)
meta_api = await MockMetaAPI.create(
session_id=session_id, address=pool.external_address
)
await MockStorageAPI.create(
address=pool.external_address,
session_id=session_id,
storage_configs={
"shared_memory": dict(),
"disk": dict(root_dirs=[tempdir]),
},
)
web_config = {
"port": get_next_port(),
"web_handlers": {
StorageWebAPIHandler.get_root_pattern(): StorageWebAPIHandler
},
}
await mo.create_actor(WebActor, web_config, address=pool.external_address)
web_storage_api = WebStorageAPI(
session_id, f'http://127.0.0.1:{web_config["port"]}', "numa-0"
)
value = np.random.rand(10, 10)
t = mt.random.rand(10, 10)
t = tile(t)
await meta_api.set_chunk_meta(
t.chunks[0], bands=[(pool.external_address, "numa-0")]
)
await web_storage_api.put(t.chunks[0].key, value)
ret_value = await web_storage_api.get(t.chunks[0].key)
np.testing.assert_array_equal(value, ret_value)
sliced_value = await web_storage_api.get(
t.chunks[0].key, conditions=[slice(3, 5), slice(None, None)]
)
np.testing.assert_array_equal(value[3:5, :], sliced_value)
infos = await web_storage_api.get_infos(t.chunks[0].key)
assert len(infos) == 1
assert infos[0].level == StorageLevel.MEMORY
assert infos[0].memory_size == t.chunks[0].nbytes
await MockStorageAPI.cleanup(pool.external_address)
await MockClusterAPI.cleanup(pool.external_address) | null |
state bar | from logging import DEBUG
import pytest
from mitmproxy.proxy import commands
from mitmproxy.proxy import events
from mitmproxy.proxy import layer
from mitmproxy.proxy.context import Context
from test.mitmproxy.proxy import tutils
class TestLayer:
def test_continue(self, tctx: Context):
class TLayer(layer.Layer):
def _handle_event(
self, event: events.Event
) -> layer.CommandGenerator[None]:
yield commands.OpenConnection(self.context.server)
yield commands.OpenConnection(self.context.server)
assert (
tutils.Playbook(TLayer(tctx))
<< commands.OpenConnection(tctx.server)
>> tutils.reply(None)
<< commands.OpenConnection(tctx.server)
>> tutils.reply(None)
)
def test_debug_messages(self, tctx: Context):
tctx.server.id = "serverid"
class TLayer(layer.Layer):
debug = " "
def _handle_event(
self, event: events.Event
) -> layer.CommandGenerator[None]:
yield from self.state(event)
def state_foo(self, event: events.Event) -> layer.CommandGenerator[None]:
assert isinstance(event, events.Start)
yield commands.OpenConnection(self.context.server)
self.state = self.METHOD_NAME
state = state_foo
def METHOD_NAME(self, event: events.Event) -> layer.CommandGenerator[None]:
assert isinstance(event, events.DataReceived)
yield commands.Log("baz")
tlayer = TLayer(tctx)
assert (
tutils.Playbook(tlayer, hooks=True, logs=True)
<< commands.Log(" >> Start({})", DEBUG)
<< commands.Log(
" << OpenConnection({'connection': Server({'id': '…rverid', 'address': None})})",
DEBUG,
)
<< commands.OpenConnection(tctx.server)
>> events.DataReceived(tctx.client, b"foo")
<< commands.Log(" >! DataReceived(client, b'foo')", DEBUG)
>> tutils.reply(None, to=-3)
<< commands.Log(
" >> Reply(OpenConnection({'connection': Server({'id': '…rverid', 'address': None, "
"'state': <ConnectionState.OPEN: 3>, 'timestamp_start': 1624544785})}), None)",
DEBUG,
)
<< commands.Log(" !> DataReceived(client, b'foo')", DEBUG)
<< commands.Log("baz")
)
assert repr(tlayer) == "TLayer(state: bar)"
def test_debug_shorten(self, tctx):
t = layer.Layer(tctx)
t.debug = " "
assert t._Layer__debug("x" * 600).message == " " + "x" * 512 + "…"
assert t._Layer__debug("x" * 600).message == " " + "x" * 256 + "…"
assert t._Layer__debug("foo").message == " foo"
class TestNextLayer:
def test_simple(self, tctx: Context):
nl = layer.NextLayer(tctx, ask_on_start=True)
nl.debug = " "
playbook = tutils.Playbook(nl, hooks=True)
assert (
playbook
<< layer.NextLayerHook(nl)
>> tutils.reply()
>> events.DataReceived(tctx.client, b"foo")
<< layer.NextLayerHook(nl)
>> tutils.reply()
>> events.DataReceived(tctx.client, b"bar")
<< layer.NextLayerHook(nl)
)
assert nl.data_client() == b"foobar"
assert nl.data_server() == b""
nl.layer = tutils.EchoLayer(tctx)
assert (
playbook
>> tutils.reply()
<< commands.SendData(tctx.client, b"foo")
<< commands.SendData(tctx.client, b"bar")
)
def test_late_hook_reply(self, tctx: Context):
"""
Properly handle case where we receive an additional event while we are waiting for
a reply from the proxy core.
"""
nl = layer.NextLayer(tctx)
playbook = tutils.Playbook(nl)
assert (
playbook
>> events.DataReceived(tctx.client, b"foo")
<< layer.NextLayerHook(nl)
>> events.DataReceived(tctx.client, b"bar")
)
assert nl.data_client() == b"foo" # "bar" is paused.
nl.layer = tutils.EchoLayer(tctx)
assert (
playbook
>> tutils.reply(to=-2)
<< commands.SendData(tctx.client, b"foo")
<< commands.SendData(tctx.client, b"bar")
)
@pytest.mark.parametrize("layer_found", [True, False])
def test_receive_close(self, tctx: Context, layer_found: bool):
"""Test that we abort a client connection which has disconnected without any layer being found."""
nl = layer.NextLayer(tctx)
playbook = tutils.Playbook(nl)
assert (
playbook
>> events.DataReceived(tctx.client, b"foo")
<< layer.NextLayerHook(nl)
>> events.ConnectionClosed(tctx.client)
)
if layer_found:
nl.layer = tutils.RecordLayer(tctx)
assert playbook >> tutils.reply(to=-2)
assert isinstance(nl.layer.event_log[-1], events.ConnectionClosed)
else:
assert (
playbook
>> tutils.reply(to=-2)
<< commands.CloseConnection(tctx.client)
<< None
)
def test_func_references(self, tctx: Context):
nl = layer.NextLayer(tctx)
playbook = tutils.Playbook(nl)
assert (
playbook
>> events.DataReceived(tctx.client, b"foo")
<< layer.NextLayerHook(nl)
)
nl.layer = tutils.EchoLayer(tctx)
handle = nl.handle_event
playbook >> tutils.reply()
playbook << commands.SendData(tctx.client, b"foo")
assert playbook
(sd,) = handle(events.DataReceived(tctx.client, b"bar"))
assert isinstance(sd, commands.SendData)
def test_repr(self, tctx: Context):
nl = layer.NextLayer(tctx)
nl.layer = tutils.EchoLayer(tctx)
assert repr(nl)
assert nl.stack_pos
assert nl.layer.stack_pos | null |
seeds | import pytest
from dbt.tests.util import run_dbt
models__sample_model_sql = """
select * from {{ source("raw", "sample_seed") }}
"""
models__schema_yml = """
version: 2
sources:
- name: raw
database: "{{ target.database }}"
schema: "{{ target.schema }}"
tables:
- name: sample_seed
columns:
- name: email
tests:
- not_null:
severity: "{{ 'error' if var('strict', false) else 'warn' }}"
models:
- name: sample_model
columns:
- name: email
tests:
- not_null:
severity: "{{ 'error' if var('strict', false) else 'warn' }}"
"""
seeds__sample_seed_csv = """id,first_name,last_name,email,gender,ip_address,updated_at
1,Judith,Kennedy,[email protected],Female,54.60.24.128,2015-12-24 12:19:28
2,Arthur,Kelly,[email protected],Male,62.56.24.215,2015-10-28 16:22:15
3,Rachel,Moreno,[email protected],Female,31.222.249.23,2016-04-05 02:05:30
4,Ralph,Turner,[email protected],Male,157.83.76.114,2016-08-08 00:06:51
5,Laura,Gonzales,[email protected],Female,30.54.105.168,2016-09-01 08:25:38
6,Katherine,Lopez,null,Female,169.138.46.89,2016-08-30 18:52:11
7,Jeremy,Hamilton,[email protected],Male,231.189.13.133,2016-07-17 02:09:46
8,Heather,Rose,[email protected],Female,87.165.201.65,2015-12-29 22:03:56
9,Gregory,Kelly,[email protected],Male,154.209.99.7,2016-03-24 21:18:16
10,Rachel,Lopez,[email protected],Female,237.165.82.71,2016-08-20 15:44:49
11,Donna,Welch,[email protected],Female,103.33.110.138,2016-02-27 01:41:48
12,Russell,Lawrence,[email protected],Male,189.115.73.4,2016-06-11 03:07:09
13,Michelle,Montgomery,[email protected],Female,243.220.95.82,2016-06-18 16:27:19
14,Walter,Castillo,null,Male,71.159.238.196,2016-10-06 01:55:44
15,Robin,Mills,[email protected],Female,172.190.5.50,2016-10-31 11:41:21
16,Raymond,Holmes,[email protected],Male,148.153.166.95,2016-10-03 08:16:38
17,Gary,Bishop,[email protected],Male,161.108.182.13,2016-08-29 19:35:20
18,Anna,Riley,[email protected],Female,253.31.108.22,2015-12-11 04:34:27
19,Sarah,Knight,[email protected],Female,222.220.3.177,2016-09-26 00:49:06
20,Phyllis,Fox,[email protected],Female,163.191.232.95,2016-08-21 10:35:19
"""
tests__sample_test_sql = """
{{ config(severity='error' if var('strict', false) else 'warn') }}
select * from {{ ref("sample_model") }} where email is null
"""
@pytest.fixture(scope="class")
def models():
return {"sample_model.sql": models__sample_model_sql, "schema.yml": models__schema_yml}
@pytest.fixture(scope="class")
def METHOD_NAME():
return {"sample_seed.csv": seeds__sample_seed_csv}
@pytest.fixture(scope="class")
def tests():
return {"null_email.sql": tests__sample_test_sql}
@pytest.fixture(scope="class")
def project_config_update():
return {
"config-version": 2,
"seed-paths": ["seeds"],
"test-paths": ["tests"],
"seeds": {
"quote_columns": False,
},
}
class TestSeverity:
@pytest.fixture(scope="class", autouse=True)
def seed_and_run(self, project):
run_dbt(["seed"])
run_dbt(["run"])
def test_generic_default(self, project):
results = run_dbt(["test", "--select", "test_type:generic"])
assert len(results) == 2
assert all([r.status == "warn" for r in results])
assert all([r.failures == 2 for r in results])
def test_generic_strict(self, project):
results = run_dbt(
["test", "--select", "test_type:generic", "--vars", '{"strict": True}'],
expect_pass=False,
)
assert len(results) == 2
assert all([r.status == "fail" for r in results])
assert all([r.failures == 2 for r in results])
def test_singular_default(self, project):
results = run_dbt(["test", "--select", "test_type:singular"])
assert len(results) == 1
assert all([r.status == "warn" for r in results])
assert all([r.failures == 2 for r in results])
def test_singular_strict(self, project):
results = run_dbt(
["test", "--select", "test_type:singular", "--vars", '{"strict": True}'],
expect_pass=False,
)
assert len(results) == 1
assert all([r.status == "fail" for r in results])
assert all([r.failures == 2 for r in results]) | null |
id | # coding=utf-8
# *** WARNING: this file was generated by pulumi. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import copy
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from ... import _utilities
from . import outputs
__all__ = [
'GetGuestConfigurationAssignmentResult',
'AwaitableGetGuestConfigurationAssignmentResult',
'get_guest_configuration_assignment',
'get_guest_configuration_assignment_output',
]
@pulumi.output_type
class GetGuestConfigurationAssignmentResult:
"""
Guest configuration assignment is an association between a machine and guest configuration.
"""
def __init__(__self__, METHOD_NAME=None, location=None, name=None, properties=None, system_data=None, type=None):
if METHOD_NAME and not isinstance(METHOD_NAME, str):
raise TypeError("Expected argument 'id' to be a str")
pulumi.set(__self__, "id", METHOD_NAME)
if location and not isinstance(location, str):
raise TypeError("Expected argument 'location' to be a str")
pulumi.set(__self__, "location", location)
if name and not isinstance(name, str):
raise TypeError("Expected argument 'name' to be a str")
pulumi.set(__self__, "name", name)
if properties and not isinstance(properties, dict):
raise TypeError("Expected argument 'properties' to be a dict")
pulumi.set(__self__, "properties", properties)
if system_data and not isinstance(system_data, dict):
raise TypeError("Expected argument 'system_data' to be a dict")
pulumi.set(__self__, "system_data", system_data)
if type and not isinstance(type, str):
raise TypeError("Expected argument 'type' to be a str")
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def METHOD_NAME(self) -> str:
"""
ARM resource id of the guest configuration assignment.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter
def location(self) -> Optional[str]:
"""
Region where the VM is located.
"""
return pulumi.get(self, "location")
@property
@pulumi.getter
def name(self) -> Optional[str]:
"""
Name of the guest configuration assignment.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter
def properties(self) -> 'outputs.GuestConfigurationAssignmentPropertiesResponse':
"""
Properties of the Guest configuration assignment.
"""
return pulumi.get(self, "properties")
@property
@pulumi.getter(name="systemData")
def system_data(self) -> 'outputs.SystemDataResponse':
"""
Azure Resource Manager metadata containing createdBy and modifiedBy information.
"""
return pulumi.get(self, "system_data")
@property
@pulumi.getter
def type(self) -> str:
"""
The type of the resource.
"""
return pulumi.get(self, "type")
class AwaitableGetGuestConfigurationAssignmentResult(GetGuestConfigurationAssignmentResult):
# pylint: disable=using-constant-test
def __await__(self):
if False:
yield self
return GetGuestConfigurationAssignmentResult(
METHOD_NAME=self.METHOD_NAME,
location=self.location,
name=self.name,
properties=self.properties,
system_data=self.system_data,
type=self.type)
def get_guest_configuration_assignment(guest_configuration_assignment_name: Optional[str] = None,
resource_group_name: Optional[str] = None,
vm_name: Optional[str] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetGuestConfigurationAssignmentResult:
"""
Get information about a guest configuration assignment
:param str guest_configuration_assignment_name: The guest configuration assignment name.
:param str resource_group_name: The resource group name.
:param str vm_name: The name of the virtual machine.
"""
__args__ = dict()
__args__['guestConfigurationAssignmentName'] = guest_configuration_assignment_name
__args__['resourceGroupName'] = resource_group_name
__args__['vmName'] = vm_name
opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)
__ret__ = pulumi.runtime.invoke('azure-native:guestconfiguration/v20220125:getGuestConfigurationAssignment', __args__, opts=opts, typ=GetGuestConfigurationAssignmentResult).value
return AwaitableGetGuestConfigurationAssignmentResult(
METHOD_NAME=pulumi.get(__ret__, 'id'),
location=pulumi.get(__ret__, 'location'),
name=pulumi.get(__ret__, 'name'),
properties=pulumi.get(__ret__, 'properties'),
system_data=pulumi.get(__ret__, 'system_data'),
type=pulumi.get(__ret__, 'type'))
@_utilities.lift_output_func(get_guest_configuration_assignment)
def get_guest_configuration_assignment_output(guest_configuration_assignment_name: Optional[pulumi.Input[str]] = None,
resource_group_name: Optional[pulumi.Input[str]] = None,
vm_name: Optional[pulumi.Input[str]] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetGuestConfigurationAssignmentResult]:
"""
Get information about a guest configuration assignment
:param str guest_configuration_assignment_name: The guest configuration assignment name.
:param str resource_group_name: The resource group name.
:param str vm_name: The name of the virtual machine.
"""
... | null |
clear markers | """
interface interactor.
"""
from .config import interface_color, pick_radius
from .interactor import BaseInteractor
from .util import clip, setpar
MAX_ROUGH=1
class InterfaceInteractor(BaseInteractor):
"""
Control the roughness of the layers.
"""
def __init__(self, profile):
BaseInteractor.__init__(self, profile)
ax = profile.axes
# markers for roughness
style = dict(linestyle = '',
transform = profile.xcoords,
marker = 's', #square
markersize = 7,
color = interface_color,
alpha = 0.6,
pickradius = pick_radius,
zorder = 8, #Prefer this to other lines
visible = False,
)
self.markers=[ax.plot([0],[0.05], label=label, **style)[0]
for label in ('interface marker L','interface marker R')]
# lines for roughness
style = dict(linestyle = '-',
transform = profile.xcoords,
marker = '',
color = interface_color,
visible = False)
self.lines=[ax.plot([0,0],[0.05,0.05], label=label, **style)[0]
for label in ('interface line L','interface line R')]
self.connect_markers(self.markers)
self._left = self._right = None
def set_layer(self):
"""
Move markers to the new layer
"""
n = self.profile.layer_num
z = self.profile.boundary[1:-1]
show_left = n is not None and n>0
show_right = n is not None and n < len(z)
if show_left:
self._left = self.profile.sample_layer(n-1).interface
else:
self._left = None
if show_right:
self._right = self.profile.sample_layer(n).interface
else:
self._right = None
#self.update_markers()
def update_markers(self):
"""
Draw the new roughness on the graph.
"""
n = self.profile.layer_num
z = self.profile.boundary[1:-1]
show_left = self._left is not None
show_right = self._right is not None
self.markers[0].set(visible=show_left)
self.lines[0].set(visible=show_left)
if show_left:
self.markers[0].set_xdata([z[n-1]+self._left.value])
self.lines[0].set_xdata([z[n-1], z[n-1]+self._left.value])
self.markers[1].set(visible=show_right)
self.lines[1].set(visible=show_right)
if show_right:
self.markers[1].set_xdata([z[n]-self._right.value])
self.lines[1].set_xdata([z[n],z[n]-self._right.value])
def METHOD_NAME(self):
"""
Remove interface markers from the graph.
"""
BaseInteractor.METHOD_NAME(self)
for h in self.lines:
h.remove()
self.lines = []
def save(self, event):
"""
Remember the interface for this layer and the next so that we
can restore on Esc.
"""
if self._left is not None:
self._left_value = self._left.value
if self._right is not None:
self._right_value = self._right.value
def restore(self, event):
"""
Restore the roughness for this layer.
"""
if self._left is not None:
self._left.value = self._left_value
if self._right is not None:
self._right.value = self._right_value
def drag(self, event):
"""
Process move to a new position, making sure that the move is allowed.
"""
z = self.profile.boundary
n = self.profile.layer_num
side = self.markers.index(event.artist)
if side == 0: # Left
limit = min(z[n]-z[n-1], z[n+1]-z[n])
v = clip(event.xdata-z[n], 0, limit/MAX_ROUGH)
setpar(self._left, v)
else: # Right
limit = min(z[n+1]-z[n], z[n+2]-z[n+1])
v = clip(z[n+1] - event.xdata, 0, limit/MAX_ROUGH)
setpar(self._right, v)
#self.update_markers() | null |
test longer | # (C) Copyright 2005-2023 Enthought, Inc., Austin, TX
# All rights reserved.
#
# This software is provided without warranty under the terms of the BSD
# license included in LICENSE.txt and may be redistributed only under
# the conditions described in the aforementioned license. The license
# is also available online at http://www.enthought.com/licenses/BSD.txt
#
# Thanks for using Enthought open source!
import unittest
from pyparsing import ParseException
from enable.savage.svg.pathdata import (
Sequence, closePath, coordinatePair, curve, ellipticalArc, horizontalLine,
lineTo, moveTo, number, quadraticBezierCurveto,
smoothQuadraticBezierCurveto, svg, verticalLine, CaselessLiteral,
)
class TestCaselessLiteral(unittest.TestCase):
def test_instantiation(self):
# regression test for https://github.com/enthought/enable/issues/887
# observed with pyparsing v >= 3
# we just test that instantiating the class doesnt raise exceptions
CaselessLiteral("test")
class TestNumber(unittest.TestCase):
parser = number
valid = ["1.e10", "1e2", "1e+4", "1e-10", "1.", "1.0", "0.1", ".2"]
invalid = ["e10", ".", "f", ""]
def testValid(self):
for num in self.valid:
self.assertEqual(float(num), self.parser.parseString(num)[0])
def testInvalid(self):
for num in self.invalid:
self.assertRaises(
ParseException, lambda: self.parser.parseString(num)
)
class TestNumberSequence(unittest.TestCase):
def testFloatsWithNoSpacing(self):
self.assertEqual(
[0.4, 0.4], list(Sequence(number).parseString("0.4.4"))
)
class TestCoords(unittest.TestCase):
def testCoordPair(self):
self.assertEqual(
coordinatePair.parseString("100 100")[0], (100.0, 100.0)
)
self.assertEqual(coordinatePair.parseString("100,2E7")[0], (100, 2e7))
def testCoordPairWithMinus(self):
self.assertEqual(
coordinatePair.parseString("100-100")[0], (100.0, -100.0)
)
def testCoordPairWithPlus(self):
self.assertEqual(
coordinatePair.parseString("100+100")[0], (100.0, 100.0)
)
def testCoordPairWithPlusAndExponent(self):
self.assertEqual(
coordinatePair.parseString("100+1e+2")[0], (100.0, 100.0)
)
def testNotAPair(self):
self.assertRaises(ParseException, coordinatePair.parseString, "100")
self
def testNoSpacing(self):
self.assertEqual(coordinatePair.parseString("-1.1.1")[0], (-1.1, 0.1))
class TestMoveTo(unittest.TestCase):
def testSimple(self):
self.assertEqual(
moveTo.parseString("M 100 100").asList()[0],
["M", [(100.0, 100.0)]],
)
def METHOD_NAME(self):
self.assertEqual(
moveTo.parseString("m 100 100 94 1e7").asList()[0],
["m", [(100.0, 100.0), (94, 1e7)]],
)
def testLine(self):
self.assertEqual(
lineTo.parseString("l 300 100").asList()[0],
["l", [(300.0, 100.0)]],
)
def testHorizonal(self):
self.assertEqual(
horizontalLine.parseString("H 100 5 20").asList()[0],
["H", [100.0, 5.0, 20.0]],
)
def testVertical(self):
self.assertEqual(
verticalLine.parseString("V 100 5 20").asList()[0],
["V", [100.0, 5.0, 20.0]],
)
class TestEllipticalArc(unittest.TestCase):
def testParse(self):
self.assertEqual(
ellipticalArc.parseString("a25,25 -30 0,1 50,-25").asList()[0],
["a", [[(25.0, 25.0), -30.0, (False, True), (50.0, -25.0)]]],
)
def testExtraArgs(self):
self.assertEqual(
ellipticalArc.parseString(
"a25,25 -30 0,1 50,-25, 10, 10"
).asList()[0],
["a", [[(25.0, 25.0), -30.0, (False, True), (50.0, -25.0)]]],
)
class TestSmoothQuadraticBezierCurveto(unittest.TestCase):
def testParse(self):
self.assertEqual(
smoothQuadraticBezierCurveto.parseString("t1000,300").asList()[0],
["t", [(1000.0, 300.0)]],
)
class TestQuadraticBezierCurveto(unittest.TestCase):
def testParse(self):
self.assertEqual(
quadraticBezierCurveto.parseString("Q1000,300 200 5").asList()[0],
["Q", [[(1000.0, 300.0), (200.0, 5.0)]]],
)
class TestCurve(unittest.TestCase):
def testParse(self):
self.assertEqual(
curve.parseString(
"C 100 200 300 400 500 600 100 200 300 400 500 600"
).asList()[0],
[
"C",
[
[(100.0, 200.0), (300.0, 400.0), (500.0, 600.0)],
[(100.0, 200.0), (300.0, 400.0), (500.0, 600.0)],
],
],
)
class TestClosePath(unittest.TestCase):
def testParse(self):
self.assertEqual(
closePath.parseString("Z").asList()[0], ("Z", (None,))
)
class TestSVG(unittest.TestCase):
def testParse(self):
path = ("M 100 100 L 300 100 L 200 300 z a 100,100 -4 0,1 25 25 "
"z T300 1000 t40, 50 h4 42 2 2,1v1,1,1 Z Q 34,10 1 1")
r = svg.parseString(path).asList()
expected = [
["M", [(100, 100)]],
["L", [(300, 100)]],
["L", [(200, 300)]],
("Z", (None,)),
["a", [[(100, 100), -4, (False, True), (25, 25)]]],
("Z", (None,)),
["T", [(300, 1000)]],
["t", [(40, 50)]],
["h", [4, 42, 2, 2, 1]],
["v", [1, 1, 1]],
("Z", (None,)),
["Q", [[(34, 10), (1, 1)]]],
]
self.assertEqual(len(r), len(expected))
for a, b in zip(expected, r):
self.assertEqual(a, b) | null |
print dict | # Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
# Copyright (c) 2019, Myrtle Software Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from enum import Enum
from metrics import word_error_rate
class Optimization(Enum):
"""Various levels of Optimization.
WARNING: This might have effect on model accuracy."""
nothing = 0
mxprO0 = 1
mxprO1 = 2
mxprO2 = 3
mxprO3 = 4
AmpOptimizations = {Optimization.mxprO0: "O0",
Optimization.mxprO1: "O1",
Optimization.mxprO2: "O2",
Optimization.mxprO3: "O3"}
def add_blank_label(labels):
if not isinstance(labels, list):
raise ValueError("labels must be a list of symbols")
labels.append("<BLANK>")
return labels
def __rnnt_decoder_predictions_tensor(tensor, labels):
"""
Takes output of greedy rnnt decoder and converts to strings.
Args:
tensor: model output tensor
label: A list of labels
Returns:
prediction
"""
hypotheses = []
labels_map = dict([(i, labels[i]) for i in range(len(labels))])
# iterate over batch
for ind in range(len(tensor)):
hypothesis = ''.join([labels_map[c] for c in tensor[ind]])
hypotheses.append(hypothesis)
return hypotheses
def __gather_predictions(predictions_list: list, labels: list) -> list:
results = []
for prediction in predictions_list:
results += __rnnt_decoder_predictions_tensor(prediction, labels=labels)
return results
def __gather_transcripts(transcript_list: list, transcript_len_list: list,
labels: list) -> list:
results = []
labels_map = dict([(i, labels[i]) for i in range(len(labels))])
for i, t in enumerate(transcript_list):
target = t.numpy().tolist()
reference = ''.join([labels_map[c] for c in target])
results.append(reference)
return results
def process_evaluation_batch(tensors: dict, global_vars: dict, labels: list):
"""
Processes results of an iteration and saves it in global_vars
Args:
tensors: dictionary with results of an evaluation iteration, e.g. loss, predictions, transcript, and output
global_vars: dictionary where processes results of iteration are saved
labels: A list of labels
"""
for kv, v in tensors.items():
if kv.startswith('predictions'):
global_vars['predictions'] += __gather_predictions(
v, labels=labels)
elif kv.startswith('transcript_length'):
transcript_len_list = v
elif kv.startswith('transcript'):
transcript_list = v
global_vars['transcripts'] += __gather_transcripts(transcript_list,
transcript_len_list,
labels=labels)
def process_evaluation_epoch(global_vars: dict, tag=None):
"""
Processes results from each worker at the end of evaluation and combine to final result
Args:
global_vars: dictionary containing information of entire evaluation
Return:
wer: final word error rate
loss: final loss
"""
hypotheses = global_vars['predictions']
references = global_vars['transcripts']
wer, scores, num_words = word_error_rate(
hypotheses=hypotheses, references=references)
return wer
def METHOD_NAME(d):
maxLen = max([len(ii) for ii in d.keys()])
fmtString = '\t%' + str(maxLen) + 's : %s'
print('Arguments:')
for keyPair in sorted(d.items()):
print(fmtString % keyPair) | null |
cleanup | #!/usr/bin/env python
import sys
from direct.showbase.ShowBase import ShowBase
from direct.showbase.InputStateGlobal import inputState
from panda3d.core import AmbientLight
from panda3d.core import DirectionalLight
from panda3d.core import LPoint3
from panda3d.core import TransformState
from panda3d.core import BitMask32
from panda3d.bullet import BulletWorld
from panda3d.bullet import BulletBoxShape
from panda3d.bullet import BulletSphereShape
from panda3d.bullet import BulletRigidBodyNode
from panda3d.bullet import BulletConeTwistConstraint
from panda3d.bullet import BulletDebugNode
class Game(ShowBase):
def __init__(self):
ShowBase.__init__(self)
base.set_background_color(0.1, 0.1, 0.8, 1)
base.set_frame_rate_meter(True)
base.cam.set_pos(0, -20, 5)
base.cam.look_at(0, 0, 0)
# Light
alight = AmbientLight('ambientLight')
alight.set_color((0.5, 0.5, 0.5, 1))
alightNP = render.attach_new_node(alight)
dlight = DirectionalLight('directionalLight')
dlight.set_direction((1, 1, -1))
dlight.set_color((0.7, 0.7, 0.7, 1))
dlightNP = render.attach_new_node(dlight)
render.clear_light()
render.set_light(alightNP)
render.set_light(dlightNP)
# Input
self.accept('escape', self.do_exit)
self.accept('r', self.do_reset)
self.accept('f1', base.toggle_wireframe)
self.accept('f2', base.toggle_texture)
self.accept('f3', self.toggle_debug)
self.accept('f5', self.do_screenshot)
self.accept('enter', self.do_shoot)
# Task
taskMgr.add(self.update, 'updateWorld')
# Physics
self.setup()
def do_exit(self):
self.METHOD_NAME()
sys.exit(1)
def do_reset(self):
self.METHOD_NAME()
self.setup()
def toggle_debug(self):
if self.debugNP.is_hidden():
self.debugNP.show()
else:
self.debugNP.hide()
def do_screenshot(self):
base.screenshot('Bullet')
def do_shoot(self):
# Get from/to points from mouse click
pMouse = base.mouseWatcherNode.get_mouse()
pFrom = LPoint3()
pTo = LPoint3()
base.camLens.extrude(pMouse, pFrom, pTo)
pFrom = render.get_relative_point(base.cam, pFrom)
pTo = render.get_relative_point(base.cam, pTo)
# Calculate initial velocity
v = pTo - pFrom
v.normalize()
v *= 100.0
# Create bullet
shape = BulletSphereShape(0.3)
body = BulletRigidBodyNode('Bullet')
bodyNP = self.worldNP.attach_new_node(body)
bodyNP.node().add_shape(shape)
bodyNP.node().set_mass(1.0)
bodyNP.node().set_linear_velocity(v)
bodyNP.node().set_ccd_motion_threshold(1e-7);
bodyNP.node().set_ccd_swept_sphere_radius(0.50);
bodyNP.set_collide_mask(BitMask32.all_on())
bodyNP.set_pos(pFrom)
visNP = loader.load_model('models/ball.egg')
visNP.set_scale(0.8)
visNP.reparent_to(bodyNP)
self.world.attach(bodyNP.node())
# Remove the bullet again after 2 seconds
taskMgr.do_method_later(2, self.do_remove, 'doRemove',
extraArgs=[bodyNP], appendTask=True)
def do_remove(self, bodyNP, task):
self.world.remove(bodyNP.node())
bodyNP.remove_node()
return task.done
def update(self, task):
dt = globalClock.get_dt()
self.world.do_physics(dt, 20, 1.0/180.0)
return task.cont
def METHOD_NAME(self):
self.worldNP.remove_node()
self.worldNP = None
self.world = None
def setup(self):
self.worldNP = render.attach_new_node('World')
# World
self.debugNP = self.worldNP.attach_new_node(BulletDebugNode('Debug'))
self.debugNP.show()
self.debugNP.node().show_wireframe(True)
self.debugNP.node().show_constraints(True)
self.debugNP.node().show_bounding_boxes(False)
self.debugNP.node().show_normals(False)
self.world = BulletWorld()
self.world.set_gravity((0, 0, -9.81))
self.world.set_debug_node(self.debugNP.node())
# Box A
shape = BulletBoxShape((0.5, 0.5, 0.5))
bodyA = BulletRigidBodyNode('Box A')
bodyNP = self.worldNP.attach_new_node(bodyA)
bodyNP.node().add_shape(shape)
bodyNP.set_collide_mask(BitMask32.all_on())
bodyNP.set_pos(-2, 0, 4)
visNP = loader.load_model('models/box.egg')
visNP.clear_model_nodes()
visNP.reparent_to(bodyNP)
self.world.attach(bodyA)
# Box B
shape = BulletBoxShape((0.5, 0.5, 0.5))
bodyB = BulletRigidBodyNode('Box B')
bodyNP = self.worldNP.attach_new_node(bodyB)
bodyNP.node().add_shape(shape)
bodyNP.node().set_mass(1.0)
bodyNP.node().set_deactivation_enabled(False)
bodyNP.set_collide_mask(BitMask32.all_on())
bodyNP.set_pos(0, 0, 0)
visNP = loader.load_model('models/box.egg')
visNP.clear_model_nodes()
visNP.reparent_to(bodyNP)
self.world.attach(bodyB)
# Cone
frameA = TransformState.make_pos_hpr((0, 0, -2), (0, 0, 90))
frameB = TransformState.make_pos_hpr((-5, 0, 0), (0, 0, 0))
cone = BulletConeTwistConstraint(bodyA, bodyB, frameA, frameB)
cone.set_debug_draw_size(2.0)
cone.set_limit(30, 45, 170, softness=1.0, bias=0.3, relaxation=8.0)
self.world.attach(cone)
game = Game()
game.run() | null |
start django | # Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See LICENSE in the project root
# for license information.
import pytest
from tests import code, debug, log, net, test_data
from tests.debug import runners, targets
from tests.patterns import some
pytestmark = pytest.mark.timeout(60)
django_server = net.WebServer(net.get_test_server_port(8000, 8100))
class paths:
django1 = test_data / "django1"
app_py = django1 / "app.py"
hello_html = django1 / "templates" / "hello.html"
bad_html = django1 / "templates" / "bad.html"
class lines:
app_py = code.get_marked_line_numbers(paths.app_py)
@pytest.fixture
@pytest.mark.parametrize("run", [runners.launch, runners.attach_connect["cli"]])
def METHOD_NAME(run):
def start(session, multiprocess=False):
# No clean way to kill Django server, expect non-zero exit code
session.expected_exit_code = some.int
session.config.update({"django": True, "subProcess": bool(multiprocess)})
args = ["runserver"]
if not multiprocess:
args += ["--noreload"]
args += ["--", str(django_server.port)]
return run(session, targets.Program(paths.app_py, args=args), cwd=paths.django1)
return start
@pytest.mark.parametrize("bp_target", ["code", "template"])
def test_django_breakpoint_no_multiproc(METHOD_NAME, bp_target):
bp_file, bp_line, bp_name = {
"code": (paths.app_py, lines.app_py["bphome"], "home"),
"template": (paths.hello_html, 8, "Django Template"),
}[bp_target]
bp_var_content = "Django-Django-Test"
with debug.Session() as session:
with METHOD_NAME(session):
breakpoints = session.set_breakpoints(bp_file, [bp_line])
for bp in breakpoints:
# They'll be verified later on for templates.
assert bp["verified"] == (bp_target == "code")
with django_server:
home_request = django_server.get("/home")
if bp_target == "template":
breakpoint_body = session.wait_for_next_event("breakpoint")
assert breakpoint_body["reason"] == "changed"
assert breakpoint_body["breakpoint"]["verified"]
session.wait_for_stop(
"breakpoint",
expected_frames=[
some.dap.frame(some.dap.source(bp_file), line=bp_line, name=bp_name)
],
)
var_content = session.get_variable("content")
assert var_content == some.dict.containing(
{
"name": "content",
"type": "str",
"value": repr(bp_var_content),
"presentationHint": {"attributes": ["rawString"]},
"evaluateName": "content",
"variablesReference": 0,
}
)
session.request_continue()
assert bp_var_content in home_request.response_text()
def test_django_template_exception_no_multiproc(METHOD_NAME):
with debug.Session() as session:
with METHOD_NAME(session):
session.request(
"setExceptionBreakpoints", {"filters": ["raised", "uncaught"]}
)
with django_server:
django_server.get("/badtemplate", log_errors=False)
stop = session.wait_for_stop(
"exception",
expected_frames=[
some.dap.frame(
some.dap.source(paths.bad_html),
line=8,
name="Django TemplateSyntaxError",
)
],
)
# Will stop once in the plugin
exception_info = session.request(
"exceptionInfo", {"threadId": stop.thread_id}
)
assert exception_info == some.dict.containing(
{
"exceptionId": some.str.ending_with("TemplateSyntaxError"),
"breakMode": "always",
"description": some.str.containing("doesnotexist"),
"details": some.dict.containing(
{
"message": some.str.containing("doesnotexist"),
"typeName": some.str.ending_with("TemplateSyntaxError"),
}
),
}
)
session.request_continue()
log.info("Exception will be reported again in {0}", paths.app_py)
session.wait_for_stop("exception")
session.request_continue()
@pytest.mark.parametrize("exc_type", ["handled", "unhandled"])
def test_django_exception_no_multiproc(METHOD_NAME, exc_type):
exc_line = lines.app_py["exc_" + exc_type]
with debug.Session() as session:
with METHOD_NAME(session):
session.request(
"setExceptionBreakpoints", {"filters": ["raised", "uncaught"]}
)
with django_server:
django_server.get("/" + exc_type)
stopped = session.wait_for_stop(
"exception",
expected_frames=[
some.dap.frame(
some.dap.source(paths.app_py),
line=exc_line,
name="bad_route_" + exc_type,
)
],
).body
assert stopped == some.dict.containing(
{
"reason": "exception",
"text": some.str.ending_with("ArithmeticError"),
"description": "Hello",
}
)
exception_info = session.request(
"exceptionInfo", {"threadId": stopped["threadId"]}
)
assert exception_info == {
"exceptionId": some.str.ending_with("ArithmeticError"),
"breakMode": "always",
"description": "Hello",
"details": {
"message": "Hello",
"typeName": some.str.ending_with("ArithmeticError"),
"source": some.path(paths.app_py),
"stackTrace": some.str,
},
}
session.request_continue()
def test_django_breakpoint_multiproc(METHOD_NAME):
bp_line = lines.app_py["bphome"]
bp_var_content = "Django-Django-Test"
with debug.Session() as parent_session:
with METHOD_NAME(parent_session, multiprocess=True):
parent_session.set_breakpoints(paths.app_py, [bp_line])
with parent_session.wait_for_next_subprocess() as child_session:
with child_session.start():
child_session.set_breakpoints(paths.app_py, [bp_line])
with django_server:
home_request = django_server.get("/home")
child_session.wait_for_stop(
"breakpoint",
expected_frames=[
some.dap.frame(
some.dap.source(paths.app_py), line=bp_line, name="home"
)
],
)
var_content = child_session.get_variable("content")
assert var_content == some.dict.containing(
{
"name": "content",
"type": "str",
"value": repr(bp_var_content),
"presentationHint": {"attributes": ["rawString"]},
"evaluateName": "content",
}
)
child_session.request_continue()
assert bp_var_content in home_request.response_text() | null |
test mcol gga ab ks | #!/usr/bin/env python
# Copyright 2022 The PySCF Developers. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# Author: Qiming Sun <[email protected]>
#
import unittest
import tempfile
import numpy
import copy
from pyscf import lib, gto, scf, dft
from pyscf import tdscf
try:
import mcfun
except ImportError:
mcfun = None
def setUpModule():
global mol, mf_lda
mol = gto.Mole()
mol.verbose = 5
mol.output = '/dev/null'
mol.atom = '''
H 0. 0. 0.
H 0. -0.7 0.7
H 0. 0.7 0.7'''
mol.basis = 'uncsto3g'
mol.spin = 1
mol.build()
mf_lda = mol.DKS().set(xc='lda,', conv_tol=1e-12,
chkfile=tempfile.NamedTemporaryFile().name).newton().run()
def tearDownModule():
global mol, mf_lda
mol.stdout.close()
del mol, mf_lda
def diagonalize(a, b, nroots=4):
nocc, nvir = a.shape[:2]
nov = nocc * nvir
a = a.reshape(nov, nov)
b = b.reshape(nov, nov)
h = numpy.block([[a , b ],
[-b.conj(),-a.conj()]])
e = numpy.linalg.eig(numpy.asarray(h))[0]
lowest_e = numpy.sort(e[e.real > 0].real)[:nroots]
lowest_e = lowest_e[lowest_e > 1e-3]
return lowest_e
class KnownValues(unittest.TestCase):
def test_tddft_lda(self):
td = mf_lda.TDDFT()
es = td.kernel(nstates=4)[0]
a,b = td.get_ab()
e_ref = diagonalize(a, b, 8)
self.assertAlmostEqual(abs(es[:3]-e_ref[:3]).max(), 0, 5)
self.assertAlmostEqual(lib.fp(es[:3] * 27.2114), 3.157449136045103, 5)
def test_tda_lda(self):
td = mf_lda.TDA()
es = td.kernel(nstates=5)[0]
a,b = td.get_ab()
nocc, nvir = a.shape[:2]
nov = nocc * nvir
e_ref = numpy.linalg.eigh(a.reshape(nov,nov))[0]
self.assertAlmostEqual(abs(es[:3]-e_ref[:3]).max(), 0, 5)
self.assertAlmostEqual(lib.fp(es[:3] * 27.2114), 3.220469947746697, 5)
def test_ab_hf(self):
mf = scf.DHF(mol).run()
self._check_against_ab_ks(mf.TDHF(), 3.162381949829394, 0.11755137875158098)
def test_col_lda_ab_ks(self):
self._check_against_ab_ks(mf_lda.TDDFT(), 2.6168030250127075, 0.10466808380307921)
def test_col_gga_ab_ks(self):
mf_b3lyp = dft.DKS(mol).set(xc='b3lyp5')
mf_b3lyp.__dict__.update(scf.chkfile.load(mf_lda.chkfile, 'scf'))
self._check_against_ab_ks(mf_b3lyp.TDDFT(), 2.568259113326634, 0.1531691970098629)
def test_col_mgga_ab_ks(self):
mf_m06l = dft.DKS(mol).set(xc='m06l')
mf_m06l.__dict__.update(scf.chkfile.load(mf_lda.chkfile, 'scf'))
self._check_against_ab_ks(mf_m06l.TDDFT(), 3.382341929143924, 0.6409502830461241)
@unittest.skipIf(mcfun is None, "mcfun library not found.")
def test_mcol_lda_ab_ks(self):
mcol_lda = dft.UDKS(mol).set(xc='lda,', collinear='mcol')
mcol_lda._numint.spin_samples = 6
mcol_lda.__dict__.update(scf.chkfile.load(mf_lda.chkfile, 'scf'))
self._check_against_ab_ks(mcol_lda.TDDFT(), 2.880950836922392, 0.45543578075726204)
@unittest.skipIf(mcfun is None, "mcfun library not found.")
def METHOD_NAME(self):
mcol_b3lyp = dft.UDKS(mol).set(xc='b3lyp5', collinear='mcol')
mcol_b3lyp._numint.spin_samples = 6
mcol_b3lyp.__dict__.update(scf.chkfile.load(mf_lda.chkfile, 'scf'))
self._check_against_ab_ks(mcol_b3lyp.TDDFT(), 2.753430274014454, 0.607433969753113)
@unittest.skipIf(mcfun is None, "mcfun library not found.")
def test_mcol_mgga_ab_ks(self):
mcol_m06l = dft.UDKS(mol).set(xc='m06l', collinear='mcol')
mcol_m06l._numint.spin_samples = 6
mcol_m06l.__dict__.update(scf.chkfile.load(mf_lda.chkfile, 'scf'))
self._check_against_ab_ks(mcol_m06l.TDDFT(), 14.934345395514491, 9.539340104227188)
def _check_against_ab_ks(self, td, refa, refb):
mf = td._scf
a, b = td.get_ab()
self.assertAlmostEqual(lib.fp(abs(a)), refa, 4)
self.assertAlmostEqual(lib.fp(abs(b)), refb, 4)
ftda = mf.TDA().gen_vind()[0]
ftdhf = td.gen_vind()[0]
n2c = mf.mo_occ.size // 2
nocc = numpy.count_nonzero(mf.mo_occ == 1)
nvir = numpy.count_nonzero(mf.mo_occ == 0) - n2c
numpy.random.seed(2)
x, y = xy = (numpy.random.random((2,nocc,nvir)) +
numpy.random.random((2,nocc,nvir)) * 1j)
ax = numpy.einsum('iajb,jb->ia', a, x)
self.assertAlmostEqual(abs(ax - ftda([x]).reshape(nocc,nvir)).max(), 0, 12)
ab1 = ax + numpy.einsum('iajb,jb->ia', b, y)
ab2 =-numpy.einsum('iajb,jb->ia', b.conj(), x)
ab2-= numpy.einsum('iajb,jb->ia', a.conj(), y)
abxy_ref = ftdhf([xy]).reshape(2,nocc,nvir)
self.assertAlmostEqual(abs(ab1 - abxy_ref[0]).max(), 0, 12)
self.assertAlmostEqual(abs(ab2 - abxy_ref[1]).max(), 0, 12)
if __name__ == "__main__":
print("Full Tests for TD-DKS")
unittest.main() | null |
delete | # Copyright (c) 2006-2009 Mitch Garnaat http://garnaat.org/
#
# Permission is hereby granted, free of charge, to any person obtaining a
# copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish, dis-
# tribute, sublicense, and/or sell copies of the Software, and to permit
# persons to whom the Software is furnished to do so, subject to the fol-
# lowing conditions:
#
# The above copyright notice and this permission notice shall be included
# in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL-
# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT
# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
# IN THE SOFTWARE.
import uuid
class OriginAccessIdentity(object):
def __init__(self, connection=None, config=None, id='',
s3_user_id='', comment=''):
self.connection = connection
self.config = config
self.id = id
self.s3_user_id = s3_user_id
self.comment = comment
self.etag = None
def startElement(self, name, attrs, connection):
if name == 'CloudFrontOriginAccessIdentityConfig':
self.config = OriginAccessIdentityConfig()
return self.config
else:
return None
def endElement(self, name, value, connection):
if name == 'Id':
self.id = value
elif name == 'S3CanonicalUserId':
self.s3_user_id = value
elif name == 'Comment':
self.comment = value
else:
setattr(self, name, value)
def update(self, comment=None):
new_config = OriginAccessIdentityConfig(self.connection,
self.config.caller_reference,
self.config.comment)
if comment is not None:
new_config.comment = comment
self.etag = self.connection.set_origin_identity_config(self.id, self.etag, new_config)
self.config = new_config
def METHOD_NAME(self):
return self.connection.delete_origin_access_identity(self.id, self.etag)
def uri(self):
return 'origin-access-identity/cloudfront/%s' % self.id
class OriginAccessIdentityConfig(object):
def __init__(self, connection=None, caller_reference='', comment=''):
self.connection = connection
if caller_reference:
self.caller_reference = caller_reference
else:
self.caller_reference = str(uuid.uuid4())
self.comment = comment
def to_xml(self):
s = '<?xml version="1.0" encoding="UTF-8"?>\n'
s += '<CloudFrontOriginAccessIdentityConfig xmlns="http://cloudfront.amazonaws.com/doc/2009-09-09/">\n'
s += ' <CallerReference>%s</CallerReference>\n' % self.caller_reference
if self.comment:
s += ' <Comment>%s</Comment>\n' % self.comment
s += '</CloudFrontOriginAccessIdentityConfig>\n'
return s
def startElement(self, name, attrs, connection):
return None
def endElement(self, name, value, connection):
if name == 'Comment':
self.comment = value
elif name == 'CallerReference':
self.caller_reference = value
else:
setattr(self, name, value)
class OriginAccessIdentitySummary(object):
def __init__(self, connection=None, id='',
s3_user_id='', comment=''):
self.connection = connection
self.id = id
self.s3_user_id = s3_user_id
self.comment = comment
self.etag = None
def startElement(self, name, attrs, connection):
return None
def endElement(self, name, value, connection):
if name == 'Id':
self.id = value
elif name == 'S3CanonicalUserId':
self.s3_user_id = value
elif name == 'Comment':
self.comment = value
else:
setattr(self, name, value)
def get_origin_access_identity(self):
return self.connection.get_origin_access_identity_info(self.id)
| null |
test run builder | import shutil
from datetime import datetime
from pathlib import Path
import pytest
from click.testing import CliRunner
from maggma.builders import CopyBuilder
from maggma.cli import run
from maggma.stores import MemoryStore, MongoStore
from monty.serialization import dumpfn
@pytest.fixture()
def mongostore():
store = MongoStore("maggma_test", "test")
store.connect()
store.remove_docs({})
yield store
store.remove_docs({})
store._collection.drop()
@pytest.fixture()
def reporting_store():
store = MongoStore("maggma_test", "reporting")
store.connect()
store.remove_docs({})
yield store
store.remove_docs({})
store._collection.drop()
def test_basic_run():
runner = CliRunner()
result = runner.invoke(run, ["--help"])
assert result.exit_code == 0
# Ensure running without any builders fail
result = runner.invoke(run)
assert result.exit_code != 0
def METHOD_NAME(mongostore):
memorystore = MemoryStore("temp")
builder = CopyBuilder(mongostore, memorystore)
mongostore.update([{mongostore.key: i, mongostore.last_updated_field: datetime.utcnow()} for i in range(10)])
runner = CliRunner()
with runner.isolated_filesystem():
dumpfn(builder, "test_builder.json")
result = runner.invoke(run, ["-v", "test_builder.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "SerialProcessor" in result.output
result = runner.invoke(run, ["-vvv", "--no_bars", "test_builder.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
result = runner.invoke(run, ["-v", "-n", "2", "test_builder.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "MultiProcessor" in result.output
result = runner.invoke(run, ["-vvv", "-n", "2", "--no_bars", "test_builder.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
def test_run_builder_chain(mongostore):
memorystore = MemoryStore("temp")
builder1 = CopyBuilder(mongostore, memorystore)
builder2 = CopyBuilder(mongostore, memorystore)
mongostore.update([{mongostore.key: i, mongostore.last_updated_field: datetime.utcnow()} for i in range(10)])
runner = CliRunner()
with runner.isolated_filesystem():
dumpfn([builder1, builder2], "test_builders.json")
result = runner.invoke(run, ["-v", "test_builders.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "SerialProcessor" in result.output
result = runner.invoke(run, ["-vvv", "--no_bars", "test_builders.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
result = runner.invoke(run, ["-v", "-n", "2", "test_builders.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "MultiProcessor" in result.output
result = runner.invoke(run, ["-vvv", "-n", "2", "--no_bars", "test_builders.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
def test_reporting(mongostore, reporting_store):
memorystore = MemoryStore("temp")
builder = CopyBuilder(mongostore, memorystore)
mongostore.update([{mongostore.key: i, mongostore.last_updated_field: datetime.utcnow()} for i in range(10)])
runner = CliRunner()
with runner.isolated_filesystem():
dumpfn(builder, "test_builder.json")
dumpfn(reporting_store, "test_reporting_store.json")
result = runner.invoke(run, ["-v", "test_builder.json", "-r", "test_reporting_store.json"])
assert result.exit_code == 0
report_docs = list(reporting_store.query())
assert len(report_docs) == 3
start_doc = next(d for d in report_docs if d["event"] == "BUILD_STARTED")
assert "sources" in start_doc
assert "targets" in start_doc
end_doc = next(d for d in report_docs if d["event"] == "BUILD_ENDED")
assert "errors" in end_doc
assert "warnings" in end_doc
update_doc = next(d for d in report_docs if d["event"] == "UPDATE")
assert "items" in update_doc
def test_python_source():
runner = CliRunner()
with runner.isolated_filesystem():
shutil.copy2(src=Path(__file__).parent / "builder_for_test.py", dst=Path(".").resolve())
result = runner.invoke(run, ["-v", "-n", "2", "builder_for_test.py"])
assert result.exit_code == 0
assert "Ended multiprocessing: DummyBuilder" in result.output
def test_python_notebook_source():
runner = CliRunner()
with runner.isolated_filesystem():
shutil.copy2(
src=Path(__file__).parent / "builder_notebook_for_test.ipynb",
dst=Path(".").resolve(),
)
result = runner.invoke(run, ["-v", "-n", "2", "builder_notebook_for_test.ipynb"])
assert result.exit_code == 0
assert "Ended multiprocessing: DummyBuilder" in result.output
def test_memray_run_builder(mongostore):
memorystore = MemoryStore("temp")
builder = CopyBuilder(mongostore, memorystore)
mongostore.update([{mongostore.key: i, mongostore.last_updated_field: datetime.utcnow()} for i in range(10)])
runner = CliRunner()
with runner.isolated_filesystem():
dumpfn(builder, "test_builder.json")
result = runner.invoke(run, ["-v", "--memray", "on", "test_builder.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "SerialProcessor" in result.output
result = runner.invoke(run, ["-vvv", "--no_bars", "--memray", "on", "test_builder.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
result = runner.invoke(run, ["-v", "-n", "2", "--memray", "on", "test_builder.json"])
assert result.exit_code == 0
assert "CopyBuilder" in result.output
assert "MultiProcessor" in result.output
result = runner.invoke(run, ["-vvv", "-n", "2", "--no_bars", "--memray", "on", "test_builder.json"])
assert result.exit_code == 0
assert "Get" not in result.output
assert "Update" not in result.output
def test_memray_user_output_dir(mongostore):
memorystore = MemoryStore("temp")
builder = CopyBuilder(mongostore, memorystore)
mongostore.update([{mongostore.key: i, mongostore.last_updated_field: datetime.utcnow()} for i in range(10)])
runner = CliRunner()
with runner.isolated_filesystem():
dumpfn(builder, "test_builder.json")
result = runner.invoke(run, ["--memray", "on", "-md", "memray_output_dir/", "test_builder.json"])
assert result.exit_code == 0
assert (Path.cwd() / "memray_output_dir").exists() is True | null |
hull convex | #!BPY
"""
Name: 'Hull'
Blender: 244
Group: 'MESH'
Submenu: 'Box' box
Submenu: 'Sphere' sphere
Submenu: 'Convex' convex
Tooltip: 'Hull Selected Objects'
"""
#Submenu: 'Cylinder' cylinder
# --------------------------------------------------------------------------
# Hull 1.1 by Amorilia
# --------------------------------------------------------------------------
# ***** BEGIN LICENSE BLOCK *****
#
# BSD License
#
# Copyright (c) 2005-2011, NIF File Format Library and Tools
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions
# are met:
# 1. Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# 2. Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# 3. The name of the NIF File Format Library and Tools project may not be
# used to endorse or promote products derived from this software
# without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
# OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
# IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
# NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#
# ***** END LICENCE BLOCK *****
# --------------------------------------------------------------------------
import bpy
import pyffi.utils.quickhull
def hull_box(ob, me, selected_only):
"""Hull mesh in a box."""
# find box hull
# todo: improve algorithm
minx = min(v.co[0] for v in me.vertices if v.sel or not selected_only)
miny = min(v.co[1] for v in me.vertices if v.sel or not selected_only)
minz = min(v.co[2] for v in me.vertices if v.sel or not selected_only)
maxx = max(v.co[0] for v in me.vertices if v.sel or not selected_only)
maxy = max(v.co[1] for v in me.vertices if v.sel or not selected_only)
maxz = max(v.co[2] for v in me.vertices if v.sel or not selected_only)
# create box
box = Blender.Mesh.New('box')
for x in [minx, maxx]:
for y in [miny, maxy]:
for z in [minz, maxz]:
box.vertices.extend(x,y,z)
box.faces.extend(
[[0,1,3,2],[6,7,5,4],[0,2,6,4],[3,1,5,7],[4,5,1,0],[7,6,2,3]])
# link box to scene and set transform
scn = Blender.Scene.GetCurrent()
boxob = scn.objects.new(box, 'box')
boxob.setMatrix(ob.getMatrix('worldspace'))
# set bounds type
boxob.display_type = 'BOUNDS'
boxob.display_bounds_type = 'BOX'
def hull_sphere(ob, me, selected_only):
"""Hull mesh in a sphere."""
# find square box hull
minx = min(v.co[0] for v in me.vertices if v.sel or not selected_only)
miny = min(v.co[1] for v in me.vertices if v.sel or not selected_only)
minz = min(v.co[2] for v in me.vertices if v.sel or not selected_only)
maxx = max(v.co[0] for v in me.vertices if v.sel or not selected_only)
maxy = max(v.co[1] for v in me.vertices if v.sel or not selected_only)
maxz = max(v.co[2] for v in me.vertices if v.sel or not selected_only)
cx = (minx+maxx)*0.5
cy = (miny+maxy)*0.5
cz = (minz+maxz)*0.5
lx = maxx-minx
ly = maxy-miny
lz = maxz-minz
l = max([lx,ly,lz])*0.5
minx = cx-l
miny = cy-l
minz = cz-l
maxx = cx+l
maxy = cy+l
maxz = cz+l
# create sphere
box = Blender.Mesh.New('sphere')
for x in [minx, maxx]:
for y in [miny, maxy]:
for z in [minz, maxz]:
box.vertices.extend(x,y,z)
box.faces.extend(
[[0,1,3,2],[6,7,5,4],[0,2,6,4],[3,1,5,7],[4,5,1,0],[7,6,2,3]])
# link box to scene and set transform
scn = Blender.Scene.GetCurrent()
boxob = scn.objects.new(box, 'sphere')
boxob.setMatrix(ob.getMatrix('worldspace'))
# set bounds type
boxob.display_type = 'BOUNDS'
boxob.display_bounds_type = 'SPHERE'
def METHOD_NAME(ob, me, selected_only, precision = 0.1):
"""Hull mesh in a convex shape."""
# find convex hull
vertices, triangles = pyffi.utils.quickhull.qhull3d(
[tuple(v.co) for v in me.vertices if v.sel or not selected_only],
precision = precision)
# create convex mesh
box = Blender.Mesh.New('convexpoly')
for vert in vertices:
box.vertices.extend(*vert)
for triangle in triangles:
box.faces.extend(triangle)
# link mesh to scene and set transform
scn = Blender.Scene.GetCurrent()
boxob = scn.objects.new(box, 'convexpoly')
boxob.setMatrix(ob.getMatrix('worldspace'))
# set bounds type
boxob.display_type = 'BOUNDS'
boxob.display_bounds_type = 'POLYHEDRON' # FIXME 'CONVEX' not in blender Python API
boxob.show_wire = True
def main(arg):
# get selected meshes
obs = [ob for ob in self.context.selected_objects if ob.type == 'MESH']
# saves editmode state and exit editmode if it is enabled
# (cannot make changes mesh data in editmode)
is_editmode = Window.EditMode()
Window.EditMode(0)
Window.WaitCursor(1)
t = sys.time()
# run script
num_affected = 0
for ob in obs:
me = ob.getData(mesh=1) # get Mesh, not NMesh
# are any vertices selected?
selected_only = is_editmode and (1 in ( vert.sel for vert in me.vertices ))
# create mesh by requested type
if arg == 'box': hull_box(ob, me, selected_only)
elif arg == 'sphere': hull_sphere(ob, me, selected_only)
elif arg == 'convex':
PREF_PRECISION = Blender.Draw.Create(0.1)
pup_block = [
('Precision', PREF_PRECISION, 0.001, 2.0, 'Maximum distance by which a vertex may fall outside the hull: larger values yield simpler hulls at the expense of missing more vertices.') ]
if not Blender.Draw.PupBlock('Convex Hull', pup_block):
return
METHOD_NAME(ob, me, selected_only, precision = PREF_PRECISION.val)
print(f"Hull finished in {(sys.time()-t):.2f} seconds")
Window.WaitCursor(0)
if is_editmode: Window.EditMode(1)
if __name__ == '__main__':
main(__script__['arg']) | null |
teardown | # Copyright 2019 ARM Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
import re
from wa import Parameter, ApkWorkload, PackageHandler, TestPackageHandler, ConfigError
from wa.utils.types import list_or_string
from wa.framework.exception import WorkloadError
class Uibenchjanktests(ApkWorkload):
name = 'uibenchjanktests'
description = """
Runs a particular test (or list of tests) of the UIBench JankTests_
test suite. The suite is provided by Google as an automated version
of the UIBench testbench for the Android UI.
The workload supports running the default set of tests without
restarting the app or running an arbitrary set of tests with
restarting the app in between each test.
.. _JankTests: https://android.googlesource.com/platform/platform_testing/+/master/tests/jank/uibench/src/com/android/uibench/janktests
"""
package_names = ['com.android.uibench.janktests']
_DUT_PACKAGE = 'com.android.test.uibench'
_DEFAULT_CLASS = 'UiBenchJankTests'
_OUTPUT_SECTION_REGEX = re.compile(
r'(\s*INSTRUMENTATION_STATUS: gfx-[\w-]+=[-+\d.]+\n)+'
r'\s*INSTRUMENTATION_STATUS_CODE: (?P<code>[-+\d]+)\n?', re.M)
_OUTPUT_GFXINFO_REGEX = re.compile(
r'INSTRUMENTATION_STATUS: (?P<name>[\w-]+)=(?P<value>[-+\d.]+)')
parameters = [
Parameter('tests', kind=list_or_string,
description="""
Tests to be run. Defaults to running every available
subtest in alphabetical order. The app will be restarted
for each subtest, unlike when using full=True.
""", default=None, aliases=['test']),
Parameter('full', kind=bool, default=False,
description="""
Runs the full suite of tests that the app defaults to
when no subtests are specified. The actual tests and their
order might depend on the version of the app. The subtests
will be run back to back without restarting the app in between.
"""),
Parameter('wait', kind=bool, default=True,
description='Forces am instrument to wait until the '
'instrumentation terminates before terminating itself. The '
'net effect is to keep the shell open until the tests have '
'finished. This flag is not required, but if you do not use '
'it, you will not see the results of your tests.'),
Parameter('raw', kind=bool, default=True,
description='Outputs results in raw format. Use this flag '
'when you want to collect performance measurements, so that '
'they are not formatted as test results. This flag is '
'designed for use with the flag -e perf true.'),
Parameter('instrument_args', kind=dict, default={},
description='Extra arguments for am instrument.'),
Parameter('no_hidden_api_checks', kind=bool, default=False,
description='Disables restrictions on the use of hidden '
'APIs.'),
]
def __init__(self, target, **kwargs):
super(Uibenchjanktests, self).__init__(target, **kwargs)
if 'iterations' not in self.instrument_args:
self.instrument_args['iterations'] = 1
self.dut_apk = PackageHandler(
self,
package_name=self._DUT_PACKAGE,
variant=self.variant,
strict=self.strict,
version=self.version,
force_install=self.force_install,
install_timeout=self.install_timeout,
uninstall=self.uninstall,
exact_abi=self.exact_abi,
prefer_host_package=self.prefer_host_package,
clear_data_on_reset=self.clear_data_on_reset)
self.apk = TestPackageHandler(
self,
package_name=self.package_name,
variant=self.variant,
strict=self.strict,
version=self.version,
force_install=self.force_install,
install_timeout=self.install_timeout,
uninstall=self.uninstall,
exact_abi=self.exact_abi,
prefer_host_package=self.prefer_host_package,
clear_data_on_reset=self.clear_data_on_reset,
instrument_args=self.instrument_args,
raw_output=self.raw,
instrument_wait=self.wait,
no_hidden_api_checks=self.no_hidden_api_checks)
def validate(self):
if self.full and self.tests is not None:
raise ConfigError("Can't select subtests while 'full' is True")
def initialize(self, context):
super(Uibenchjanktests, self).initialize(context)
self.dut_apk.initialize(context)
self.dut_apk.initialize_package(context)
self.output = {}
# Full run specified, don't select subtests
if self.full:
self.apk.args['class'] = '{}.{}'.format(
self.package_names[0], self._DEFAULT_CLASS
)
return
self.available_tests = {
test: cl for test, cl in self.apk.apk_info.methods
if test.startswith('test')
}
# default to running all tests in alphabetical order
# pylint: disable=access-member-before-definition
if not self.tests:
self.tests = sorted(self.available_tests.keys())
# raise error if any of the tests are not available
elif any([t not in self.available_tests for t in self.tests]):
msg = 'Unknown test(s) specified. Known tests: {}'
known_tests = '\n'.join(self.available_tests.keys())
raise ValueError(msg.format(known_tests))
def run(self, context):
# Full run, just run the activity directly
if self.full:
self.apk.start_activity()
self.apk.wait_instrument_over()
self.output['full'] = self.apk.instrument_output
return
for test in self.tests:
self.apk.args['class'] = '{}.{}#{}'.format(
self.package_names[0],
self.available_tests[test], test
)
self.apk.setup(context)
self.apk.start_activity()
try:
self.apk.wait_instrument_over()
except WorkloadError as e:
self.logger.warning(str(e))
self.output[test] = self.apk.instrument_output
def update_output(self, context):
super(Uibenchjanktests, self).update_output(context)
for test, test_output in self.output.items():
for section in self._OUTPUT_SECTION_REGEX.finditer(test_output):
if int(section.group('code')) != -1:
msg = 'Run failed (INSTRUMENTATION_STATUS_CODE: {}). See log.'
raise RuntimeError(msg.format(section.group('code')))
for metric in self._OUTPUT_GFXINFO_REGEX.finditer(section.group()):
context.add_metric(metric.group('name'), metric.group('value'),
classifiers={'test_name': test})
def METHOD_NAME(self, context):
super(Uibenchjanktests, self).METHOD_NAME(context)
self.dut_apk.METHOD_NAME() | null |
on touch move | """
Drag Behavior
=============
The :class:`~kivy.uix.behaviors.drag.DragBehavior`
`mixin <https://en.wikipedia.org/wiki/Mixin>`_ class provides Drag behavior.
When combined with a widget, dragging in the rectangle defined by the
:attr:`~kivy.uix.behaviors.drag.DragBehavior.drag_rectangle` will drag the
widget.
Example
-------
The following example creates a draggable label::
from kivy.uix.label import Label
from kivy.app import App
from kivy.uix.behaviors import DragBehavior
from kivy.lang import Builder
# You could also put the following in your kv file...
kv = '''
<DragLabel>:
# Define the properties for the DragLabel
drag_rectangle: self.x, self.y, self.width, self.height
drag_timeout: 10000000
drag_distance: 0
FloatLayout:
# Define the root widget
DragLabel:
size_hint: 0.25, 0.2
text: 'Drag me'
'''
class DragLabel(DragBehavior, Label):
pass
class TestApp(App):
def build(self):
return Builder.load_string(kv)
TestApp().run()
"""
__all__ = ('DragBehavior', )
from kivy.clock import Clock
from kivy.properties import NumericProperty, ReferenceListProperty
from kivy.config import Config
from kivy.metrics import sp
from functools import partial
# When we are generating documentation, Config doesn't exist
_scroll_timeout = _scroll_distance = 0
if Config:
_scroll_timeout = Config.getint('widgets', 'scroll_timeout')
_scroll_distance = Config.getint('widgets', 'scroll_distance')
class DragBehavior(object):
'''
The DragBehavior `mixin <https://en.wikipedia.org/wiki/Mixin>`_ provides
Drag behavior. When combined with a widget, dragging in the rectangle
defined by :attr:`drag_rectangle` will drag the widget. Please see
the :mod:`drag behaviors module <kivy.uix.behaviors.drag>` documentation
for more information.
.. versionadded:: 1.8.0
'''
drag_distance = NumericProperty(_scroll_distance)
'''Distance to move before dragging the :class:`DragBehavior`, in pixels.
As soon as the distance has been traveled, the :class:`DragBehavior` will
start to drag, and no touch event will be dispatched to the children.
It is advisable that you base this value on the dpi of your target device's
screen.
:attr:`drag_distance` is a :class:`~kivy.properties.NumericProperty` and
defaults to the `scroll_distance` as defined in the user
:class:`~kivy.config.Config` (20 pixels by default).
'''
drag_timeout = NumericProperty(_scroll_timeout)
'''Timeout allowed to trigger the :attr:`drag_distance`, in milliseconds.
If the user has not moved :attr:`drag_distance` within the timeout,
dragging will be disabled, and the touch event will be dispatched to the
children.
:attr:`drag_timeout` is a :class:`~kivy.properties.NumericProperty` and
defaults to the `scroll_timeout` as defined in the user
:class:`~kivy.config.Config` (55 milliseconds by default).
'''
drag_rect_x = NumericProperty(0)
'''X position of the axis aligned bounding rectangle where dragging
is allowed (in window coordinates).
:attr:`drag_rect_x` is a :class:`~kivy.properties.NumericProperty` and
defaults to 0.
'''
drag_rect_y = NumericProperty(0)
'''Y position of the axis aligned bounding rectangle where dragging
is allowed (in window coordinates).
:attr:`drag_rect_Y` is a :class:`~kivy.properties.NumericProperty` and
defaults to 0.
'''
drag_rect_width = NumericProperty(100)
'''Width of the axis aligned bounding rectangle where dragging is allowed.
:attr:`drag_rect_width` is a :class:`~kivy.properties.NumericProperty` and
defaults to 100.
'''
drag_rect_height = NumericProperty(100)
'''Height of the axis aligned bounding rectangle where dragging is allowed.
:attr:`drag_rect_height` is a :class:`~kivy.properties.NumericProperty` and
defaults to 100.
'''
drag_rectangle = ReferenceListProperty(drag_rect_x, drag_rect_y,
drag_rect_width, drag_rect_height)
'''Position and size of the axis aligned bounding rectangle where dragging
is allowed.
:attr:`drag_rectangle` is a :class:`~kivy.properties.ReferenceListProperty`
of (:attr:`drag_rect_x`, :attr:`drag_rect_y`, :attr:`drag_rect_width`,
:attr:`drag_rect_height`) properties.
'''
def __init__(self, **kwargs):
self._drag_touch = None
super(DragBehavior, self).__init__(**kwargs)
def _get_uid(self, prefix='sv'):
return '{0}.{1}'.format(prefix, self.uid)
def on_touch_down(self, touch):
xx, yy, w, h = self.drag_rectangle
x, y = touch.pos
if not self.collide_point(x, y):
touch.ud[self._get_uid('svavoid')] = True
return super(DragBehavior, self).on_touch_down(touch)
if self._drag_touch or ('button' in touch.profile and
touch.button.startswith('scroll')) or\
not ((xx < x <= xx + w) and (yy < y <= yy + h)):
return super(DragBehavior, self).on_touch_down(touch)
# no mouse scrolling, so the user is going to drag with this touch.
self._drag_touch = touch
uid = self._get_uid()
touch.grab(self)
touch.ud[uid] = {
'mode': 'unknown',
'dx': 0,
'dy': 0}
Clock.schedule_once(self._change_touch_mode,
self.drag_timeout / 1000.)
return True
def METHOD_NAME(self, touch):
if self._get_uid('svavoid') in touch.ud or\
self._drag_touch is not touch:
return super(DragBehavior, self).METHOD_NAME(touch) or\
self._get_uid() in touch.ud
if touch.grab_current is not self:
return True
uid = self._get_uid()
ud = touch.ud[uid]
mode = ud['mode']
if mode == 'unknown':
ud['dx'] += abs(touch.dx)
ud['dy'] += abs(touch.dy)
if ud['dx'] > sp(self.drag_distance):
mode = 'drag'
if ud['dy'] > sp(self.drag_distance):
mode = 'drag'
ud['mode'] = mode
if mode == 'drag':
self.x += touch.dx
self.y += touch.dy
return True
def on_touch_up(self, touch):
if self._get_uid('svavoid') in touch.ud:
return super(DragBehavior, self).on_touch_up(touch)
if self._drag_touch and self in [x() for x in touch.grab_list]:
touch.ungrab(self)
self._drag_touch = None
ud = touch.ud[self._get_uid()]
if ud['mode'] == 'unknown':
super(DragBehavior, self).on_touch_down(touch)
Clock.schedule_once(partial(self._do_touch_up, touch), .1)
else:
if self._drag_touch is not touch:
super(DragBehavior, self).on_touch_up(touch)
return self._get_uid() in touch.ud
def _do_touch_up(self, touch, *largs):
super(DragBehavior, self).on_touch_up(touch)
# don't forget about grab event!
for x in touch.grab_list[:]:
touch.grab_list.remove(x)
x = x()
if not x:
continue
touch.grab_current = x
super(DragBehavior, self).on_touch_up(touch)
touch.grab_current = None
def _change_touch_mode(self, *largs):
if not self._drag_touch:
return
uid = self._get_uid()
touch = self._drag_touch
ud = touch.ud[uid]
if ud['mode'] != 'unknown':
return
touch.ungrab(self)
self._drag_touch = None
touch.push()
touch.apply_transform_2d(self.parent.to_widget)
super(DragBehavior, self).on_touch_down(touch)
touch.pop()
return | null |
clear reactions | import logging
import re
import sys
import os
from difflib import get_close_matches
from enum import IntEnum
from logging.handlers import RotatingFileHandler
from string import Formatter
import discord
from discord.ext import commands
import _string
try:
from colorama import Fore, Style
except ImportError:
Fore = Style = type("Dummy", (object,), {"__getattr__": lambda self, item: ""})()
if ".heroku" in os.environ.get("PYTHONHOME", ""):
# heroku
Fore = Style = type("Dummy", (object,), {"__getattr__": lambda self, item: ""})()
class PermissionLevel(IntEnum):
OWNER = 5
ADMINISTRATOR = 4
ADMIN = 4
MODERATOR = 3
MOD = 3
SUPPORTER = 2
RESPONDER = 2
REGULAR = 1
INVALID = -1
class InvalidConfigError(commands.BadArgument):
def __init__(self, msg, *args):
super().__init__(msg, *args)
self.msg = msg
@property
def embed(self):
# Single reference of Color.red()
return discord.Embed(title="Error", description=self.msg, color=discord.Color.red())
class ModmailLogger(logging.Logger):
@staticmethod
def _debug_(*msgs):
return f'{Fore.CYAN}{" ".join(msgs)}{Style.RESET_ALL}'
@staticmethod
def _info_(*msgs):
return f'{Fore.LIGHTMAGENTA_EX}{" ".join(msgs)}{Style.RESET_ALL}'
@staticmethod
def _error_(*msgs):
return f'{Fore.RED}{" ".join(msgs)}{Style.RESET_ALL}'
def debug(self, msg, *args, **kwargs):
if self.isEnabledFor(logging.DEBUG):
self._log(logging.DEBUG, self._debug_(msg), args, **kwargs)
def info(self, msg, *args, **kwargs):
if self.isEnabledFor(logging.INFO):
self._log(logging.INFO, self._info_(msg), args, **kwargs)
def warning(self, msg, *args, **kwargs):
if self.isEnabledFor(logging.WARNING):
self._log(logging.WARNING, self._error_(msg), args, **kwargs)
def error(self, msg, *args, **kwargs):
if self.isEnabledFor(logging.ERROR):
self._log(logging.ERROR, self._error_(msg), args, **kwargs)
def critical(self, msg, *args, **kwargs):
if self.isEnabledFor(logging.CRITICAL):
self._log(logging.CRITICAL, self._error_(msg), args, **kwargs)
def line(self, level="info"):
if level == "info":
level = logging.INFO
elif level == "debug":
level = logging.DEBUG
else:
level = logging.INFO
if self.isEnabledFor(level):
self._log(
level,
Fore.BLACK + Style.BRIGHT + "-------------------------" + Style.RESET_ALL,
[],
)
logging.setLoggerClass(ModmailLogger)
log_level = logging.INFO
loggers = set()
ch = logging.StreamHandler(stream=sys.stdout)
ch.setLevel(log_level)
formatter = logging.Formatter(
"%(asctime)s %(name)s[%(lineno)d] - %(levelname)s: %(message)s", datefmt="%m/%d/%y %H:%M:%S"
)
ch.setFormatter(formatter)
ch_debug = None
def getLogger(name=None) -> ModmailLogger:
logger = logging.getLogger(name)
logger.setLevel(log_level)
logger.addHandler(ch)
if ch_debug is not None:
logger.addHandler(ch_debug)
loggers.add(logger)
return logger
class FileFormatter(logging.Formatter):
ansi_escape = re.compile(r"\x1B\[[0-?]*[ -/]*[@-~]")
def format(self, record):
record.msg = self.ansi_escape.sub("", record.msg)
return super().format(record)
def configure_logging(name, level=None):
global ch_debug, log_level
ch_debug = RotatingFileHandler(name, mode="a+", maxBytes=48000, backupCount=1, encoding="utf-8")
formatter_debug = FileFormatter(
"%(asctime)s %(name)s[%(lineno)d] - %(levelname)s: %(message)s",
datefmt="%Y-%m-%d %H:%M:%S",
)
ch_debug.setFormatter(formatter_debug)
ch_debug.setLevel(logging.DEBUG)
if level is not None:
log_level = level
ch.setLevel(log_level)
for logger in loggers:
logger.setLevel(log_level)
logger.addHandler(ch_debug)
class _Default:
pass
Default = _Default()
class SafeFormatter(Formatter):
def get_field(self, field_name, args, kwargs):
first, rest = _string.formatter_field_name_split(field_name)
try:
obj = self.get_value(first, args, kwargs)
except (IndexError, KeyError):
return "<Invalid>", first
# loop through the rest of the field_name, doing
# getattr or getitem as needed
# stops when reaches the depth of 2 or starts with _.
try:
for n, (is_attr, i) in enumerate(rest):
if n >= 2:
break
if is_attr:
if str(i).startswith("_"):
break
obj = getattr(obj, i)
else:
obj = obj[i]
else:
return obj, first
except (IndexError, KeyError):
pass
return "<Invalid>", first
class UnseenFormatter(Formatter):
def get_value(self, key, args, kwds):
if isinstance(key, str):
try:
return kwds[key]
except KeyError:
return "{" + key + "}"
else:
return super().get_value(key, args, kwds)
class SimilarCategoryConverter(commands.CategoryChannelConverter):
async def convert(self, ctx, argument):
bot = ctx.bot
guild = ctx.guild
try:
return await super().convert(ctx, argument)
except commands.ChannelNotFound:
if guild:
categories = {c.name.casefold(): c for c in guild.categories}
else:
categories = {
c.name.casefold(): c
for c in bot.get_all_channels()
if isinstance(c, discord.CategoryChannel)
}
result = get_close_matches(argument.casefold(), categories.keys(), n=1, cutoff=0.75)
if result:
result = categories[result[0]]
if not isinstance(result, discord.CategoryChannel):
raise commands.ChannelNotFound(argument)
return result
class DummyMessage:
"""
A class mimicking the original :class:discord.Message
where all functions that require an actual message to exist
is replaced with a dummy function
"""
def __init__(self, message):
if message:
message.attachments = []
self._message = message
def __getattr__(self, name: str):
return getattr(self._message, name)
def __bool__(self):
return bool(self._message)
async def delete(self, *, delay=None):
return
async def edit(self, **fields):
return
async def add_reaction(self, emoji):
return
async def remove_reaction(self, emoji):
return
async def clear_reaction(self, emoji):
return
async def METHOD_NAME(self):
return
async def pin(self, *, reason=None):
return
async def unpin(self, *, reason=None):
return
async def publish(self):
return
async def ack(self):
return
class DMDisabled(IntEnum):
NONE = 0
NEW_THREADS = 1
ALL_THREADS = 2
class HostingMethod(IntEnum):
HEROKU = 0
PM2 = 1
SYSTEMD = 2
SCREEN = 3
DOCKER = 4
OTHER = 5 | null |
wrap py func | # Copyright 2017 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
"""Pyfunc creation utilities."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from collections import namedtuple
from tensorflow.python.framework import dtypes
from tensorflow.python.framework import tensor_util
from tensorflow.python.ops import script_ops
class MatchDType(namedtuple('MatchDType', ('arg_number',))):
"""Allows matching the dtype of an argument.
Used in conjunction with function calls. For example, MatchDType(0) will
match the DType of the first argument.
"""
pass
def METHOD_NAME(f, return_dtypes, args, kwargs=None, use_dummy_return=False):
"""Helper that wraps a callable to py_func.
The helper passes tensor arguments through the py_func interface. Non-tensor
arguments are allowed, and will be passed to f directly. Note that non-tensor
arguments are captured by f will not update every time the wrapper is
called (this is consistent with its argument list, which only includes
the tensor arguments). In general, it's safest not to reuse this wrapper.
Args:
f: Callable
return_dtypes: None, individual of tuple/list of DType or MatchDType, the
data type for each of f's return value(s). Set to None if f has no
return values or use_dummy_return is True. Use MatchDType to define a
dtype identical to that of `i`th argument (argument 0 is the first);
an argument must of Tensor type if it is to be used with MatchDType.
args: Positional arguments for f, as list or tuple.
kwargs: Keyword arguments for f, as dict with string keys. May be None.
use_dummy_return: If True, the function will return a dummy value of 1
and discard its actual return value.
Returns:
The return values of f converted to tensor.
Raises:
ValueError: if any of the arguments are incorrect.
"""
if return_dtypes and use_dummy_return:
raise ValueError('if use_dummy_return is True, return_dtypes must be empty')
tensor_args = []
tensor_args_idx = {}
# Of the positional arguments, only grab the tensor ones to be passed through
# the py_func.
n_args = len(args)
arg_is_tensor = tuple(map(tensor_util.is_tensor, args))
for i in range(n_args):
if arg_is_tensor[i]:
tensor_args_idx[i] = len(tensor_args)
tensor_args.append(args[i])
# We essentially take the tensor kwargs, if any, and add them to the list of
# positional arguments. The kwargs are then reconstructed inside the py_func.
#
# For example, if
#
# args = [Tensor(1), 'foo']
# kwargs = {'a': Tensor(2), 'b': 'bar'}
#
# Then
#
# tensor_args = (Tensor(1), Tensor(2))
# kwarg_keys = ('a', 'b')
if kwargs:
kwarg_keys = tuple(kwargs.keys())
kwarg_is_tensor = {k: tensor_util.is_tensor(kwargs[k]) for k in kwarg_keys}
for k in kwarg_keys:
if kwarg_is_tensor[k]:
tensor_args_idx[k] = len(tensor_args)
tensor_args.append(kwargs[k])
else:
kwarg_keys = ()
# Set up return dtypes.
def match_arg_dtype(arg_number):
arg = args[arg_number]
if not arg_is_tensor[arg_number]:
raise ValueError(
'argument %d was used with MatchDType and must be a tf.Tensor, but '
'was %s instead' % (arg_number, type(arg)))
return arg.dtype
if return_dtypes:
if isinstance(return_dtypes, MatchDType):
return_dtypes = match_arg_dtype(return_dtypes.arg_number)
elif isinstance(return_dtypes, (list, tuple)):
return_dtypes = tuple(
match_arg_dtype(a.arg_number) if isinstance(a, MatchDType) else a
for a in return_dtypes)
else:
assert isinstance(return_dtypes, dtypes.DType)
def f_wrapper(*tensor_args):
f_args = tuple(tensor_args[tensor_args_idx[i]] if arg_is_tensor[i] else a
for i, a in enumerate(args))
f_kwargs = {
k: tensor_args[tensor_args_idx[k]] if kwarg_is_tensor[k] else kwargs[k]
for i, k in enumerate(kwarg_keys)
}
retval = f(*f_args, **f_kwargs)
return 1 if use_dummy_return else retval
if use_dummy_return:
return_dtypes = dtypes.int32
return script_ops.eager_py_func(f_wrapper, tensor_args, return_dtypes) | null |
sku | # coding=utf-8
# *** WARNING: this file was generated by pulumi. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import copy
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from ... import _utilities
from . import outputs
__all__ = [
'GetNotificationHubResult',
'AwaitableGetNotificationHubResult',
'get_notification_hub',
'get_notification_hub_output',
]
@pulumi.output_type
class GetNotificationHubResult:
"""
Notification Hub Resource.
"""
def __init__(__self__, id=None, location=None, name=None, properties=None, METHOD_NAME=None, system_data=None, tags=None, type=None):
if id and not isinstance(id, str):
raise TypeError("Expected argument 'id' to be a str")
pulumi.set(__self__, "id", id)
if location and not isinstance(location, str):
raise TypeError("Expected argument 'location' to be a str")
pulumi.set(__self__, "location", location)
if name and not isinstance(name, str):
raise TypeError("Expected argument 'name' to be a str")
pulumi.set(__self__, "name", name)
if properties and not isinstance(properties, dict):
raise TypeError("Expected argument 'properties' to be a dict")
pulumi.set(__self__, "properties", properties)
if METHOD_NAME and not isinstance(METHOD_NAME, dict):
raise TypeError("Expected argument 'sku' to be a dict")
pulumi.set(__self__, "sku", METHOD_NAME)
if system_data and not isinstance(system_data, dict):
raise TypeError("Expected argument 'system_data' to be a dict")
pulumi.set(__self__, "system_data", system_data)
if tags and not isinstance(tags, dict):
raise TypeError("Expected argument 'tags' to be a dict")
pulumi.set(__self__, "tags", tags)
if type and not isinstance(type, str):
raise TypeError("Expected argument 'type' to be a str")
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def id(self) -> str:
"""
Fully qualified resource ID for the resource. E.g. "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}"
"""
return pulumi.get(self, "id")
@property
@pulumi.getter
def location(self) -> str:
"""
The geo-location where the resource lives
"""
return pulumi.get(self, "location")
@property
@pulumi.getter
def name(self) -> str:
"""
The name of the resource
"""
return pulumi.get(self, "name")
@property
@pulumi.getter
def properties(self) -> 'outputs.NotificationHubPropertiesResponse':
"""
NotificationHub properties.
"""
return pulumi.get(self, "properties")
@property
@pulumi.getter
def METHOD_NAME(self) -> Optional['outputs.SkuResponse']:
"""
The Sku description for a namespace
"""
return pulumi.get(self, "sku")
@property
@pulumi.getter(name="systemData")
def system_data(self) -> 'outputs.SystemDataResponse':
"""
Azure Resource Manager metadata containing createdBy and modifiedBy information.
"""
return pulumi.get(self, "system_data")
@property
@pulumi.getter
def tags(self) -> Optional[Mapping[str, str]]:
"""
Resource tags.
"""
return pulumi.get(self, "tags")
@property
@pulumi.getter
def type(self) -> str:
"""
The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or "Microsoft.Storage/storageAccounts"
"""
return pulumi.get(self, "type")
class AwaitableGetNotificationHubResult(GetNotificationHubResult):
# pylint: disable=using-constant-test
def __await__(self):
if False:
yield self
return GetNotificationHubResult(
id=self.id,
location=self.location,
name=self.name,
properties=self.properties,
METHOD_NAME=self.METHOD_NAME,
system_data=self.system_data,
tags=self.tags,
type=self.type)
def get_notification_hub(namespace_name: Optional[str] = None,
notification_hub_name: Optional[str] = None,
resource_group_name: Optional[str] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetNotificationHubResult:
"""
Notification Hub Resource.
:param str namespace_name: Namespace name
:param str notification_hub_name: Notification Hub name
:param str resource_group_name: The name of the resource group. The name is case insensitive.
"""
__args__ = dict()
__args__['namespaceName'] = namespace_name
__args__['notificationHubName'] = notification_hub_name
__args__['resourceGroupName'] = resource_group_name
opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)
__ret__ = pulumi.runtime.invoke('azure-native:notificationhubs/v20230901:getNotificationHub', __args__, opts=opts, typ=GetNotificationHubResult).value
return AwaitableGetNotificationHubResult(
id=pulumi.get(__ret__, 'id'),
location=pulumi.get(__ret__, 'location'),
name=pulumi.get(__ret__, 'name'),
properties=pulumi.get(__ret__, 'properties'),
METHOD_NAME=pulumi.get(__ret__, 'sku'),
system_data=pulumi.get(__ret__, 'system_data'),
tags=pulumi.get(__ret__, 'tags'),
type=pulumi.get(__ret__, 'type'))
@_utilities.lift_output_func(get_notification_hub)
def get_notification_hub_output(namespace_name: Optional[pulumi.Input[str]] = None,
notification_hub_name: Optional[pulumi.Input[str]] = None,
resource_group_name: Optional[pulumi.Input[str]] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetNotificationHubResult]:
"""
Notification Hub Resource.
:param str namespace_name: Namespace name
:param str notification_hub_name: Notification Hub name
:param str resource_group_name: The name of the resource group. The name is case insensitive.
"""
... | null |
add | """
Manage groups on Solaris
.. important::
If you feel that Salt should be using this module to manage groups on a
minion, and it is using a different module (or gives an error similar to
*'group.info' is not available*), see :ref:`here
<module-provider-override>`.
"""
import logging
try:
import grp
except ImportError:
pass
log = logging.getLogger(__name__)
# Define the module's virtual name
__virtualname__ = "group"
def __virtual__():
"""
Set the group module if the kernel is AIX
"""
if __grains__["kernel"] == "AIX":
return __virtualname__
return (
False,
"The aix_group execution module failed to load: only available on AIX systems.",
)
def METHOD_NAME(name, gid=None, system=False, root=None, **kwargs):
"""
Add the specified group
CLI Example:
.. code-block:: bash
salt '*' group.add foo 3456
"""
cmd = "mkgroup "
if system and root is not None:
cmd += "-a "
if gid:
cmd += "id={} ".format(gid)
cmd += name
ret = __salt__["cmd.run_all"](cmd, python_shell=False)
return not ret["retcode"]
def delete(name):
"""
Remove the named group
CLI Example:
.. code-block:: bash
salt '*' group.delete foo
"""
ret = __salt__["cmd.run_all"]("rmgroup {}".format(name), python_shell=False)
return not ret["retcode"]
def info(name):
"""
Return information about a group
CLI Example:
.. code-block:: bash
salt '*' group.info foo
"""
try:
grinfo = grp.getgrnam(name)
except KeyError:
return {}
else:
return {
"name": grinfo.gr_name,
"passwd": grinfo.gr_passwd,
"gid": grinfo.gr_gid,
"members": grinfo.gr_mem,
}
def getent(refresh=False):
"""
Return info on all groups
CLI Example:
.. code-block:: bash
salt '*' group.getent
"""
if "group.getent" in __context__ and not refresh:
return __context__["group.getent"]
ret = []
for grinfo in grp.getgrall():
ret.append(info(grinfo.gr_name))
__context__["group.getent"] = ret
return ret
def chgid(name, gid):
"""
Change the gid for a named group
CLI Example:
.. code-block:: bash
salt '*' group.chgid foo 4376
"""
pre_gid = __salt__["file.group_to_gid"](name)
if gid == pre_gid:
return True
cmd = "chgroup id={} {}".format(gid, name)
__salt__["cmd.run"](cmd, python_shell=False)
post_gid = __salt__["file.group_to_gid"](name)
if post_gid != pre_gid:
return post_gid == gid
return False
def adduser(name, username, root=None):
"""
Add a user in the group.
CLI Example:
.. code-block:: bash
salt '*' group.adduser foo bar
Verifies if a valid username 'bar' as a member of an existing group 'foo',
if not then adds it.
"""
cmd = "chgrpmem -m + {} {}".format(username, name)
retcode = __salt__["cmd.retcode"](cmd, python_shell=False)
return not retcode
def deluser(name, username, root=None):
"""
Remove a user from the group.
CLI Example:
.. code-block:: bash
salt '*' group.deluser foo bar
Removes a member user 'bar' from a group 'foo'. If group is not present
then returns True.
"""
grp_info = __salt__["group.info"](name)
try:
if username in grp_info["members"]:
cmd = "chgrpmem -m - {} {}".format(username, name)
ret = __salt__["cmd.run"](cmd, python_shell=False)
return not ret["retcode"]
else:
return True
except Exception: # pylint: disable=broad-except
return True
def members(name, members_list, root=None):
"""
Replaces members of the group with a provided list.
CLI Example:
.. code-block:: bash
salt '*' group.members foo 'user1,user2,user3,...'
Replaces a membership list for a local group 'foo'.
foo:x:1234:user1,user2,user3,...
"""
cmd = "chgrpmem -m = {} {}".format(members_list, name)
retcode = __salt__["cmd.retcode"](cmd, python_shell=False)
return not retcode | null |
wait on systemd | import queue
import subprocess
import threading
from steps.common import debug
class Podman:
def __init__(self, context, container_name):
self.context = context
self.container_name = container_name
self.container = None
debug(self.context, "Podman.__init__()")
self.new_container()
def __del__(self):
debug(self.context, "Podman.__del__()")
try:
self.kill()
except Exception:
pass
def kill(self):
debug(self.context, "Podman.kill()")
if not self.container:
return
self.container.kill()
self.container = None
def new_container(self):
debug(self.context, "Podman.new_container()")
# no need to stop the running container
# becuse the new container replaces an old container with the identical name
self.container = Container(self.context, name=self.container_name)
class ThreadedPodman:
def __init__(self, context, container_name_prefix, max_containers=1):
self.context = context
self.container = None
debug(self.context, "ThreadedPodman.__init__()")
self.max_containers = max_containers
self.container_name_prefix = container_name_prefix
self.container_name_num = 0
# produce new containers
self.container_producer_queue = queue.Queue(maxsize=self.max_containers)
self.container_producer_queue_is_stopping = threading.Event()
self.container_producer_queue_is_stopped = threading.Event()
self.container_producer_thread = threading.Thread(target=self.container_producer, daemon=True)
self.container_producer_thread.start()
# consume (kill) used containers
self.container_consumer_queue = queue.Queue()
self.container_consumer_thread = threading.Thread(target=self.container_consumer, daemon=True)
self.container_consumer_thread.start()
self.new_container()
def __del__(self):
debug(self.context, "ThreadedPodman.__del__()")
try:
self.kill()
except Exception:
pass
def kill(self):
debug(self.context, "ThreadedPodman.kill()")
self.container_producer_queue_is_stopping.set()
container = getattr(self, "container", None)
if container:
self.container_consumer_queue.put(container)
self.container = None
while not self.container_producer_queue_is_stopped.is_set():
try:
container = self.container_producer_queue.get(block=True, timeout=1)
self.container_consumer_queue.put(container)
except queue.Empty:
continue
# 'None' is a signal to finish processing the queue
self.container_consumer_queue.put(None)
self.container_producer_thread.join()
self.container_consumer_thread.join()
def container_producer(self):
while not self.container_producer_queue_is_stopping.is_set():
if self.container_name_prefix:
self.container_name_num += 1
container_name = f"{self.container_name_prefix}{self.container_name_num}"
else:
container_name = None
container = Container(self.context, name=container_name)
self.container_producer_queue.put(container, block=True)
self.container_producer_queue_is_stopped.set()
def container_consumer(self):
while True:
container = self.container_consumer_queue.get(block=True)
if container is None:
break
container.kill()
def new_container(self):
debug(self.context, "ThreadedPodman.new_container()")
if getattr(self, "container", None):
self.container_consumer_queue.put(self.container)
self.container = self.container_producer_queue.get(block=True)
debug(self.context, f"> {self.container}")
class Container:
def __init__(self, context, name=None):
self.context = context
debug(self.context, "Container.__init__()")
self.container_name = name
self.container_id = None
self.port = None
self.start()
def __del__(self):
try:
self.kill()
except Exception:
pass
def __repr__(self):
result = super().__repr__()
result += f"(port:{self.port}, id:{self.container_id}, name:{self.container_name})"
return result
def _run(self, args, check=True):
cmd = ["podman"] + args
debug(self.context, "Running command:", cmd)
proc = subprocess.run(
cmd,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding="utf-8",
check=check,
)
debug(self.context, "> return code:", proc.returncode)
debug(self.context, "> stdout:", proc.stdout)
debug(self.context, "> stderr:", proc.stderr)
return proc
def start(self):
debug(self.context, "Container.start()")
args = [
"run",
"--hostname", "obs-server-behave",
]
if self.container_name:
args += [
"--name", self.container_name,
"--replace",
"--stop-signal", "SIGKILL",
]
args += [
"--rm",
"--detach",
"--interactive",
"--tty",
"-p", "443",
"obs-server"
]
proc = self._run(args)
lines = proc.stdout.strip().splitlines()
self.container_id = lines[-1]
self.METHOD_NAME()
self.port = self.get_port()
def kill(self):
if not self.container_id:
return
debug(self.context, "Container.kill()")
args = ["kill", self.container_id]
self._run(args)
self.container_id = None
def restart(self):
debug(self.context, "Container.restart()")
self.kill()
self.start()
def METHOD_NAME(self):
args = [
"exec",
self.container_id,
"/usr/bin/systemctl", "is-system-running", "--wait"
]
self._run(args, check=False)
def get_port(self):
args = ["port", self.container_id]
proc = self._run(args)
lines = proc.stdout.strip().splitlines()
for line in lines:
if line.startswith("443/tcp"):
# return <port> from: "443/tcp -> 0.0.0.0:<port>"
return line.split(":")[-1]
raise RuntimeError(f"Could not determine port of container {self.container_id}") | null |
l2 norm scale | # Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
# Copyright (c) 2017-2019, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from paddle import fluid
from paddle.fluid.param_attr import ParamAttr
class VGG(object):
"""
VGG, see https://arxiv.org/abs/1409.1556
"""
def __init__(self):
super(VGG, self).__init__()
def __call__(self, input):
layers = []
layers += self._vgg_block(input)
layers += self._add_extras_block(layers[-1])
norm_cfg = [20., -1, -1, -1, -1, -1]
for k, v in enumerate(layers):
if not norm_cfg[k] == -1:
layers[k] = self.METHOD_NAME(v, init_scale=norm_cfg[k])
return layers
def _vgg_block(self, input):
num_layers = [2, 2, 3, 3, 3]
vgg_base = [64, 128, 256, 512, 512]
conv = input
layers = []
for k, v in enumerate(vgg_base):
conv = self._conv_block(
conv, v, num_layers[k], name="conv{}_".format(k + 1))
layers.append(conv)
if k == 4:
conv = self._pooling_block(conv, 3, 1, pool_padding=1)
else:
conv = self._pooling_block(conv, 2, 2)
fc6 = self._conv_layer(conv, 1024, 3, 1, 6, dilation=6, name="fc6")
fc7 = self._conv_layer(fc6, 1024, 1, 1, 0, name="fc7")
return [layers[3], fc7]
def _add_extras_block(self, input):
cfg = [[256, 512, 1, 2, 3], [128, 256, 1, 2, 3],
[128, 256, 0, 1, 3], [128, 256, 0, 1, 3]]
conv = input
layers = []
for k, v in enumerate(cfg):
conv = self._extra_block(
conv, v[0], v[1], v[2], v[3], v[4],
name="conv{}_".format(6 + k))
layers.append(conv)
return layers
def _conv_block(self, input, num_filter, groups, name=None):
conv = input
for i in range(groups):
conv = self._conv_layer(
input=conv,
num_filters=num_filter,
filter_size=3,
stride=1,
padding=1,
act='relu',
name=name + str(i + 1))
return conv
def _extra_block(self,
input,
num_filters1,
num_filters2,
padding_size,
stride_size,
filter_size,
name=None):
# 1x1 conv
conv_1 = self._conv_layer(
input=input,
num_filters=int(num_filters1),
filter_size=1,
stride=1,
act='relu',
padding=0,
name=name + "1")
# 3x3 conv
conv_2 = self._conv_layer(
input=conv_1,
num_filters=int(num_filters2),
filter_size=filter_size,
stride=stride_size,
act='relu',
padding=padding_size,
name=name + "2")
return conv_2
def _conv_layer(self,
input,
num_filters,
filter_size,
stride,
padding,
dilation=1,
act='relu',
use_cudnn=True,
name=None):
conv = fluid.layers.conv2d(
input=input,
num_filters=num_filters,
filter_size=filter_size,
stride=stride,
padding=padding,
dilation=dilation,
act=act,
use_cudnn=use_cudnn,
param_attr=ParamAttr(name=name + "_weights"),
bias_attr=ParamAttr(name=name + "_biases"),
name=name + '.conv2d.output.1')
return conv
def _pooling_block(self,
conv,
pool_size,
pool_stride,
pool_padding=0,
ceil_mode=True):
pool = fluid.layers.pool2d(
input=conv,
pool_size=pool_size,
pool_type='max',
pool_stride=pool_stride,
pool_padding=pool_padding,
ceil_mode=ceil_mode)
return pool
def METHOD_NAME(self, input, init_scale=1.0, channel_shared=False):
from paddle.fluid.layer_helper import LayerHelper
from paddle.fluid.initializer import Constant
helper = LayerHelper("Scale")
l2_norm = fluid.layers.l2_normalize(
input, axis=1) # l2 norm along channel
shape = [1] if channel_shared else [input.shape[1]]
scale = helper.create_parameter(
attr=helper.param_attr,
shape=shape,
dtype=input.dtype,
default_initializer=Constant(init_scale))
out = fluid.layers.elementwise_mul(
x=l2_norm, y=scale, axis=-1 if channel_shared else 1,
name="conv4_3_norm_scale")
return out
class SSD(object):
"""
Single Shot MultiBox Detector, see https://arxiv.org/abs/1512.02325
"""
def __init__(self, num_classes=81):
super(SSD, self).__init__()
self.backbone = VGG()
self.num_classes = num_classes
def __call__(self, image, gt_box, gt_label):
body_feats = self.backbone(image)
locs, confs, box, box_var = fluid.layers.multi_box_head(
inputs=body_feats,
image=image,
num_classes=self.num_classes,
min_ratio=15,
max_ratio=90,
base_size=300,
min_sizes=[30.0, 60.0, 111.0, 162.0, 213.0, 264.0],
max_sizes=[60.0, 111.0, 162.0, 213.0, 264.0, 315.0],
aspect_ratios=[[2.], [2., 3.], [2., 3.], [2., 3.], [2.], [2.]],
steps=[8, 16, 32, 64, 100, 300],
offset=0.5,
flip=True,
min_max_aspect_ratios_order=False,
kernel_size=3,
pad=1)
loss = fluid.layers.ssd_loss(locs, confs, gt_box, gt_label, box,
box_var)
loss = fluid.layers.reduce_sum(loss)
return loss | null |
up | import os
import click
from aimcore.cli.utils import set_log_level, start_uvicorn_app
from aimcore.cli.ui.utils import build_db_upgrade_command, get_free_port_num
from aimcore.web.configs import (
AIM_UI_BASE_PATH,
AIM_UI_DEFAULT_HOST,
AIM_UI_DEFAULT_PORT,
AIM_UI_MOUNTED_REPO_PATH,
AIM_UI_TELEMETRY_KEY,
AIM_PROXY_URL,
AIM_PROFILER_KEY
)
from aim._sdk.configs import AIM_ENV_MODE_KEY
from aim._sdk.repo import Repo
from aimcore.web.utils import exec_cmd
from aimcore.web.utils import ShellCommandException
from aim._ext.tracking import analytics
@click.command('ui')
@click.option('-h', '--host', default=AIM_UI_DEFAULT_HOST, type=str)
@click.option('-p', '--port', default=AIM_UI_DEFAULT_PORT, type=int)
@click.option('-w', '--workers', default=1, type=int)
@click.option('--uds', required=False, type=click.Path(exists=False,
file_okay=True,
dir_okay=False,
readable=True))
@click.option('--repo', required=False, default=os.getcwd(), type=click.Path(exists=True,
file_okay=False,
dir_okay=True,
writable=True))
@click.option('--package', '--pkg', required=False, default='', show_default='asp', type=str)
@click.option('--dev', is_flag=True, default=False)
@click.option('--ssl-keyfile', required=False, type=click.Path(exists=True,
file_okay=True,
dir_okay=False,
readable=True))
@click.option('--ssl-certfile', required=False, type=click.Path(exists=True,
file_okay=True,
dir_okay=False,
readable=True))
@click.option('--base-path', required=False, default='', type=str)
@click.option('--profiler', is_flag=True, default=False)
@click.option('--log-level', required=False, default='', type=str)
@click.option('-y', '--yes', is_flag=True, help='Automatically confirm prompt')
def ui(dev, host, port, workers, uds,
repo,
package,
ssl_keyfile, ssl_certfile,
base_path,
profiler, log_level, yes):
"""
Start Aim UI with the --repo repository.
"""
if not log_level:
log_level = 'debug' if dev else 'warning'
set_log_level(log_level)
os.environ[AIM_ENV_MODE_KEY] = 'dev' if dev else 'prod'
if base_path:
# process `base_path` as ui requires leading slash
if base_path.endswith('/'):
base_path = base_path[:-1]
if base_path and not base_path.startswith('/'):
base_path = f'/{base_path}'
os.environ[AIM_UI_BASE_PATH] = base_path
if not Repo.exists(repo):
init_repo = yes or click.confirm(f'\'{repo}\' is not a valid Aim repository. Do you want to initialize it?')
if not init_repo:
click.echo('To initialize repo please run the following command:')
click.secho('aim init', fg='yellow')
return
Repo.init(repo)
repo_inst = Repo.from_path(repo, read_only=True)
os.environ[AIM_UI_MOUNTED_REPO_PATH] = repo
dev_package_dir = repo_inst.dev_package_dir
if package:
repo_inst.set_active_package(pkg_name=package)
try:
db_cmd = build_db_upgrade_command()
exec_cmd(db_cmd, stream_output=True)
except ShellCommandException:
click.echo('Failed to initialize Aim DB. '
'Please see the logs above for details.')
exit(1)
if port == 0:
try:
port = get_free_port_num()
except Exception:
pass
if not dev and os.getenv(AIM_UI_TELEMETRY_KEY, 1) == '0':
click.echo(f'"{AIM_UI_TELEMETRY_KEY}" is ignored. Read how to opt-out here: '
f'https://aimstack.readthedocs.io/en/latest/community/telemetry.html')
if dev:
analytics.dev_mode = True
click.secho('Running Aim UI on repo `{}`'.format(repo_inst), fg='yellow')
if uds:
click.echo('Aim UI running on {}'.format(uds))
else:
scheme = 'https' if ssl_keyfile or ssl_certfile else 'http'
click.echo('Open {}://{}:{}{}'.format(scheme, host, port, base_path), err=True)
proxy_url = os.environ.get(AIM_PROXY_URL)
if proxy_url:
click.echo(f'Proxy {proxy_url}{base_path}/')
click.echo('Press Ctrl+C to exit')
analytics.track_event(event_name='[Aim UI] Start UI')
if profiler:
os.environ[AIM_PROFILER_KEY] = '1'
if dev:
import aim
import aimstack
import aimcore
reload_dirs = [os.path.dirname(aim.__file__), os.path.dirname(aimcore.__file__), os.path.dirname(aimstack.__file__), dev_package_dir]
else:
reload_dirs = []
try:
start_uvicorn_app('aimcore.web.run:app',
host=host, port=port, workers=workers, uds=uds,
ssl_keyfile=ssl_keyfile, ssl_certfile=ssl_certfile, log_level=log_level,
reload=dev, reload_dirs=reload_dirs)
except Exception:
click.echo('Failed to run Aim UI. '
'Please see the logs above for details.')
exit(1)
@click.command('up', context_settings={'ignore_unknown_options': True, 'allow_extra_args': True}, hidden=True)
def METHOD_NAME(**kwargs):
click.secho('Command `aim up` has been removed. Use `aim ui` instead.', fg='yellow') | null |
train forward | # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import paddle
import paddle.nn as nn
import paddle.nn.functional as F
import math
import net
class DygraphModel():
# define model
def create_model(self, config):
emb_path = config.get("hyper_parameters.emb_path")
vocab_size = config.get("hyper_parameters.vocab_size")
emb_size = config.get("hyper_parameters.emb_size")
kernel_num = config.get("hyper_parameters.kernel_num")
conv_filter = config.get("hyper_parameters.conv_filter")
conv_act = config.get("hyper_parameters.conv_act")
hidden_size = config.get("hyper_parameters.hidden_size")
out_size = config.get("hyper_parameters.out_size")
pool_size = config.get("hyper_parameters.pool_size")
pool_stride = config.get("hyper_parameters.pool_stride")
pool_padding = config.get("hyper_parameters.pool_padding")
pool_type = config.get("hyper_parameters.pool_type")
hidden_act = config.get("hyper_parameters.hidden_act")
pyramid_model = net.MatchPyramidLayer(
emb_path, vocab_size, emb_size, kernel_num, conv_filter, conv_act,
hidden_size, out_size, pool_size, pool_stride, pool_padding,
pool_type, hidden_act)
return pyramid_model
# define feeds which convert numpy of batch data to paddle.tensor
def create_feeds(self, batch_data, sentence_left_size,
sentence_right_size):
sentence_left = paddle.to_tensor(batch_data[0].numpy().astype('int64')
.reshape(-1, sentence_left_size))
sentence_right = paddle.to_tensor(batch_data[1].numpy().astype('int64')
.reshape(-1, sentence_right_size))
return [sentence_left, sentence_right]
# define loss function by predicts and label
def create_loss(self, prediction):
pos = paddle.slice(
prediction, axes=[0, 1], starts=[0, 0], ends=[64, 1])
neg = paddle.slice(
prediction, axes=[0, 1], starts=[64, 0], ends=[128, 1])
loss_part1 = paddle.subtract(
paddle.full(
shape=[64, 1], fill_value=1.0, dtype='float32'), pos)
loss_part2 = paddle.add(loss_part1, neg)
loss_part3 = paddle.maximum(
paddle.full(
shape=[64, 1], fill_value=0.0, dtype='float32'),
loss_part2)
avg_cost = paddle.mean(loss_part3)
return avg_cost
# define optimizer
def create_optimizer(self, dy_model, config):
lr = config.get("hyper_parameters.optimizer.learning_rate", 0.001)
optimizer = paddle.optimizer.Adam(
learning_rate=lr, parameters=dy_model.parameters())
return optimizer
# define metrics such as auc/acc
# multi-task need to define multi metric
def create_metrics(self):
metrics_list_name = []
metrics_list = []
return metrics_list, metrics_list_name
# construct train forward phase
def METHOD_NAME(self, dy_model, metrics_list, batch_data, config):
sentence_left_size = config.get("hyper_parameters.sentence_left_size")
sentence_right_size = config.get(
"hyper_parameters.sentence_right_size")
batch_size = config.get("runner.train_batch_size", 128)
inputs = self.create_feeds(batch_data, sentence_left_size,
sentence_right_size)
prediction = dy_model.forward(inputs)
loss = self.create_loss(prediction)
# update metrics
print_dict = {"loss": loss}
return loss, metrics_list, print_dict
def infer_forward(self, dy_model, metrics_list, batch_data, config):
sentence_left_size = config.get("hyper_parameters.sentence_left_size")
sentence_right_size = config.get(
"hyper_parameters.sentence_right_size")
batch_size = config.get("runner.infer_batch_size", 128)
inputs = self.create_feeds(batch_data, sentence_left_size,
sentence_right_size)
prediction = dy_model.forward(inputs)
# update metrics
print_dict = {"prediction": prediction}
return metrics_list, print_dict | null |
get config settings | # Copyright 2015 Red Hat Inc.
#
# SPDX-License-Identifier: Apache-2.0
"""Bandit is a tool designed to find common security issues in Python code."""
import argparse
import importlib
import logging
import os
import sys
import yaml
from bandit.core import extension_loader
PROG_NAME = "bandit_conf_generator"
LOG = logging.getLogger(__name__)
template = """
### Bandit config file generated from:
# '{cli}'
### This config may optionally select a subset of tests to run or skip by
### filling out the 'tests' and 'skips' lists given below. If no tests are
### specified for inclusion then it is assumed all tests are desired. The skips
### set will remove specific tests from the include set. This can be controlled
### using the -t/-s CLI options. Note that the same test ID should not appear
### in both 'tests' and 'skips', this would be nonsensical and is detected by
### Bandit at runtime.
# Available tests:
{test_list}
# (optional) list included test IDs here, eg '[B101, B406]':
{test}
# (optional) list skipped test IDs here, eg '[B101, B406]':
{skip}
### (optional) plugin settings - some test plugins require configuration data
### that may be given here, per-plugin. All bandit test plugins have a built in
### set of sensible defaults and these will be used if no configuration is
### provided. It is not necessary to provide settings for every (or any) plugin
### if the defaults are acceptable.
{settings}
"""
def init_logger():
"""Init logger."""
LOG.handlers = []
log_level = logging.INFO
log_format_string = "[%(levelname)5s]: %(message)s"
logging.captureWarnings(True)
LOG.setLevel(log_level)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter(log_format_string))
LOG.addHandler(handler)
def parse_args():
"""Parse arguments."""
help_description = """Bandit Config Generator
This tool is used to generate an optional profile. The profile may be used
to include or skip tests and override values for plugins.
When used to store an output profile, this tool will output a template that
includes all plugins and their default settings. Any settings which aren't
being overridden can be safely removed from the profile and default values
will be used. Bandit will prefer settings from the profile over the built
in values."""
parser = argparse.ArgumentParser(
description=help_description,
formatter_class=argparse.RawTextHelpFormatter,
)
parser.add_argument(
"--show-defaults",
dest="show_defaults",
action="store_true",
help="show the default settings values for each "
"plugin but do not output a profile",
)
parser.add_argument(
"-o",
"--out",
dest="output_file",
action="store",
help="output file to save profile",
)
parser.add_argument(
"-t",
"--tests",
dest="tests",
action="store",
default=None,
type=str,
help="list of test names to run",
)
parser.add_argument(
"-s",
"--skip",
dest="skips",
action="store",
default=None,
type=str,
help="list of test names to skip",
)
args = parser.parse_args()
if not args.output_file and not args.show_defaults:
parser.print_help()
parser.exit(1)
return args
def METHOD_NAME():
"""Get configuration settings."""
config = {}
for plugin in extension_loader.MANAGER.plugins:
fn_name = plugin.name
function = plugin.plugin
# if a function takes config...
if hasattr(function, "_takes_config"):
fn_module = importlib.import_module(function.__module__)
# call the config generator if it exists
if hasattr(fn_module, "gen_config"):
config[fn_name] = fn_module.gen_config(function._takes_config)
return yaml.safe_dump(config, default_flow_style=False)
def main():
"""Config generator to write configuration file."""
init_logger()
args = parse_args()
yaml_settings = METHOD_NAME()
if args.show_defaults:
print(yaml_settings)
if args.output_file:
if os.path.exists(os.path.abspath(args.output_file)):
LOG.error("File %s already exists, exiting", args.output_file)
sys.exit(2)
try:
with open(args.output_file, "w") as f:
skips = args.skips.split(",") if args.skips else []
tests = args.tests.split(",") if args.tests else []
for skip in skips:
if not extension_loader.MANAGER.check_id(skip):
raise RuntimeError(f"unknown ID in skips: {skip}")
for test in tests:
if not extension_loader.MANAGER.check_id(test):
raise RuntimeError(f"unknown ID in tests: {test}")
tpl = "# {0} : {1}"
test_list = [
tpl.format(t.plugin._test_id, t.name)
for t in extension_loader.MANAGER.plugins
]
others = [
tpl.format(k, v["name"])
for k, v in (
extension_loader.MANAGER.blacklist_by_id.items()
)
]
test_list.extend(others)
test_list.sort()
contents = template.format(
cli=" ".join(sys.argv),
settings=yaml_settings,
test_list="\n".join(test_list),
skip="skips: " + str(skips) if skips else "skips:",
test="tests: " + str(tests) if tests else "tests:",
)
f.write(contents)
except OSError:
LOG.error("Unable to open %s for writing", args.output_file)
except Exception as e:
LOG.error("Error: %s", e)
else:
LOG.info("Successfully wrote profile: %s", args.output_file)
return 0
if __name__ == "__main__":
sys.exit(main()) | null |
test future cancel no underlying future | import os
import concurrent.futures as cf
import pytest
from parsl.app.errors import AppException
from parsl.executors.flux.executor import (
FluxExecutor,
FluxFutureWrapper,
_complete_future,
)
try:
import flux.job.executor # noqa: F401
except ImportError:
FLUX_AVAIL = False
else:
FLUX_AVAIL = True
require_flux = pytest.mark.skipif(
not FLUX_AVAIL, reason="Flux not available, test will fail"
)
ERRMSG = "Some error message"
def multiply(x, y):
return x * y
def bad_foo():
raise ValueError(ERRMSG)
@require_flux
@pytest.mark.local
def test_multiply():
with FluxExecutor() as executor:
executor.start()
futures = [executor.submit(multiply, {}, i, 7) for i in range(5)]
for i, future in enumerate(futures):
assert future.result() == i * 7
assert future.done()
assert future.exception() is None
assert isinstance(future, FluxFutureWrapper)
@require_flux
@pytest.mark.local
def test_except():
with FluxExecutor() as executor:
executor.start()
future = executor.submit(bad_foo, {})
with pytest.raises(ValueError, match=ERRMSG):
future.result()
assert isinstance(future.exception(), ValueError)
@require_flux
@pytest.mark.local
@pytest.mark.skipif(
not hasattr(os, "sched_getaffinity") or len(os.sched_getaffinity(0)) < 2,
reason="Not Linux or too few CPUs",
)
def test_affinity():
with FluxExecutor() as executor:
executor.start()
future = executor.submit(os.sched_getaffinity, {"cores_per_task": 2}, 0)
assert len(future.result()) > 1
@require_flux
@pytest.mark.local
def test_cancel():
with FluxExecutor() as executor:
executor.start()
futures = [executor.submit(multiply, {}, i, 9) for i in range(3)]
for i, future in enumerate(futures):
if future.cancel():
assert future.cancelled()
assert future.done()
with pytest.raises(cf.CancelledError):
future.exception()
with pytest.raises(cf.CancelledError):
future.result()
else:
assert future.running()
assert future.done()
assert not future.cancelled()
assert future.result() == i * 9
@pytest.mark.local
def test_future_cancel():
underlying_future = cf.Future()
wrapper_future = FluxFutureWrapper()
wrapper_future._flux_future = underlying_future
assert not wrapper_future.done()
assert not wrapper_future.running()
assert not wrapper_future.cancelled()
assert wrapper_future.cancel() # should cancel underlying future
assert wrapper_future.cancel() # try again for good measure
assert wrapper_future.cancelled()
assert wrapper_future.done()
assert underlying_future.cancelled()
assert underlying_future.done()
@pytest.mark.local
def test_future_running():
underlying_future = cf.Future()
wrapper_future = FluxFutureWrapper()
assert not underlying_future.running()
assert underlying_future.set_running_or_notify_cancel()
assert underlying_future.running()
assert not wrapper_future.running()
wrapper_future._flux_future = underlying_future
assert wrapper_future.running()
@pytest.mark.local
def test_future_callback_returncode():
testfile = ".fluxexecutortest.txt"
returncode = 1
underlying_future = cf.Future()
wrapper_future = FluxFutureWrapper()
wrapper_future._flux_future = underlying_future
underlying_future.add_done_callback(
lambda fut: _complete_future(testfile, wrapper_future, fut)
)
underlying_future.set_result(returncode)
assert wrapper_future.done()
assert isinstance(wrapper_future.exception(), AppException)
@pytest.mark.local
def test_future_callback_nofile():
testfile = ".fluxexecutortest.txt"
returncode = 0
if os.path.isfile(testfile):
os.remove(testfile)
underlying_future = cf.Future()
wrapper_future = FluxFutureWrapper()
wrapper_future._flux_future = underlying_future
underlying_future.add_done_callback(
lambda fut: _complete_future(testfile, wrapper_future, fut)
)
underlying_future.set_result(returncode)
assert wrapper_future.done()
assert isinstance(wrapper_future.exception(), FileNotFoundError)
@pytest.mark.local
def test_future_callback_flux_exception():
underlying_future = cf.Future()
wrapper_future = FluxFutureWrapper()
wrapper_future._flux_future = underlying_future
underlying_future.add_done_callback(
lambda fut: _complete_future(".fluxexecutortest.txt", wrapper_future, fut)
)
underlying_future.set_exception(ValueError())
assert wrapper_future.done()
assert isinstance(wrapper_future.exception(), ValueError)
@pytest.mark.local
def METHOD_NAME():
wrapper_future = FluxFutureWrapper()
assert wrapper_future.cancel()
assert wrapper_future.cancelled()
assert not wrapper_future.running() | null |
set | """
Support for Debconf
"""
import logging
import os
import re
import salt.utils.files
import salt.utils.path
import salt.utils.stringutils
import salt.utils.versions
log = logging.getLogger(__name__)
__func_alias__ = {"set_": "set"}
# Define the module's virtual name
__virtualname__ = "debconf"
def __virtual__():
"""
Confirm this module is on a Debian based system and that debconf-utils
is installed.
"""
if __grains__["os_family"] != "Debian":
return (
False,
"The debconfmod module could not be loaded: unsupported OS family",
)
if salt.utils.path.which("debconf-get-selections") is None:
return (
False,
"The debconfmod module could not be loaded: "
"debconf-utils is not installed.",
)
return __virtualname__
def _unpack_lines(out):
"""
Unpack the debconf lines
"""
rexp = (
"(?ms)"
"^(?P<package>[^#]\\S+)[\t ]+"
"(?P<question>\\S+)[\t ]+"
"(?P<type>\\S+)[\t ]+"
"(?P<value>[^\n]*)$"
)
lines = re.findall(rexp, out)
return lines
def get_selections(fetchempty=True):
"""
Answers to debconf questions for all packages in the following format::
{'package': [['question', 'type', 'value'], ...]}
CLI Example:
.. code-block:: bash
salt '*' debconf.get_selections
"""
selections = {}
cmd = "debconf-get-selections"
out = __salt__["cmd.run_stdout"](cmd)
lines = _unpack_lines(out)
for line in lines:
package, question, type_, value = line
if fetchempty or value:
(selections.setdefault(package, []).append([question, type_, value]))
return selections
def show(name):
"""
Answers to debconf questions for a package in the following format::
[['question', 'type', 'value'], ...]
If debconf doesn't know about a package, we return None.
CLI Example:
.. code-block:: bash
salt '*' debconf.show <package name>
"""
selections = get_selections()
result = selections.get(name)
return result
def _set_file(path):
"""
Execute the set selections command for debconf
"""
cmd = "debconf-set-selections {}".format(path)
__salt__["cmd.run_stdout"](cmd, python_shell=False)
def METHOD_NAME(package, question, type, value, *extra):
"""
Set answers to debconf questions for a package.
CLI Example:
.. code-block:: bash
salt '*' debconf.set <package> <question> <type> <value> [<value> ...]
"""
if extra:
value = " ".join((value,) + tuple(extra))
fd_, fname = salt.utils.files.mkstemp(prefix="salt-", close_fd=False)
line = "{} {} {} {}".format(package, question, type, value)
os.write(fd_, salt.utils.stringutils.to_bytes(line))
os.close(fd_)
_set_file(fname)
os.unlink(fname)
return True
def set_template(path, template, context, defaults, saltenv="base", **kwargs):
"""
Set answers to debconf questions from a template.
path
location of the file containing the package selections
template
template format
context
variables to add to the template environment
default
default values for the template environment
CLI Example:
.. code-block:: bash
salt '*' debconf.set_template salt://pathto/pkg.selections.jinja jinja None None
"""
path = __salt__["cp.get_template"](
path=path,
dest=None,
template=template,
saltenv=saltenv,
context=context,
defaults=defaults,
**kwargs
)
return set_file(path, saltenv, **kwargs)
def set_file(path, saltenv="base", **kwargs):
"""
Set answers to debconf questions from a file.
CLI Example:
.. code-block:: bash
salt '*' debconf.set_file salt://pathto/pkg.selections
"""
if "__env__" in kwargs:
# "env" is not supported; Use "saltenv".
kwargs.pop("__env__")
path = __salt__["cp.cache_file"](path, saltenv)
if path:
_set_file(path)
return True
return False | null |
setup | """ Substitution of osparc variables and secrets
"""
import logging
from collections.abc import Callable, Mapping
from copy import deepcopy
from typing import Any
from fastapi import FastAPI
from models_library.projects import ProjectID
from models_library.projects_nodes_io import NodeID
from models_library.services import ServiceKey, ServiceVersion
from models_library.users import UserID
from models_library.utils.specs_substitution import SpecsSubstitutionsResolver
from pydantic import EmailStr
from ..utils.db import get_repository
from ..utils.osparc_variables import (
ContextDict,
OsparcVariablesTable,
resolve_variables_from_context,
)
from .db.repositories.services_environments import ServicesEnvironmentsRepository
_logger = logging.getLogger(__name__)
async def substitute_vendor_secrets_in_specs(
app: FastAPI,
specs: dict[str, Any],
*,
service_key: ServiceKey,
service_version: ServiceVersion,
) -> dict[str, Any]:
assert specs # nosec
resolver = SpecsSubstitutionsResolver(specs, upgrade=False)
repo = get_repository(app, ServicesEnvironmentsRepository)
if any(repo.is_vendor_secret_identifier(idr) for idr in resolver.get_identifiers()):
# checks before to avoid unnecesary calls to pg
vendor_secrets = await repo.get_vendor_secrets(
service_key=service_key, service_version=service_version
)
# resolve substitutions
resolver.set_substitutions(environs=vendor_secrets)
new_specs: dict[str, Any] = resolver.run()
return new_specs
return deepcopy(specs)
async def resolve_and_substitute_session_variables_in_specs(
app: FastAPI,
specs: dict[str, Any],
*,
user_id: UserID,
product_name: str,
project_id: ProjectID,
node_id: NodeID,
) -> dict[str, Any]:
assert specs # nosec
table: OsparcVariablesTable = app.state.session_variables_table
resolver = SpecsSubstitutionsResolver(specs, upgrade=False)
if requested := set(resolver.get_identifiers()):
available = set(table.variables_names())
if identifiers := available.intersection(requested):
environs = await resolve_variables_from_context(
table.copy(include=identifiers),
context=ContextDict(
app=app,
user_id=user_id,
product_name=product_name,
project_id=project_id,
node_id=node_id,
),
)
resolver.set_substitutions(environs=environs)
new_specs: dict[str, Any] = resolver.run()
return new_specs
return deepcopy(specs)
async def resolve_and_substitute_lifespan_variables_in_specs(
_app: FastAPI,
_specs: dict[str, Any],
*,
_callbacks_registry: Mapping[str, Callable],
):
raise NotImplementedError
async def _request_user_email(app: FastAPI, user_id: UserID) -> EmailStr:
repo = get_repository(app, ServicesEnvironmentsRepository)
return await repo.get_user_email(user_id=user_id)
async def _request_user_role(app: FastAPI, user_id: UserID):
repo = get_repository(app, ServicesEnvironmentsRepository)
return await repo.get_user_role(user_id=user_id)
def _setup_session_osparc_variables(app: FastAPI):
app.state.session_variables_table = table = OsparcVariablesTable()
# Registers some session osparc_variables
# WARNING: context_name needs to match session_context!
for name, context_name in [
("OSPARC_VARIABLE_PRODUCT_NAME", "product_name"),
("OSPARC_VARIABLE_STUDY_UUID", "project_id"),
("OSPARC_VARIABLE_NODE_ID", "node_id"),
]:
table.register_from_context(name, context_name)
table.register_from_handler("OSPARC_VARIABLE_USER_EMAIL")(_request_user_email)
table.register_from_handler("OSPARC_VARIABLE_USER_ROLE")(_request_user_role)
_logger.debug(
"Registered session_variables_table=%s", sorted(table.variables_names())
)
def METHOD_NAME(app: FastAPI):
"""
**o2sparc variables and secrets** are identifiers-value maps that are substituted on the service specs (e.g. docker-compose).
- **vendor secrets**: information set by a vendor on the platform. e.g. a vendor service license
- **session variables**: some session information as "current user email" or the "current product name"
- **lifespan variables**: produced before a service is started and cleaned up after it finishes (e.g. API tokens )
"""
def on_startup() -> None:
_setup_session_osparc_variables(app)
app.add_event_handler("startup", on_startup) | null |
finalize submodules | from collections.abc import Iterable
from ..._utils import flatten, deprecated
from ...hdl import dsl, ir
__all__ = ["Module", "FinalizeError"]
def _flat_list(e):
if isinstance(e, Iterable):
return list(flatten(e))
else:
return [e]
class CompatFinalizeError(Exception):
pass
FinalizeError = CompatFinalizeError
class _CompatModuleProxy:
def __init__(self, cm):
object.__setattr__(self, "_cm", cm)
class _CompatModuleComb(_CompatModuleProxy):
@deprecated("instead of `self.comb +=`, use `m.d.comb +=`")
def __iadd__(self, assigns):
self._cm._module._add_statement(assigns, domain=None, depth=0, compat_mode=True)
return self
class _CompatModuleSyncCD:
def __init__(self, cm, cd):
self._cm = cm
self._cd = cd
@deprecated("instead of `self.sync.<domain> +=`, use `m.d.<domain> +=`")
def __iadd__(self, assigns):
self._cm._module._add_statement(assigns, domain=self._cd, depth=0, compat_mode=True)
return self
class _CompatModuleSync(_CompatModuleProxy):
@deprecated("instead of `self.sync +=`, use `m.d.sync +=`")
def __iadd__(self, assigns):
self._cm._module._add_statement(assigns, domain="sync", depth=0, compat_mode=True)
return self
def __getattr__(self, name):
return _CompatModuleSyncCD(self._cm, name)
def __setattr__(self, name, value):
if not isinstance(value, _CompatModuleSyncCD):
raise AttributeError("Attempted to assign sync property - use += instead")
class _CompatModuleSpecials(_CompatModuleProxy):
@deprecated("instead of `self.specials.<name> =`, use `m.submodules.<name> =`")
def __setattr__(self, name, value):
self._cm._submodules.append((name, value))
setattr(self._cm, name, value)
@deprecated("instead of `self.specials +=`, use `m.submodules +=`")
def __iadd__(self, other):
self._cm._submodules += [(None, e) for e in _flat_list(other)]
return self
class _CompatModuleSubmodules(_CompatModuleProxy):
@deprecated("instead of `self.submodules.<name> =`, use `m.submodules.<name> =`")
def __setattr__(self, name, value):
self._cm._submodules.append((name, value))
setattr(self._cm, name, value)
@deprecated("instead of `self.submodules +=`, use `m.submodules +=`")
def __iadd__(self, other):
self._cm._submodules += [(None, e) for e in _flat_list(other)]
return self
class _CompatModuleClockDomains(_CompatModuleProxy):
@deprecated("instead of `self.clock_domains.<name> =`, use `m.domains.<name> =`")
def __setattr__(self, name, value):
self.__iadd__(value)
setattr(self._cm, name, value)
@deprecated("instead of `self.clock_domains +=`, use `m.domains +=`")
def __iadd__(self, other):
self._cm._module.domains += _flat_list(other)
return self
class CompatModule(ir.Elaboratable):
_MustUse__silence = True
# Actually returns another Amaranth Elaboratable (amaranth.dsl.Module), not a Fragment.
def get_fragment(self):
assert not self.get_fragment_called
self.get_fragment_called = True
self.finalize()
return self._module
def elaborate(self, platform):
if not self.get_fragment_called:
self.get_fragment()
return self._module
def __getattr__(self, name):
if name == "comb":
return _CompatModuleComb(self)
elif name == "sync":
return _CompatModuleSync(self)
elif name == "specials":
return _CompatModuleSpecials(self)
elif name == "submodules":
return _CompatModuleSubmodules(self)
elif name == "clock_domains":
return _CompatModuleClockDomains(self)
elif name == "finalized":
self.finalized = False
return self.finalized
elif name == "_module":
self._module = dsl.Module()
return self._module
elif name == "_submodules":
self._submodules = []
return self._submodules
elif name == "_clock_domains":
self._clock_domains = []
return self._clock_domains
elif name == "get_fragment_called":
self.get_fragment_called = False
return self.get_fragment_called
else:
raise AttributeError("'{}' object has no attribute '{}'"
.format(type(self).__name__, name))
def finalize(self, *args, **kwargs):
def METHOD_NAME():
for name, submodule in self._submodules:
if not hasattr(submodule, "finalize"):
continue
if submodule.finalized:
continue
submodule.finalize(*args, **kwargs)
if not self.finalized:
self.finalized = True
METHOD_NAME()
self.do_finalize(*args, **kwargs)
METHOD_NAME()
for name, submodule in self._submodules:
self._module._add_submodule(submodule, name)
def do_finalize(self):
pass
Module = CompatModule | null |
import success message | """Display information to help diagnose install issues.
Currently shows
- OS info
- Python info
- Gtk3 availability (gi.repository vs pgi)
This can be complemented by running the unittests.
"""
import os
import platform
import sys
import traceback
from collections import namedtuple
COL_WIDTH = 10
AvailableModules = namedtuple("AvailableModules", "gi pgi meta pango pubsub rsvg vext")
def display_platform():
# environment info
is_virtualenv = "VIRTUAL_ENV" in os.environ
# operating system info
def linux_distribution():
try:
return platform.linux_distribution()
except:
return "N/A"
# operating system info
def platform_dist():
try:
return platform.dist()
except:
return "N/A"
print(
"""Python:
sys.executable: %s
virtualenv: %s
version: %s
dist: %s
linux_distribution: %s
system: %s
machine: %s
platform: %s
version: %s
mac_ver: %s
win32_ver: %s
"""
% (
sys.executable,
is_virtualenv or "no",
" ".join(sys.version.split("\n")),
str(" ".join(platform_dist())),
" ".join(linux_distribution()),
platform.system(),
platform.machine(),
platform.platform(),
platform.version(),
platform.mac_ver(),
platform.win32_ver(),
),
)
def METHOD_NAME(module, name):
return "\n".join(
[
f" import {name.ljust(COL_WIDTH)} [success]:",
" " + module.__file__,
],
)
def import_fail_message(mn, reason):
return "\n".join(
[f" import {mn.ljust(COL_WIDTH)} [failed]:", " " + reason],
)
def test_import(name, failmsg=None, gi_require=None, gi=None):
if all([gi_require, gi]):
try:
gi.require_version(*gi_require)
except ValueError as ex:
# import will be attempted anyway
err = ex.args[0]
print(import_fail_message(name, err))
return
try:
module = __import__(name)
print(METHOD_NAME(module, name))
return module
except ImportError as e:
print(import_fail_message(name, failmsg or str(e)))
except Exception as e:
print(import_fail_message(name, str(e)))
def test_imports():
"""Attempt to import dependencies."""
print("Test Imports:")
# gtk
gi = test_import("gi")
pgi = test_import("pgi")
_gi = gi or pgi
if gi:
pango = test_import("gi.repository.Pango", gi_require=("Pango", "1.0"), gi=_gi)
else:
print(" No gi.repository.Pango implementation, text will not be available")
pango = None
# virtualenv help
vext = test_import("vext")
# internal dependencies
pubsub = test_import("pubsub")
meta = test_import("meta")
rsvg = test_import(
"gi.repository.Rsvg", "SVG Support unavailable", gi_require=("Rsvg", "2.0"),
)
return (
test_import("shoebot"),
AvailableModules(
gi=gi, pgi=pgi, meta=meta, pubsub=pubsub, rsvg=rsvg, vext=vext, pango=pango,
),
)
def shoebot_example(**shoebot_kwargs):
"""Decorator to run some code in a bot instance."""
def decorator(f):
def run():
from shoebot.util import ShoebotInstallError
print(f" Shoebot - {f.__name__.replace('_', ' ')}:")
try:
import shoebot
outputfile = f"/tmp/shoebot-{f.__name__}.png"
bot = shoebot.create_bot(outputfile=outputfile)
f(bot)
bot.finish()
print(f" [passed] : {outputfile}")
print("")
except ShoebotInstallError as e:
print(" [failed]", e.args[0])
print("")
except Exception:
print(" [failed] - traceback:")
for line in traceback.format_exc().splitlines():
print(f" {line}")
print("")
return run
return decorator
@shoebot_example()
def standard_module_example(bot):
bot.size(640, 480)
bot.fill(1, 0.5, 0.1)
bot.rect(10, 10, 100, 100)
@shoebot_example()
def module_using_text(bot):
bot.size(640, 480)
bot.stroke(0)
bot.text("Should work with gi not pgi", 0, 0)
def display_graphics_implementation():
print("Graphics Implementation:")
try:
from shoebot.core.backend import driver
for k, v in list(driver.get_libs().items()):
print(f" {k}: {v}")
except Exception as e:
raise
def diagnose():
display_platform()
shoebot_module, available_modules = test_imports()
if not shoebot_module:
print("Skipping shoebot module tests.")
return
display_graphics_implementation()
try:
import shoebot
except ImportError as e:
print("Cannot 'import shoebot'")
traceback.print_exc()
return False
print("\nShoebot Tests:")
# shoebot itself
standard_module_example()
# shoebot with text (will fail under pypy or pgi)
if available_modules.pango:
module_using_text()
if __name__ == "__main__":
diagnose() | null |
rebuild ctype | #
# Module which supports allocation of ctypes objects from shared memory
#
# multiprocessing/sharedctypes.py
#
# Copyright (c) 2006-2008, R Oudkerk
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions
# are met:
#
# 1. Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# 2. Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
# 3. Neither the name of author nor the names of any contributors may be
# used to endorse or promote products derived from this software
# without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
# ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
# OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
# OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
# SUCH DAMAGE.
#
import sys
import ctypes
import weakref
from multiprocessing import heap, RLock
from multiprocessing.forking import assert_spawning, ForkingPickler
__all__ = ['RawValue', 'RawArray', 'Value', 'Array', 'copy', 'synchronized']
#
#
#
typecode_to_type = {
'c': ctypes.c_char,
'b': ctypes.c_byte, 'B': ctypes.c_ubyte,
'h': ctypes.c_short, 'H': ctypes.c_ushort,
'i': ctypes.c_int, 'I': ctypes.c_uint,
'l': ctypes.c_long, 'L': ctypes.c_ulong,
'f': ctypes.c_float, 'd': ctypes.c_double
}
try:
typecode_to_type['u'] = ctypes.c_wchar
except AttributeError:
pass
#
#
#
def _new_value(type_):
size = ctypes.sizeof(type_)
wrapper = heap.BufferWrapper(size)
return METHOD_NAME(type_, wrapper, None)
def RawValue(typecode_or_type, *args):
'''
Returns a ctypes object allocated from shared memory
'''
type_ = typecode_to_type.get(typecode_or_type, typecode_or_type)
obj = _new_value(type_)
ctypes.memset(ctypes.addressof(obj), 0, ctypes.sizeof(obj))
obj.__init__(*args)
return obj
def RawArray(typecode_or_type, size_or_initializer):
'''
Returns a ctypes array allocated from shared memory
'''
type_ = typecode_to_type.get(typecode_or_type, typecode_or_type)
if isinstance(size_or_initializer, (int, long)):
type_ = type_ * size_or_initializer
obj = _new_value(type_)
ctypes.memset(ctypes.addressof(obj), 0, ctypes.sizeof(obj))
return obj
else:
type_ = type_ * len(size_or_initializer)
result = _new_value(type_)
result.__init__(*size_or_initializer)
return result
def Value(typecode_or_type, *args, **kwds):
'''
Return a synchronization wrapper for a Value
'''
lock = kwds.pop('lock', None)
if kwds:
raise ValueError('unrecognized keyword argument(s): %s' % kwds.keys())
obj = RawValue(typecode_or_type, *args)
if lock is False:
return obj
if lock in (True, None):
lock = RLock()
if not hasattr(lock, 'acquire'):
raise AttributeError("'%r' has no method 'acquire'" % lock)
return synchronized(obj, lock)
def Array(typecode_or_type, size_or_initializer, **kwds):
'''
Return a synchronization wrapper for a RawArray
'''
lock = kwds.pop('lock', None)
if kwds:
raise ValueError('unrecognized keyword argument(s): %s' % kwds.keys())
obj = RawArray(typecode_or_type, size_or_initializer)
if lock is False:
return obj
if lock in (True, None):
lock = RLock()
if not hasattr(lock, 'acquire'):
raise AttributeError("'%r' has no method 'acquire'" % lock)
return synchronized(obj, lock)
def copy(obj):
new_obj = _new_value(type(obj))
ctypes.pointer(new_obj)[0] = obj
return new_obj
def synchronized(obj, lock=None):
assert not isinstance(obj, SynchronizedBase), 'object already synchronized'
if isinstance(obj, ctypes._SimpleCData):
return Synchronized(obj, lock)
elif isinstance(obj, ctypes.Array):
if obj._type_ is ctypes.c_char:
return SynchronizedString(obj, lock)
return SynchronizedArray(obj, lock)
else:
cls = type(obj)
try:
scls = class_cache[cls]
except KeyError:
names = [field[0] for field in cls._fields_]
d = dict((name, make_property(name)) for name in names)
classname = 'Synchronized' + cls.__name__
scls = class_cache[cls] = type(classname, (SynchronizedBase,), d)
return scls(obj, lock)
#
# Functions for pickling/unpickling
#
def reduce_ctype(obj):
assert_spawning(obj)
if isinstance(obj, ctypes.Array):
return METHOD_NAME, (obj._type_, obj._wrapper, obj._length_)
else:
return METHOD_NAME, (type(obj), obj._wrapper, None)
def METHOD_NAME(type_, wrapper, length):
if length is not None:
type_ = type_ * length
ForkingPickler.register(type_, reduce_ctype)
obj = type_.from_address(wrapper.get_address())
obj._wrapper = wrapper
return obj
#
# Function to create properties
#
def make_property(name):
try:
return prop_cache[name]
except KeyError:
d = {}
exec template % ((name,)*7) in d
prop_cache[name] = d[name]
return d[name]
template = '''
def get%s(self):
self.acquire()
try:
return self._obj.%s
finally:
self.release()
def set%s(self, value):
self.acquire()
try:
self._obj.%s = value
finally:
self.release()
%s = property(get%s, set%s)
'''
prop_cache = {}
class_cache = weakref.WeakKeyDictionary()
#
# Synchronized wrappers
#
class SynchronizedBase(object):
def __init__(self, obj, lock=None):
self._obj = obj
self._lock = lock or RLock()
self.acquire = self._lock.acquire
self.release = self._lock.release
def __reduce__(self):
assert_spawning(self)
return synchronized, (self._obj, self._lock)
def get_obj(self):
return self._obj
def get_lock(self):
return self._lock
def __repr__(self):
return '<%s wrapper for %s>' % (type(self).__name__, self._obj)
class Synchronized(SynchronizedBase):
value = make_property('value')
class SynchronizedArray(SynchronizedBase):
def __len__(self):
return len(self._obj)
def __getitem__(self, i):
self.acquire()
try:
return self._obj[i]
finally:
self.release()
def __setitem__(self, i, value):
self.acquire()
try:
self._obj[i] = value
finally:
self.release()
def __getslice__(self, start, stop):
self.acquire()
try:
return self._obj[start:stop]
finally:
self.release()
def __setslice__(self, start, stop, values):
self.acquire()
try:
self._obj[start:stop] = values
finally:
self.release()
class SynchronizedString(SynchronizedArray):
value = make_property('value')
raw = make_property('raw') | null |
test default realizations | from argparse import Namespace
from uuid import UUID
import pytest
from ert.cli import model_factory
from ert.libres_facade import LibresFacade
from ert.run_models import (
EnsembleExperiment,
EnsembleSmoother,
IteratedEnsembleSmoother,
MultipleDataAssimilation,
SingleTestRun,
)
@pytest.mark.parametrize(
"target_case, format_mode, expected",
[
("test", False, "test"),
(None, False, "default_smoother_update"),
(None, True, "default_%d"),
],
)
def test_target_case_name(target_case, expected, format_mode, poly_case):
ert = poly_case
args = Namespace(current_case="default", target_case=target_case)
assert (
model_factory._target_case_name(ert, args, format_mode=format_mode) == expected
)
def METHOD_NAME(poly_case):
facade = LibresFacade(poly_case)
args = Namespace(realizations=None)
assert (
model_factory._realizations(args, facade.get_ensemble_size())
== [True] * facade.get_ensemble_size()
)
def test_custom_realizations(poly_case):
facade = LibresFacade(poly_case)
args = Namespace(realizations="0-4,7,8")
ensemble_size = facade.get_ensemble_size()
active_mask = [False] * ensemble_size
active_mask[0] = True
active_mask[1] = True
active_mask[2] = True
active_mask[3] = True
active_mask[4] = True
active_mask[7] = True
active_mask[8] = True
assert model_factory._realizations(args, ensemble_size) == active_mask
def test_setup_single_test_run(poly_case, storage):
ert = poly_case
model = model_factory._setup_single_test_run(
ert, storage, Namespace(current_case="default"), UUID(int=0)
)
assert isinstance(model, SingleTestRun)
assert len(model._simulation_arguments.keys()) == 3
assert "active_realizations" in model._simulation_arguments
def test_setup_ensemble_experiment(poly_case, storage):
ert = poly_case
args = Namespace(realizations=None, iter_num=1, current_case="default")
model = model_factory._setup_ensemble_experiment(
ert,
storage,
args,
UUID(int=0),
)
assert isinstance(model, EnsembleExperiment)
assert len(model._simulation_arguments.keys()) == 4
assert "active_realizations" in model._simulation_arguments
def test_setup_ensemble_smoother(poly_case, storage):
ert = poly_case
args = Namespace(
realizations="0-4,7,8", current_case="default", target_case="test_case"
)
model = model_factory._setup_ensemble_smoother(
ert,
storage,
args,
UUID(int=0),
)
assert isinstance(model, EnsembleSmoother)
assert len(model._simulation_arguments.keys()) == 5
assert "active_realizations" in model._simulation_arguments
assert "target_case" in model._simulation_arguments
assert "analysis_module" in model._simulation_arguments
def test_setup_multiple_data_assimilation(poly_case, storage):
ert = poly_case
args = Namespace(
realizations="0-4,7,8",
weights="6,4,2",
current_case="default",
target_case="test_case_%d",
start_iteration="0",
restart_run=False,
prior_ensemble="default",
)
model = model_factory._setup_multiple_data_assimilation(
ert,
storage,
args,
UUID(int=0),
)
assert isinstance(model, MultipleDataAssimilation)
assert len(model._simulation_arguments.keys()) == 8
assert "active_realizations" in model._simulation_arguments
assert "target_case" in model._simulation_arguments
assert "analysis_module" in model._simulation_arguments
assert "weights" in model._simulation_arguments
def test_setup_iterative_ensemble_smoother(poly_case, storage):
ert = poly_case
args = Namespace(
realizations="0-4,7,8",
current_case="default",
target_case="test_case_%d",
num_iterations="10",
)
model = model_factory._setup_iterative_ensemble_smoother(
ert,
storage,
args,
UUID(int=0),
)
assert isinstance(model, IteratedEnsembleSmoother)
assert len(model._simulation_arguments.keys()) == 6
assert "active_realizations" in model._simulation_arguments
assert "target_case" in model._simulation_arguments
assert "analysis_module" in model._simulation_arguments
assert "num_iterations" in model._simulation_arguments
assert LibresFacade(ert).get_number_of_iterations() == 10 | null |
run | """Support for OK-formatted test files"""
import os
import io
import doctest
import warnings
import pathlib
from contextlib import redirect_stderr, redirect_stdout
from textwrap import dedent
from .abstract_test import TestFile, TestCase, TestCaseResult
from ..utils import hide_outputs
def run_doctest(name, doctest_string, global_environment):
"""
Run a single test with given ``global_environment``. Returns ``(True, '')`` if the doctest passes.
Returns ``(False, failure_message)`` if the doctest fails.
Args:
name (``str``): name of doctest
doctest_string (``str``): doctest in string form
global_environment (``dict``): global environment resulting from the execution of a python
script/notebook
Returns:
``tuple`` of (``bool``, ``str``): results from running the test
"""
examples = doctest.DocTestParser().parse(
doctest_string,
name
)
test = doctest.DocTest(
[e for e in examples if isinstance(e, doctest.Example)],
global_environment,
name,
None,
None,
doctest_string
)
doctestrunner = doctest.DocTestRunner(verbose=True)
runresults = io.StringIO()
with redirect_stdout(runresults), redirect_stderr(runresults), hide_outputs():
doctestrunner.METHOD_NAME(test, clear_globs=False)
with open(os.devnull, 'w') as f, redirect_stderr(f), redirect_stdout(f):
result = doctestrunner.summarize(verbose=True)
# An individual test can only pass or fail
if result.failed == 0:
return (True, '')
else:
return False, runresults.getvalue()
class OKTestFile(TestFile):
"""
A single OK-formatted test file for Otter.
Args:
name (``str``): the name of test file
path (``str``): the path to the test file
test_cases (``list`` of ``TestCase``): a list of parsed tests to be run
value (``int``, optional): the point value of this test, defaults to 1
all_or_nothing (``bool``, optional): whether the test should be graded all-or-nothing across
cases
Attributes:
name (``str``): the name of test file
path (``str``): the path to the test file
test_cases (``list`` of ``TestCase``): a list of parsed tests to be run
value (``int``): the point value of this test, defaults to 1
all_or_nothing (``bool``): whether the test should be graded all-or-nothing across
cases
passed_all (``bool``): whether all of the test cases were passed
test_case_results (``list`` of ``TestCaseResult``): a list of results for the test cases in
``test_cases``
grade (``float``): the percentage of ``points`` earned for this test file as a decimal
"""
def METHOD_NAME(self, global_environment):
"""
Run the test cases on ``global_environment``, saving the results in
``self.test_case_results``.
Arguments:
``global_environment`` (``dict``): result of executing a Python notebook/script
"""
for i, test_case in enumerate(self.test_cases):
passed, result = run_doctest(self.name + ' ' + str(i), test_case.body, global_environment)
if passed:
result = '✅ Test case passed'
else:
result = '❌ Test case failed\n' + result
self.test_case_results.append(TestCaseResult(
test_case = test_case,
message = result,
passed = passed,
))
@classmethod
def from_spec(cls, test_spec, path=""):
"""
Parse an OK-formatted ``dict`` and return an ``OKTestFile``.
Args:
test_spec (``dict[str: object]``): the OK-formatted ``dict``
path (``str``, optional): the path to the test file this ``dict`` was parsed from
Returns:
``OKTestFile``: the new ``OKTestFile`` object created from the given file
"""
# Make sure there is a name
assert 'name' in test_spec
# Do not support multiple suites in the same file
assert len(test_spec['suites']) == 1
test_suite = test_spec['suites'][0]
# Only support doctest. I am unsure if other tests are implemented
assert test_suite.get('type', 'doctest') == 'doctest'
# Not setup and teardown supported
assert not bool(test_suite.get('setup'))
assert not bool(test_suite.get('teardown'))
test_cases = []
for i, test_case in enumerate(test_spec['suites'][0]['cases']):
test_cases.append(TestCase(
name = test_case.get('name', f"{test_spec['name']} - {i + 1}"),
body = dedent(test_case['code']),
hidden = test_case.get('hidden', True),
points = test_case.get('points', None),
success_message = test_case.get('success_message', None),
failure_message = test_case.get('failure_message', None)
))
# resolve point values for each test case
spec_pts = test_spec.get('points', None)
test_cases = cls.resolve_test_file_points(spec_pts, test_cases)
# convert path into PurePosixPath for test name
path = str(pathlib.Path(path).as_posix())
# grab whether the tests are all-or-nothing
all_or_nothing = test_spec.get('all_or_nothing', True)
return cls(test_spec['name'], path, test_cases, all_or_nothing)
@classmethod
def from_file(cls, path):
"""
Parse an OK-formatted test file & return an ``OKTestFile``.
Args:
path (``str``): the path to the test file
Returns:
``OKTestFile``: the new ``OKTestFile`` object created from the given file
"""
# ok test files are python files, with a global 'test' defined
test_globals = {}
with open(path) as f:
exec(f.read(), test_globals)
test_spec = test_globals['test']
return cls.from_spec(test_spec, path=path) | null |
biosample characterization | import pytest
@pytest.fixture
def biosample_characterization_no_review(testapp, award, lab, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def METHOD_NAME(testapp, award, lab, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def biosample_characterization_2nd_opinion(testapp, award, lab, submitter, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
'review': {
'status': 'requires secondary opinion',
'lab': lab['@id'],
'reviewed_by': submitter['@id'],
},
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def biosample_characterization_exempt(testapp, award, lab, submitter, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
'review': {
'status': 'exempt from standards',
'lab': lab['@id'],
'reviewed_by': submitter['@id'],
},
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def biosample_characterization_not_compliant(testapp, award, lab, submitter, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
'review': {
'status': 'not compliant',
'lab': lab['@id'],
'reviewed_by': submitter['@id'],
},
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def biosample_characterization_compliant(testapp, award, lab, submitter, biosample, attachment):
item = {
'characterizes': biosample['@id'],
'award': award['@id'],
'lab': lab['@id'],
'attachment': attachment,
'review': {
'status': 'compliant',
'lab': lab['@id'],
'reviewed_by': submitter['@id'],
},
}
return testapp.post_json('/biosample_characterization', item).json['@graph'][0]
@pytest.fixture
def biosample_characterization_1(biosample_characterization_base):
item = biosample_characterization_base.copy()
item.update({
'schema_version': '2',
'status': 'APPROVED',
'characterization_method': 'immunofluorescence',
})
return item
@pytest.fixture
def biosample_characterization_2(biosample_characterization_base):
item = biosample_characterization_base.copy()
item.update({
'schema_version': '3',
'status': 'IN PROGRESS',
'award': '1a4d6443-8e29-4b4a-99dd-f93e72d42418'
})
return item
@pytest.fixture
def antibody_characterization_3(antibody_characterization):
item = antibody_characterization.copy()
item.update({
'schema_version': '4',
'characterization_method': 'immunoblot',
})
return item
@pytest.fixture
def biosample_characterization_4(root, METHOD_NAME, publication):
item = root.get_by_uuid(METHOD_NAME['uuid'])
properties = item.properties.copy()
properties.update({
'schema_version': '4',
'references': [publication['identifiers'][0]],
})
return properties
@pytest.fixture
def antibody_characterization_10(antibody_characterization_1):
item = antibody_characterization_1.copy()
item.update({
'status': 'pending dcc review',
'characterization_method': 'immunoprecipitation followed by mass spectrometry',
'comment': 'We tried really hard to characterize this antibody.',
'notes': 'Your plea has been noted.'
})
return item
@pytest.fixture
def antibody_characterization_11(antibody_characterization):
item = antibody_characterization.copy()
item.update({
'characterization_reviews': [{
'biosample_term_name': 'K562',
'biosample_term_id': 'EFO:0002067',
'lane_status': 'exempt from standards',
'biosample_type': 'immortalized cell line',
'lane': 2,
'organism': '/organisms/human/'
}]
})
return item
@pytest.fixture
def antibody_characterization_13(antibody_characterization):
item = antibody_characterization.copy()
item.update({
'characterization_reviews': [{
'biosample_term_name': 'HUES62',
'biosample_term_id': 'EFO:0007087',
'lane_status': 'exempt from standards',
'biosample_type': 'induced pluripotent stem cell line',
'lane': 2,
'organism': '/organisms/human/'
}]
})
return item
@pytest.fixture
def antibody_characterization_14(antibody_characterization):
item = antibody_characterization.copy()
item.update({
'characterization_reviews': [{
'biosample_term_name': 'A549',
'biosample_term_id': 'EFO:0001086',
'lane_status': 'exempt from standards',
'biosample_type': 'cell line',
'lane': 2,
'organism': '/organisms/human/'
}]
})
return item | null |
target positive classification loss | import numpy as np
import pytest
from paz.optimization.losses.multi_box_loss import MultiBoxLoss
@pytest.fixture
def loss():
return MultiBoxLoss()
@pytest.fixture
def y_true():
y_true = np.array(
[[38.38629, 48.666668, 10.362101, 11.512976, 0., 1.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0.],
[27.143208, 34.41253, 8.629259, 9.7801285, 1., 0.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0.],
[27.143208, 68.82506, 8.629259, 13.245829, 1., 0.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0., 0., 0., 0., 0., 0.,
0.]], dtype='float32')
return np.expand_dims(y_true, axis=0)
@pytest.fixture
def y_pred():
y_pred = np.array(
[[36.99653894, 46.4176432, 10.35266677, 10.1656072, 0.05621409,
0.98060555, 0.01017545, 0.03181472, 0.02227341, 0.00503445,
0.00746015, 0.15980312, 0.10174269, 0.01672697, 0.0111077,
0.02144868, 0.07346129, 0.03899017, 0.01071656, 0.03946776,
0.0830264, 0.06763985, 0.04077367, 0.07804006, 0.04347721],
[26.614379, 32.0909085, 4.2000501, 7.0869583, 0.0423508, 0.91125538,
0.04441671, 0.03053759, 0.07411292, 0.03454058, 0.04849431,
0.0592223, 0.0134144, 0.09800261, 0.0433236, 0.04486571,
0.01135817, 0.08123691, 0.02096761, 0.03070671, 0.04680151,
0.12522466, 0.06783583, 0.05873021, 0.01260151],
[2.16936564, 4.4787911, 6.314962, 4.42737758, 0.83406942,
0.04166197, 0.01605819, 0.04750001, 0.01329675, 0.0126452,
0.02085183, 0.0172693, 0.03088947, 0.02661936, 0.01231482,
0.04099588, 0.02453831, 0.07038483, 0.06579002, 0.13424149,
0.04614118, 0.03297557, 0.1374058, 0.15315633, 0.02119431]],
dtype='float32')
return np.expand_dims(y_pred, axis=0)
@pytest.fixture
def target_multibox_loss():
return 6.8489789962768555
@pytest.fixture
def target_smooth_l1_loss():
return np.array([[3.5220284, 8.989227, 98.507996]], dtype='float32')
@pytest.fixture
def target_cross_entropy_loss():
return np.array([[0.019584997, 3.161768, 0.18143862]], dtype='float32')
@pytest.fixture
def target_localization_loss():
return np.array(3.4861877, dtype='float32')
@pytest.fixture
def METHOD_NAME():
return np.array(0.019584997, dtype='float32')
@pytest.fixture
def target_negative_classification_loss():
return np.array(3.3432066, dtype='float32')
def test_multiboxloss(y_true, y_pred, loss, target_multibox_loss):
total_loss = loss.compute_loss(y_true, y_pred)
assert (float(total_loss) == target_multibox_loss)
def test_smooth_l1_loss(y_true, y_pred, loss, target_smooth_l1_loss):
smooth_l1_loss = loss._smooth_l1(y_true, y_pred)
smooth_l1_loss = np.asarray(smooth_l1_loss, dtype='float32')
print(smooth_l1_loss - target_smooth_l1_loss)
assert np.allclose(smooth_l1_loss, target_smooth_l1_loss)
def test_cross_entropy_loss(y_true, y_pred, loss, target_cross_entropy_loss):
cross_entropy_loss = loss._cross_entropy(y_true, y_pred)
cross_entropy_loss = np.asarray(cross_entropy_loss, dtype='float32')
assert np.allclose(cross_entropy_loss, target_cross_entropy_loss)
def test_localization_loss(y_true, y_pred, loss, target_localization_loss):
localization_loss = loss.localization(y_true, y_pred)
localization_loss = np.asarray(localization_loss, dtype='float32')
assert np.allclose(localization_loss, target_localization_loss)
def test_positive_classification_loss(y_true, y_pred, loss,
METHOD_NAME):
positive_classification_loss = loss.positive_classification(y_true, y_pred)
positive_classification_loss = np.asarray(
positive_classification_loss, dtype='float32')
assert np.allclose(
positive_classification_loss, METHOD_NAME)
def test_negative_classification_loss(y_true, y_pred, loss,
target_negative_classification_loss):
negative_classification_loss = loss.negative_classification(y_true, y_pred)
negative_classification_loss = np.asarray(
negative_classification_loss, dtype='float32')
assert np.allclose(
negative_classification_loss, target_negative_classification_loss) | null |
get eeprom dom raw | # Platform-specific SFP transceiver interface for SONiC
# This plugin supports QSFP and SFP.
try:
import time
from sonic_sfp.sfputilbase import SfpUtilBase
except ImportError as e:
raise ImportError("%s - required module not found" % str(e))
class SfpUtil(SfpUtilBase):
"""Platform-specific SfpUtil class"""
PORT_START = 0
PORT_END = 34
QSFP_PORT_START = 0
QSFP_PORT_END = 32
SFP_PORT_START = 33
SFP_PORT_END = 34
EEPROM_OFFSET = 11
PORT_INFO_PATH = '/sys/class/t7132s_cpld'
_port_name = ""
_port_to_eeprom_mapping = {}
_port_to_i2cbus_mapping = {}
_port_to_offset = [11, 30, 12, 29, 13, 28, 14, 27, 15, 34,
16, 33, 17, 32, 18, 31, 19, 38, 20, 37,
21, 36, 22, 35, 23, 42, 24, 41, 25, 40,
26, 39,
43, 44]
_global_port_pres_dict = {}
@property
def port_start(self):
return self.PORT_START
@property
def port_end(self):
return self.PORT_END
@property
def qsfp_ports(self):
return []
@property
def qsfp_ports(self):
return list(range(self.QSFP_PORT_START, self.QSFP_PORT_END))
@property
def port_to_eeprom_mapping(self):
return self._port_to_eeprom_mapping
@property
def port_to_i2cbus_mapping(self):
return self._port_to_i2cbus_mapping
def get_port_name(self, port_num):
if port_num in self.qsfp_ports:
self._port_name = "QSFP" + str(port_num + 1)
else:
self._port_name = "SFP" + str(port_num - self.QSFP_PORT_END + 1)
return self._port_name
def METHOD_NAME(self, port_num):
if port_num in self.qsfp_ports:
# QSFP DOM EEPROM is also at addr 0x50 and thus also stored in eeprom_ifraw
return None
else:
# Read dom eeprom at addr 0x51
return self._read_eeprom_devid(port_num, self.DOM_EEPROM_ADDR, 256)
def __init__(self):
# Override port_to_eeprom_mapping for class initialization
eeprom_path = '/sys/bus/i2c/devices/i2c-{0}/{0}-0050/eeprom'
for x in range(self.PORT_START, self.PORT_END):
self.port_to_i2cbus_mapping[x] = (x + self.EEPROM_OFFSET)
self.port_to_eeprom_mapping[x] = eeprom_path.format(
self._port_to_offset[x])
self.init_global_port_presence()
SfpUtilBase.__init__(self)
def init_global_port_presence(self):
for port_num in range(self.port_start, (self.port_end)):
presence = self.get_presence(port_num)
if(presence):
self._global_port_pres_dict[port_num] = '1'
else:
self._global_port_pres_dict[port_num] = '0'
def get_presence(self, port_num):
# Check for invalid port_num
if port_num not in list(range(self.port_start, self.port_end + 1)):
return False
# Get path for access port presence status
port_name = self.get_port_name(port_num)
sysfs_filename = "qsfp_modprs" if port_num in self.qsfp_ports else "sfp_modabs"
reg_path = "/".join([self.PORT_INFO_PATH, port_name, sysfs_filename])
# Read status
try:
reg_file = open(reg_path)
content = reg_file.readline().rstrip()
reg_value = int(content)
except IOError as e:
print("Error: unable to open file: %s" % str(e))
return False
# Module present is active low
if reg_value == 0:
return True
return False
def get_low_power_mode(self, port_num):
# Check for invalid QSFP port_num
if port_num not in self.qsfp_ports:
return False
try:
port_name = self.get_port_name(port_num)
reg_file = open("/".join([self.PORT_INFO_PATH,
port_name, "qsfp_lpmode"]))
except IOError as e:
print("Error: unable to open file: %s" % str(e))
return False
# Read status
content = reg_file.readline().rstrip()
reg_value = int(content)
# low power mode is active high
if reg_value == 0:
return False
return True
def set_low_power_mode(self, port_num, lpmode):
# Check for invalid QSFP port_num
if port_num not in self.qsfp_ports:
return False
try:
port_name = self.get_port_name(port_num)
reg_file = open("/".join([self.PORT_INFO_PATH,
port_name, "qsfp_lpmode"]), "r+")
except IOError as e:
print("Error: unable to open file: %s" % str(e))
return False
content = hex(lpmode)
reg_file.seek(0)
reg_file.write(content)
reg_file.close()
return True
def reset(self, port_num):
# Check for invalid QSFP port_num
if port_num not in self.qsfp_ports:
return False
try:
port_name = self.get_port_name(port_num)
reg_file = open("/".join([self.PORT_INFO_PATH,
port_name, "qsfp_reset"]), "w")
except IOError as e:
print("Error: unable to open file: %s" % str(e))
return False
# Convert our register value back to a hex string and write back
reg_file.seek(0)
reg_file.write(hex(0))
reg_file.close()
# Sleep 1 second to allow it to settle
time.sleep(1)
# Flip the bit back high and write back to the register to take port out of reset
try:
reg_file = open(
"/".join([self.PORT_INFO_PATH, port_name, "qsfp_reset"]), "w")
except IOError as e:
print("Error: unable to open file: %s" % str(e))
return False
reg_file.seek(0)
reg_file.write(hex(1))
reg_file.close()
return True
def get_transceiver_change_event(self):
port_dict = {}
while True:
for port_num in range(self.port_start, (self.port_end + 1)):
presence = self.get_presence(port_num)
if(presence and self._global_port_pres_dict[port_num] == '0'):
self._global_port_pres_dict[port_num] = '1'
port_dict[port_num] = '1'
elif(not presence and
self._global_port_pres_dict[port_num] == '1'):
self._global_port_pres_dict[port_num] = '0'
port_dict[port_num] = '0'
if(len(port_dict) > 0):
return True, port_dict
time.sleep(0.5) | null |
impute inactive | from __future__ import annotations
from abc import abstractmethod
from typing import Any
import numpy as np
import sklearn.gaussian_process
from ConfigSpace import ConfigurationSpace
from sklearn.gaussian_process import GaussianProcessRegressor
from sklearn.gaussian_process.kernels import Kernel, KernelOperator
from smac.model.abstract_model import AbstractModel
from smac.model.gaussian_process.priors.abstract_prior import AbstractPrior
from smac.model.gaussian_process.priors.tophat_prior import SoftTopHatPrior, TophatPrior
__copyright__ = "Copyright 2022, automl.org"
__license__ = "3-clause BSD"
class AbstractGaussianProcess(AbstractModel):
"""Abstract base class for all Gaussian process models.
Parameters
----------
configspace : ConfigurationSpace
kernel : Kernel
Kernel which is used for the Gaussian process.
instance_features : dict[str, list[int | float]] | None, defaults to None
Features (list of int or floats) of the instances (str). The features are incorporated into the X data,
on which the model is trained on.
pca_components : float, defaults to 7
Number of components to keep when using PCA to reduce dimensionality of instance features.
seed : int
"""
def __init__(
self,
configspace: ConfigurationSpace,
kernel: Kernel,
instance_features: dict[str, list[int | float]] | None = None,
pca_components: int | None = 7,
seed: int = 0,
):
super().__init__(
configspace=configspace,
instance_features=instance_features,
pca_components=pca_components,
seed=seed,
)
self._kernel = kernel
self._gp = self._get_gaussian_process()
@property
def meta(self) -> dict[str, Any]: # noqa: D102
meta = super().meta
meta.update({"kernel": self._kernel.meta})
return meta
@abstractmethod
def _get_gaussian_process(self) -> GaussianProcessRegressor:
"""Generates a Gaussian process."""
raise NotImplementedError()
def _normalize(self, y: np.ndarray) -> np.ndarray:
"""Normalize data to zero mean unit standard deviation.
Parameters
----------
y : np.ndarray
Target values for the Gaussian process.
Returns
-------
normalized_y : np.ndarray
Normalized y values.
"""
self.mean_y_ = np.mean(y)
self.std_y_ = np.std(y)
if self.std_y_ == 0:
self.std_y_ = 1
return (y - self.mean_y_) / self.std_y_
def _untransform_y(
self,
y: np.ndarray,
var: np.ndarray | None = None,
) -> np.ndarray | tuple[np.ndarray, np.ndarray]:
"""Transform zero mean unit standard deviation data into the regular space.
Warning
-------
This function should be used after a prediction with the Gaussian process which was
trained on normalized data.
Parameters
----------
y : np.ndarray
Normalized data.
var : np.ndarray | None, defaults to None
Normalized variance.
Returns
-------
untransformed_y : np.ndarray | tuple[np.ndarray, np.ndarray]
"""
y = y * self.std_y_ + self.mean_y_
if var is not None:
var = var * self.std_y_**2
return y, var # type: ignore
return y
def _get_all_priors(
self,
add_bound_priors: bool = True,
add_soft_bounds: bool = False,
) -> list[list[AbstractPrior]]:
"""Returns all priors."""
# Obtain a list of all priors for each tunable hyperparameter of the kernel
all_priors = []
to_visit = []
to_visit.append(self._gp.kernel.k1)
to_visit.append(self._gp.kernel.k2)
while len(to_visit) > 0:
current_param = to_visit.pop(0)
if isinstance(current_param, KernelOperator):
to_visit.insert(0, current_param.k1)
to_visit.insert(1, current_param.k2)
continue
elif isinstance(current_param, Kernel):
hps = current_param.hyperparameters
assert len(hps) == 1
hp = hps[0]
if hp.fixed:
continue
bounds = hps[0].bounds
for i in range(hps[0].n_elements):
priors_for_hp = []
if current_param.prior is not None:
priors_for_hp.append(current_param.prior)
if add_bound_priors:
if add_soft_bounds:
priors_for_hp.append(
SoftTopHatPrior(
lower_bound=bounds[i][0],
upper_bound=bounds[i][1],
seed=self._rng.randint(0, 2**20),
exponent=2,
)
)
else:
priors_for_hp.append(
TophatPrior(
lower_bound=bounds[i][0],
upper_bound=bounds[i][1],
seed=self._rng.randint(0, 2**20),
)
)
all_priors.append(priors_for_hp)
return all_priors
def _set_has_conditions(self) -> None:
"""Sets `has_conditions` on `current_param`."""
has_conditions = len(self._configspace.get_conditions()) > 0
to_visit = []
to_visit.append(self._kernel)
while len(to_visit) > 0:
current_param = to_visit.pop(0)
if isinstance(current_param, sklearn.gaussian_process.kernels.KernelOperator):
to_visit.insert(0, current_param.k1)
to_visit.insert(1, current_param.k2)
current_param.has_conditions = has_conditions
elif isinstance(current_param, sklearn.gaussian_process.kernels.Kernel):
current_param.has_conditions = has_conditions
else:
raise ValueError(current_param)
def METHOD_NAME(self, X: np.ndarray) -> np.ndarray:
"""Imputes inactives."""
X = X.copy()
X[~np.isfinite(X)] = -1
return X | null |
get video url | # -*- coding: utf-8 -*-
# Copyright: (c) 2017, SylvainCecchetto
# GNU General Public License v2.0+ (see LICENSE.txt or https://www.gnu.org/licenses/gpl-2.0.txt)
# This file is part of Catch-up TV & More
from __future__ import unicode_literals
from builtins import str
import re
from codequick import Listitem, Resolver, Route, utils
import urlquick
from resources.lib import resolver_proxy
from resources.lib.menu_utils import item_post_treatment
# TO DO
# Play Spanish Videos
URL_ROOT = utils.urljoin_partial('https://www.tetesaclaques.tv')
@Route.register
def website_root(plugin, item_id, **kwargs):
"""Add modes in the listing"""
resp = urlquick.get(URL_ROOT(''))
root = resp.parse("li", attrs={"id": "menu-videos"})
for category in root.iterfind(".//li"):
if 'clips_espagnol' not in category.find('.//a').get('href'):
item = Listitem()
if 'personnages' in category.find('.//a').get('href'):
value_next = 'list_shows'
else:
value_next = 'list_videos_1'
item.label = category.find('.//a').text
category_url = URL_ROOT(category.find('.//a').get('href'))
item.set_callback(eval(value_next),
item_id=item_id,
category_url=category_url,
page=1)
item_post_treatment(item)
yield item
@Route.register
def list_shows(plugin, item_id, category_url, page, **kwargs):
"""Build categories listing"""
resp = urlquick.get(category_url)
root = resp.parse("div", attrs={"class": "personnages"})
for personnage in root.iterfind(".//a"):
item = Listitem()
item.label = personnage.get('title')
item.art['thumb'] = item.art['landscape'] = URL_ROOT(personnage.find('.//img').get('src'))
show_url = URL_ROOT(personnage.get('href'))
item.set_callback(list_videos_2,
item_id=item_id,
category_url=show_url)
item_post_treatment(item)
yield item
@Route.register
def list_videos_1(plugin, item_id, category_url, page, **kwargs):
"""Build videos listing"""
resp = urlquick.get(category_url + '/par_date/%s' % str(page))
at_least_one_item = False
if 'serietele' in category_url or 'zarbi' in category_url:
root = resp.parse("div", attrs={"class": "serieTele"})
for episode in root.iterfind(".//div"):
if episode.find('.//a') is not None and \
episode.find(".//img[@class='thumb']") is not None:
at_least_one_item = True
item = Listitem()
item.label = episode.find(".//span[@class='saison-episode']"
).text.strip() + ' ' + episode.find(
'.//img').get('alt')
video_url = URL_ROOT(episode.find('.//a').get('href'))
item.art['thumb'] = item.art['landscape'] = URL_ROOT(episode.find('.//img').get('src'))
item.set_callback(METHOD_NAME,
item_id=item_id,
video_url=video_url)
item_post_treatment(item,
is_playable=True,
is_downloadable=True)
yield item
else:
root = resp.parse()
for episode in root.iterfind(".//a[@class='lienThumbCollection']"):
at_least_one_item = True
item = Listitem()
item.label = episode.find('.//img').get('alt')
video_url = URL_ROOT(episode.get('href'))
item.art['thumb'] = item.art['landscape'] = URL_ROOT(episode.find('.//img').get('src'))
item.set_callback(METHOD_NAME,
item_id=item_id,
video_url=video_url)
item_post_treatment(item, is_playable=True, is_downloadable=True)
yield item
if at_least_one_item:
# More videos...
yield Listitem.next_page(item_id=item_id,
category_url=category_url,
page=page + 1)
else:
plugin.notify(plugin.localize(30718), '')
yield False
@Route.register
def list_videos_2(plugin, item_id, category_url, **kwargs):
"""Build videos listing"""
resp = urlquick.get(category_url)
root = resp.parse()
for episode in root.iterfind(".//a[@class='lienThumbCollection']"):
item = Listitem()
item.label = episode.find('.//img').get('alt')
video_url = URL_ROOT(episode.get('href'))
item.art['thumb'] = item.art['landscape'] = URL_ROOT(episode.find('.//img').get('src'))
item.set_callback(METHOD_NAME,
item_id=item_id,
video_url=video_url)
item_post_treatment(item, is_playable=True, is_downloadable=True)
yield item
@Resolver.register
def METHOD_NAME(plugin,
item_id,
video_url,
download_mode=False,
**kwargs):
"""Get video URL and start video player"""
video_html = urlquick.get(video_url).text
if re.compile('AtedraVideo.video_id = "(.*?)"').findall(video_html):
video_id = re.compile('AtedraVideo.video_id = "(.*?)"').findall(
video_html)[0]
else:
# TO DO Espagnol Video / Return 404 (TO REMOVE)
return False
return resolver_proxy.get_stream_youtube(plugin, video_id, download_mode) | null |
getpwent r | #
# Module for simulation of utility "getent passwd -s sss" from coreutils
#
# Copyright (c) 2016 Red Hat, Inc.
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
from ctypes import (c_int, c_char_p, c_ulong, POINTER,
Structure, create_string_buffer, get_errno)
from sssd_nss import NssReturnCode, SssdNssError, nss_sss_ctypes_loader
PASSWD_BUFLEN = 1024
class Passwd(Structure):
_fields_ = [("pw_name", c_char_p),
("pw_passwd", c_char_p),
("pw_uid", c_int),
("pw_gid", c_int),
("pw_gecos", c_char_p),
("pw_dir", c_char_p),
("pw_shell", c_char_p)]
def set_user_dict(res, result_p):
if res != NssReturnCode.SUCCESS:
return dict()
user_dict = dict()
user_dict['name'] = result_p[0].pw_name.decode('utf-8')
user_dict['passwd'] = result_p[0].pw_passwd.decode('utf-8')
user_dict['uid'] = result_p[0].pw_uid
user_dict['gid'] = result_p[0].pw_gid
user_dict['gecos'] = result_p[0].pw_gecos.decode('utf-8')
user_dict['dir'] = result_p[0].pw_dir.decode('utf-8')
user_dict['shell'] = result_p[0].pw_shell.decode('utf-8')
return user_dict
def getpwnam_r(name, result_p, buffer_p, buflen):
"""
ctypes wrapper for:
enum nss_status _nss_sss_getpwnam_r(const char *name,
struct passwd *result,
char *buffer,
size_t buflen,
int *errnop)
"""
func = nss_sss_ctypes_loader("_nss_sss_getpwnam_r")
func.restype = c_int
func.argtypes = [c_char_p, POINTER(Passwd),
c_char_p, c_ulong, POINTER(c_int)]
errno = POINTER(c_int)(c_int(0))
name = name.encode('utf-8')
res = func(c_char_p(name), result_p, buffer_p, buflen, errno)
return (int(res), int(errno[0]), result_p)
def getpwuid_r(uid, result_p, buffer_p, buflen):
"""
ctypes wrapper for:
enum nss_status _nss_sss_getpwuid_r(uid_t uid,
struct passwd *result,
char *buffer,
size_t buflen,
int *errnop)
"""
func = nss_sss_ctypes_loader("_nss_sss_getpwuid_r")
func.restype = c_int
func.argtypes = [c_ulong, POINTER(Passwd),
c_char_p, c_ulong, POINTER(c_int)]
errno = POINTER(c_int)(c_int(0))
res = func(uid, result_p, buffer_p, buflen, errno)
return (int(res), int(errno[0]), result_p)
def setpwent():
"""
ctypes wrapper for:
void setpwent(void)
"""
func = nss_sss_ctypes_loader("_nss_sss_setpwent")
func.argtypes = []
res = func()
assert res == NssReturnCode.SUCCESS
errno = get_errno()
if errno != 0:
raise SssdNssError(errno, "setpwent")
def endpwent():
"""
ctypes wrapper for:
void endpwent(void)
"""
func = nss_sss_ctypes_loader("_nss_sss_endpwent")
func.argtypes = []
res = func()
assert res == NssReturnCode.SUCCESS
errno = get_errno()
if errno != 0:
raise SssdNssError(errno, "endpwent")
def METHOD_NAME(result_p, buffer_p, buflen):
"""
ctypes wrapper for:
enum nss_status _nss_sss_getpwent_r(struct passwd *result,
char *buffer, size_t buflen,
int *errnop)
"""
func = nss_sss_ctypes_loader("_nss_sss_getpwent_r")
func.restype = c_int
func.argtypes = [POINTER(Passwd), c_char_p, c_ulong, POINTER(c_int)]
errno = POINTER(c_int)(c_int(0))
res = func(result_p, buffer_p, buflen, errno)
return (int(res), int(errno[0]), result_p)
def getpwent():
result = Passwd()
result_p = POINTER(Passwd)(result)
buff = create_string_buffer(PASSWD_BUFLEN)
res, errno, result_p = METHOD_NAME(result_p, buff, PASSWD_BUFLEN)
if errno != 0:
raise SssdNssError(errno, "getpwent_r")
user_dict = set_user_dict(res, result_p)
return res, user_dict
def call_sssd_getpwnam(name):
"""
A Python wrapper to retrieve a user by name. Returns:
(res, user_dict)
if res is NssReturnCode.SUCCESS, then user_dict contains the keys
corresponding to the C passwd structure fields. Otherwise, the dictionary
is empty and errno indicates the error code
"""
result = Passwd()
result_p = POINTER(Passwd)(result)
buff = create_string_buffer(PASSWD_BUFLEN)
res, errno, result_p = getpwnam_r(name, result_p, buff, PASSWD_BUFLEN)
if errno != 0:
raise SssdNssError(errno, "getpwnam_r")
user_dict = set_user_dict(res, result_p)
return res, user_dict
def call_sssd_getpwuid(uid):
"""
A Python wrapper to retrieve a user by UID. Returns:
(res, user_dict)
if res is NssReturnCode.SUCCESS, then user_dict contains the keys
corresponding to the C passwd structure fields. Otherwise, the dictionary
is empty and errno indicates the error code
"""
result = Passwd()
result_p = POINTER(Passwd)(result)
buff = create_string_buffer(PASSWD_BUFLEN)
res, errno, result_p = getpwuid_r(uid, result_p, buff, PASSWD_BUFLEN)
if errno != 0:
raise SssdNssError(errno, "getpwuid_r")
user_dict = set_user_dict(res, result_p)
return res, user_dict
def call_sssd_enumeration():
"""
enumerate users from sssd module only
"""
setpwent()
user_list = []
res, user = getpwent()
while res == NssReturnCode.SUCCESS:
user_list.append(user)
res, user = getpwent()
endpwent()
return user_list | null |
run patten check | # -*- mode: python -*-
# =============================================================================
# @@-COPYRIGHT-START-@@
#
# Copyright (c) 2023, Qualcomm Innovation Center, Inc. All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# 1. Redistributions of source code must retain the above copyright notice,
# this list of conditions and the following disclaimer.
#
# 2. Redistributions in binary form must reproduce the above copyright notice,
# this list of conditions and the following disclaimer in the documentation
# and/or other materials provided with the distribution.
#
# 3. Neither the name of the copyright holder nor the names of its contributors
# may be used to endorse or promote products derived from this software
# without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
# POSSIBILITY OF SUCH DAMAGE.
#
# SPDX-License-Identifier: BSD-3-Clause
#
# @@-COPYRIGHT-END-@@
# =============================================================================
""" Utility for checking model architechture prior to using AIMET feature. """
from typing import Callable, Union, Tuple
import torch
from aimet_common.utils import AimetLogger
from aimet_torch.meta.connectedgraph import ConnectedGraph
from aimet_torch.arch_checker.arch_checker_rules import TorchActivations
from aimet_torch.arch_checker.arch_checker_utils import (ArchCheckerReport,
OpStructure,
check_type_deco,
get_node_check_dict,
get_pattern_check_list,
check_module)
logger = AimetLogger.get_area_logger(AimetLogger.LogAreas.Utils)
class ArchChecker:
"""
ArchChecker object to check pytorch model architechture and suggest better architechture prior
to training.
"""
_node_check_dict = get_node_check_dict()
_pattern_checks = get_pattern_check_list()
_arch_checker_report = ArchCheckerReport()
@staticmethod
def add_node_check(check_target_type: torch.nn.Module, arch_check: Callable):
"""
Add extra checks for node checks in architecture checker.
:param check_target_type: layer type to be checked.
:param arch_check: node checker function.
"""
# All TorchActivations are included to one type.
if isinstance(check_target_type, TorchActivations):
# Init TorchActivations check if not exist.
if TorchActivations not in ArchChecker._node_check_dict:
ArchChecker._node_check_dict[TorchActivations] = []
if check_target_type is TorchActivations:
ArchChecker._node_check_dict[TorchActivations].append(arch_check)
else:
# Add check_type_deco wrapper if check_target_type is a subclass of TorchActivations.
ArchChecker._node_check_dict[TorchActivations].append(check_type_deco(check_target_type)(arch_check))
else:
if check_target_type in ArchChecker._node_check_dict:
ArchChecker._node_check_dict[check_target_type].append(arch_check)
else:
ArchChecker._node_check_dict[check_target_type] = [arch_check]
@staticmethod
def add_pattern_check(arch_check: Callable):
"""
Add extra checks for pattern checks in architecture checker.
:param arch_check: pattern checker function.
:param check_kind: Check kind. Different check kind requires different arguments.
"""
ArchChecker._pattern_checks.append(arch_check)
@staticmethod
def check_model_arch(model: torch.nn.Module, dummy_input: Union[torch.Tensor, Tuple], result_dir: str = None)-> ArchCheckerReport:
"""
Check each node in the model using checks in _node_check_dict. Record only the nodes and
failed tests.
:param model: Torch model to be checked.
:param dummy_input: A dummy input to the model. Can be a Tensor or a Tuple of Tensors
:return arch_checker_report: {op.dotted_name_op: NodeErrorReportObject }
"""
def run_node_checks():
"""
Walk through connected_graph and applies node checks on each node.
"""
for op in connected_graph.ordered_ops:
module = op.get_module()
if module and isinstance(module, tuple(ArchChecker._node_check_dict.keys())):
if isinstance(module, TorchActivations):
checks = ArchChecker._node_check_dict[TorchActivations]
else:
checks = ArchChecker._node_check_dict[type(module)]
failed_checks_set = check_module(module, checks)
if failed_checks_set:
ArchChecker._arch_checker_report.update_raw_report(op, failed_checks_set)
logger.info("Graph/Node: %s: %s fails check: %s", op.dotted_name, module,
failed_checks_set)
def METHOD_NAME():
"""
Applies pattern checks on connected graph.
"""
for _check in ArchChecker._pattern_checks:
failed_check_ops = _check(connected_graph)
if failed_check_ops:
# Pattern check that marks structure returns List[List[Op]]
# Transform List[List[Op]] to List[OpStructure]
if isinstance(failed_check_ops[0], list):
failed_check_ops = [OpStructure(_op_tuple) for _op_tuple in failed_check_ops]
ArchChecker._arch_checker_report.update_raw_report(failed_check_ops, _check.__name__)
for op in failed_check_ops:
logger.info("Graph/Node: %s: %s fails check: %s", op.dotted_name, op.get_module(), {_check.__name__})
connected_graph = ConnectedGraph(model, dummy_input)
# Run all node checkes
logger.info("Running node checkes.")
run_node_checks()
# Run all pattern checkes
logger.info("Running pattern checkes.")
METHOD_NAME()
if result_dir is not None:
ArchChecker.set_export_dir(result_dir)
ArchChecker._arch_checker_report.export_to_html()
@staticmethod
def set_export_dir(dir_path: str):
""" Set export dir. """
ArchChecker._arch_checker_report.result_dir = dir_path | null |
dense strategy rocm | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
"""Definition of ROCm operator strategy."""
# pylint: disable=invalid-name,unused-argument,unused-wildcard-import,wildcard-import
from tvm import topi
from tvm.te import SpecializedCondition
from tvm.contrib.thrust import can_use_rocthrust
from tvm.contrib import miopen
from .generic import *
from .. import op as _op
from .cuda import batch_matmul_strategy_cuda, conv2d_strategy_cuda, dense_strategy_cuda
@conv2d_strategy.register("rocm")
def conv2d_strategy_rocm(attrs, inputs, out_type, target):
"""conv2d rocm strategy"""
groups = attrs.groups
layout = attrs.data_layout
padding = attrs.get_int_tuple("padding")
strategy = conv2d_strategy_cuda(attrs, inputs, out_type, target)
# add miopen implementation
if (
"miopen" in target.libs
and groups == 1
and layout == "NCHW"
and padding[0] == padding[2]
and padding[1] == padding[3]
):
strategy.add_implementation(
wrap_compute_conv2d(topi.rocm.conv2d_nchw_miopen, need_data_layout=True),
wrap_topi_schedule(topi.rocm.schedule_conv2d_nchw_miopen),
name="conv2d_nchw_miopen.rocm",
plevel=50,
)
return strategy
@dense_strategy.register("rocm")
def METHOD_NAME(attrs, inputs, out_type, target):
"""Dense strategy for ROCM"""
assert len(inputs[0].shape) == 2 and len(inputs[1].shape) == 2, "Only support 2-dim dense"
strategy = dense_strategy_cuda(attrs, inputs, out_type, target)
if target.kind.name == "rocm" and "rocblas" in target.libs:
assert out_type.dtype == inputs[0].dtype, "Mixed precision not supported."
strategy.add_implementation(
wrap_compute_dense(topi.rocm.dense_rocblas),
wrap_topi_schedule(topi.rocm.schedule_dense_rocblas),
name="dense_rocblas.rocm",
plevel=15,
)
return strategy
@batch_matmul_strategy.register("rocm")
def batch_matmul_strategy_rocm(attrs, inputs, out_type, target):
"""Batch matmul strategy for ROCM"""
strategy = batch_matmul_strategy_cuda(attrs, inputs, out_type, target)
if target.kind.name == "rocm" and "rocblas" in target.libs:
assert out_type.dtype == inputs[0].dtype, "Mixed precision not supported."
strategy.add_implementation(
wrap_compute_batch_matmul(topi.rocm.batch_matmul_rocblas),
wrap_topi_schedule(topi.rocm.schedule_batch_matmul_rocblas),
name="batch_matmul_rocblas.rocm",
plevel=12,
)
return strategy
@argsort_strategy.register(["rocm"])
def argsort_strategy_cuda(attrs, inputs, out_type, target):
"""argsort rocm strategy"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_argsort(topi.cuda.argsort),
wrap_topi_schedule(topi.cuda.schedule_argsort),
name="argsort.rocm",
)
if can_use_rocthrust(target, "tvm.contrib.thrust.sort"):
strategy.add_implementation(
wrap_compute_argsort(topi.cuda.argsort_thrust),
wrap_topi_schedule(topi.cuda.schedule_argsort),
name="argsort_thrust.rocm",
plevel=15,
)
return strategy
@scatter_elements_strategy.register(["rocm"])
def scatter_elements_cuda(attrs, inputs, out_type, target):
"""scatter rocm strategy"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_scatter_elements(topi.cuda.scatter_elements),
wrap_topi_schedule(topi.cuda.schedule_extern),
name="scatter_elements.rocm",
plevel=10,
)
rank = len(inputs[0].shape)
with SpecializedCondition(rank == 1 and attrs.reduction == "update"):
if can_use_rocthrust(target, "tvm.contrib.thrust.stable_sort_by_key"):
strategy.add_implementation(
wrap_compute_scatter_elements(topi.cuda.scatter_via_sort),
wrap_topi_schedule(topi.cuda.schedule_scatter_via_sort),
name="scatter_via_sort.rocm",
plevel=9, # use the sequential version by default
)
return strategy
@sort_strategy.register(["rocm"])
def sort_strategy_cuda(attrs, inputs, out_type, target):
"""sort rocm strategy"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_sort(topi.cuda.sort),
wrap_topi_schedule(topi.cuda.schedule_sort),
name="sort.rocm",
)
if can_use_rocthrust(target, "tvm.contrib.thrust.sort"):
strategy.add_implementation(
wrap_compute_sort(topi.cuda.sort_thrust),
wrap_topi_schedule(topi.cuda.schedule_sort),
name="sort_thrust.cuda",
plevel=15,
)
return strategy
@topk_strategy.register(["rocm"])
def topk_strategy_cuda(attrs, inputs, out_type, target):
"""topk rocm strategy"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_topk(topi.cuda.topk),
wrap_topi_schedule(topi.cuda.schedule_topk),
name="topk.rocm",
)
if can_use_rocthrust(target, "tvm.contrib.thrust.sort"):
strategy.add_implementation(
wrap_compute_topk(topi.cuda.topk_thrust),
wrap_topi_schedule(topi.cuda.schedule_topk),
name="topk_thrust.rocm",
plevel=15,
)
return strategy
@softmax_strategy.register(["rocm"])
def softmax_strategy_rocm(attrs, inputs, out_type, target):
"""rocm strategy for softmax"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_softmax(topi.nn.softmax),
wrap_topi_schedule(topi.cuda.schedule_softmax),
name="softmax.rocm",
)
if "miopen" in target.libs:
strategy.add_implementation(
wrap_compute_softmax(miopen.softmax),
wrap_topi_schedule(topi.generic.schedule_extern),
name="softmax.miopen",
plevel=15,
)
return strategy
@log_softmax_strategy.register(["rocm"])
def log_softmax_strategy_rocm(attrs, inputs, out_type, target):
"""rocm strategy for log softmax"""
strategy = _op.OpStrategy()
strategy.add_implementation(
wrap_compute_softmax(topi.nn.log_softmax),
wrap_topi_schedule(topi.cuda.schedule_softmax),
name="log_softmax.rocm",
)
if "miopen" in target.libs:
strategy.add_implementation(
wrap_compute_softmax(miopen.log_softmax),
wrap_topi_schedule(topi.generic.schedule_extern),
name="log_softmax.miopen",
plevel=15,
)
return strategy | null |
test calibration repr | # Copyright 2019 The Cirq Developers
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import datetime
import pytest
import matplotlib as mpl
import matplotlib.pyplot as plt
from google.protobuf.text_format import Merge
import cirq
import cirq_google as cg
from cirq_google.api import v2
_CALIBRATION_DATA = Merge(
"""
timestamp_ms: 1562544000021,
metrics: [{
name: 'two_qubit_xeb',
targets: ['0_0', '0_1'],
values: [{
double_val: .9999
}]
}, {
name: 'two_qubit_xeb',
targets: ['0_0', '1_0'],
values: [{
double_val: .9998
}]
}, {
name: 't1',
targets: ['0_0'],
values: [{
double_val: 321
}]
}, {
name: 't1',
targets: ['0_1'],
values: [{
double_val: 911
}]
}, {
name: 't1',
targets: ['1_0'],
values: [{
double_val: 505
}]
}, {
name: 'globalMetric',
values: [{
int32_val: 12300
}]
}]
""",
v2.metrics_pb2.MetricsSnapshot(),
)
def test_calibration_metrics_dictionary():
calibration = cg.Calibration(_CALIBRATION_DATA)
t1s = calibration['t1']
assert t1s == {
(cirq.GridQubit(0, 0),): [321],
(cirq.GridQubit(0, 1),): [911],
(cirq.GridQubit(1, 0),): [505],
}
assert len(calibration) == 3
assert 't1' in calibration
assert 't2' not in calibration
for qubits, values in t1s.items():
assert len(qubits) == 1
assert len(values) == 1
with pytest.raises(TypeError, match="was 1"):
_ = calibration[1]
with pytest.raises(KeyError, match='not-it'):
_ = calibration['not-it']
def test_calibration_str():
calibration = cg.Calibration(_CALIBRATION_DATA)
assert str(calibration) == "Calibration(keys=['globalMetric', 't1', 'two_qubit_xeb'])"
def METHOD_NAME():
calibration = cg.Calibration(_CALIBRATION_DATA)
cirq.testing.assert_equivalent_repr(calibration, setup_code="import cirq\nimport cirq_google")
def test_calibration_timestamp_str():
calibration = cg.Calibration(_CALIBRATION_DATA)
assert calibration.timestamp_str(tz=datetime.timezone.utc) == '2019-07-08 00:00:00.021021+00:00'
assert (
calibration.timestamp_str(tz=datetime.timezone(datetime.timedelta(hours=1)))
== '2019-07-08 01:00:00.021021+01:00'
)
def test_to_proto():
calibration = cg.Calibration(_CALIBRATION_DATA)
assert calibration == cg.Calibration(calibration.to_proto())
invalid_value = cg.Calibration(metrics={'metric': {(cirq.GridQubit(1, 1),): [1.1, {}]}})
with pytest.raises(ValueError, match='Unsupported metric value'):
invalid_value.to_proto()
def test_value_to_float():
assert cg.Calibration.value_to_float([1.1]) == 1.1
assert cg.Calibration.value_to_float([0.7, 0.5]) == 0.7
assert cg.Calibration.value_to_float([7]) == 7
with pytest.raises(ValueError, match='was empty'):
cg.Calibration.value_to_float([])
with pytest.raises(ValueError, match='could not convert string to float'):
cg.Calibration.value_to_float(['went for a walk'])
def test_calibrations_with_string_key():
calibration = cg.Calibration(metrics={'metric1': {('alpha',): [0.1]}})
expected_proto = Merge(
"""
metrics: [{
name: 'metric1'
targets: ['alpha']
values: [{double_val: 0.1}]
}]
""",
v2.metrics_pb2.MetricsSnapshot(),
)
assert expected_proto == calibration.to_proto()
assert calibration == cg.Calibration(expected_proto)
assert calibration == cg.Calibration(calibration.to_proto())
def test_key_to_qubits():
qubits = tuple([cirq.GridQubit(1, 1), cirq.GridQubit(1, 2)])
assert cg.Calibration.key_to_qubit(qubits) == cirq.GridQubit(1, 1)
assert cg.Calibration.key_to_qubits(qubits) == (cirq.GridQubit(1, 1), cirq.GridQubit(1, 2))
with pytest.raises(ValueError, match='was not a tuple of qubits'):
cg.Calibration.key_to_qubit('alpha')
with pytest.raises(ValueError, match='was not a tuple of grid qubits'):
cg.Calibration.key_to_qubits('alpha')
def test_calibration_heatmap():
calibration = cg.Calibration(_CALIBRATION_DATA)
heatmap = calibration.heatmap('t1')
figure = mpl.figure.Figure()
axes = figure.add_subplot(111)
heatmap.plot(axes)
assert axes.get_title() == 'T1'
heatmap = calibration.heatmap('two_qubit_xeb')
figure = mpl.figure.Figure()
axes = figure.add_subplot(999)
heatmap.plot(axes)
assert axes.get_title() == 'Two Qubit Xeb'
with pytest.raises(ValueError, match="one or two qubits.*multi_qubit"):
multi_qubit_data = Merge(
"""metrics: [{
name: 'multi_qubit',
targets: ['0_0', '0_1', '1_0'],
values: [{double_val: 0.999}]}]""",
v2.metrics_pb2.MetricsSnapshot(),
)
cg.Calibration(multi_qubit_data).heatmap('multi_qubit')
with pytest.raises(ValueError, match="single metric values.*multi_value"):
multi_qubit_data = Merge(
"""metrics: [{
name: 'multi_value',
targets: ['0_0'],
values: [{double_val: 0.999}, {double_val: 0.001}]}]""",
v2.metrics_pb2.MetricsSnapshot(),
)
cg.Calibration(multi_qubit_data).heatmap('multi_value')
@pytest.mark.usefixtures('closefigures')
def test_calibration_plot_histograms():
calibration = cg.Calibration(_CALIBRATION_DATA)
_, ax = plt.subplots(1, 1)
calibration.plot_histograms(['t1', 'two_qubit_xeb'], ax, labels=['T1', 'XEB'])
assert len(ax.get_lines()) == 4
with pytest.raises(ValueError, match="single metric values.*multi_value"):
multi_qubit_data = Merge(
"""metrics: [{
name: 'multi_value',
targets: ['0_0'],
values: [{double_val: 0.999}, {double_val: 0.001}]}]""",
v2.metrics_pb2.MetricsSnapshot(),
)
cg.Calibration(multi_qubit_data).plot_histograms('multi_value')
@pytest.mark.usefixtures('closefigures')
def test_calibration_plot():
calibration = cg.Calibration(_CALIBRATION_DATA)
_, axs = calibration.plot('two_qubit_xeb')
assert axs[0].get_title() == 'Two Qubit Xeb'
assert len(axs[1].get_lines()) == 2 | null |
test pluto verify underflow | import pytest
hardware = ["pluto", "pluto_rev_c"]
classname = "adi.Pluto"
#########################################
@pytest.mark.iio_hardware(hardware)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize(
"attr, start, stop, step, tol, repeats",
[
("tx_hardwaregain_chan0", -89.75, 0.0, 0.25, 0, 100),
("rx_lo", 325000000, 3800000000, 1, 8, 100),
("tx_lo", 325000000, 3800000000, 1, 8, 100),
("sample_rate", 2084000, 61440000, 1, 4, 100),
("loopback", 0, 0, 1, 0, 0),
("loopback", 1, 1, 1, 0, 0),
("loopback", 2, 2, 1, 0, 0),
],
)
def test_pluto_attr(
test_attribute_single_value,
iio_uri,
classname,
attr,
start,
stop,
step,
tol,
repeats,
):
test_attribute_single_value(
iio_uri, classname, attr, start, stop, step, tol, repeats
)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
def test_pluto_rx_data(test_dma_rx, iio_uri, classname, channel):
test_dma_rx(iio_uri, classname, channel)
#########################################
@pytest.mark.iio_hardware(hardware)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
def test_pluto_tx_data(test_dma_tx, iio_uri, classname, channel):
test_dma_tx(iio_uri, classname, channel)
#########################################
@pytest.mark.iio_hardware(hardware)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize(
"param_set",
[
dict(
tx_lo=1000000000,
rx_lo=1000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
)
],
)
def test_pluto_cyclic_buffers(
test_cyclic_buffer, iio_uri, classname, channel, param_set
):
test_cyclic_buffer(iio_uri, classname, channel, param_set)
#########################################
@pytest.mark.iio_hardware(hardware)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize(
"param_set",
[
dict(
tx_lo=1000000000,
rx_lo=1000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
)
],
)
def test_pluto_cyclic_buffers_exception(
test_cyclic_buffer_exception, iio_uri, classname, channel, param_set
):
test_cyclic_buffer_exception(iio_uri, classname, channel, param_set)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
def test_pluto_loopback(test_dma_loopback, iio_uri, classname, channel):
test_dma_loopback(iio_uri, classname, channel)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize(
"param_set",
[
dict(
tx_lo=1000000000,
rx_lo=1000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
)
],
)
@pytest.mark.parametrize("sfdr_min", [40])
def test_pluto_sfdr(test_sfdr, iio_uri, classname, channel, param_set, sfdr_min):
test_sfdr(iio_uri, classname, channel, param_set, sfdr_min)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize("frequency, scale", [(1000000, 1)])
@pytest.mark.parametrize(
"param_set",
[
dict(
tx_lo=1000000000,
rx_lo=1000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-30,
sample_rate=4000000,
)
],
)
@pytest.mark.parametrize("peak_min", [-40])
def test_pluto_dds_loopback(
test_dds_loopback,
iio_uri,
classname,
param_set,
channel,
frequency,
scale,
peak_min,
):
test_dds_loopback(
iio_uri, classname, param_set, channel, frequency, scale, peak_min
)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize(
"param_set",
[
dict(
tx_lo=1000000000,
rx_lo=1000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
),
dict(
tx_lo=2000000000,
rx_lo=2000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
),
dict(
tx_lo=3000000000,
rx_lo=3000000000,
gain_control_mode_chan0="slow_attack",
tx_hardwaregain_chan0=-20,
sample_rate=4000000,
),
],
)
def test_pluto_iq_loopback(test_iq_loopback, iio_uri, classname, channel, param_set):
test_iq_loopback(iio_uri, classname, channel, param_set)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
def test_pluto_loopback_zeros(test_dma_dac_zeros, iio_uri, classname, channel):
test_dma_dac_zeros(iio_uri, classname, channel)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize("buffer_size", [2 ** 20])
@pytest.mark.parametrize("sample_rate", [600e3])
def test_pluto_verify_overflow(
test_verify_overflow, iio_uri, classname, channel, buffer_size, sample_rate
):
test_verify_overflow(iio_uri, classname, channel, buffer_size, sample_rate)
#########################################
@pytest.mark.iio_hardware(hardware, True)
@pytest.mark.parametrize("classname", [(classname)])
@pytest.mark.parametrize("channel", [0])
@pytest.mark.parametrize("buffer_size", [2 ** 20])
@pytest.mark.parametrize("sample_rate", [600e3])
def METHOD_NAME(
test_verify_underflow, iio_uri, classname, channel, buffer_size, sample_rate
):
test_verify_underflow(iio_uri, classname, channel, buffer_size, sample_rate) | null |
get num psus | #
# psuutil.py
# Platform-specific PSU status interface for SONiC
#
import logging
import os.path
try:
from sonic_psu.psu_base import PsuBase
except ImportError as e:
raise ImportError(str(e) + "- required module not found")
class PsuUtil(PsuBase):
"""Platform-specific PSUutil class"""
HWMON_PATH = '/sys/class/hwmon/hwmon1/'
PSU1_PREFIX = 'power43_'
PSU2_PREFIX = 'power54_'
MAX_PSUS = 2
def __init__(self):
PsuBase.__init__(self)
# Get sysfs attribute
def get_attr_value(self, attr_path):
retval = 'ERR'
if (not os.path.isfile(attr_path)):
return retval
try:
with open(attr_path, 'r') as fd:
retval = fd.read()
except Exception:
logging.error("Unable to open ", attr_path, " file !")
retval = retval.rstrip('\r\n')
return retval
def get_attr_filename(self, index, attr):
if (index == 1):
attr_file = self.PSU1_PREFIX + attr
elif (index == 2):
attr_file = self.PSU2_PREFIX + attr
else:
logging.error("Invalid PSU number:", index)
return ''
return attr_file
def METHOD_NAME(self):
"""
Retrieves the number of PSUs available on the device
:return: An integer, the number of PSUs available on the device
"""
return self.MAX_PSUS
def get_psu_status(self, index):
"""
Retrieves the oprational status of power supply unit (PSU) defined
by index <index>
:param index: An integer, index of the PSU of which to query status
:return: Boolean, True if PSU is operating properly, False if PSU is\
faulty
"""
status = False
attr_filename = self.get_attr_filename(index, 'input')
if attr_filename == '':
return status
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
attr_value = float(attr_value)
# Check PSU status
if (attr_value != 0.0):
status = True
return status
def get_psu_presence(self, index):
"""
Retrieves the presence status of power supply unit (PSU) defined
by index <index>
:param index: An integer, index of the PSU of which to query status
:return: Boolean, True if PSU is plugged, False if not
"""
status = False
attr_filename = self.get_attr_filename(index, 'present')
if attr_filename == '':
return status
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
attr_value = int(attr_value, 16)
# Check PSU status
if (attr_value == 1):
status = True
return status
def get_powergood_status(self, index):
status = False
attr_filename = self.get_attr_filename(index, 'input')
if attr_filename == '':
return status
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
attr_value = float(attr_value)
# Check PSU status
if (attr_value != 0.0):
status = True
return status
def get_model(self, index):
attr_filename = self.get_attr_filename(index, 'model')
if attr_filename == '':
return None
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
return attr_value.rstrip()
def get_mfr_id(self, index):
attr_filename = self.get_attr_filename(index, 'mfrid')
if attr_filename == '':
return None
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
return attr_value.rstrip()
def get_serial(self, index):
attr_filename = self.get_attr_filename(index, 'sn')
if attr_filename == '':
return None
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
return attr_value.rstrip()
def get_direction(self, index):
if (index == 1):
direction_file = 'fan41_direction'
elif (index == 2):
direction_file = 'fan52_direction'
else:
logging.error("Invalid PSU number:", index)
return None
direction = self.get_attr_value(self.HWMON_PATH + direction_file)
direction = direction.rstrip()
"""
1: FB 2: BF
Since the fan is at rear of the switch, FB means Exhaust; BF means Intake
"""
if direction == '2':
return "INTAKE"
else:
return "EXHAUST"
def get_output_voltage(self, index):
if (index == 1):
attr_file = 'in48_input'
elif (index == 2):
attr_file = 'in59_input'
else:
logging.error("Invalid PSU number:", index)
return 0.0
voltage = self.get_attr_value(self.HWMON_PATH + attr_file)
voltage = voltage.rstrip()
if (voltage != 'ERR'):
voltage, dummy = voltage.split('.', 1)
else:
return 0.0
return float(voltage)/1000
def get_output_current(self, index):
if (index == 1):
attr_file = 'curr40_input'
elif (index == 2):
attr_file = 'curr51_input'
else:
logging.error("Invalid PSU number:", index)
return 0.0
current = self.get_attr_value(self.HWMON_PATH + attr_file)
current = current.rstrip()
if (current != 'ERR'):
current, dummy = current.split('.',1)
else:
return 0.0
return float(current)/1000
def get_output_power(self, index):
attr_filename = self.get_attr_filename(index, 'input')
if attr_filename == '':
return 0.0
attr_path = self.HWMON_PATH + attr_filename
attr_value = self.get_attr_value(attr_path)
if (attr_value != 'ERR'):
attr_value = float(attr_value)
else:
return 0.0
return float(attr_value/1000)
def get_fan_rpm(self, index, fan_idx):
if (index == 1):
rpm_file = 'fan41_input'
elif (index == 2):
rpm_file = 'fan52_input'
else:
logging.error("Invalid PSU number:", index)
return 0
rpm = self.get_attr_value(self.HWMON_PATH + rpm_file)
rpm = rpm.rstrip()
if (rpm != 'ERR'):
rpm = float(rpm)
else:
return 0
return int(rpm) | null |
update compute cost | import copy
from typing import List
from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, ShardingStrategy, TrainCycleItem
from .strategy_generator import FollowingStrategyGenerator
__all__ = ['UnaryElementwiseGenerator']
class UnaryElementwiseGenerator(FollowingStrategyGenerator):
"""
UnaryElementwiseGenerator which deals with the sharding strategies of UnaryElementwiseOp.
"""
def validate(self) -> bool:
return super().validate()
def METHOD_NAME(self, strategy: ShardingStrategy):
compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20)
strategy.compute_cost = compute_cost
def update_memory_cost(self, strategy: ShardingStrategy):
'''
Compute the memory cost per device with this specific strategy.
'''
forward_size_mapping = {
'input': self._compute_size_in_bytes(strategy, "input"),
'output': self._compute_size_in_bytes(strategy, "output")
}
backward_size_mapping = copy.deepcopy(forward_size_mapping)
backward_size_mapping.pop("output")
# compute fwd cost incurred
# fwd_cost = input + output
fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)])
fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)])
fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost)
# compute bwd cost incurred
# bwd_cost = input_grad
bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)])
bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)])
bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost)
# compute total cost
total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost,
parameter=fwd_parameter_cost + bwd_parameter_cost)
memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost)
strategy.memory_cost = memory_cost
def collate_strategies(self) -> List[ShardingStrategy]:
strategy_list = []
# For element-wise function, we keep the sharding spec of output node same as
# the input. Therefore, the different strategies of input node with same
# output sharding spec will generate same strategy for element-wise function.
for index, strategy in enumerate(self.predecessor_node.strategies_vector):
dim_partition_dict_mapping = {}
communication_action_mapping = {}
input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]]
dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict
dim_partition_dict_for_output = copy.deepcopy(dim_partition_dict_for_input)
dim_partition_dict_mapping = {
"input": dim_partition_dict_for_input,
"output": dim_partition_dict_for_output,
}
sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping)
# add index into name to pass the duplicated check
# we keep same strategies with different name for node merging, and it will not increase the searching space,
# because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node.
name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}'
strategy = self.get_sharding_strategy(name=name,
sharding_spec_mapping=sharding_spec_mapping,
communication_action_mapping=communication_action_mapping)
strategy_list.append(strategy)
return strategy_list | null |
send error | import os
import signal
import subprocess
import sys
from contextlib import contextmanager
from io import TextIOWrapper
from subprocess import Popen
from time import sleep
from typing import Any, Iterator, NamedTuple, Optional, Sequence, Tuple
import dagster._check as check
from dagster._core.errors import DagsterError
from dagster._serdes.serdes import (
deserialize_value,
serialize_value,
whitelist_for_serdes,
)
from dagster._utils.error import (
ExceptionInfo,
SerializableErrorInfo,
serializable_error_info_from_exc_info,
)
def write_unary_input(input_file: str, obj: NamedTuple) -> None:
check.str_param(input_file, "input_file")
check.not_none_param(obj, "obj")
with open(os.path.abspath(input_file), "w", encoding="utf8") as fp:
fp.write(serialize_value(obj))
def read_unary_input(input_file: str) -> Tuple[object, ...]:
check.str_param(input_file, "input_file")
with open(os.path.abspath(input_file), "r", encoding="utf8") as fp:
return deserialize_value(fp.read(), NamedTuple)
def ipc_write_unary_response(output_file: str, obj: NamedTuple) -> None:
check.not_none_param(obj, "obj")
with ipc_write_stream(output_file) as stream:
stream.send(obj)
def read_unary_response(
output_file: str, timeout: int = 30, ipc_process: "Optional[Popen[bytes]]" = None
) -> Optional[NamedTuple]:
messages = list(ipc_read_event_stream(output_file, timeout=timeout, ipc_process=ipc_process))
check.invariant(len(messages) == 1)
return messages[0]
@whitelist_for_serdes
class IPCStartMessage(NamedTuple("_IPCStartMessage", [])):
def __new__(cls):
return super(IPCStartMessage, cls).__new__(cls)
@whitelist_for_serdes
class IPCErrorMessage(
NamedTuple(
"_IPCErrorMessage",
[("serializable_error_info", SerializableErrorInfo), ("message", Optional[str])],
)
):
"""This represents a user error encountered during the IPC call. This indicates a business
logic error, rather than a protocol. Consider this a "task failed successfully"
use case.
"""
def __new__(cls, serializable_error_info: SerializableErrorInfo, message: Optional[str]):
return super(IPCErrorMessage, cls).__new__(
cls,
serializable_error_info=check.inst_param(
serializable_error_info, "serializable_error_info", SerializableErrorInfo
),
message=check.opt_str_param(message, "message"),
)
@whitelist_for_serdes
class IPCEndMessage(NamedTuple("_IPCEndMessage", [])):
def __new__(cls):
return super(IPCEndMessage, cls).__new__(cls)
class DagsterIPCProtocolError(DagsterError):
"""This indicates that something went wrong with the protocol. E.g. the
process being called did not emit an IPCStartMessage first.
"""
def __init__(self, message: str):
self.message = message
super(DagsterIPCProtocolError, self).__init__(message)
class FileBasedWriteStream:
def __init__(self, file_path: str):
check.str_param("file_path", file_path)
self._file_path = file_path
def send(self, dagster_named_tuple: NamedTuple) -> None:
_send(self._file_path, dagster_named_tuple)
def send_error(self, exc_info: ExceptionInfo, message: Optional[str] = None) -> None:
METHOD_NAME(self._file_path, exc_info, message=message)
def _send(file_path: str, obj: NamedTuple) -> None:
with open(os.path.abspath(file_path), "a+", encoding="utf8") as fp:
fp.write(serialize_value(obj) + "\n")
def METHOD_NAME(file_path: str, exc_info: ExceptionInfo, message: Optional[str]) -> None:
return _send(
file_path,
IPCErrorMessage(
serializable_error_info=serializable_error_info_from_exc_info(exc_info), message=message
),
)
@contextmanager
def ipc_write_stream(file_path: str) -> Iterator[FileBasedWriteStream]:
check.str_param("file_path", file_path)
_send(file_path, IPCStartMessage())
try:
yield FileBasedWriteStream(file_path)
except Exception:
METHOD_NAME(file_path, sys.exc_info(), message=None)
finally:
_send(file_path, IPCEndMessage())
def _process_line(file_pointer: TextIOWrapper, sleep_interval: float = 0.1) -> Optional[NamedTuple]:
while True:
line = file_pointer.readline()
if line:
return deserialize_value(line.rstrip(), NamedTuple)
sleep(sleep_interval)
def _poll_process(ipc_process: "Optional[Popen[bytes]]") -> None:
if not ipc_process:
return
if ipc_process.poll() is not None:
raise DagsterIPCProtocolError(
"Process exited with return code {return_code} while waiting for events".format(
return_code=ipc_process.returncode
)
)
def ipc_read_event_stream(
file_path: str, timeout: int = 30, ipc_process: "Optional[Popen[bytes]]" = None
) -> Iterator[Optional[NamedTuple]]:
# Wait for file to be ready
sleep_interval = 0.1
elapsed_time = 0
while elapsed_time < timeout and not os.path.exists(file_path):
_poll_process(ipc_process)
elapsed_time += sleep_interval
sleep(sleep_interval)
if not os.path.exists(file_path):
raise DagsterIPCProtocolError(
"Timeout: read stream has not received any data in {timeout} seconds".format(
timeout=timeout
)
)
with open(os.path.abspath(file_path), "r", encoding="utf8") as file_pointer:
message = _process_line(file_pointer)
while elapsed_time < timeout and message is None:
_poll_process(ipc_process)
elapsed_time += sleep_interval
sleep(sleep_interval)
message = _process_line(file_pointer)
# Process start message
if not isinstance(message, IPCStartMessage):
raise DagsterIPCProtocolError(
"Attempted to read stream at file {file_path}, but first message was not an "
"IPCStartMessage".format(file_path=file_path)
)
message = _process_line(file_pointer)
while not isinstance(message, IPCEndMessage):
if message is None:
_poll_process(ipc_process)
yield message
message = _process_line(file_pointer)
# Windows subprocess termination utilities
# https://stefan.sofa-rockers.org/2013/08/15/handling-sub-process-hierarchies-python-linux-os-x/
def open_ipc_subprocess(parts: Sequence[str], **kwargs: Any) -> "Popen[bytes]":
"""Sets the correct flags to support graceful termination."""
check.list_param(parts, "parts", str)
creationflags = 0
if sys.platform == "win32":
creationflags = subprocess.CREATE_NEW_PROCESS_GROUP
return subprocess.Popen( # type: ignore # (unclear whether this is actually guaranteed to return Popen[bytes])
parts,
creationflags=creationflags,
**kwargs,
)
def interrupt_ipc_subprocess(proc: "Popen[bytes]") -> None:
"""Send CTRL_BREAK on Windows, SIGINT on other platforms."""
if sys.platform == "win32":
proc.send_signal(signal.CTRL_BREAK_EVENT)
else:
proc.send_signal(signal.SIGINT)
def interrupt_ipc_subprocess_pid(pid: int) -> None:
"""Send CTRL_BREAK on Windows, SIGINT on other platforms."""
check.int_param(pid, "pid")
if sys.platform == "win32":
os.kill(pid, signal.CTRL_BREAK_EVENT)
else:
os.kill(pid, signal.SIGINT) | null |
arn | # coding=utf-8
# *** WARNING: this file was generated by the Pulumi Terraform Bridge (tfgen) Tool. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import copy
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from .. import _utilities
__all__ = [
'GetParameterResult',
'AwaitableGetParameterResult',
'get_parameter',
'get_parameter_output',
]
@pulumi.output_type
class GetParameterResult:
"""
A collection of values returned by getParameter.
"""
def __init__(__self__, METHOD_NAME=None, id=None, insecure_value=None, name=None, type=None, value=None, version=None, with_decryption=None):
if METHOD_NAME and not isinstance(METHOD_NAME, str):
raise TypeError("Expected argument 'arn' to be a str")
pulumi.set(__self__, "arn", METHOD_NAME)
if id and not isinstance(id, str):
raise TypeError("Expected argument 'id' to be a str")
pulumi.set(__self__, "id", id)
if insecure_value and not isinstance(insecure_value, str):
raise TypeError("Expected argument 'insecure_value' to be a str")
pulumi.set(__self__, "insecure_value", insecure_value)
if name and not isinstance(name, str):
raise TypeError("Expected argument 'name' to be a str")
pulumi.set(__self__, "name", name)
if type and not isinstance(type, str):
raise TypeError("Expected argument 'type' to be a str")
pulumi.set(__self__, "type", type)
if value and not isinstance(value, str):
raise TypeError("Expected argument 'value' to be a str")
pulumi.set(__self__, "value", value)
if version and not isinstance(version, int):
raise TypeError("Expected argument 'version' to be a int")
pulumi.set(__self__, "version", version)
if with_decryption and not isinstance(with_decryption, bool):
raise TypeError("Expected argument 'with_decryption' to be a bool")
pulumi.set(__self__, "with_decryption", with_decryption)
@property
@pulumi.getter
def METHOD_NAME(self) -> str:
return pulumi.get(self, "arn")
@property
@pulumi.getter
def id(self) -> str:
"""
The provider-assigned unique ID for this managed resource.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter(name="insecureValue")
def insecure_value(self) -> str:
return pulumi.get(self, "insecure_value")
@property
@pulumi.getter
def name(self) -> str:
return pulumi.get(self, "name")
@property
@pulumi.getter
def type(self) -> str:
return pulumi.get(self, "type")
@property
@pulumi.getter
def value(self) -> str:
return pulumi.get(self, "value")
@property
@pulumi.getter
def version(self) -> int:
return pulumi.get(self, "version")
@property
@pulumi.getter(name="withDecryption")
def with_decryption(self) -> Optional[bool]:
return pulumi.get(self, "with_decryption")
class AwaitableGetParameterResult(GetParameterResult):
# pylint: disable=using-constant-test
def __await__(self):
if False:
yield self
return GetParameterResult(
METHOD_NAME=self.METHOD_NAME,
id=self.id,
insecure_value=self.insecure_value,
name=self.name,
type=self.type,
value=self.value,
version=self.version,
with_decryption=self.with_decryption)
def get_parameter(name: Optional[str] = None,
with_decryption: Optional[bool] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetParameterResult:
"""
Provides an SSM Parameter data source.
## Example Usage
```python
import pulumi
import pulumi_aws as aws
foo = aws.ssm.get_parameter(name="foo")
```
> **Note:** The unencrypted value of a SecureString will be stored in the raw state as plain-text.
:param str name: Name of the parameter.
:param bool with_decryption: Whether to return decrypted `SecureString` value. Defaults to `true`.
In addition to all arguments above, the following attributes are exported:
"""
__args__ = dict()
__args__['name'] = name
__args__['withDecryption'] = with_decryption
opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)
__ret__ = pulumi.runtime.invoke('aws:ssm/getParameter:getParameter', __args__, opts=opts, typ=GetParameterResult).value
return AwaitableGetParameterResult(
METHOD_NAME=pulumi.get(__ret__, 'arn'),
id=pulumi.get(__ret__, 'id'),
insecure_value=pulumi.get(__ret__, 'insecure_value'),
name=pulumi.get(__ret__, 'name'),
type=pulumi.get(__ret__, 'type'),
value=pulumi.get(__ret__, 'value'),
version=pulumi.get(__ret__, 'version'),
with_decryption=pulumi.get(__ret__, 'with_decryption'))
@_utilities.lift_output_func(get_parameter)
def get_parameter_output(name: Optional[pulumi.Input[str]] = None,
with_decryption: Optional[pulumi.Input[Optional[bool]]] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetParameterResult]:
"""
Provides an SSM Parameter data source.
## Example Usage
```python
import pulumi
import pulumi_aws as aws
foo = aws.ssm.get_parameter(name="foo")
```
> **Note:** The unencrypted value of a SecureString will be stored in the raw state as plain-text.
:param str name: Name of the parameter.
:param bool with_decryption: Whether to return decrypted `SecureString` value. Defaults to `true`.
In addition to all arguments above, the following attributes are exported:
"""
... | null |
properties | # coding=utf-8
# *** WARNING: this file was generated by pulumi. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import copy
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from ... import _utilities
from . import outputs
__all__ = [
'GetContainerRegistryResult',
'AwaitableGetContainerRegistryResult',
'get_container_registry',
'get_container_registry_output',
]
@pulumi.output_type
class GetContainerRegistryResult:
"""
Container registry resource payload.
"""
def __init__(__self__, id=None, name=None, METHOD_NAME=None, system_data=None, type=None):
if id and not isinstance(id, str):
raise TypeError("Expected argument 'id' to be a str")
pulumi.set(__self__, "id", id)
if name and not isinstance(name, str):
raise TypeError("Expected argument 'name' to be a str")
pulumi.set(__self__, "name", name)
if METHOD_NAME and not isinstance(METHOD_NAME, dict):
raise TypeError("Expected argument 'properties' to be a dict")
pulumi.set(__self__, "properties", METHOD_NAME)
if system_data and not isinstance(system_data, dict):
raise TypeError("Expected argument 'system_data' to be a dict")
pulumi.set(__self__, "system_data", system_data)
if type and not isinstance(type, str):
raise TypeError("Expected argument 'type' to be a str")
pulumi.set(__self__, "type", type)
@property
@pulumi.getter
def id(self) -> str:
"""
Fully qualified resource Id for the resource.
"""
return pulumi.get(self, "id")
@property
@pulumi.getter
def name(self) -> str:
"""
The name of the resource.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter
def METHOD_NAME(self) -> 'outputs.ContainerRegistryPropertiesResponse':
"""
Properties of the container registry resource payload.
"""
return pulumi.get(self, "properties")
@property
@pulumi.getter(name="systemData")
def system_data(self) -> 'outputs.SystemDataResponse':
"""
Metadata pertaining to creation and last modification of the resource.
"""
return pulumi.get(self, "system_data")
@property
@pulumi.getter
def type(self) -> str:
"""
The type of the resource.
"""
return pulumi.get(self, "type")
class AwaitableGetContainerRegistryResult(GetContainerRegistryResult):
# pylint: disable=using-constant-test
def __await__(self):
if False:
yield self
return GetContainerRegistryResult(
id=self.id,
name=self.name,
METHOD_NAME=self.METHOD_NAME,
system_data=self.system_data,
type=self.type)
def get_container_registry(container_registry_name: Optional[str] = None,
resource_group_name: Optional[str] = None,
service_name: Optional[str] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetContainerRegistryResult:
"""
Get the container registries resource.
:param str container_registry_name: The name of the container registry.
:param str resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal.
:param str service_name: The name of the Service resource.
"""
__args__ = dict()
__args__['containerRegistryName'] = container_registry_name
__args__['resourceGroupName'] = resource_group_name
__args__['serviceName'] = service_name
opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)
__ret__ = pulumi.runtime.invoke('azure-native:appplatform/v20230701preview:getContainerRegistry', __args__, opts=opts, typ=GetContainerRegistryResult).value
return AwaitableGetContainerRegistryResult(
id=pulumi.get(__ret__, 'id'),
name=pulumi.get(__ret__, 'name'),
METHOD_NAME=pulumi.get(__ret__, 'properties'),
system_data=pulumi.get(__ret__, 'system_data'),
type=pulumi.get(__ret__, 'type'))
@_utilities.lift_output_func(get_container_registry)
def get_container_registry_output(container_registry_name: Optional[pulumi.Input[str]] = None,
resource_group_name: Optional[pulumi.Input[str]] = None,
service_name: Optional[pulumi.Input[str]] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetContainerRegistryResult]:
"""
Get the container registries resource.
:param str container_registry_name: The name of the container registry.
:param str resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal.
:param str service_name: The name of the Service resource.
"""
... | null |
test get zone normal | """
:codeauthor: Jayesh Kariya <[email protected]>
"""
import pytest
import salt.modules.win_timezone as win_timezone
from salt.exceptions import CommandExecutionError
from tests.support.mock import MagicMock, patch
pytestmark = [
pytest.mark.skipif(not win_timezone.HAS_PYTZ, reason="This test requires pytz"),
]
@pytest.fixture
def configure_loader_modules():
return {
win_timezone: {
"__opts__": {},
"__salt__": {},
"__utils__": {},
},
}
def METHOD_NAME():
"""
Test if it get current timezone (i.e. Asia/Calcutta)
"""
mock_read_ok = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "India Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_ok}):
assert win_timezone.get_zone() == "Asia/Calcutta"
def test_get_zone_normal_dstoff():
"""
Test if it gets current timezone with dst off (i.e. America/Denver)
"""
mock_read_ok = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "Mountain Standard Time_dstoff",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_ok}):
assert win_timezone.get_zone() == "America/Denver"
def test_get_zone_normal_dstoff_issue():
"""
Test regression with dstoff fix stripping unwanted characters
"""
mock_read_ok = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "FLE Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_ok}):
assert win_timezone.get_zone() == "Europe/Kiev"
@pytest.mark.parametrize("timezone", win_timezone.mapper.list_win())
def test_get_zone_all(timezone):
"""
Test all Win zones are properly resolved and none returns Unknown
"""
mock_read_ok = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": timezone,
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_ok}):
assert win_timezone.get_zone() != "Unknown"
def test_get_zone_unknown():
"""
Test get_zone with unknown timezone (i.e. Indian Standard Time)
"""
mock_read_error = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "Indian Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_error}):
assert win_timezone.get_zone() == "Unknown"
def test_get_zone_error():
"""
Test get_zone when it encounters an error
"""
mock_read_fatal = MagicMock(
return_value={"pid": 78, "retcode": 1, "stderr": "", "stdout": ""}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read_fatal}):
with pytest.raises(CommandExecutionError):
win_timezone.get_zone()
def test_get_offset():
"""
Test if it get current numeric timezone offset from UCT (i.e. +0530)
"""
mock_read = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "India Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read}):
assert win_timezone.get_offset() == "+0530"
def test_get_zonecode():
"""
Test if it get current timezone (i.e. PST, MDT, etc)
"""
mock_read = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "India Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read}):
assert win_timezone.get_zonecode() == "IST"
def test_set_zone():
"""
Test if it unlinks, then symlinks /etc/localtime to the set timezone.
"""
mock_write = MagicMock(
return_value={"pid": 78, "retcode": 0, "stderr": "", "stdout": ""}
)
mock_read = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "India Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_write}), patch.dict(
win_timezone.__salt__, {"cmd.run_all": mock_read}
):
assert win_timezone.set_zone("Asia/Calcutta")
def test_zone_compare():
"""
Test if it checks the md5sum between the given timezone, and
the one set in /etc/localtime. Returns True if they match,
and False if not. Mostly useful for running state checks.
"""
mock_read = MagicMock(
return_value={
"pid": 78,
"retcode": 0,
"stderr": "",
"stdout": "India Standard Time",
}
)
with patch.dict(win_timezone.__salt__, {"cmd.run_all": mock_read}):
assert win_timezone.zone_compare("Asia/Calcutta")
def test_get_hwclock():
"""
Test if it get current hardware clock setting (UTC or localtime)
"""
assert win_timezone.get_hwclock() == "localtime"
def test_set_hwclock():
"""
Test if it sets the hardware clock to be either UTC or localtime
"""
assert not win_timezone.set_hwclock("UTC") | null |
prepare request | # pylint: disable=too-many-lines
# coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
import sys
from typing import Any, Callable, Dict, Iterable, Optional, TypeVar
import urllib.parse
from azure.core.exceptions import (
ClientAuthenticationError,
HttpResponseError,
ResourceExistsError,
ResourceNotFoundError,
ResourceNotModifiedError,
map_error,
)
from azure.core.paging import ItemPaged
from azure.core.pipeline import PipelineResponse
from azure.core.pipeline.transport import HttpResponse
from azure.core.rest import HttpRequest
from azure.core.tracing.decorator import distributed_trace
from azure.core.utils import case_insensitive_dict
from azure.mgmt.core.exceptions import ARMErrorFormat
from .. import models as _models
from .._serialization import Serializer
from .._vendor import _convert_request
if sys.version_info >= (3, 8):
from typing import Literal # pylint: disable=no-name-in-module, ungrouped-imports
else:
from typing_extensions import Literal # type: ignore # pylint: disable=ungrouped-imports
T = TypeVar("T")
ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]]
_SERIALIZER = Serializer()
_SERIALIZER.client_side_validation = False
def build_list_request(**kwargs: Any) -> HttpRequest:
_headers = case_insensitive_dict(kwargs.pop("headers", {}) or {})
_params = case_insensitive_dict(kwargs.pop("params", {}) or {})
api_version: Literal["2022-11-01-preview"] = kwargs.pop(
"api_version", _params.pop("api-version", "2022-11-01-preview")
)
accept = _headers.pop("Accept", "application/json")
# Construct URL
_url = kwargs.pop("template_url", "/providers/Microsoft.ServiceLinker/operations")
# Construct parameters
_params["api-version"] = _SERIALIZER.query("api_version", api_version, "str")
# Construct headers
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs)
class Operations:
"""
.. warning::
**DO NOT** instantiate this class directly.
Instead, you should access the following operations through
:class:`~azure.mgmt.servicelinker.ServiceLinkerManagementClient`'s
:attr:`operations` attribute.
"""
models = _models
def __init__(self, *args, **kwargs):
input_args = list(args)
self._client = input_args.pop(0) if input_args else kwargs.pop("client")
self._config = input_args.pop(0) if input_args else kwargs.pop("config")
self._serialize = input_args.pop(0) if input_args else kwargs.pop("serializer")
self._deserialize = input_args.pop(0) if input_args else kwargs.pop("deserializer")
@distributed_trace
def list(self, **kwargs: Any) -> Iterable["_models.Operation"]:
"""Lists the available ServiceLinker REST API operations.
:keyword callable cls: A custom type or function that will be passed the direct response
:return: An iterator like instance of either Operation or the result of cls(response)
:rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.servicelinker.models.Operation]
:raises ~azure.core.exceptions.HttpResponseError:
"""
_headers = kwargs.pop("headers", {}) or {}
_params = case_insensitive_dict(kwargs.pop("params", {}) or {})
api_version: Literal["2022-11-01-preview"] = kwargs.pop(
"api_version", _params.pop("api-version", self._config.api_version)
)
cls: ClsType[_models.OperationListResult] = kwargs.pop("cls", None)
error_map = {
401: ClientAuthenticationError,
404: ResourceNotFoundError,
409: ResourceExistsError,
304: ResourceNotModifiedError,
}
error_map.update(kwargs.pop("error_map", {}) or {})
def METHOD_NAME(next_link=None):
if not next_link:
request = build_list_request(
api_version=api_version,
template_url=self.list.metadata["url"],
headers=_headers,
params=_params,
)
request = _convert_request(request)
request.url = self._client.format_url(request.url)
else:
# make call to next link with the client's api-version
_parsed_next_link = urllib.parse.urlparse(next_link)
_next_request_params = case_insensitive_dict(
{
key: [urllib.parse.quote(v) for v in value]
for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items()
}
)
_next_request_params["api-version"] = self._config.api_version
request = HttpRequest(
"GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params
)
request = _convert_request(request)
request.url = self._client.format_url(request.url)
request.method = "GET"
return request
def extract_data(pipeline_response):
deserialized = self._deserialize("OperationListResult", pipeline_response)
list_of_elem = deserialized.value
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return deserialized.next_link or None, iter(list_of_elem)
def get_next(next_link=None):
request = METHOD_NAME(next_link)
pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access
request, stream=False, **kwargs
)
response = pipeline_response.http_response
if response.status_code not in [200]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response)
raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)
return pipeline_response
return ItemPaged(get_next, extract_data)
list.metadata = {"url": "/providers/Microsoft.ServiceLinker/operations"} | null |
publish | import asyncio
import pickle
try: # pragma: no cover
from redis import asyncio as aioredis
from redis.exceptions import RedisError
except ImportError: # pragma: no cover
try:
import aioredis
from aioredis.exceptions import RedisError
except ImportError:
aioredis = None
RedisError = None
from .asyncio_pubsub_manager import AsyncPubSubManager
class AsyncRedisManager(AsyncPubSubManager): # pragma: no cover
"""Redis based client manager for asyncio servers.
This class implements a Redis backend for event sharing across multiple
processes.
To use a Redis backend, initialize the :class:`AsyncServer` instance as
follows::
url = 'redis://hostname:port/0'
server = socketio.AsyncServer(
client_manager=socketio.AsyncRedisManager(url))
:param url: The connection URL for the Redis server. For a default Redis
store running on the same host, use ``redis://``. To use an
SSL connection, use ``rediss://``.
:param channel: The channel name on which the server sends and receives
notifications. Must be the same in all the servers.
:param write_only: If set to ``True``, only initialize to emit events. The
default of ``False`` initializes the class for emitting
and receiving.
:param redis_options: additional keyword arguments to be passed to
``aioredis.from_url()``.
"""
name = 'aioredis'
def __init__(self, url='redis://localhost:6379/0', channel='socketio',
write_only=False, logger=None, redis_options=None):
if aioredis is None:
raise RuntimeError('Redis package is not installed '
'(Run "pip install redis" in your virtualenv).')
if not hasattr(aioredis.Redis, 'from_url'):
raise RuntimeError('Version 2 of aioredis package is required.')
self.redis_url = url
self.redis_options = redis_options or {}
self._redis_connect()
super().__init__(channel=channel, write_only=write_only, logger=logger)
def _redis_connect(self):
self.redis = aioredis.Redis.from_url(self.redis_url,
**self.redis_options)
self.pubsub = self.redis.pubsub(ignore_subscribe_messages=True)
async def METHOD_NAME(self, data):
retry = True
while True:
try:
if not retry:
self._redis_connect()
return await self.redis.publish(
self.channel, pickle.dumps(data))
except RedisError:
if retry:
self._get_logger().error('Cannot publish to redis... '
'retrying')
retry = False
else:
self._get_logger().error('Cannot publish to redis... '
'giving up')
break
async def _redis_listen_with_retries(self):
retry_sleep = 1
connect = False
while True:
try:
if connect:
self._redis_connect()
await self.pubsub.subscribe(self.channel)
retry_sleep = 1
async for message in self.pubsub.listen():
yield message
except RedisError:
self._get_logger().error('Cannot receive from redis... '
'retrying in '
'{} secs'.format(retry_sleep))
connect = True
await asyncio.sleep(retry_sleep)
retry_sleep *= 2
if retry_sleep > 60:
retry_sleep = 60
async def _listen(self):
channel = self.channel.encode('utf-8')
await self.pubsub.subscribe(self.channel)
async for message in self._redis_listen_with_retries():
if message['channel'] == channel and \
message['type'] == 'message' and 'data' in message:
yield message['data']
await self.pubsub.unsubscribe(self.channel) | null |
test withdraw twice | # These tests are auto-generated with test data from:
# https://github.com/exercism/problem-specifications/tree/main/exercises/bank-account/canonical-data.json
# File last updated on 2023-07-20
import unittest
from bank_account import (
BankAccount,
)
class BankAccountTest(unittest.TestCase):
def test_newly_opened_account_has_zero_balance(self):
account = BankAccount()
account.open()
self.assertEqual(account.get_balance(), 0)
def test_single_deposit(self):
account = BankAccount()
account.open()
account.deposit(100)
self.assertEqual(account.get_balance(), 100)
def test_multiple_deposits(self):
account = BankAccount()
account.open()
account.deposit(100)
account.deposit(50)
self.assertEqual(account.get_balance(), 150)
def test_withdraw_once(self):
account = BankAccount()
account.open()
account.deposit(100)
account.withdraw(75)
self.assertEqual(account.get_balance(), 25)
def METHOD_NAME(self):
account = BankAccount()
account.open()
account.deposit(100)
account.withdraw(80)
account.withdraw(20)
self.assertEqual(account.get_balance(), 0)
def test_can_do_multiple_operations_sequentially(self):
account = BankAccount()
account.open()
account.deposit(100)
account.deposit(110)
account.withdraw(200)
account.deposit(60)
account.withdraw(50)
self.assertEqual(account.get_balance(), 20)
def test_cannot_check_balance_of_closed_account(self):
account = BankAccount()
account.open()
account.close()
with self.assertRaises(ValueError) as err:
account.get_balance()
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account not open")
def test_cannot_deposit_into_closed_account(self):
account = BankAccount()
account.open()
account.close()
with self.assertRaises(ValueError) as err:
account.deposit(50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account not open")
def test_cannot_deposit_into_unopened_account(self):
account = BankAccount()
with self.assertRaises(ValueError) as err:
account.deposit(50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account not open")
def test_cannot_withdraw_from_closed_account(self):
account = BankAccount()
account.open()
account.close()
with self.assertRaises(ValueError) as err:
account.withdraw(50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account not open")
def test_cannot_close_an_account_that_was_not_opened(self):
account = BankAccount()
with self.assertRaises(ValueError) as err:
account.close()
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account not open")
def test_cannot_open_an_already_opened_account(self):
account = BankAccount()
account.open()
with self.assertRaises(ValueError) as err:
account.open()
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "account already open")
def test_reopened_account_does_not_retain_balance(self):
account = BankAccount()
account.open()
account.deposit(50)
account.close()
account.open()
self.assertEqual(account.get_balance(), 0)
def test_cannot_withdraw_more_than_deposited(self):
account = BankAccount()
account.open()
account.deposit(25)
with self.assertRaises(ValueError) as err:
account.withdraw(50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "amount must be less than balance")
def test_cannot_withdraw_negative(self):
account = BankAccount()
account.open()
account.deposit(100)
with self.assertRaises(ValueError) as err:
account.withdraw(-50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "amount must be greater than 0")
def test_cannot_deposit_negative(self):
account = BankAccount()
account.open()
with self.assertRaises(ValueError) as err:
account.deposit(-50)
self.assertEqual(type(err.exception), ValueError)
self.assertEqual(err.exception.args[0], "amount must be greater than 0") | null |
add publication trustlevels | # automatically generated by the FlatBuffers compiler, do not modify
# namespace: proto
import flatbuffers
from flatbuffers.compat import import_numpy
np = import_numpy()
class SubscriberFeatures(object):
__slots__ = ['_tab']
@classmethod
def GetRootAs(cls, buf, offset=0):
n = flatbuffers.encode.Get(flatbuffers.packer.uoffset, buf, offset)
x = SubscriberFeatures()
x.Init(buf, n + offset)
return x
@classmethod
def GetRootAsSubscriberFeatures(cls, buf, offset=0):
"""This method is deprecated. Please switch to GetRootAs."""
return cls.GetRootAs(buf, offset)
# SubscriberFeatures
def Init(self, buf, pos):
self._tab = flatbuffers.table.Table(buf, pos)
# SubscriberFeatures
def PublisherIdentification(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(4))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def PatternBasedSubscription(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(6))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def PublicationTrustlevels(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(8))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def SubscriptionRevocation(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(10))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def EventHistory(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(12))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def AcknowledgeSubscriberReceived(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(14))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def PayloadTransparency(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(16))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# SubscriberFeatures
def PayloadEncryptionCryptobox(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(18))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
def SubscriberFeaturesStart(builder): builder.StartObject(8)
def Start(builder):
return SubscriberFeaturesStart(builder)
def SubscriberFeaturesAddPublisherIdentification(builder, publisherIdentification): builder.PrependBoolSlot(0, publisherIdentification, 0)
def AddPublisherIdentification(builder, publisherIdentification):
return SubscriberFeaturesAddPublisherIdentification(builder, publisherIdentification)
def SubscriberFeaturesAddPatternBasedSubscription(builder, patternBasedSubscription): builder.PrependBoolSlot(1, patternBasedSubscription, 0)
def AddPatternBasedSubscription(builder, patternBasedSubscription):
return SubscriberFeaturesAddPatternBasedSubscription(builder, patternBasedSubscription)
def SubscriberFeaturesAddPublicationTrustlevels(builder, publicationTrustlevels): builder.PrependBoolSlot(2, publicationTrustlevels, 0)
def METHOD_NAME(builder, publicationTrustlevels):
return SubscriberFeaturesAddPublicationTrustlevels(builder, publicationTrustlevels)
def SubscriberFeaturesAddSubscriptionRevocation(builder, subscriptionRevocation): builder.PrependBoolSlot(3, subscriptionRevocation, 0)
def AddSubscriptionRevocation(builder, subscriptionRevocation):
return SubscriberFeaturesAddSubscriptionRevocation(builder, subscriptionRevocation)
def SubscriberFeaturesAddEventHistory(builder, eventHistory): builder.PrependBoolSlot(4, eventHistory, 0)
def AddEventHistory(builder, eventHistory):
return SubscriberFeaturesAddEventHistory(builder, eventHistory)
def SubscriberFeaturesAddAcknowledgeSubscriberReceived(builder, acknowledgeSubscriberReceived): builder.PrependBoolSlot(5, acknowledgeSubscriberReceived, 0)
def AddAcknowledgeSubscriberReceived(builder, acknowledgeSubscriberReceived):
return SubscriberFeaturesAddAcknowledgeSubscriberReceived(builder, acknowledgeSubscriberReceived)
def SubscriberFeaturesAddPayloadTransparency(builder, payloadTransparency): builder.PrependBoolSlot(6, payloadTransparency, 0)
def AddPayloadTransparency(builder, payloadTransparency):
return SubscriberFeaturesAddPayloadTransparency(builder, payloadTransparency)
def SubscriberFeaturesAddPayloadEncryptionCryptobox(builder, payloadEncryptionCryptobox): builder.PrependBoolSlot(7, payloadEncryptionCryptobox, 0)
def AddPayloadEncryptionCryptobox(builder, payloadEncryptionCryptobox):
return SubscriberFeaturesAddPayloadEncryptionCryptobox(builder, payloadEncryptionCryptobox)
def SubscriberFeaturesEnd(builder): return builder.EndObject()
def End(builder):
return SubscriberFeaturesEnd(builder | null |
send signal | import os
import re
import pwd, grp
from . import interfaces
import tuned.logs
import tuned.consts as consts
from inspect import ismethod
import socket
import json
import select
log = tuned.logs.get()
class UnixSocketExporter(interfaces.ExporterInterface):
"""
Export method calls through Unix Domain Socket Interface.
We take a method to be exported and create a simple wrapper function
to call it. This is required as we need the original function to be
bound to the original object instance. While the wrapper will be bound
to an object we dynamically construct.
"""
def __init__(self, socket_path=consts.CFG_DEF_UNIX_SOCKET_PATH,
signal_paths=consts.CFG_DEF_UNIX_SOCKET_SIGNAL_PATHS,
ownership=consts.CFG_DEF_UNIX_SOCKET_OWNERSHIP,
permissions=consts.CFG_DEF_UNIX_SOCKET_PERMISIONS,
connections_backlog=consts.CFG_DEF_UNIX_SOCKET_CONNECTIONS_BACKLOG):
self._socket_path = socket_path
self._socket_object = None
self._socket_signal_paths = re.split(r",;", signal_paths) if signal_paths else []
self._socket_signal_objects = []
self._ownership = [-1, -1]
if ownership:
ownership = ownership.split()
for i, o in enumerate(ownership[:2]):
try:
self._ownership[i] = int(o)
except ValueError:
try:
# user
if i == 0:
self._ownership[i] = pwd.getpwnam(o).pw_uid
# group
else:
self._ownership[i] = grp.getgrnam(o).gr_gid
except KeyError:
log.error("%s '%s' does not exists, leaving default" % ("User" if i == 0 else "Group", o))
self._permissions = permissions
self._connections_backlog = connections_backlog
self._unix_socket_methods = {}
self._signals = set()
self._conn = None
self._channel = None
def running(self):
return self._socket_object is not None
def export(self, method, in_signature, out_signature):
if not ismethod(method):
raise Exception("Only bound methods can be exported.")
method_name = method.__name__
if method_name in self._unix_socket_methods:
raise Exception("Method with this name (%s) is already exported." % method_name)
class wrapper(object):
def __init__(self, in_signature, out_signature):
self._in_signature = in_signature
self._out_signature = out_signature
def __call__(self, *args, **kwargs):
return method(*args, **kwargs)
self._unix_socket_methods[method_name] = wrapper(in_signature, out_signature)
def signal(self, method, out_signature):
if not ismethod(method):
raise Exception("Only bound methods can be exported.")
method_name = method.__name__
if method_name in self._unix_socket_methods:
raise Exception("Method with this name (%s) is already exported." % method_name)
class wrapper(object):
def __init__(self, out_signature):
self._out_signature = out_signature
def __call__(self, *args, **kwargs):
return method(*args, **kwargs)
self._unix_socket_methods[method_name] = wrapper(out_signature)
self._signals.add(method_name)
def METHOD_NAME(self, signal, *args, **kwargs):
if not signal in self._signals:
raise Exception("Signal '%s' doesn't exist." % signal)
for p in self._socket_signal_paths:
log.debug("Sending signal on socket %s" % p)
try:
s = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
s.setblocking(False)
s.connect(p)
self._send_data(s, {"jsonrpc": "2.0", "method": signal, "params": args})
s.close()
except OSError as e:
log.warning("Error while sending signal '%s' to socket '%s': %s" % (signal, p, e))
def register_signal_path(self, path):
self._socket_signal_paths.append(path)
def _construct_socket_object(self):
if self._socket_path:
if os.path.exists(self._socket_path):
os.unlink(self._socket_path)
self._socket_object = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
self._socket_object.bind(self._socket_path)
self._socket_object.listen(self._connections_backlog)
os.chown(self._socket_path, self._ownership[0], self._ownership[1])
if self._permissions:
os.chmod(self._socket_path, self._permissions)
def start(self):
if self.running():
return
self.stop()
self._construct_socket_object()
def stop(self):
if self._socket_object:
self._socket_object.close()
def _send_data(self, s, data):
log.debug("Sending socket data: %s)" % data)
try:
s.send(json.dumps(data).encode("utf-8"))
except Exception as e:
log.warning("Failed to send data '%s': %s" % (data, e))
def _create_response(self, data, id, error=False):
res = {
"jsonrpc": "2.0",
"id": id
}
if error:
res["error"] = data
else:
res["result"] = data
return res
def _create_error_responce(self, code, message, id=None, data=None):
return self._create_response({
"code": code,
"message": message,
"data": data,
}, error=True, id=id)
def _create_result_response(self, result, id):
return self._create_response(result, id)
def _check_id(self, data):
if data.get("id"):
return data
return None
def _process_request(self, req):
if type(req) != dict or req.get("jsonrpc") != "2.0" or not req.get("method"):
return self._create_error_responce(-32600, "Invalid Request")
id = req.get("id")
ret = None
if req["method"] not in self._unix_socket_methods:
return self._check_id(self._create_error_responce(-32601, "Method not found", id))
try:
if not req.get("params"):
ret = self._unix_socket_methods[req["method"]]()
elif type(req["params"]) in (list, tuple):
ret = self._unix_socket_methods[req["method"]](*req["params"])
elif type(req["params"]) == dict:
ret = self._unix_socket_methods[req["method"]](**req["params"])
else:
return self._check_id(self._create_error_responce(-32600, "Invalid Request", id))
except TypeError as e:
return self._check_id(self._create_error_responce(-32602, "Invalid params", id, str(e)))
except Exception as e:
return self._check_id(self._create_error_responce(1, "Error", id, str(e)))
return self._check_id(self._create_result_response(ret, id))
def period_check(self):
"""
Periodically checks socket object for new calls. This allows to function without special thread.
Interface is according JSON-RPC 2.0 Specification (see https://www.jsonrpc.org/specification)
Example calls:
printf '[{"jsonrpc": "2.0", "method": "active_profile", "id": 1}, {"jsonrpc": "2.0", "method": "profiles", "id": 2}]' | nc -U /run/tuned/tuned.sock
printf '{"jsonrpc": "2.0", "method": "switch_profile", "params": {"profile_name": "balanced"}, "id": 1}' | nc -U /run/tuned/tuned.sock
"""
if not self.running():
return
while True:
r, _, _ = select.select([self._socket_object], (), (), 0)
if r:
conn, _ = self._socket_object.accept()
try:
data = ""
while True:
rec_data = conn.recv(4096).decode()
if not rec_data:
break
data += rec_data
except Exception as e:
log.error("Failed to load data of message: %s" % e)
continue
if data:
try:
data = json.loads(data)
except Exception as e:
log.error("Failed to load json data '%s': %s" % (data, e))
self._send_data(conn, self._create_error_responce(-32700, "Parse error", str(e)))
continue
if type(data) not in (tuple, list, dict):
log.error("Wrong format of call")
self._send_data(conn, self._create_error_responce(-32700, "Parse error", str(e)))
continue
if type(data) in (tuple, list):
if len(data) == 0:
self._send_data(conn, self._create_error_responce(-32600, "Invalid Request", str(e)))
continue
res = []
for req in data:
r = self._process_request(req)
if r:
res.append(r)
if res:
self._send_data(conn, res)
else:
res = self._process_request(data)
if r:
self._send_data(conn, res)
else:
return
| null |
test sythesize sequences | import datetime
import pandas as pd
from deepecho import load_demo
from sdv.datasets.demo import download_demo
from sdv.metadata import SingleTableMetadata
from sdv.sequential import PARSynthesizer
def _get_par_data_and_metadata():
date = datetime.datetime.strptime('2020-01-01', '%Y-%m-%d')
data = pd.DataFrame({
'column1': [1.0, 2.0, 1.5, 1.3],
'date': [date, date, date, date],
'column2': ['b', 'a', 'a', 'c'],
'entity': [1, 1, 2, 2],
'context': ['a', 'a', 'b', 'b']
})
metadata = SingleTableMetadata()
metadata.detect_from_dataframe(data)
metadata.update_column('entity', sdtype='id')
metadata.set_sequence_key('entity')
metadata.set_sequence_index('date')
return data, metadata
def test_par():
"""Test the ``PARSynthesizer`` end to end."""
# Setup
data = load_demo()
data['date'] = pd.to_datetime(data['date'])
metadata = SingleTableMetadata()
metadata.detect_from_dataframe(data)
metadata.update_column('store_id', sdtype='id')
metadata.set_sequence_key('store_id')
metadata.set_sequence_index('date')
model = PARSynthesizer(
metadata=metadata,
context_columns=['region'],
epochs=1,
)
# Run
model.fit(data)
sampled = model.sample(100)
# Assert
assert sampled.shape == data.shape
assert (sampled.dtypes == data.dtypes).all()
assert (sampled.notna().sum(axis=1) != 0).all()
def test_column_after_date_simple():
"""Test that adding a column after the ``sequence_index`` column works."""
# Setup
date = datetime.datetime.strptime('2020-01-01', '%Y-%m-%d')
data = pd.DataFrame({
'col': [1, 1],
'date': [date, date],
'col2': ['hello', 'world'],
})
metadata = SingleTableMetadata()
metadata.detect_from_dataframe(data)
metadata.update_column('col', sdtype='id')
metadata.set_sequence_key('col')
metadata.set_sequence_index('date')
# Run
model = PARSynthesizer(metadata=metadata, epochs=1)
model.fit(data)
sampled = model.sample(1)
# Assert
assert sampled.shape == data.shape
assert (sampled.dtypes == data.dtypes).all()
assert (sampled.notna().sum(axis=1) != 0).all()
def test_column_after_date_complex():
"""Test that adding multiple columns after the ``sequence_index`` column works."""
# Setup
data, metadata = _get_par_data_and_metadata()
# Run
model = PARSynthesizer(metadata=metadata, context_columns=['context'], epochs=1)
model.fit(data)
sampled = model.sample(2)
# Assert
assert sampled.shape == data.shape
assert (sampled.dtypes == data.dtypes).all()
assert (sampled.notna().sum(axis=1) != 0).all()
def test_save_and_load(tmp_path):
"""Test that synthesizers can be saved and loaded properly."""
# Setup
_, metadata = _get_par_data_and_metadata()
instance = PARSynthesizer(metadata)
synthesizer_path = tmp_path / 'synthesizer.pkl'
instance.save(synthesizer_path)
# Run
loaded_instance = PARSynthesizer.load(synthesizer_path)
# Assert
assert isinstance(loaded_instance, PARSynthesizer)
assert metadata == instance.metadata
def METHOD_NAME(tmp_path):
"""End to end test for synthesizing sequences.
The following functionalities are being tested:
* Fit a ``PARSynthesizer`` with the demo dataset.
* Fit a ``PARSynthesizer`` with custom context.
* Sample from the model.
* Conditionally sample from the model.
* Save and Load.
"""
# Setup
real_data, metadata = download_demo(
modality='sequential',
dataset_name='nasdaq100_2019'
)
real_data[real_data['Symbol'] == 'AMZN']['Sector'].unique()
synthesizer = PARSynthesizer(
metadata,
epochs=5,
context_columns=['Sector', 'Industry']
)
custom_synthesizer = PARSynthesizer(
metadata,
epochs=5,
context_columns=['Sector', 'Industry'],
verbose=True
)
scenario_context = pd.DataFrame(data={
'Symbol': ['COMPANY-A', 'COMPANY-B', 'COMPANY-C', 'COMPANY-D', 'COMPANY-E'],
'Sector': ['Technology'] * 2 + ['Consumer Services'] * 3,
'Industry': [
'Computer Manufacturing', 'Computer Software: Prepackaged Software',
'Hotels/Resorts', 'Restaurants', 'Clothing/Shoe/Accessory Stores'
]
})
# Run - Fit
synthesizer.fit(real_data)
custom_synthesizer.fit(real_data)
# Run - Sample
synthetic_data = synthesizer.sample(num_sequences=10)
custom_synthetic_data = custom_synthesizer.sample(num_sequences=3, sequence_length=2)
custom_synthetic_data_conditional = custom_synthesizer.sample_sequential_columns(
context_columns=scenario_context,
sequence_length=2
)
# Save and Load
model_path = tmp_path / 'my_synthesizer.pkl'
synthesizer.save(model_path)
loaded_synthesizer = PARSynthesizer.load(model_path)
loaded_sample = loaded_synthesizer.sample(100)
# Assert
assert all(custom_synthetic_data_conditional['Symbol'].value_counts() == 2)
companies = ['COMPANY-A', 'COMPANY-B', 'COMPANY-C', 'COMPANY-D', 'COMPANY-E']
assert companies in custom_synthetic_data_conditional['Symbol'].unique()
assert custom_synthetic_data_conditional['Sector'].value_counts()['Technology'] == 4
assert custom_synthetic_data_conditional['Sector'].value_counts()['Consumer Services'] == 6
industries = [
'Computer Manufacturing',
'Computer Software: Prepackaged Software',
'Hotels/Resorts',
'Restaurants',
'Clothing/Shoe/Accessory Stores'
]
assert industries in custom_synthetic_data_conditional['Industry'].unique()
assert model_path.exists()
assert model_path.is_file()
assert loaded_synthesizer.get_info() == synthesizer.get_info()
assert loaded_synthesizer.metadata.to_dict() == metadata.to_dict()
synthesizer.validate(synthetic_data)
synthesizer.validate(custom_synthetic_data)
synthesizer.validate(custom_synthetic_data_conditional)
synthesizer.validate(loaded_sample)
loaded_synthesizer.validate(synthetic_data)
loaded_synthesizer.validate(loaded_sample) | null |
from server address and domain | # -*- test-case-name: twisted.names.test.test_names -*-
# Copyright (c) Twisted Matrix Laboratories.
# See LICENSE for details.
__all__ = ["SecondaryAuthority", "SecondaryAuthorityService"]
from twisted.application import service
from twisted.internet import defer, task
from twisted.names import client, common, dns, resolve
from twisted.names.authority import FileAuthority
from twisted.python import failure, log
from twisted.python.compat import nativeString
class SecondaryAuthorityService(service.Service):
"""
A service that keeps one or more authorities up to date by doing hourly
zone transfers from a master.
@ivar primary: IP address of the master.
@type primary: L{str}
@ivar domains: An authority for each domain mirrored from the master.
@type domains: L{list} of L{SecondaryAuthority}
"""
calls = None
_port = 53
def __init__(self, primary, domains):
"""
@param primary: The IP address of the server from which to perform
zone transfers.
@type primary: L{str}
@param domains: A sequence of domain names for which to perform
zone transfers.
@type domains: L{list} of L{bytes}
"""
self.primary = nativeString(primary)
self.domains = [SecondaryAuthority(primary, d) for d in domains]
@classmethod
def fromServerAddressAndDomains(cls, serverAddress, domains):
"""
Construct a new L{SecondaryAuthorityService} from a tuple giving a
server address and a C{str} giving the name of a domain for which this
is an authority.
@param serverAddress: A two-tuple, the first element of which is a
C{str} giving an IP address and the second element of which is a
C{int} giving a port number. Together, these define where zone
transfers will be attempted from.
@param domains: Domain names for which to perform zone transfers.
@type domains: sequence of L{bytes}
@return: A new instance of L{SecondaryAuthorityService}.
"""
primary, port = serverAddress
service = cls(primary, [])
service._port = port
service.domains = [
SecondaryAuthority.METHOD_NAME(serverAddress, d)
for d in domains
]
return service
def getAuthority(self):
"""
Get a resolver for the transferred domains.
@rtype: L{ResolverChain}
"""
return resolve.ResolverChain(self.domains)
def startService(self):
service.Service.startService(self)
self.calls = [task.LoopingCall(d.transfer) for d in self.domains]
i = 0
from twisted.internet import reactor
for c in self.calls:
# XXX Add errbacks, respect proper timeouts
reactor.callLater(i, c.start, 60 * 60)
i += 1
def stopService(self):
service.Service.stopService(self)
for c in self.calls:
c.stop()
class SecondaryAuthority(FileAuthority):
"""
An Authority that keeps itself updated by performing zone transfers.
@ivar primary: The IP address of the server from which zone transfers will
be attempted.
@type primary: L{str}
@ivar _port: The port number of the server from which zone transfers will
be attempted.
@type _port: L{int}
@ivar domain: The domain for which this is the secondary authority.
@type domain: L{bytes}
@ivar _reactor: The reactor to use to perform the zone transfers, or
L{None} to use the global reactor.
"""
transferring = False
soa = records = None
_port = 53
_reactor = None
def __init__(self, primaryIP, domain):
"""
@param domain: The domain for which this will be the secondary
authority.
@type domain: L{bytes} or L{str}
"""
# Yep. Skip over FileAuthority.__init__. This is a hack until we have
# a good composition-based API for the complicated DNS record lookup
# logic we want to share.
common.ResolverBase.__init__(self)
self.primary = nativeString(primaryIP)
self.domain = dns.domainString(domain)
@classmethod
def METHOD_NAME(cls, serverAddress, domain):
"""
Construct a new L{SecondaryAuthority} from a tuple giving a server
address and a C{bytes} giving the name of a domain for which this is an
authority.
@param serverAddress: A two-tuple, the first element of which is a
C{str} giving an IP address and the second element of which is a
C{int} giving a port number. Together, these define where zone
transfers will be attempted from.
@param domain: A C{bytes} giving the domain to transfer.
@type domain: L{bytes}
@return: A new instance of L{SecondaryAuthority}.
"""
primary, port = serverAddress
secondary = cls(primary, domain)
secondary._port = port
return secondary
def transfer(self):
"""
Attempt a zone transfer.
@returns: A L{Deferred} that fires with L{None} when attempted zone
transfer has completed.
"""
# FIXME: This logic doesn't avoid duplicate transfers
# https://twistedmatrix.com/trac/ticket/9754
if self.transferring: # <-- never true
return
self.transfering = True # <-- speling
reactor = self._reactor
if reactor is None:
from twisted.internet import reactor
resolver = client.Resolver(
servers=[(self.primary, self._port)], reactor=reactor
)
return (
resolver.lookupZone(self.domain)
.addCallback(self._cbZone)
.addErrback(self._ebZone)
)
def _lookup(self, name, cls, type, timeout=None):
if not self.soa or not self.records:
# No transfer has occurred yet. Fail non-authoritatively so that
# the caller can try elsewhere.
return defer.fail(failure.Failure(dns.DomainError(name)))
return FileAuthority._lookup(self, name, cls, type, timeout)
def _cbZone(self, zone):
ans, _, _ = zone
self.records = r = {}
for rec in ans:
if not self.soa and rec.type == dns.SOA:
self.soa = (rec.name.name.lower(), rec.payload)
else:
r.setdefault(rec.name.name.lower(), []).append(rec.payload)
def _ebZone(self, failure):
log.msg(
"Updating %s from %s failed during zone transfer"
% (self.domain, self.primary)
)
log.err(failure)
def update(self):
self.transfer().addCallbacks(self._cbTransferred, self._ebTransferred)
def _cbTransferred(self, result):
self.transferring = False
def _ebTransferred(self, failure):
self.transferred = False
log.msg(
"Transferring %s from %s failed after zone transfer"
% (self.domain, self.primary)
)
log.err(failure) | null |
test policy state spec is empty | # coding=utf-8
# Copyright 2020 The TF-Agents Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Test for tf_agents.utils.random_py_policy."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
import tensorflow as tf # pylint: disable=g-explicit-tensorflow-version-import
from tf_agents.policies import random_py_policy
from tf_agents.specs import array_spec
from tf_agents.trajectories import time_step
from tf_agents.utils import test_utils
class RandomPyPolicyTest(test_utils.TestCase):
def setUp(self):
super(RandomPyPolicyTest, self).setUp()
self._time_step_spec = time_step.time_step_spec(
observation_spec=array_spec.ArraySpec((1,), np.int32)
)
self._time_step = time_step.restart(observation=np.array([1]))
def testGeneratesActions(self):
action_spec = [
array_spec.BoundedArraySpec((2, 3), np.int32, -10, 10),
array_spec.BoundedArraySpec((1, 2), np.int32, -10, 10),
]
policy = random_py_policy.RandomPyPolicy(
time_step_spec=self._time_step_spec, action_spec=action_spec
)
action_step = policy.action(self._time_step)
tf.nest.assert_same_structure(action_spec, action_step.action)
self.assertTrue(np.all(action_step.action[0] >= -10))
self.assertTrue(np.all(action_step.action[0] <= 10))
self.assertTrue(np.all(action_step.action[1] >= -10))
self.assertTrue(np.all(action_step.action[1] <= 10))
def testGeneratesBatchedActions(self):
action_spec = [
array_spec.BoundedArraySpec((2, 3), np.int32, -10, 10),
array_spec.BoundedArraySpec((1, 2), np.int32, -10, 10),
]
policy = random_py_policy.RandomPyPolicy(
time_step_spec=self._time_step_spec,
action_spec=action_spec,
outer_dims=(3,),
)
action_step = policy.action(self._time_step)
tf.nest.assert_same_structure(action_spec, action_step.action)
self.assertEqual((3, 2, 3), action_step.action[0].shape)
self.assertEqual((3, 1, 2), action_step.action[1].shape)
self.assertTrue(np.all(action_step.action[0] >= -10))
self.assertTrue(np.all(action_step.action[0] <= 10))
self.assertTrue(np.all(action_step.action[1] >= -10))
self.assertTrue(np.all(action_step.action[1] <= 10))
def testGeneratesBatchedActionsWithoutSpecifyingOuterDims(self):
action_spec = [
array_spec.BoundedArraySpec((2, 3), np.int32, -10, 10),
array_spec.BoundedArraySpec((1, 2), np.int32, -10, 10),
]
time_step_spec = time_step.time_step_spec(
observation_spec=array_spec.ArraySpec((1,), np.int32)
)
policy = random_py_policy.RandomPyPolicy(
time_step_spec=time_step_spec, action_spec=action_spec
)
action_step = policy.action(
time_step.restart(np.array([[1], [2], [3]], dtype=np.int32))
)
tf.nest.assert_same_structure(action_spec, action_step.action)
self.assertEqual((3, 2, 3), action_step.action[0].shape)
self.assertEqual((3, 1, 2), action_step.action[1].shape)
self.assertTrue(np.all(action_step.action[0] >= -10))
self.assertTrue(np.all(action_step.action[0] <= 10))
self.assertTrue(np.all(action_step.action[1] >= -10))
self.assertTrue(np.all(action_step.action[1] <= 10))
def METHOD_NAME(self):
policy = random_py_policy.RandomPyPolicy(
time_step_spec=self._time_step_spec, action_spec=[]
)
self.assertEqual(policy.policy_state_spec, ())
def testMasking(self):
batch_size = 1000
time_step_spec = time_step.time_step_spec(
observation_spec=array_spec.ArraySpec((1,), np.int32)
)
action_spec = array_spec.BoundedArraySpec((), np.int64, -5, 5)
# We create a fixed mask here for testing purposes. Normally the mask would
# be part of the observation.
mask = [0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0]
np_mask = np.array(mask)
batched_mask = np.array([mask for _ in range(batch_size)])
policy = random_py_policy.RandomPyPolicy(
time_step_spec=time_step_spec,
action_spec=action_spec,
observation_and_action_constraint_splitter=(
lambda obs: (obs, batched_mask)
),
)
my_time_step = time_step.restart(time_step_spec, batch_size)
action_step = policy.action(my_time_step)
tf.nest.assert_same_structure(action_spec, action_step.action)
# Sample from the policy 1000 times, and ensure that actions considered
# invalid according to the mask are never chosen.
action_ = self.evaluate(action_step.action)
self.assertTrue(np.all(action_ >= -5))
self.assertTrue(np.all(action_ <= 5))
self.assertAllEqual(
np_mask[action_ - action_spec.minimum], np.ones([batch_size])
)
# Ensure that all valid actions occur somewhere within the batch. Because we
# sample 1000 times, the chance of this failing for any particular action is
# (2/3)^1000, roughly 1e-176.
for index in range(action_spec.minimum, action_spec.maximum + 1):
if np_mask[index - action_spec.minimum]:
self.assertIn(index, action_)
if __name__ == '__main__':
test_utils.main() | null |
build list by location request | # pylint: disable=too-many-lines
# coding=utf-8
# --------------------------------------------------------------------------
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License. See License.txt in the project root for license information.
# Code generated by Microsoft (R) AutoRest Code Generator.
# Changes may cause incorrect behavior and will be lost if the code is regenerated.
# --------------------------------------------------------------------------
from typing import Any, Callable, Dict, Iterable, Optional, TypeVar
import urllib.parse
from azure.core.exceptions import (
ClientAuthenticationError,
HttpResponseError,
ResourceExistsError,
ResourceNotFoundError,
ResourceNotModifiedError,
map_error,
)
from azure.core.paging import ItemPaged
from azure.core.pipeline import PipelineResponse
from azure.core.pipeline.transport import HttpResponse
from azure.core.rest import HttpRequest
from azure.core.tracing.decorator import distributed_trace
from azure.core.utils import case_insensitive_dict
from azure.mgmt.core.exceptions import ARMErrorFormat
from .. import models as _models
from ..._serialization import Serializer
from .._vendor import _convert_request
T = TypeVar("T")
ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]]
_SERIALIZER = Serializer()
_SERIALIZER.client_side_validation = False
def METHOD_NAME(location: str, subscription_id: str, **kwargs: Any) -> HttpRequest:
_headers = case_insensitive_dict(kwargs.pop("headers", {}) or {})
_params = case_insensitive_dict(kwargs.pop("params", {}) or {})
api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2021-02-01"))
accept = _headers.pop("Accept", "application/json")
# Construct URL
_url = kwargs.pop(
"template_url", "/subscriptions/{subscriptionId}/providers/Microsoft.Storage/locations/{location}/usages"
) # pylint: disable=line-too-long
path_format_arguments = {
"subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1),
"location": _SERIALIZER.url("location", location, "str"),
}
_url: str = _url.format(**path_format_arguments) # type: ignore
# Construct parameters
_params["api-version"] = _SERIALIZER.query("api_version", api_version, "str")
# Construct headers
_headers["Accept"] = _SERIALIZER.header("accept", accept, "str")
return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs)
class UsagesOperations:
"""
.. warning::
**DO NOT** instantiate this class directly.
Instead, you should access the following operations through
:class:`~azure.mgmt.storage.v2021_02_01.StorageManagementClient`'s
:attr:`usages` attribute.
"""
models = _models
def __init__(self, *args, **kwargs):
input_args = list(args)
self._client = input_args.pop(0) if input_args else kwargs.pop("client")
self._config = input_args.pop(0) if input_args else kwargs.pop("config")
self._serialize = input_args.pop(0) if input_args else kwargs.pop("serializer")
self._deserialize = input_args.pop(0) if input_args else kwargs.pop("deserializer")
self._api_version = input_args.pop(0) if input_args else kwargs.pop("api_version")
@distributed_trace
def list_by_location(self, location: str, **kwargs: Any) -> Iterable["_models.Usage"]:
"""Gets the current usage count and the limit for the resources of the location under the
subscription.
:param location: The location of the Azure Storage resource. Required.
:type location: str
:keyword callable cls: A custom type or function that will be passed the direct response
:return: An iterator like instance of either Usage or the result of cls(response)
:rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.storage.v2021_02_01.models.Usage]
:raises ~azure.core.exceptions.HttpResponseError:
"""
_headers = kwargs.pop("headers", {}) or {}
_params = case_insensitive_dict(kwargs.pop("params", {}) or {})
api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._api_version or "2021-02-01"))
cls: ClsType[_models.UsageListResult] = kwargs.pop("cls", None)
error_map = {
401: ClientAuthenticationError,
404: ResourceNotFoundError,
409: ResourceExistsError,
304: ResourceNotModifiedError,
}
error_map.update(kwargs.pop("error_map", {}) or {})
def prepare_request(next_link=None):
if not next_link:
request = METHOD_NAME(
location=location,
subscription_id=self._config.subscription_id,
api_version=api_version,
template_url=self.list_by_location.metadata["url"],
headers=_headers,
params=_params,
)
request = _convert_request(request)
request.url = self._client.format_url(request.url)
else:
# make call to next link with the client's api-version
_parsed_next_link = urllib.parse.urlparse(next_link)
_next_request_params = case_insensitive_dict(
{
key: [urllib.parse.quote(v) for v in value]
for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items()
}
)
_next_request_params["api-version"] = self._config.api_version
request = HttpRequest(
"GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params
)
request = _convert_request(request)
request.url = self._client.format_url(request.url)
request.method = "GET"
return request
def extract_data(pipeline_response):
deserialized = self._deserialize("UsageListResult", pipeline_response)
list_of_elem = deserialized.value
if cls:
list_of_elem = cls(list_of_elem) # type: ignore
return None, iter(list_of_elem)
def get_next(next_link=None):
request = prepare_request(next_link)
_stream = False
pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access
request, stream=_stream, **kwargs
)
response = pipeline_response.http_response
if response.status_code not in [200]:
map_error(status_code=response.status_code, response=response, error_map=error_map)
raise HttpResponseError(response=response, error_format=ARMErrorFormat)
return pipeline_response
return ItemPaged(get_next, extract_data)
list_by_location.metadata = {
"url": "/subscriptions/{subscriptionId}/providers/Microsoft.Storage/locations/{location}/usages"
} | null |
perform http01 | """Example ACME-V2 API for HTTP-01 challenge.
Brief:
This a complete usage example of the python-acme API.
Limitations of this example:
- Works for only one Domain name
- Performs only HTTP-01 challenge
- Uses ACME-v2
Workflow:
(Account creation)
- Create account key
- Register account and accept TOS
(Certificate actions)
- Select HTTP-01 within offered challenges by the CA server
- Set up http challenge resource
- Set up standalone web server
- Create domain private key and CSR
- Issue certificate
- Renew certificate
- Revoke certificate
(Account update actions)
- Change contact information
- Deactivate Account
"""
from contextlib import contextmanager
from cryptography.hazmat.backends import default_backend
from cryptography.hazmat.primitives.asymmetric import rsa
import josepy as jose
import OpenSSL
from acme import challenges
from acme import client
from acme import crypto_util
from acme import errors
from acme import messages
from acme import standalone
# Constants:
# This is the staging point for ACME-V2 within Let's Encrypt.
DIRECTORY_URL = 'https://acme-staging-v02.api.letsencrypt.org/directory'
USER_AGENT = 'python-acme-example'
# Account key size
ACC_KEY_BITS = 2048
# Certificate private key size
CERT_PKEY_BITS = 2048
# Domain name for the certificate.
DOMAIN = 'client.example.com'
# If you are running Boulder locally, it is possible to configure any port
# number to execute the challenge, but real CA servers will always use port
# 80, as described in the ACME specification.
PORT = 80
# Useful methods and classes:
def new_csr_comp(domain_name, pkey_pem=None):
"""Create certificate signing request."""
if pkey_pem is None:
# Create private key.
pkey = OpenSSL.crypto.PKey()
pkey.generate_key(OpenSSL.crypto.TYPE_RSA, CERT_PKEY_BITS)
pkey_pem = OpenSSL.crypto.dump_privatekey(OpenSSL.crypto.FILETYPE_PEM,
pkey)
csr_pem = crypto_util.make_csr(pkey_pem, [domain_name])
return pkey_pem, csr_pem
def select_http01_chall(orderr):
"""Extract authorization resource from within order resource."""
# Authorization Resource: authz.
# This object holds the offered challenges by the server and their status.
authz_list = orderr.authorizations
for authz in authz_list:
# Choosing challenge.
# authz.body.challenges is a set of ChallengeBody objects.
for i in authz.body.challenges:
# Find the supported challenge.
if isinstance(i.chall, challenges.HTTP01):
return i
raise Exception('HTTP-01 challenge was not offered by the CA server.')
@contextmanager
def challenge_server(http_01_resources):
"""Manage standalone server set up and shutdown."""
# Setting up a fake server that binds at PORT and any address.
address = ('', PORT)
try:
servers = standalone.HTTP01DualNetworkedServers(address,
http_01_resources)
# Start client standalone web server.
servers.serve_forever()
yield servers
finally:
# Shutdown client web server and unbind from PORT
servers.shutdown_and_server_close()
def METHOD_NAME(client_acme, challb, orderr):
"""Set up standalone webserver and perform HTTP-01 challenge."""
response, validation = challb.response_and_validation(client_acme.net.key)
resource = standalone.HTTP01RequestHandler.HTTP01Resource(
chall=challb.chall, response=response, validation=validation)
with challenge_server({resource}):
# Let the CA server know that we are ready for the challenge.
client_acme.answer_challenge(challb, response)
# Wait for challenge status and then issue a certificate.
# It is possible to set a deadline time.
finalized_orderr = client_acme.poll_and_finalize(orderr)
return finalized_orderr.fullchain_pem
# Main examples:
def example_http():
"""This example executes the whole process of fulfilling a HTTP-01
challenge for one specific domain.
The workflow consists of:
(Account creation)
- Create account key
- Register account and accept TOS
(Certificate actions)
- Select HTTP-01 within offered challenges by the CA server
- Set up http challenge resource
- Set up standalone web server
- Create domain private key and CSR
- Issue certificate
- Renew certificate
- Revoke certificate
(Account update actions)
- Change contact information
- Deactivate Account
"""
# Create account key
acc_key = jose.JWKRSA(
key=rsa.generate_private_key(public_exponent=65537,
key_size=ACC_KEY_BITS,
backend=default_backend()))
# Register account and accept TOS
net = client.ClientNetwork(acc_key, user_agent=USER_AGENT)
directory = client.ClientV2.get_directory(DIRECTORY_URL, net)
client_acme = client.ClientV2(directory, net=net)
# Terms of Service URL is in client_acme.directory.meta.terms_of_service
# Registration Resource: regr
# Creates account with contact information.
email = ('[email protected]')
regr = client_acme.new_account(
messages.NewRegistration.from_data(
email=email, terms_of_service_agreed=True))
# Create domain private key and CSR
pkey_pem, csr_pem = new_csr_comp(DOMAIN)
# Issue certificate
orderr = client_acme.new_order(csr_pem)
# Select HTTP-01 within offered challenges by the CA server
challb = select_http01_chall(orderr)
# The certificate is ready to be used in the variable "fullchain_pem".
fullchain_pem = METHOD_NAME(client_acme, challb, orderr)
# Renew certificate
_, csr_pem = new_csr_comp(DOMAIN, pkey_pem)
orderr = client_acme.new_order(csr_pem)
challb = select_http01_chall(orderr)
# Performing challenge
fullchain_pem = METHOD_NAME(client_acme, challb, orderr)
# Revoke certificate
fullchain_com = jose.ComparableX509(
OpenSSL.crypto.load_certificate(
OpenSSL.crypto.FILETYPE_PEM, fullchain_pem))
try:
client_acme.revoke(fullchain_com, 0) # revocation reason = 0
except errors.ConflictError:
# Certificate already revoked.
pass
# Query registration status.
client_acme.net.account = regr
try:
regr = client_acme.query_registration(regr)
except errors.Error as err:
if err.typ == messages.ERROR_PREFIX + 'unauthorized':
# Status is deactivated.
pass
raise
# Change contact information
email = '[email protected]'
regr = client_acme.update_registration(
regr.update(
body=regr.body.update(
contact=('mailto:' + email,)
)
)
)
# Deactivate account/registration
regr = client_acme.deactivate_registration(regr)
if __name__ == "__main__":
example_http() | null |
check single value default metric | #!/usr/bin/env python
# -*- coding: utf-8 -*-
#
# Copyright 2019 The FATE Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from federatedml.util import consts, LOGGER
from federatedml.param.base_param import BaseParam
class EvaluateParam(BaseParam):
"""
Define the evaluation method of binary/multiple classification and regression
Parameters
----------
eval_type : {'binary', 'regression', 'multi'}
support 'binary' for HomoLR, HeteroLR and Secureboosting,
support 'regression' for Secureboosting,
'multi' is not support these version
unfold_multi_result : bool
unfold multi result and get several one-vs-rest binary classification results
pos_label : int or float or str
specify positive label type, depend on the data's label. this parameter effective only for 'binary'
need_run: bool, default True
Indicate if this module needed to be run
"""
def __init__(self, eval_type="binary", pos_label=1, need_run=True, metrics=None,
run_clustering_arbiter_metric=False, unfold_multi_result=False):
super().__init__()
self.eval_type = eval_type
self.pos_label = pos_label
self.need_run = need_run
self.metrics = metrics
self.unfold_multi_result = unfold_multi_result
self.run_clustering_arbiter_metric = run_clustering_arbiter_metric
self.default_metrics = {
consts.BINARY: consts.ALL_BINARY_METRICS,
consts.MULTY: consts.ALL_MULTI_METRICS,
consts.REGRESSION: consts.ALL_REGRESSION_METRICS,
consts.CLUSTERING: consts.ALL_CLUSTER_METRICS
}
self.allowed_metrics = {
consts.BINARY: consts.ALL_BINARY_METRICS,
consts.MULTY: consts.ALL_MULTI_METRICS,
consts.REGRESSION: consts.ALL_REGRESSION_METRICS,
consts.CLUSTERING: consts.ALL_CLUSTER_METRICS
}
def _use_single_value_default_metrics(self):
self.default_metrics = {
consts.BINARY: consts.DEFAULT_BINARY_METRIC,
consts.MULTY: consts.DEFAULT_MULTI_METRIC,
consts.REGRESSION: consts.DEFAULT_REGRESSION_METRIC,
consts.CLUSTERING: consts.DEFAULT_CLUSTER_METRIC
}
def _check_valid_metric(self, metrics_list):
metric_list = consts.ALL_METRIC_NAME
alias_name: dict = consts.ALIAS
full_name_list = []
metrics_list = [str.lower(i) for i in metrics_list]
for metric in metrics_list:
if metric in metric_list:
if metric not in full_name_list:
full_name_list.append(metric)
continue
valid_flag = False
for alias, full_name in alias_name.items():
if metric in alias:
if full_name not in full_name_list:
full_name_list.append(full_name)
valid_flag = True
break
if not valid_flag:
raise ValueError('metric {} is not supported'.format(metric))
allowed_metrics = self.allowed_metrics[self.eval_type]
for m in full_name_list:
if m not in allowed_metrics:
raise ValueError('metric {} is not used for {} task'.format(m, self.eval_type))
if consts.RECALL in full_name_list and consts.PRECISION not in full_name_list:
full_name_list.append(consts.PRECISION)
if consts.RECALL not in full_name_list and consts.PRECISION in full_name_list:
full_name_list.append(consts.RECALL)
return full_name_list
def check(self):
descr = "evaluate param's "
self.eval_type = self.check_and_change_lower(self.eval_type,
[consts.BINARY, consts.MULTY, consts.REGRESSION,
consts.CLUSTERING],
descr)
if type(self.pos_label).__name__ not in ["str", "float", "int"]:
raise ValueError(
"evaluate param's pos_label {} not supported, should be str or float or int type".format(
self.pos_label))
if type(self.need_run).__name__ != "bool":
raise ValueError(
"evaluate param's need_run {} not supported, should be bool".format(
self.need_run))
if self.metrics is None or len(self.metrics) == 0:
self.metrics = self.default_metrics[self.eval_type]
LOGGER.warning('use default metric {} for eval type {}'.format(self.metrics, self.eval_type))
self.check_boolean(self.unfold_multi_result, 'multi_result_unfold')
self.metrics = self._check_valid_metric(self.metrics)
return True
def METHOD_NAME(self):
self._use_single_value_default_metrics()
# in validation strategy, psi f1-score and confusion-mat pr-quantile are not supported in cur version
if self.metrics is None or len(self.metrics) == 0:
self.metrics = self.default_metrics[self.eval_type]
LOGGER.warning('use default metric {} for eval type {}'.format(self.metrics, self.eval_type))
ban_metric = [consts.PSI, consts.F1_SCORE, consts.CONFUSION_MAT, consts.QUANTILE_PR]
for metric in self.metrics:
if metric in ban_metric:
self.metrics.remove(metric)
self.check() | null |
test last | # Licensed under the GPLv3 - see LICENSE
import sys
from importlib import reload
from importlib.metadata import EntryPoint, entry_points
import pytest
from .. import io as bio, vdif, base
class TestExistingIOFormat:
def setup_method(self):
dir(bio) # Ensure entries are loaded.
self.vdif_entry = bio._entries['vdif']
def teardown_method(self):
bio._entries['vdif'] = self.vdif_entry
bio._bad_entries.discard('vdif')
dir(bio) # does update.
def test_io_entry_point(self):
assert hasattr(bio, 'vdif')
assert 'vdif' in bio._entries
assert 'vdif' in bio.FORMATS
del bio.vdif
assert 'vdif' not in bio.__dict__
assert 'vdif' in bio.FORMATS
assert 'vdif' in dir(bio)
assert hasattr(bio, 'vdif')
assert 'vdif' in bio.__dict__
assert 'vdif' in bio.FORMATS
def test_fake_bad_vdif(self):
assert bio.vdif is vdif
del bio.vdif
bio._entries['vdif'] = EntryPoint('vdif', 'bad.vdif', '')
with pytest.raises(AttributeError, match='not loadable'):
bio.vdif
assert 'vdif' not in dir(bio)
assert 'vdif' in bio._bad_entries
# Does not auto-reload since already known as bad.
with pytest.raises(AttributeError, match='has no attribute'):
bio.vdif
# But will reload if we reload and thus start over.
reload(bio)
assert bio.vdif is vdif
assert 'vdif' in bio.FORMATS
class TestNewIOFormats:
def setup_entry(self, entry):
self.added = entry.name
bio._entries[entry.name] = entry
bio.FORMATS.append(entry.name)
def teardown_method(self):
bio._entries.pop(self.added, None)
if self.added in bio.FORMATS:
bio.FORMATS.remove(self.added)
bio._bad_entries.discard(self.added)
bio.__dict__.pop(self.added, None)
def test_find_new(self):
self.setup_entry(EntryPoint('new', 'baseband.vdif', ''))
assert 'new' in dir(bio)
assert 'new' in bio.FORMATS
assert bio.new is vdif
# Check that it comes back if we remove it from the module.
bio.__dict__.pop('new', None)
assert 'new' not in bio.__dict__
assert 'new' in bio.FORMATS
assert 'new' in dir(bio)
assert bio.new is vdif
def test_cannot_getattr_bad(self):
self.setup_entry(EntryPoint('bad', 'really_bad', ''))
assert 'bad' in dir(bio)
assert 'bad' in bio.FORMATS
with pytest.raises(AttributeError, match='not loadable'):
bio.bad
assert 'bad' not in dir(bio)
assert 'bad' not in bio.FORMATS
with pytest.raises(AttributeError, match='has no attribute'):
bio.bad
def test_not_hasattr_bad(self):
self.setup_entry(EntryPoint('bad', 'really_bad', ''))
assert 'bad' in dir(bio)
assert not hasattr(bio, 'bad')
assert 'bad' not in dir(bio)
class TestTasks:
def setup_method(self):
"""Remove baseband.tasks so we can reload and find test entry point."""
# Also ensures that stuff we add gets removed.
sys.modules.pop('baseband.tasks', None)
def test_first(self):
import baseband.tasks as tasks
assert 'vdif_payload_module' not in dir(tasks)
def test_task_discovery(self, tmpdir, monkeypatch):
with open(tmpdir.mkdir('task_tests-0.1.dist-info')
.join('entry_points.txt'), 'wt') as fw:
fw.write('[baseband.tasks]\n'
'vdif_payload_module = baseband.vdif.payload\n'
'vdif_header_all = baseband.vdif.header:__all__\n'
'_nomod = baseband.base.utils:__all__\n')
monkeypatch.syspath_prepend(str(tmpdir))
# Now (re-)import tasks
import baseband.tasks as tasks
# We loaded just the vdif module.
assert 'vdif_payload_module' in dir(tasks)
assert 'vdif' not in dir(tasks)
assert 'payload' not in dir(tasks)
assert tasks.vdif_payload_module is vdif.payload
assert 'VDIFPayload' not in dir(tasks)
# But helpers and everything in it.
assert 'vdif_header_all' in dir(tasks)
assert 'header' not in dir(tasks)
assert 'VDIFHeader' in dir(tasks)
assert tasks.vdif_header_all is vdif.header
assert tasks.VDIFHeader is vdif.header.VDIFHeader
# And what's in utils, but not the name.
assert '_nomod' not in dir(tasks)
assert 'CRC' in dir(tasks)
assert tasks.CRC is base.utils.CRC
def test_bad_task_definition(self, tmpdir, monkeypatch):
with open(tmpdir.mkdir('bad_task_tests-0.1.dist-info')
.join('entry_points.txt'), 'wt') as fw:
fw.write('[baseband.tasks]\n'
'vdif_payload_module = baseband.vdif.payload\n'
'utils = baseband.base.utils.__all__\n' # typo: . not :
'does_not_exist = baseband.does_not_exist\n')
monkeypatch.syspath_prepend(str(tmpdir))
import baseband.tasks as tasks
assert tasks.vdif_payload_module is vdif.payload
assert not hasattr(tasks, 'utils')
assert 'utils' not in dir(tasks)
assert 'does_not_exist' not in dir(tasks)
assert len(tasks._bad_entries) == 2
assert (set(entry.name for entry in tasks._bad_entries)
== {'utils', 'does_not_exist'})
@pytest.mark.xfail(entry_points(group="baseband.tasks")
if sys.version_info >= (3, 10) else
entry_points().get("baseband.tasks", []),
reason='cannot test for lack of entry points')
def test_message_on_empty_tasks(self):
import baseband.tasks as tasks
with pytest.raises(AttributeError, match='No.*entry points found'):
tasks.does_not_exist
def METHOD_NAME(self):
import baseband.tasks as tasks
assert 'vdif_payload_module' not in dir(tasks) | null |
create module obj | # ----------------------------------------------------------------------
#
# Brad T. Aagaard, U.S. Geological Survey
# Charles A. Williams, GNS Science
# Matthew G. Knepley, University at Buffalo
#
# This code was developed as part of the Computational Infrastructure
# for Geodynamics (http://geodynamics.org).
#
# Copyright (c) 2010-2022 University of California, Davis
#
# See LICENSE.md for license information.
#
# ----------------------------------------------------------------------
#
# @file pylith/bc/NeumannTimeDependent.py
#
# @brief Python object for managing a time-dependent Neumann (natural) boundary condition.
#
# Factory: boundary_condition
from .BoundaryCondition import BoundaryCondition
from .bc import NeumannTimeDependent as ModuleNeumannTimeDependent
from pylith.utils.NullComponent import NullComponent
def validateDir(value):
"""Validate direction.
"""
msg = "Direction must be a 3 component vector (list)."
if not isinstance(value, list):
raise ValueError(msg)
if 3 != len(value):
raise ValueError(msg)
try:
nums = list(map(float, value))
except:
raise ValueError(msg)
return nums
class NeumannTimeDependent(BoundaryCondition, ModuleNeumannTimeDependent):
"""
Neumann time-dependent boundary condition. Implements `BoundaryCondition`.
This boundary condition applies a Neumann boundary condition for a single solution subfield on a boundary.
To apply Neumann boundary conditions for multiple solution subfields on a boundary, use multiple Neumann boundary conditions.
:::{important}
The components are specified in the local normal-tangential coordinate system for the boundary. Ambiguities in specifying the shear (tangential) tractions in 3D problems are resolved using the `ref_dir_1` and `ref_dir_2` properties.
The first tangential direction is $\\vec{z} \\times \\vec{r}_1$ unless these are colinear, then $\\vec{r}_2$ (`ref_dir_2`) is used.
The second tangential direction is $\\vec{n} \\times \\vec{t}_1$.
:::
:::{seealso}
See [`AuxSubfieldsTimeDependent` Component](AuxSubfieldsTimeDependent.md) for the functional form of the time depenence.
:::
"""
DOC_CONFIG = {
"cfg": """
# Neumann (traction) boundary condition in 2D on -y boundary.
[pylithapp.problem.bc.bc_yneg]
label = boundary_yneg
field = displacement
scale_name = pressure
use_initial = False
use_time_history = True
db_auxiliary_field = spatialdata.spatialdb.UniformDB
db_auxiliary_field.description = Displacement Neumann BC +y boundary
db_auxiliary_field.values = [time_history_amplitude_tangential, time_history_amplitude_normal, time_history_start_time]
db_auxiliary_field.data = [2.0*MPa, -1.0*MPa, 0.0]
time_history = spatialdata.spatialdb.TimeHistory
time_history.description = Impulse time history
time_history.filename = impulse.timedb
""",
}
import pythia.pyre.inventory
scaleName = pythia.pyre.inventory.str("scale_name", default="pressure",
validator=pythia.pyre.inventory.choice(["length", "time", "pressure", "density", "velocity"]))
scaleName.meta['tip'] = "Type of scale for nondimensionalizing Neumann boundary condition ('pressure' for elasticity)."
useInitial = pythia.pyre.inventory.bool("use_initial", default=True)
useInitial.meta['tip'] = "Use initial term in time-dependent expression."
useRate = pythia.pyre.inventory.bool("use_rate", default=False)
useRate.meta['tip'] = "Use rate term in time-dependent expression."
useTimeHistory = pythia.pyre.inventory.bool("use_time_history", default=False)
useTimeHistory.meta['tip'] = "Use time history term in time-dependent expression."
dbTimeHistory = pythia.pyre.inventory.facility("time_history", factory=NullComponent, family="temporal_database")
dbTimeHistory.meta['tip'] = "Time history with normalized amplitude as a function of time."
refDir1 = pythia.pyre.inventory.list("ref_dir_1", default=[0.0, 0.0, 1.0], validator=validateDir)
refDir1.meta['tip'] = "First choice for reference direction to discriminate among tangential directions in 3D."
refDir2 = pythia.pyre.inventory.list("ref_dir_2", default=[0.0, 1.0, 0.0], validator=validateDir)
refDir2.meta['tip'] = "Second choice for reference direction to discriminate among tangential directions in 3D."
def __init__(self, name="neumanntimedependent"):
"""Constructor.
"""
BoundaryCondition.__init__(self, name)
return
def _defaults(self):
from .AuxSubfieldsTimeDependent import AuxSubfieldsTimeDependent
self.auxiliarySubfields = AuxSubfieldsTimeDependent("auxiliary_subfields")
def preinitialize(self, problem):
"""Do pre-initialization setup.
"""
from pylith.mpi.Communicator import mpi_is_root
if mpi_is_root():
self._info.log(
"Performing minimal initialization of time-dependent Neumann boundary condition '%s'." % self.aliases[-1])
BoundaryCondition.preinitialize(self, problem)
ModuleNeumannTimeDependent.setScaleName(self, self.scaleName)
ModuleNeumannTimeDependent.setRefDir1(self, self.refDir1)
ModuleNeumannTimeDependent.setRefDir2(self, self.refDir2)
ModuleNeumannTimeDependent.useInitial(self, self.useInitial)
ModuleNeumannTimeDependent.useRate(self, self.useRate)
ModuleNeumannTimeDependent.useTimeHistory(self, self.useTimeHistory)
if not isinstance(self.dbTimeHistory, NullComponent):
ModuleNeumannTimeDependent.setTimeHistoryDB(self, self.dbTimeHistory)
return
def _validate(self, context):
if self.inventory.useTimeHistory and isinstance(self.inventory.dbTimeHistory, NullComponent):
trait = self.inventory.getTrait("time_history")
self._validationError(context, trait,
f"Missing time history database for time-dependent Neumann boundary condition '{self.aliases[-1]}'.")
if not self.inventory.useTimeHistory and not isinstance(self.inventory.dbTimeHistory, NullComponent):
self._warning.log(
f"Time history for time-dependent Neumann boundary condition '{self.aliases[-1]}' not enabled. Ignoring provided time history database.")
def _validationError(self, context, trait, msg):
from pythia.pyre.inventory.Item import Item
error = ValueError(msg)
descriptor = self.getTraitDescriptor(trait.name)
context.error(error, items=[Item(trait, descriptor)])
def _configure(self):
"""Setup members using inventory.
"""
BoundaryCondition._configure(self)
return
def METHOD_NAME(self):
"""Create handle to corresponding C++ object.
"""
ModuleNeumannTimeDependent.__init__(self)
return
# Factories
def boundary_condition():
"""Factory associated with NeumannTimeDependent.
"""
return NeumannTimeDependent()
# End of file | null |