repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 75
19.8k
| code_tokens
sequence | docstring
stringlengths 3
17.3k
| docstring_tokens
sequence | sha
stringlengths 40
40
| url
stringlengths 87
242
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
daskos/mentor | mentor/proxies/scheduler.py | SchedulerDriverProxy.acknowledge | def acknowledge(self, status):
"""Acknowledges the status update.
This should only be called once the status update is processed durably
by the scheduler.
Not that explicit acknowledgements must be requested via the constructor
argument, otherwise a call to this method will cause the driver to
crash.
"""
logging.info('Acknowledges status update {}'.format(status))
return self.driver.acknowledgeStatusUpdate(encode(status)) | python | def acknowledge(self, status):
"""Acknowledges the status update.
This should only be called once the status update is processed durably
by the scheduler.
Not that explicit acknowledgements must be requested via the constructor
argument, otherwise a call to this method will cause the driver to
crash.
"""
logging.info('Acknowledges status update {}'.format(status))
return self.driver.acknowledgeStatusUpdate(encode(status)) | [
"def",
"acknowledge",
"(",
"self",
",",
"status",
")",
":",
"logging",
".",
"info",
"(",
"'Acknowledges status update {}'",
".",
"format",
"(",
"status",
")",
")",
"return",
"self",
".",
"driver",
".",
"acknowledgeStatusUpdate",
"(",
"encode",
"(",
"status",
")",
")"
] | Acknowledges the status update.
This should only be called once the status update is processed durably
by the scheduler.
Not that explicit acknowledgements must be requested via the constructor
argument, otherwise a call to this method will cause the driver to
crash. | [
"Acknowledges",
"the",
"status",
"update",
"."
] | b5fd64e3a3192f5664fa5c03e8517cacb4e0590f | https://github.com/daskos/mentor/blob/b5fd64e3a3192f5664fa5c03e8517cacb4e0590f/mentor/proxies/scheduler.py#L221-L232 | train |
daskos/mentor | mentor/proxies/scheduler.py | SchedulerDriverProxy.message | def message(self, executor_id, slave_id, message):
"""Sends a message from the framework to one of its executors.
These messages are best effort; do not expect a framework message to be
retransmitted in any reliable fashion.
"""
logging.info('Sends message `{}` to executor `{}` on slave `{}`'.format(
message, executor_id, slave_id))
return self.driver.sendFrameworkMessage(encode(executor_id),
encode(slave_id),
message) | python | def message(self, executor_id, slave_id, message):
"""Sends a message from the framework to one of its executors.
These messages are best effort; do not expect a framework message to be
retransmitted in any reliable fashion.
"""
logging.info('Sends message `{}` to executor `{}` on slave `{}`'.format(
message, executor_id, slave_id))
return self.driver.sendFrameworkMessage(encode(executor_id),
encode(slave_id),
message) | [
"def",
"message",
"(",
"self",
",",
"executor_id",
",",
"slave_id",
",",
"message",
")",
":",
"logging",
".",
"info",
"(",
"'Sends message `{}` to executor `{}` on slave `{}`'",
".",
"format",
"(",
"message",
",",
"executor_id",
",",
"slave_id",
")",
")",
"return",
"self",
".",
"driver",
".",
"sendFrameworkMessage",
"(",
"encode",
"(",
"executor_id",
")",
",",
"encode",
"(",
"slave_id",
")",
",",
"message",
")"
] | Sends a message from the framework to one of its executors.
These messages are best effort; do not expect a framework message to be
retransmitted in any reliable fashion. | [
"Sends",
"a",
"message",
"from",
"the",
"framework",
"to",
"one",
"of",
"its",
"executors",
"."
] | b5fd64e3a3192f5664fa5c03e8517cacb4e0590f | https://github.com/daskos/mentor/blob/b5fd64e3a3192f5664fa5c03e8517cacb4e0590f/mentor/proxies/scheduler.py#L234-L244 | train |
virtuald/pygi-composite-templates | gi_composites.py | _connect_func | def _connect_func(builder, obj, signal_name, handler_name,
connect_object, flags, cls):
'''Handles GtkBuilder signal connect events'''
if connect_object is None:
extra = ()
else:
extra = (connect_object,)
# The handler name refers to an attribute on the template instance,
# so ask GtkBuilder for the template instance
template_inst = builder.get_object(cls.__gtype_name__)
if template_inst is None: # This should never happen
errmsg = "Internal error: cannot find template instance! obj: %s; " \
"signal: %s; handler: %s; connect_obj: %s; class: %s" % \
(obj, signal_name, handler_name, connect_object, cls)
warnings.warn(errmsg, GtkTemplateWarning)
return
handler = getattr(template_inst, handler_name)
if flags == GObject.ConnectFlags.AFTER:
obj.connect_after(signal_name, handler, *extra)
else:
obj.connect(signal_name, handler, *extra)
template_inst.__connected_template_signals__.add(handler_name) | python | def _connect_func(builder, obj, signal_name, handler_name,
connect_object, flags, cls):
'''Handles GtkBuilder signal connect events'''
if connect_object is None:
extra = ()
else:
extra = (connect_object,)
# The handler name refers to an attribute on the template instance,
# so ask GtkBuilder for the template instance
template_inst = builder.get_object(cls.__gtype_name__)
if template_inst is None: # This should never happen
errmsg = "Internal error: cannot find template instance! obj: %s; " \
"signal: %s; handler: %s; connect_obj: %s; class: %s" % \
(obj, signal_name, handler_name, connect_object, cls)
warnings.warn(errmsg, GtkTemplateWarning)
return
handler = getattr(template_inst, handler_name)
if flags == GObject.ConnectFlags.AFTER:
obj.connect_after(signal_name, handler, *extra)
else:
obj.connect(signal_name, handler, *extra)
template_inst.__connected_template_signals__.add(handler_name) | [
"def",
"_connect_func",
"(",
"builder",
",",
"obj",
",",
"signal_name",
",",
"handler_name",
",",
"connect_object",
",",
"flags",
",",
"cls",
")",
":",
"if",
"connect_object",
"is",
"None",
":",
"extra",
"=",
"(",
")",
"else",
":",
"extra",
"=",
"(",
"connect_object",
",",
")",
"# The handler name refers to an attribute on the template instance,",
"# so ask GtkBuilder for the template instance",
"template_inst",
"=",
"builder",
".",
"get_object",
"(",
"cls",
".",
"__gtype_name__",
")",
"if",
"template_inst",
"is",
"None",
":",
"# This should never happen",
"errmsg",
"=",
"\"Internal error: cannot find template instance! obj: %s; \"",
"\"signal: %s; handler: %s; connect_obj: %s; class: %s\"",
"%",
"(",
"obj",
",",
"signal_name",
",",
"handler_name",
",",
"connect_object",
",",
"cls",
")",
"warnings",
".",
"warn",
"(",
"errmsg",
",",
"GtkTemplateWarning",
")",
"return",
"handler",
"=",
"getattr",
"(",
"template_inst",
",",
"handler_name",
")",
"if",
"flags",
"==",
"GObject",
".",
"ConnectFlags",
".",
"AFTER",
":",
"obj",
".",
"connect_after",
"(",
"signal_name",
",",
"handler",
",",
"*",
"extra",
")",
"else",
":",
"obj",
".",
"connect",
"(",
"signal_name",
",",
"handler",
",",
"*",
"extra",
")",
"template_inst",
".",
"__connected_template_signals__",
".",
"add",
"(",
"handler_name",
")"
] | Handles GtkBuilder signal connect events | [
"Handles",
"GtkBuilder",
"signal",
"connect",
"events"
] | a22be54ea95b8125b36deaa3ce7171e84158d486 | https://github.com/virtuald/pygi-composite-templates/blob/a22be54ea95b8125b36deaa3ce7171e84158d486/gi_composites.py#L36-L63 | train |
virtuald/pygi-composite-templates | gi_composites.py | _register_template | def _register_template(cls, template_bytes):
'''Registers the template for the widget and hooks init_template'''
# This implementation won't work if there are nested templates, but
# we can't do that anyways due to PyGObject limitations so it's ok
if not hasattr(cls, 'set_template'):
raise TypeError("Requires PyGObject 3.13.2 or greater")
cls.set_template(template_bytes)
bound_methods = set()
bound_widgets = set()
# Walk the class, find marked callbacks and child attributes
for name in dir(cls):
o = getattr(cls, name, None)
if inspect.ismethod(o):
if hasattr(o, '_gtk_callback'):
bound_methods.add(name)
# Don't need to call this, as connect_func always gets called
#cls.bind_template_callback_full(name, o)
elif isinstance(o, _Child):
cls.bind_template_child_full(name, True, 0)
bound_widgets.add(name)
# Have to setup a special connect function to connect at template init
# because the methods are not bound yet
cls.set_connect_func(_connect_func, cls)
cls.__gtemplate_methods__ = bound_methods
cls.__gtemplate_widgets__ = bound_widgets
base_init_template = cls.init_template
cls.init_template = lambda s: _init_template(s, cls, base_init_template) | python | def _register_template(cls, template_bytes):
'''Registers the template for the widget and hooks init_template'''
# This implementation won't work if there are nested templates, but
# we can't do that anyways due to PyGObject limitations so it's ok
if not hasattr(cls, 'set_template'):
raise TypeError("Requires PyGObject 3.13.2 or greater")
cls.set_template(template_bytes)
bound_methods = set()
bound_widgets = set()
# Walk the class, find marked callbacks and child attributes
for name in dir(cls):
o = getattr(cls, name, None)
if inspect.ismethod(o):
if hasattr(o, '_gtk_callback'):
bound_methods.add(name)
# Don't need to call this, as connect_func always gets called
#cls.bind_template_callback_full(name, o)
elif isinstance(o, _Child):
cls.bind_template_child_full(name, True, 0)
bound_widgets.add(name)
# Have to setup a special connect function to connect at template init
# because the methods are not bound yet
cls.set_connect_func(_connect_func, cls)
cls.__gtemplate_methods__ = bound_methods
cls.__gtemplate_widgets__ = bound_widgets
base_init_template = cls.init_template
cls.init_template = lambda s: _init_template(s, cls, base_init_template) | [
"def",
"_register_template",
"(",
"cls",
",",
"template_bytes",
")",
":",
"# This implementation won't work if there are nested templates, but",
"# we can't do that anyways due to PyGObject limitations so it's ok",
"if",
"not",
"hasattr",
"(",
"cls",
",",
"'set_template'",
")",
":",
"raise",
"TypeError",
"(",
"\"Requires PyGObject 3.13.2 or greater\"",
")",
"cls",
".",
"set_template",
"(",
"template_bytes",
")",
"bound_methods",
"=",
"set",
"(",
")",
"bound_widgets",
"=",
"set",
"(",
")",
"# Walk the class, find marked callbacks and child attributes",
"for",
"name",
"in",
"dir",
"(",
"cls",
")",
":",
"o",
"=",
"getattr",
"(",
"cls",
",",
"name",
",",
"None",
")",
"if",
"inspect",
".",
"ismethod",
"(",
"o",
")",
":",
"if",
"hasattr",
"(",
"o",
",",
"'_gtk_callback'",
")",
":",
"bound_methods",
".",
"add",
"(",
"name",
")",
"# Don't need to call this, as connect_func always gets called",
"#cls.bind_template_callback_full(name, o)",
"elif",
"isinstance",
"(",
"o",
",",
"_Child",
")",
":",
"cls",
".",
"bind_template_child_full",
"(",
"name",
",",
"True",
",",
"0",
")",
"bound_widgets",
".",
"add",
"(",
"name",
")",
"# Have to setup a special connect function to connect at template init",
"# because the methods are not bound yet",
"cls",
".",
"set_connect_func",
"(",
"_connect_func",
",",
"cls",
")",
"cls",
".",
"__gtemplate_methods__",
"=",
"bound_methods",
"cls",
".",
"__gtemplate_widgets__",
"=",
"bound_widgets",
"base_init_template",
"=",
"cls",
".",
"init_template",
"cls",
".",
"init_template",
"=",
"lambda",
"s",
":",
"_init_template",
"(",
"s",
",",
"cls",
",",
"base_init_template",
")"
] | Registers the template for the widget and hooks init_template | [
"Registers",
"the",
"template",
"for",
"the",
"widget",
"and",
"hooks",
"init_template"
] | a22be54ea95b8125b36deaa3ce7171e84158d486 | https://github.com/virtuald/pygi-composite-templates/blob/a22be54ea95b8125b36deaa3ce7171e84158d486/gi_composites.py#L66-L101 | train |
virtuald/pygi-composite-templates | gi_composites.py | _init_template | def _init_template(self, cls, base_init_template):
'''This would be better as an override for Gtk.Widget'''
# TODO: could disallow using a metaclass.. but this is good enough
# .. if you disagree, feel free to fix it and issue a PR :)
if self.__class__ is not cls:
raise TypeError("Inheritance from classes with @GtkTemplate decorators "
"is not allowed at this time")
connected_signals = set()
self.__connected_template_signals__ = connected_signals
base_init_template(self)
for name in self.__gtemplate_widgets__:
widget = self.get_template_child(cls, name)
self.__dict__[name] = widget
if widget is None:
# Bug: if you bind a template child, and one of them was
# not present, then the whole template is broken (and
# it's not currently possible for us to know which
# one is broken either -- but the stderr should show
# something useful with a Gtk-CRITICAL message)
raise AttributeError("A missing child widget was set using "
"GtkTemplate.Child and the entire "
"template is now broken (widgets: %s)" %
', '.join(self.__gtemplate_widgets__))
for name in self.__gtemplate_methods__.difference(connected_signals):
errmsg = ("Signal '%s' was declared with @GtkTemplate.Callback " +
"but was not present in template") % name
warnings.warn(errmsg, GtkTemplateWarning) | python | def _init_template(self, cls, base_init_template):
'''This would be better as an override for Gtk.Widget'''
# TODO: could disallow using a metaclass.. but this is good enough
# .. if you disagree, feel free to fix it and issue a PR :)
if self.__class__ is not cls:
raise TypeError("Inheritance from classes with @GtkTemplate decorators "
"is not allowed at this time")
connected_signals = set()
self.__connected_template_signals__ = connected_signals
base_init_template(self)
for name in self.__gtemplate_widgets__:
widget = self.get_template_child(cls, name)
self.__dict__[name] = widget
if widget is None:
# Bug: if you bind a template child, and one of them was
# not present, then the whole template is broken (and
# it's not currently possible for us to know which
# one is broken either -- but the stderr should show
# something useful with a Gtk-CRITICAL message)
raise AttributeError("A missing child widget was set using "
"GtkTemplate.Child and the entire "
"template is now broken (widgets: %s)" %
', '.join(self.__gtemplate_widgets__))
for name in self.__gtemplate_methods__.difference(connected_signals):
errmsg = ("Signal '%s' was declared with @GtkTemplate.Callback " +
"but was not present in template") % name
warnings.warn(errmsg, GtkTemplateWarning) | [
"def",
"_init_template",
"(",
"self",
",",
"cls",
",",
"base_init_template",
")",
":",
"# TODO: could disallow using a metaclass.. but this is good enough",
"# .. if you disagree, feel free to fix it and issue a PR :)",
"if",
"self",
".",
"__class__",
"is",
"not",
"cls",
":",
"raise",
"TypeError",
"(",
"\"Inheritance from classes with @GtkTemplate decorators \"",
"\"is not allowed at this time\"",
")",
"connected_signals",
"=",
"set",
"(",
")",
"self",
".",
"__connected_template_signals__",
"=",
"connected_signals",
"base_init_template",
"(",
"self",
")",
"for",
"name",
"in",
"self",
".",
"__gtemplate_widgets__",
":",
"widget",
"=",
"self",
".",
"get_template_child",
"(",
"cls",
",",
"name",
")",
"self",
".",
"__dict__",
"[",
"name",
"]",
"=",
"widget",
"if",
"widget",
"is",
"None",
":",
"# Bug: if you bind a template child, and one of them was",
"# not present, then the whole template is broken (and",
"# it's not currently possible for us to know which",
"# one is broken either -- but the stderr should show",
"# something useful with a Gtk-CRITICAL message)",
"raise",
"AttributeError",
"(",
"\"A missing child widget was set using \"",
"\"GtkTemplate.Child and the entire \"",
"\"template is now broken (widgets: %s)\"",
"%",
"', '",
".",
"join",
"(",
"self",
".",
"__gtemplate_widgets__",
")",
")",
"for",
"name",
"in",
"self",
".",
"__gtemplate_methods__",
".",
"difference",
"(",
"connected_signals",
")",
":",
"errmsg",
"=",
"(",
"\"Signal '%s' was declared with @GtkTemplate.Callback \"",
"+",
"\"but was not present in template\"",
")",
"%",
"name",
"warnings",
".",
"warn",
"(",
"errmsg",
",",
"GtkTemplateWarning",
")"
] | This would be better as an override for Gtk.Widget | [
"This",
"would",
"be",
"better",
"as",
"an",
"override",
"for",
"Gtk",
".",
"Widget"
] | a22be54ea95b8125b36deaa3ce7171e84158d486 | https://github.com/virtuald/pygi-composite-templates/blob/a22be54ea95b8125b36deaa3ce7171e84158d486/gi_composites.py#L104-L136 | train |
mikeboers/PyHAML | haml/util.py | extract_haml | def extract_haml(fileobj, keywords, comment_tags, options):
""" babel translation token extract function for haml files """
import haml
from mako import lexer, parsetree
from mako.ext.babelplugin import extract_nodes
encoding = options.get('input_encoding', options.get('encoding', None))
template_node = lexer.Lexer(haml.preprocessor(fileobj.read()), input_encoding=encoding).parse()
for extracted in extract_nodes(template_node.get_children(), keywords, comment_tags, options):
yield extracted | python | def extract_haml(fileobj, keywords, comment_tags, options):
""" babel translation token extract function for haml files """
import haml
from mako import lexer, parsetree
from mako.ext.babelplugin import extract_nodes
encoding = options.get('input_encoding', options.get('encoding', None))
template_node = lexer.Lexer(haml.preprocessor(fileobj.read()), input_encoding=encoding).parse()
for extracted in extract_nodes(template_node.get_children(), keywords, comment_tags, options):
yield extracted | [
"def",
"extract_haml",
"(",
"fileobj",
",",
"keywords",
",",
"comment_tags",
",",
"options",
")",
":",
"import",
"haml",
"from",
"mako",
"import",
"lexer",
",",
"parsetree",
"from",
"mako",
".",
"ext",
".",
"babelplugin",
"import",
"extract_nodes",
"encoding",
"=",
"options",
".",
"get",
"(",
"'input_encoding'",
",",
"options",
".",
"get",
"(",
"'encoding'",
",",
"None",
")",
")",
"template_node",
"=",
"lexer",
".",
"Lexer",
"(",
"haml",
".",
"preprocessor",
"(",
"fileobj",
".",
"read",
"(",
")",
")",
",",
"input_encoding",
"=",
"encoding",
")",
".",
"parse",
"(",
")",
"for",
"extracted",
"in",
"extract_nodes",
"(",
"template_node",
".",
"get_children",
"(",
")",
",",
"keywords",
",",
"comment_tags",
",",
"options",
")",
":",
"yield",
"extracted"
] | babel translation token extract function for haml files | [
"babel",
"translation",
"token",
"extract",
"function",
"for",
"haml",
"files"
] | 9ecb7c85349948428474869aad5b8d1c7de8dbed | https://github.com/mikeboers/PyHAML/blob/9ecb7c85349948428474869aad5b8d1c7de8dbed/haml/util.py#L3-L13 | train |
openvax/isovar | isovar/read_helpers.py | get_single_allele_from_reads | def get_single_allele_from_reads(allele_reads):
"""
Given a sequence of AlleleRead objects, which are expected to all have
the same allele, return that allele.
"""
allele_reads = list(allele_reads)
if len(allele_reads) == 0:
raise ValueError("Expected non-empty list of AlleleRead objects")
seq = allele_reads[0].allele
if any(read.allele != seq for read in allele_reads):
raise ValueError("Expected all AlleleRead objects to have same allele '%s', got %s" % (
seq, allele_reads))
return seq | python | def get_single_allele_from_reads(allele_reads):
"""
Given a sequence of AlleleRead objects, which are expected to all have
the same allele, return that allele.
"""
allele_reads = list(allele_reads)
if len(allele_reads) == 0:
raise ValueError("Expected non-empty list of AlleleRead objects")
seq = allele_reads[0].allele
if any(read.allele != seq for read in allele_reads):
raise ValueError("Expected all AlleleRead objects to have same allele '%s', got %s" % (
seq, allele_reads))
return seq | [
"def",
"get_single_allele_from_reads",
"(",
"allele_reads",
")",
":",
"allele_reads",
"=",
"list",
"(",
"allele_reads",
")",
"if",
"len",
"(",
"allele_reads",
")",
"==",
"0",
":",
"raise",
"ValueError",
"(",
"\"Expected non-empty list of AlleleRead objects\"",
")",
"seq",
"=",
"allele_reads",
"[",
"0",
"]",
".",
"allele",
"if",
"any",
"(",
"read",
".",
"allele",
"!=",
"seq",
"for",
"read",
"in",
"allele_reads",
")",
":",
"raise",
"ValueError",
"(",
"\"Expected all AlleleRead objects to have same allele '%s', got %s\"",
"%",
"(",
"seq",
",",
"allele_reads",
")",
")",
"return",
"seq"
] | Given a sequence of AlleleRead objects, which are expected to all have
the same allele, return that allele. | [
"Given",
"a",
"sequence",
"of",
"AlleleRead",
"objects",
"which",
"are",
"expected",
"to",
"all",
"have",
"the",
"same",
"allele",
"return",
"that",
"allele",
"."
] | b39b684920e3f6b344851d6598a1a1c67bce913b | https://github.com/openvax/isovar/blob/b39b684920e3f6b344851d6598a1a1c67bce913b/isovar/read_helpers.py#L25-L39 | train |
mikeboers/PyHAML | haml/nodes.py | Base.iter_all_children | def iter_all_children(self):
'''Return an iterator that yields every node which is a child of this one.
This includes inline children, and control structure `else` clauses.
'''
if self.inline_child:
yield self.inline_child
for x in self.children:
yield x | python | def iter_all_children(self):
'''Return an iterator that yields every node which is a child of this one.
This includes inline children, and control structure `else` clauses.
'''
if self.inline_child:
yield self.inline_child
for x in self.children:
yield x | [
"def",
"iter_all_children",
"(",
"self",
")",
":",
"if",
"self",
".",
"inline_child",
":",
"yield",
"self",
".",
"inline_child",
"for",
"x",
"in",
"self",
".",
"children",
":",
"yield",
"x"
] | Return an iterator that yields every node which is a child of this one.
This includes inline children, and control structure `else` clauses. | [
"Return",
"an",
"iterator",
"that",
"yields",
"every",
"node",
"which",
"is",
"a",
"child",
"of",
"this",
"one",
"."
] | 9ecb7c85349948428474869aad5b8d1c7de8dbed | https://github.com/mikeboers/PyHAML/blob/9ecb7c85349948428474869aad5b8d1c7de8dbed/haml/nodes.py#L18-L27 | train |
pyviz/imagen | imagen/transferfn/__init__.py | TransferFn.initialize | def initialize(self, **kwargs):
"""
Transfer functions may need additional information before the
supplied numpy array can be modified in place. For instance,
transfer functions may have state which needs to be allocated
in memory with a certain size. In other cases, the transfer
function may need to know about the coordinate system
associated with the input data.
"""
if not set(kwargs.keys()).issuperset(self.init_keys):
raise Exception("TransferFn needs to be initialized with %s"
% ','.join(repr(el) for el in self.init_keys)) | python | def initialize(self, **kwargs):
"""
Transfer functions may need additional information before the
supplied numpy array can be modified in place. For instance,
transfer functions may have state which needs to be allocated
in memory with a certain size. In other cases, the transfer
function may need to know about the coordinate system
associated with the input data.
"""
if not set(kwargs.keys()).issuperset(self.init_keys):
raise Exception("TransferFn needs to be initialized with %s"
% ','.join(repr(el) for el in self.init_keys)) | [
"def",
"initialize",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"set",
"(",
"kwargs",
".",
"keys",
"(",
")",
")",
".",
"issuperset",
"(",
"self",
".",
"init_keys",
")",
":",
"raise",
"Exception",
"(",
"\"TransferFn needs to be initialized with %s\"",
"%",
"','",
".",
"join",
"(",
"repr",
"(",
"el",
")",
"for",
"el",
"in",
"self",
".",
"init_keys",
")",
")"
] | Transfer functions may need additional information before the
supplied numpy array can be modified in place. For instance,
transfer functions may have state which needs to be allocated
in memory with a certain size. In other cases, the transfer
function may need to know about the coordinate system
associated with the input data. | [
"Transfer",
"functions",
"may",
"need",
"additional",
"information",
"before",
"the",
"supplied",
"numpy",
"array",
"can",
"be",
"modified",
"in",
"place",
".",
"For",
"instance",
"transfer",
"functions",
"may",
"have",
"state",
"which",
"needs",
"to",
"be",
"allocated",
"in",
"memory",
"with",
"a",
"certain",
"size",
".",
"In",
"other",
"cases",
"the",
"transfer",
"function",
"may",
"need",
"to",
"know",
"about",
"the",
"coordinate",
"system",
"associated",
"with",
"the",
"input",
"data",
"."
] | 53c5685c880f54b42795964d8db50b02e8590e88 | https://github.com/pyviz/imagen/blob/53c5685c880f54b42795964d8db50b02e8590e88/imagen/transferfn/__init__.py#L31-L42 | train |
pyviz/imagen | imagen/transferfn/__init__.py | TransferFnWithState.override_plasticity_state | def override_plasticity_state(self, new_plasticity_state):
"""
Temporarily disable plasticity of internal state.
This function should be implemented by all subclasses so that
after a call, the output should always be the same for any
given input pattern, and no call should have any effect that
persists after restore_plasticity_state() is called.
By default, simply saves a copy of the 'plastic' parameter to
an internal stack (so that it can be restored by
restore_plasticity_state()), and then sets the plastic
parameter to the given value (True or False).
"""
self._plasticity_setting_stack.append(self.plastic)
self.plastic=new_plasticity_state | python | def override_plasticity_state(self, new_plasticity_state):
"""
Temporarily disable plasticity of internal state.
This function should be implemented by all subclasses so that
after a call, the output should always be the same for any
given input pattern, and no call should have any effect that
persists after restore_plasticity_state() is called.
By default, simply saves a copy of the 'plastic' parameter to
an internal stack (so that it can be restored by
restore_plasticity_state()), and then sets the plastic
parameter to the given value (True or False).
"""
self._plasticity_setting_stack.append(self.plastic)
self.plastic=new_plasticity_state | [
"def",
"override_plasticity_state",
"(",
"self",
",",
"new_plasticity_state",
")",
":",
"self",
".",
"_plasticity_setting_stack",
".",
"append",
"(",
"self",
".",
"plastic",
")",
"self",
".",
"plastic",
"=",
"new_plasticity_state"
] | Temporarily disable plasticity of internal state.
This function should be implemented by all subclasses so that
after a call, the output should always be the same for any
given input pattern, and no call should have any effect that
persists after restore_plasticity_state() is called.
By default, simply saves a copy of the 'plastic' parameter to
an internal stack (so that it can be restored by
restore_plasticity_state()), and then sets the plastic
parameter to the given value (True or False). | [
"Temporarily",
"disable",
"plasticity",
"of",
"internal",
"state",
"."
] | 53c5685c880f54b42795964d8db50b02e8590e88 | https://github.com/pyviz/imagen/blob/53c5685c880f54b42795964d8db50b02e8590e88/imagen/transferfn/__init__.py#L70-L85 | train |
pyblish/pyblish-houdini | pyblish_houdini/lib.py | register_host | def register_host():
"""Register supported hosts"""
pyblish.api.register_host("hython")
pyblish.api.register_host("hpython")
pyblish.api.register_host("houdini") | python | def register_host():
"""Register supported hosts"""
pyblish.api.register_host("hython")
pyblish.api.register_host("hpython")
pyblish.api.register_host("houdini") | [
"def",
"register_host",
"(",
")",
":",
"pyblish",
".",
"api",
".",
"register_host",
"(",
"\"hython\"",
")",
"pyblish",
".",
"api",
".",
"register_host",
"(",
"\"hpython\"",
")",
"pyblish",
".",
"api",
".",
"register_host",
"(",
"\"houdini\"",
")"
] | Register supported hosts | [
"Register",
"supported",
"hosts"
] | 661b08696f04b4c5d8b03aa0c75cba3ca72f1e8d | https://github.com/pyblish/pyblish-houdini/blob/661b08696f04b4c5d8b03aa0c75cba3ca72f1e8d/pyblish_houdini/lib.py#L80-L84 | train |
pyblish/pyblish-houdini | pyblish_houdini/lib.py | maintained_selection | def maintained_selection():
"""Maintain selection during context
Example:
>>> with maintained_selection():
... # Modify selection
... node.setSelected(on=False, clear_all_selected=True)
>>> # Selection restored
"""
previous_selection = hou.selectedNodes()
try:
yield
finally:
if previous_selection:
for node in previous_selection:
node.setSelected(on=True)
else:
for node in previous_selection:
node.setSelected(on=False) | python | def maintained_selection():
"""Maintain selection during context
Example:
>>> with maintained_selection():
... # Modify selection
... node.setSelected(on=False, clear_all_selected=True)
>>> # Selection restored
"""
previous_selection = hou.selectedNodes()
try:
yield
finally:
if previous_selection:
for node in previous_selection:
node.setSelected(on=True)
else:
for node in previous_selection:
node.setSelected(on=False) | [
"def",
"maintained_selection",
"(",
")",
":",
"previous_selection",
"=",
"hou",
".",
"selectedNodes",
"(",
")",
"try",
":",
"yield",
"finally",
":",
"if",
"previous_selection",
":",
"for",
"node",
"in",
"previous_selection",
":",
"node",
".",
"setSelected",
"(",
"on",
"=",
"True",
")",
"else",
":",
"for",
"node",
"in",
"previous_selection",
":",
"node",
".",
"setSelected",
"(",
"on",
"=",
"False",
")"
] | Maintain selection during context
Example:
>>> with maintained_selection():
... # Modify selection
... node.setSelected(on=False, clear_all_selected=True)
>>> # Selection restored | [
"Maintain",
"selection",
"during",
"context"
] | 661b08696f04b4c5d8b03aa0c75cba3ca72f1e8d | https://github.com/pyblish/pyblish-houdini/blob/661b08696f04b4c5d8b03aa0c75cba3ca72f1e8d/pyblish_houdini/lib.py#L104-L124 | train |
portfoliome/postpy | postpy/sql.py | execute_transaction | def execute_transaction(conn, statements: Iterable):
"""Execute several statements in single DB transaction."""
with conn:
with conn.cursor() as cursor:
for statement in statements:
cursor.execute(statement)
conn.commit() | python | def execute_transaction(conn, statements: Iterable):
"""Execute several statements in single DB transaction."""
with conn:
with conn.cursor() as cursor:
for statement in statements:
cursor.execute(statement)
conn.commit() | [
"def",
"execute_transaction",
"(",
"conn",
",",
"statements",
":",
"Iterable",
")",
":",
"with",
"conn",
":",
"with",
"conn",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"for",
"statement",
"in",
"statements",
":",
"cursor",
".",
"execute",
"(",
"statement",
")",
"conn",
".",
"commit",
"(",
")"
] | Execute several statements in single DB transaction. | [
"Execute",
"several",
"statements",
"in",
"single",
"DB",
"transaction",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L10-L17 | train |
portfoliome/postpy | postpy/sql.py | execute_transactions | def execute_transactions(conn, statements: Iterable):
"""Execute several statements each as a single DB transaction."""
with conn.cursor() as cursor:
for statement in statements:
try:
cursor.execute(statement)
conn.commit()
except psycopg2.ProgrammingError:
conn.rollback() | python | def execute_transactions(conn, statements: Iterable):
"""Execute several statements each as a single DB transaction."""
with conn.cursor() as cursor:
for statement in statements:
try:
cursor.execute(statement)
conn.commit()
except psycopg2.ProgrammingError:
conn.rollback() | [
"def",
"execute_transactions",
"(",
"conn",
",",
"statements",
":",
"Iterable",
")",
":",
"with",
"conn",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"for",
"statement",
"in",
"statements",
":",
"try",
":",
"cursor",
".",
"execute",
"(",
"statement",
")",
"conn",
".",
"commit",
"(",
")",
"except",
"psycopg2",
".",
"ProgrammingError",
":",
"conn",
".",
"rollback",
"(",
")"
] | Execute several statements each as a single DB transaction. | [
"Execute",
"several",
"statements",
"each",
"as",
"a",
"single",
"DB",
"transaction",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L20-L29 | train |
portfoliome/postpy | postpy/sql.py | execute_closing_transaction | def execute_closing_transaction(statements: Iterable):
"""Open a connection, commit a transaction, and close it."""
with closing(connect()) as conn:
with conn.cursor() as cursor:
for statement in statements:
cursor.execute(statement) | python | def execute_closing_transaction(statements: Iterable):
"""Open a connection, commit a transaction, and close it."""
with closing(connect()) as conn:
with conn.cursor() as cursor:
for statement in statements:
cursor.execute(statement) | [
"def",
"execute_closing_transaction",
"(",
"statements",
":",
"Iterable",
")",
":",
"with",
"closing",
"(",
"connect",
"(",
")",
")",
"as",
"conn",
":",
"with",
"conn",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"for",
"statement",
"in",
"statements",
":",
"cursor",
".",
"execute",
"(",
"statement",
")"
] | Open a connection, commit a transaction, and close it. | [
"Open",
"a",
"connection",
"commit",
"a",
"transaction",
"and",
"close",
"it",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L32-L38 | train |
portfoliome/postpy | postpy/sql.py | select | def select(conn, query: str, params=None, name=None, itersize=5000):
"""Return a select statement's results as a namedtuple.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server.
"""
with conn.cursor(name, cursor_factory=NamedTupleCursor) as cursor:
cursor.itersize = itersize
cursor.execute(query, params)
for result in cursor:
yield result | python | def select(conn, query: str, params=None, name=None, itersize=5000):
"""Return a select statement's results as a namedtuple.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server.
"""
with conn.cursor(name, cursor_factory=NamedTupleCursor) as cursor:
cursor.itersize = itersize
cursor.execute(query, params)
for result in cursor:
yield result | [
"def",
"select",
"(",
"conn",
",",
"query",
":",
"str",
",",
"params",
"=",
"None",
",",
"name",
"=",
"None",
",",
"itersize",
"=",
"5000",
")",
":",
"with",
"conn",
".",
"cursor",
"(",
"name",
",",
"cursor_factory",
"=",
"NamedTupleCursor",
")",
"as",
"cursor",
":",
"cursor",
".",
"itersize",
"=",
"itersize",
"cursor",
".",
"execute",
"(",
"query",
",",
"params",
")",
"for",
"result",
"in",
"cursor",
":",
"yield",
"result"
] | Return a select statement's results as a namedtuple.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server. | [
"Return",
"a",
"select",
"statement",
"s",
"results",
"as",
"a",
"namedtuple",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L41-L58 | train |
portfoliome/postpy | postpy/sql.py | select_dict | def select_dict(conn, query: str, params=None, name=None, itersize=5000):
"""Return a select statement's results as dictionary.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server.
"""
with conn.cursor(name, cursor_factory=RealDictCursor) as cursor:
cursor.itersize = itersize
cursor.execute(query, params)
for result in cursor:
yield result | python | def select_dict(conn, query: str, params=None, name=None, itersize=5000):
"""Return a select statement's results as dictionary.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server.
"""
with conn.cursor(name, cursor_factory=RealDictCursor) as cursor:
cursor.itersize = itersize
cursor.execute(query, params)
for result in cursor:
yield result | [
"def",
"select_dict",
"(",
"conn",
",",
"query",
":",
"str",
",",
"params",
"=",
"None",
",",
"name",
"=",
"None",
",",
"itersize",
"=",
"5000",
")",
":",
"with",
"conn",
".",
"cursor",
"(",
"name",
",",
"cursor_factory",
"=",
"RealDictCursor",
")",
"as",
"cursor",
":",
"cursor",
".",
"itersize",
"=",
"itersize",
"cursor",
".",
"execute",
"(",
"query",
",",
"params",
")",
"for",
"result",
"in",
"cursor",
":",
"yield",
"result"
] | Return a select statement's results as dictionary.
Parameters
----------
conn : database connection
query : select query string
params : query parameters.
name : server side cursor name. defaults to client side.
itersize : number of records fetched by server. | [
"Return",
"a",
"select",
"statement",
"s",
"results",
"as",
"dictionary",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L61-L78 | train |
portfoliome/postpy | postpy/sql.py | select_each | def select_each(conn, query: str, parameter_groups, name=None):
"""Run select query for each parameter set in single transaction."""
with conn:
with conn.cursor(name=name) as cursor:
for parameters in parameter_groups:
cursor.execute(query, parameters)
yield cursor.fetchone() | python | def select_each(conn, query: str, parameter_groups, name=None):
"""Run select query for each parameter set in single transaction."""
with conn:
with conn.cursor(name=name) as cursor:
for parameters in parameter_groups:
cursor.execute(query, parameters)
yield cursor.fetchone() | [
"def",
"select_each",
"(",
"conn",
",",
"query",
":",
"str",
",",
"parameter_groups",
",",
"name",
"=",
"None",
")",
":",
"with",
"conn",
":",
"with",
"conn",
".",
"cursor",
"(",
"name",
"=",
"name",
")",
"as",
"cursor",
":",
"for",
"parameters",
"in",
"parameter_groups",
":",
"cursor",
".",
"execute",
"(",
"query",
",",
"parameters",
")",
"yield",
"cursor",
".",
"fetchone",
"(",
")"
] | Run select query for each parameter set in single transaction. | [
"Run",
"select",
"query",
"for",
"each",
"parameter",
"set",
"in",
"single",
"transaction",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L81-L88 | train |
portfoliome/postpy | postpy/sql.py | query_columns | def query_columns(conn, query, name=None):
"""Lightweight query to retrieve column list of select query.
Notes
-----
Strongly urged to specify a cursor name for performance.
"""
with conn.cursor(name) as cursor:
cursor.itersize = 1
cursor.execute(query)
cursor.fetchmany(0)
column_names = [column.name for column in cursor.description]
return column_names | python | def query_columns(conn, query, name=None):
"""Lightweight query to retrieve column list of select query.
Notes
-----
Strongly urged to specify a cursor name for performance.
"""
with conn.cursor(name) as cursor:
cursor.itersize = 1
cursor.execute(query)
cursor.fetchmany(0)
column_names = [column.name for column in cursor.description]
return column_names | [
"def",
"query_columns",
"(",
"conn",
",",
"query",
",",
"name",
"=",
"None",
")",
":",
"with",
"conn",
".",
"cursor",
"(",
"name",
")",
"as",
"cursor",
":",
"cursor",
".",
"itersize",
"=",
"1",
"cursor",
".",
"execute",
"(",
"query",
")",
"cursor",
".",
"fetchmany",
"(",
"0",
")",
"column_names",
"=",
"[",
"column",
".",
"name",
"for",
"column",
"in",
"cursor",
".",
"description",
"]",
"return",
"column_names"
] | Lightweight query to retrieve column list of select query.
Notes
-----
Strongly urged to specify a cursor name for performance. | [
"Lightweight",
"query",
"to",
"retrieve",
"column",
"list",
"of",
"select",
"query",
"."
] | fe26199131b15295fc5f669a0ad2a7f47bf490ee | https://github.com/portfoliome/postpy/blob/fe26199131b15295fc5f669a0ad2a7f47bf490ee/postpy/sql.py#L91-L105 | train |
openvax/isovar | isovar/reference_coding_sequence_key.py | ReferenceCodingSequenceKey.from_variant_and_transcript | def from_variant_and_transcript(
cls,
variant,
transcript,
context_size):
"""
Extracts the reference sequence around a variant locus on a particular
transcript and determines the reading frame at the start of that
sequence context.
Parameters
----------
variant : varcode.Variant
transcript : pyensembl.Transcript
context_size : int
Returns SequenceKeyWithReadingFrame object or None if Transcript lacks
coding sequence, protein sequence or annotated start/stop codons.
"""
if not transcript.contains_start_codon:
logger.info(
"Expected transcript %s for variant %s to have start codon",
transcript.name,
variant)
return None
if not transcript.contains_stop_codon:
logger.info(
"Expected transcript %s for variant %s to have stop codon",
transcript.name,
variant)
return None
if not transcript.protein_sequence:
logger.info(
"Expected transript %s for variant %s to have protein sequence",
transcript.name,
variant)
return None
sequence_key = ReferenceSequenceKey.from_variant_and_transcript(
variant=variant,
transcript=transcript,
context_size=context_size)
if sequence_key is None:
logger.info(
"No sequence key for variant %s on transcript %s",
variant,
transcript.name)
return None
return cls.from_variant_and_transcript_and_sequence_key(
variant=variant,
transcript=transcript,
sequence_key=sequence_key) | python | def from_variant_and_transcript(
cls,
variant,
transcript,
context_size):
"""
Extracts the reference sequence around a variant locus on a particular
transcript and determines the reading frame at the start of that
sequence context.
Parameters
----------
variant : varcode.Variant
transcript : pyensembl.Transcript
context_size : int
Returns SequenceKeyWithReadingFrame object or None if Transcript lacks
coding sequence, protein sequence or annotated start/stop codons.
"""
if not transcript.contains_start_codon:
logger.info(
"Expected transcript %s for variant %s to have start codon",
transcript.name,
variant)
return None
if not transcript.contains_stop_codon:
logger.info(
"Expected transcript %s for variant %s to have stop codon",
transcript.name,
variant)
return None
if not transcript.protein_sequence:
logger.info(
"Expected transript %s for variant %s to have protein sequence",
transcript.name,
variant)
return None
sequence_key = ReferenceSequenceKey.from_variant_and_transcript(
variant=variant,
transcript=transcript,
context_size=context_size)
if sequence_key is None:
logger.info(
"No sequence key for variant %s on transcript %s",
variant,
transcript.name)
return None
return cls.from_variant_and_transcript_and_sequence_key(
variant=variant,
transcript=transcript,
sequence_key=sequence_key) | [
"def",
"from_variant_and_transcript",
"(",
"cls",
",",
"variant",
",",
"transcript",
",",
"context_size",
")",
":",
"if",
"not",
"transcript",
".",
"contains_start_codon",
":",
"logger",
".",
"info",
"(",
"\"Expected transcript %s for variant %s to have start codon\"",
",",
"transcript",
".",
"name",
",",
"variant",
")",
"return",
"None",
"if",
"not",
"transcript",
".",
"contains_stop_codon",
":",
"logger",
".",
"info",
"(",
"\"Expected transcript %s for variant %s to have stop codon\"",
",",
"transcript",
".",
"name",
",",
"variant",
")",
"return",
"None",
"if",
"not",
"transcript",
".",
"protein_sequence",
":",
"logger",
".",
"info",
"(",
"\"Expected transript %s for variant %s to have protein sequence\"",
",",
"transcript",
".",
"name",
",",
"variant",
")",
"return",
"None",
"sequence_key",
"=",
"ReferenceSequenceKey",
".",
"from_variant_and_transcript",
"(",
"variant",
"=",
"variant",
",",
"transcript",
"=",
"transcript",
",",
"context_size",
"=",
"context_size",
")",
"if",
"sequence_key",
"is",
"None",
":",
"logger",
".",
"info",
"(",
"\"No sequence key for variant %s on transcript %s\"",
",",
"variant",
",",
"transcript",
".",
"name",
")",
"return",
"None",
"return",
"cls",
".",
"from_variant_and_transcript_and_sequence_key",
"(",
"variant",
"=",
"variant",
",",
"transcript",
"=",
"transcript",
",",
"sequence_key",
"=",
"sequence_key",
")"
] | Extracts the reference sequence around a variant locus on a particular
transcript and determines the reading frame at the start of that
sequence context.
Parameters
----------
variant : varcode.Variant
transcript : pyensembl.Transcript
context_size : int
Returns SequenceKeyWithReadingFrame object or None if Transcript lacks
coding sequence, protein sequence or annotated start/stop codons. | [
"Extracts",
"the",
"reference",
"sequence",
"around",
"a",
"variant",
"locus",
"on",
"a",
"particular",
"transcript",
"and",
"determines",
"the",
"reading",
"frame",
"at",
"the",
"start",
"of",
"that",
"sequence",
"context",
"."
] | b39b684920e3f6b344851d6598a1a1c67bce913b | https://github.com/openvax/isovar/blob/b39b684920e3f6b344851d6598a1a1c67bce913b/isovar/reference_coding_sequence_key.py#L150-L207 | train |
theno/fabsetup | setup.py | create_readme_with_long_description | def create_readme_with_long_description():
'''Try to convert content of README.md into rst format using pypandoc,
write it into README and return it.
If pypandoc cannot be imported write content of README.md unchanged into
README and return it.
'''
this_dir = os.path.abspath(os.path.dirname(__file__))
readme_md = os.path.join(this_dir, 'README.md')
readme = os.path.join(this_dir, 'README')
if os.path.exists(readme_md):
# this is the case when running `python setup.py sdist`
if os.path.exists(readme):
os.remove(readme)
try:
import pypandoc
long_description = pypandoc.convert(readme_md, 'rst', format='md')
except(ImportError):
with open(readme_md, encoding='utf-8') as in_:
long_description = in_.read()
with open(readme, 'w') as out:
out.write(long_description)
else:
# this is in case of `pip install fabsetup-x.y.z.tar.gz`
with open(readme, encoding='utf-8') as in_:
long_description = in_.read()
return long_description | python | def create_readme_with_long_description():
'''Try to convert content of README.md into rst format using pypandoc,
write it into README and return it.
If pypandoc cannot be imported write content of README.md unchanged into
README and return it.
'''
this_dir = os.path.abspath(os.path.dirname(__file__))
readme_md = os.path.join(this_dir, 'README.md')
readme = os.path.join(this_dir, 'README')
if os.path.exists(readme_md):
# this is the case when running `python setup.py sdist`
if os.path.exists(readme):
os.remove(readme)
try:
import pypandoc
long_description = pypandoc.convert(readme_md, 'rst', format='md')
except(ImportError):
with open(readme_md, encoding='utf-8') as in_:
long_description = in_.read()
with open(readme, 'w') as out:
out.write(long_description)
else:
# this is in case of `pip install fabsetup-x.y.z.tar.gz`
with open(readme, encoding='utf-8') as in_:
long_description = in_.read()
return long_description | [
"def",
"create_readme_with_long_description",
"(",
")",
":",
"this_dir",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
")",
"readme_md",
"=",
"os",
".",
"path",
".",
"join",
"(",
"this_dir",
",",
"'README.md'",
")",
"readme",
"=",
"os",
".",
"path",
".",
"join",
"(",
"this_dir",
",",
"'README'",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"readme_md",
")",
":",
"# this is the case when running `python setup.py sdist`",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"readme",
")",
":",
"os",
".",
"remove",
"(",
"readme",
")",
"try",
":",
"import",
"pypandoc",
"long_description",
"=",
"pypandoc",
".",
"convert",
"(",
"readme_md",
",",
"'rst'",
",",
"format",
"=",
"'md'",
")",
"except",
"(",
"ImportError",
")",
":",
"with",
"open",
"(",
"readme_md",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"in_",
":",
"long_description",
"=",
"in_",
".",
"read",
"(",
")",
"with",
"open",
"(",
"readme",
",",
"'w'",
")",
"as",
"out",
":",
"out",
".",
"write",
"(",
"long_description",
")",
"else",
":",
"# this is in case of `pip install fabsetup-x.y.z.tar.gz`",
"with",
"open",
"(",
"readme",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"in_",
":",
"long_description",
"=",
"in_",
".",
"read",
"(",
")",
"return",
"long_description"
] | Try to convert content of README.md into rst format using pypandoc,
write it into README and return it.
If pypandoc cannot be imported write content of README.md unchanged into
README and return it. | [
"Try",
"to",
"convert",
"content",
"of",
"README",
".",
"md",
"into",
"rst",
"format",
"using",
"pypandoc",
"write",
"it",
"into",
"README",
"and",
"return",
"it",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/setup.py#L13-L44 | train |
sloria/pypi-cli | pypi_cli.py | stat | def stat(package, graph):
"""Print download statistics for a package.
\b
Example:
pypi stat requests
"""
client = requests.Session()
for name_or_url in package:
package = get_package(name_or_url, client)
if not package:
secho(u'Invalid name or URL: "{name}"'.format(name=name_or_url),
fg='red', file=sys.stderr)
continue
try:
version_downloads = package.version_downloads
except NotFoundError:
secho(u'No versions found for "{0}". '
u'Skipping. . .'.format(package.name),
fg='red', file=sys.stderr)
continue
echo(u"Fetching statistics for '{url}'. . .".format(
url=package.package_url))
min_ver, min_downloads = package.min_version
max_ver, max_downloads = package.max_version
if min_ver is None or max_ver is None:
raise click.ClickException('Package has no releases')
avg_downloads = package.average_downloads
total = package.downloads
echo()
header = u'Download statistics for {name}'.format(name=package.name)
echo_header(header)
if graph:
echo()
echo('Downloads by version')
echo(package.chart())
echo()
echo("Min downloads: {min_downloads:12,} ({min_ver})".format(
**locals()))
echo("Max downloads: {max_downloads:12,} ({max_ver})".format(
**locals()))
echo("Avg downloads: {avg_downloads:12,}".format(**locals()))
echo("Total downloads: {total:12,}".format(**locals()))
echo()
echo_download_summary(package)
echo() | python | def stat(package, graph):
"""Print download statistics for a package.
\b
Example:
pypi stat requests
"""
client = requests.Session()
for name_or_url in package:
package = get_package(name_or_url, client)
if not package:
secho(u'Invalid name or URL: "{name}"'.format(name=name_or_url),
fg='red', file=sys.stderr)
continue
try:
version_downloads = package.version_downloads
except NotFoundError:
secho(u'No versions found for "{0}". '
u'Skipping. . .'.format(package.name),
fg='red', file=sys.stderr)
continue
echo(u"Fetching statistics for '{url}'. . .".format(
url=package.package_url))
min_ver, min_downloads = package.min_version
max_ver, max_downloads = package.max_version
if min_ver is None or max_ver is None:
raise click.ClickException('Package has no releases')
avg_downloads = package.average_downloads
total = package.downloads
echo()
header = u'Download statistics for {name}'.format(name=package.name)
echo_header(header)
if graph:
echo()
echo('Downloads by version')
echo(package.chart())
echo()
echo("Min downloads: {min_downloads:12,} ({min_ver})".format(
**locals()))
echo("Max downloads: {max_downloads:12,} ({max_ver})".format(
**locals()))
echo("Avg downloads: {avg_downloads:12,}".format(**locals()))
echo("Total downloads: {total:12,}".format(**locals()))
echo()
echo_download_summary(package)
echo() | [
"def",
"stat",
"(",
"package",
",",
"graph",
")",
":",
"client",
"=",
"requests",
".",
"Session",
"(",
")",
"for",
"name_or_url",
"in",
"package",
":",
"package",
"=",
"get_package",
"(",
"name_or_url",
",",
"client",
")",
"if",
"not",
"package",
":",
"secho",
"(",
"u'Invalid name or URL: \"{name}\"'",
".",
"format",
"(",
"name",
"=",
"name_or_url",
")",
",",
"fg",
"=",
"'red'",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"continue",
"try",
":",
"version_downloads",
"=",
"package",
".",
"version_downloads",
"except",
"NotFoundError",
":",
"secho",
"(",
"u'No versions found for \"{0}\". '",
"u'Skipping. . .'",
".",
"format",
"(",
"package",
".",
"name",
")",
",",
"fg",
"=",
"'red'",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"continue",
"echo",
"(",
"u\"Fetching statistics for '{url}'. . .\"",
".",
"format",
"(",
"url",
"=",
"package",
".",
"package_url",
")",
")",
"min_ver",
",",
"min_downloads",
"=",
"package",
".",
"min_version",
"max_ver",
",",
"max_downloads",
"=",
"package",
".",
"max_version",
"if",
"min_ver",
"is",
"None",
"or",
"max_ver",
"is",
"None",
":",
"raise",
"click",
".",
"ClickException",
"(",
"'Package has no releases'",
")",
"avg_downloads",
"=",
"package",
".",
"average_downloads",
"total",
"=",
"package",
".",
"downloads",
"echo",
"(",
")",
"header",
"=",
"u'Download statistics for {name}'",
".",
"format",
"(",
"name",
"=",
"package",
".",
"name",
")",
"echo_header",
"(",
"header",
")",
"if",
"graph",
":",
"echo",
"(",
")",
"echo",
"(",
"'Downloads by version'",
")",
"echo",
"(",
"package",
".",
"chart",
"(",
")",
")",
"echo",
"(",
")",
"echo",
"(",
"\"Min downloads: {min_downloads:12,} ({min_ver})\"",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"echo",
"(",
"\"Max downloads: {max_downloads:12,} ({max_ver})\"",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"echo",
"(",
"\"Avg downloads: {avg_downloads:12,}\"",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"echo",
"(",
"\"Total downloads: {total:12,}\"",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"echo",
"(",
")",
"echo_download_summary",
"(",
"package",
")",
"echo",
"(",
")"
] | Print download statistics for a package.
\b
Example:
pypi stat requests | [
"Print",
"download",
"statistics",
"for",
"a",
"package",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L97-L143 | train |
sloria/pypi-cli | pypi_cli.py | browse | def browse(package, homepage):
"""Browse to a package's PyPI or project homepage."""
p = Package(package)
try:
if homepage:
secho(u'Opening homepage for "{0}"...'.format(package), bold=True)
url = p.home_page
else:
secho(u'Opening PyPI page for "{0}"...'.format(package), bold=True)
url = p.package_url
except NotFoundError:
abort_not_found(package)
click.launch(url) | python | def browse(package, homepage):
"""Browse to a package's PyPI or project homepage."""
p = Package(package)
try:
if homepage:
secho(u'Opening homepage for "{0}"...'.format(package), bold=True)
url = p.home_page
else:
secho(u'Opening PyPI page for "{0}"...'.format(package), bold=True)
url = p.package_url
except NotFoundError:
abort_not_found(package)
click.launch(url) | [
"def",
"browse",
"(",
"package",
",",
"homepage",
")",
":",
"p",
"=",
"Package",
"(",
"package",
")",
"try",
":",
"if",
"homepage",
":",
"secho",
"(",
"u'Opening homepage for \"{0}\"...'",
".",
"format",
"(",
"package",
")",
",",
"bold",
"=",
"True",
")",
"url",
"=",
"p",
".",
"home_page",
"else",
":",
"secho",
"(",
"u'Opening PyPI page for \"{0}\"...'",
".",
"format",
"(",
"package",
")",
",",
"bold",
"=",
"True",
")",
"url",
"=",
"p",
".",
"package_url",
"except",
"NotFoundError",
":",
"abort_not_found",
"(",
"package",
")",
"click",
".",
"launch",
"(",
"url",
")"
] | Browse to a package's PyPI or project homepage. | [
"Browse",
"to",
"a",
"package",
"s",
"PyPI",
"or",
"project",
"homepage",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L157-L169 | train |
sloria/pypi-cli | pypi_cli.py | search | def search(query, n_results, web):
"""Search for a pypi package.
\b
Examples:
\b
pypi search requests
pypi search 'requests oauth'
pypi search requests -n 20
pypi search 'requests toolbelt' --web
"""
if web:
secho(u'Opening search page for "{0}"...'.format(query), bold=True)
url = SEARCH_URL.format(query=urlquote(query))
click.launch(url)
else:
searcher = Searcher()
results = searcher.search(query, n=n_results)
first_line = style(u'Search results for "{0}"\n'.format(query),
bold=True)
echo_via_pager(
first_line +
'\n'.join([format_result(result) for result in results])
) | python | def search(query, n_results, web):
"""Search for a pypi package.
\b
Examples:
\b
pypi search requests
pypi search 'requests oauth'
pypi search requests -n 20
pypi search 'requests toolbelt' --web
"""
if web:
secho(u'Opening search page for "{0}"...'.format(query), bold=True)
url = SEARCH_URL.format(query=urlquote(query))
click.launch(url)
else:
searcher = Searcher()
results = searcher.search(query, n=n_results)
first_line = style(u'Search results for "{0}"\n'.format(query),
bold=True)
echo_via_pager(
first_line +
'\n'.join([format_result(result) for result in results])
) | [
"def",
"search",
"(",
"query",
",",
"n_results",
",",
"web",
")",
":",
"if",
"web",
":",
"secho",
"(",
"u'Opening search page for \"{0}\"...'",
".",
"format",
"(",
"query",
")",
",",
"bold",
"=",
"True",
")",
"url",
"=",
"SEARCH_URL",
".",
"format",
"(",
"query",
"=",
"urlquote",
"(",
"query",
")",
")",
"click",
".",
"launch",
"(",
"url",
")",
"else",
":",
"searcher",
"=",
"Searcher",
"(",
")",
"results",
"=",
"searcher",
".",
"search",
"(",
"query",
",",
"n",
"=",
"n_results",
")",
"first_line",
"=",
"style",
"(",
"u'Search results for \"{0}\"\\n'",
".",
"format",
"(",
"query",
")",
",",
"bold",
"=",
"True",
")",
"echo_via_pager",
"(",
"first_line",
"+",
"'\\n'",
".",
"join",
"(",
"[",
"format_result",
"(",
"result",
")",
"for",
"result",
"in",
"results",
"]",
")",
")"
] | Search for a pypi package.
\b
Examples:
\b
pypi search requests
pypi search 'requests oauth'
pypi search requests -n 20
pypi search 'requests toolbelt' --web | [
"Search",
"for",
"a",
"pypi",
"package",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L192-L216 | train |
sloria/pypi-cli | pypi_cli.py | info | def info(package, long_description, classifiers, license):
"""Get info about a package or packages.
"""
client = requests.Session()
for name_or_url in package:
package = get_package(name_or_url, client)
if not package:
secho(u'Invalid name or URL: "{name}"'.format(name=name_or_url),
fg='red', file=sys.stderr)
continue
# Name and summary
try:
info = package.data['info']
except NotFoundError:
secho(u'No versions found for "{0}". '
u'Skipping. . .'.format(package.name),
fg='red', file=sys.stderr)
continue
echo_header(name_or_url)
if package.summary:
echo(package.summary)
# Version info
echo()
echo('Latest release: {version:12}'.format(version=info['version']))
# Long description
if long_description:
echo()
echo(package.description)
# Download info
echo()
echo_download_summary(package)
# Author info
echo()
author, author_email = package.author, package.author_email
if author:
echo(u'Author: {author:12}'.format(**locals()))
if author_email:
echo(u'Author email: {author_email:12}'.format(**locals()))
# Maintainer info
maintainer, maintainer_email = (package.maintainer,
package.maintainer_email)
if maintainer or maintainer_email:
echo()
if maintainer:
echo(u'Maintainer: {maintainer:12}'.format(**locals()))
if maintainer_email:
echo(u'Maintainer email: {maintainer_email:12}'.format(**locals()))
# URLS
echo()
echo(u'PyPI URL: {pypi_url:12}'.format(pypi_url=package.package_url))
if package.home_page:
echo(u'Home Page: {home_page:12}'.format(
home_page=package.home_page))
if package.docs_url:
echo(u'Documentation: {docs_url:12}'.format(
docs_url=package.docs_url))
# Classifiers
if classifiers:
echo()
echo(u'Classifiers: ')
for each in info.get('classifiers', []):
echo('\t' + each)
if license and package.license:
echo()
echo(u'License: ', nl=False)
# license may be just a name, e.g. 'BSD' or the full license text
# If a new line is found in the text, print a new line
if package.license.find('\n') >= 0 or len(package.license) > 80:
echo()
echo(package.license)
echo() | python | def info(package, long_description, classifiers, license):
"""Get info about a package or packages.
"""
client = requests.Session()
for name_or_url in package:
package = get_package(name_or_url, client)
if not package:
secho(u'Invalid name or URL: "{name}"'.format(name=name_or_url),
fg='red', file=sys.stderr)
continue
# Name and summary
try:
info = package.data['info']
except NotFoundError:
secho(u'No versions found for "{0}". '
u'Skipping. . .'.format(package.name),
fg='red', file=sys.stderr)
continue
echo_header(name_or_url)
if package.summary:
echo(package.summary)
# Version info
echo()
echo('Latest release: {version:12}'.format(version=info['version']))
# Long description
if long_description:
echo()
echo(package.description)
# Download info
echo()
echo_download_summary(package)
# Author info
echo()
author, author_email = package.author, package.author_email
if author:
echo(u'Author: {author:12}'.format(**locals()))
if author_email:
echo(u'Author email: {author_email:12}'.format(**locals()))
# Maintainer info
maintainer, maintainer_email = (package.maintainer,
package.maintainer_email)
if maintainer or maintainer_email:
echo()
if maintainer:
echo(u'Maintainer: {maintainer:12}'.format(**locals()))
if maintainer_email:
echo(u'Maintainer email: {maintainer_email:12}'.format(**locals()))
# URLS
echo()
echo(u'PyPI URL: {pypi_url:12}'.format(pypi_url=package.package_url))
if package.home_page:
echo(u'Home Page: {home_page:12}'.format(
home_page=package.home_page))
if package.docs_url:
echo(u'Documentation: {docs_url:12}'.format(
docs_url=package.docs_url))
# Classifiers
if classifiers:
echo()
echo(u'Classifiers: ')
for each in info.get('classifiers', []):
echo('\t' + each)
if license and package.license:
echo()
echo(u'License: ', nl=False)
# license may be just a name, e.g. 'BSD' or the full license text
# If a new line is found in the text, print a new line
if package.license.find('\n') >= 0 or len(package.license) > 80:
echo()
echo(package.license)
echo() | [
"def",
"info",
"(",
"package",
",",
"long_description",
",",
"classifiers",
",",
"license",
")",
":",
"client",
"=",
"requests",
".",
"Session",
"(",
")",
"for",
"name_or_url",
"in",
"package",
":",
"package",
"=",
"get_package",
"(",
"name_or_url",
",",
"client",
")",
"if",
"not",
"package",
":",
"secho",
"(",
"u'Invalid name or URL: \"{name}\"'",
".",
"format",
"(",
"name",
"=",
"name_or_url",
")",
",",
"fg",
"=",
"'red'",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"continue",
"# Name and summary",
"try",
":",
"info",
"=",
"package",
".",
"data",
"[",
"'info'",
"]",
"except",
"NotFoundError",
":",
"secho",
"(",
"u'No versions found for \"{0}\". '",
"u'Skipping. . .'",
".",
"format",
"(",
"package",
".",
"name",
")",
",",
"fg",
"=",
"'red'",
",",
"file",
"=",
"sys",
".",
"stderr",
")",
"continue",
"echo_header",
"(",
"name_or_url",
")",
"if",
"package",
".",
"summary",
":",
"echo",
"(",
"package",
".",
"summary",
")",
"# Version info",
"echo",
"(",
")",
"echo",
"(",
"'Latest release: {version:12}'",
".",
"format",
"(",
"version",
"=",
"info",
"[",
"'version'",
"]",
")",
")",
"# Long description",
"if",
"long_description",
":",
"echo",
"(",
")",
"echo",
"(",
"package",
".",
"description",
")",
"# Download info",
"echo",
"(",
")",
"echo_download_summary",
"(",
"package",
")",
"# Author info",
"echo",
"(",
")",
"author",
",",
"author_email",
"=",
"package",
".",
"author",
",",
"package",
".",
"author_email",
"if",
"author",
":",
"echo",
"(",
"u'Author: {author:12}'",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"if",
"author_email",
":",
"echo",
"(",
"u'Author email: {author_email:12}'",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"# Maintainer info",
"maintainer",
",",
"maintainer_email",
"=",
"(",
"package",
".",
"maintainer",
",",
"package",
".",
"maintainer_email",
")",
"if",
"maintainer",
"or",
"maintainer_email",
":",
"echo",
"(",
")",
"if",
"maintainer",
":",
"echo",
"(",
"u'Maintainer: {maintainer:12}'",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"if",
"maintainer_email",
":",
"echo",
"(",
"u'Maintainer email: {maintainer_email:12}'",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")",
"# URLS",
"echo",
"(",
")",
"echo",
"(",
"u'PyPI URL: {pypi_url:12}'",
".",
"format",
"(",
"pypi_url",
"=",
"package",
".",
"package_url",
")",
")",
"if",
"package",
".",
"home_page",
":",
"echo",
"(",
"u'Home Page: {home_page:12}'",
".",
"format",
"(",
"home_page",
"=",
"package",
".",
"home_page",
")",
")",
"if",
"package",
".",
"docs_url",
":",
"echo",
"(",
"u'Documentation: {docs_url:12}'",
".",
"format",
"(",
"docs_url",
"=",
"package",
".",
"docs_url",
")",
")",
"# Classifiers",
"if",
"classifiers",
":",
"echo",
"(",
")",
"echo",
"(",
"u'Classifiers: '",
")",
"for",
"each",
"in",
"info",
".",
"get",
"(",
"'classifiers'",
",",
"[",
"]",
")",
":",
"echo",
"(",
"'\\t'",
"+",
"each",
")",
"if",
"license",
"and",
"package",
".",
"license",
":",
"echo",
"(",
")",
"echo",
"(",
"u'License: '",
",",
"nl",
"=",
"False",
")",
"# license may be just a name, e.g. 'BSD' or the full license text",
"# If a new line is found in the text, print a new line",
"if",
"package",
".",
"license",
".",
"find",
"(",
"'\\n'",
")",
">=",
"0",
"or",
"len",
"(",
"package",
".",
"license",
")",
">",
"80",
":",
"echo",
"(",
")",
"echo",
"(",
"package",
".",
"license",
")",
"echo",
"(",
")"
] | Get info about a package or packages. | [
"Get",
"info",
"about",
"a",
"package",
"or",
"packages",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L227-L306 | train |
sloria/pypi-cli | pypi_cli.py | bargraph | def bargraph(data, max_key_width=30):
"""Return a bar graph as a string, given a dictionary of data."""
lines = []
max_length = min(max(len(key) for key in data.keys()), max_key_width)
max_val = max(data.values())
max_val_length = max(
len(_style_value(val))
for val in data.values())
term_width = get_terminal_size()[0]
max_bar_width = term_width - MARGIN - (max_length + 3 + max_val_length + 3)
template = u"{key:{key_width}} [ {value:{val_width}} ] {bar}"
for key, value in data.items():
try:
bar = int(math.ceil(max_bar_width * value / max_val)) * TICK
except ZeroDivisionError:
bar = ''
line = template.format(
key=key[:max_length],
value=_style_value(value),
bar=bar,
key_width=max_length,
val_width=max_val_length
)
lines.append(line)
return '\n'.join(lines) | python | def bargraph(data, max_key_width=30):
"""Return a bar graph as a string, given a dictionary of data."""
lines = []
max_length = min(max(len(key) for key in data.keys()), max_key_width)
max_val = max(data.values())
max_val_length = max(
len(_style_value(val))
for val in data.values())
term_width = get_terminal_size()[0]
max_bar_width = term_width - MARGIN - (max_length + 3 + max_val_length + 3)
template = u"{key:{key_width}} [ {value:{val_width}} ] {bar}"
for key, value in data.items():
try:
bar = int(math.ceil(max_bar_width * value / max_val)) * TICK
except ZeroDivisionError:
bar = ''
line = template.format(
key=key[:max_length],
value=_style_value(value),
bar=bar,
key_width=max_length,
val_width=max_val_length
)
lines.append(line)
return '\n'.join(lines) | [
"def",
"bargraph",
"(",
"data",
",",
"max_key_width",
"=",
"30",
")",
":",
"lines",
"=",
"[",
"]",
"max_length",
"=",
"min",
"(",
"max",
"(",
"len",
"(",
"key",
")",
"for",
"key",
"in",
"data",
".",
"keys",
"(",
")",
")",
",",
"max_key_width",
")",
"max_val",
"=",
"max",
"(",
"data",
".",
"values",
"(",
")",
")",
"max_val_length",
"=",
"max",
"(",
"len",
"(",
"_style_value",
"(",
"val",
")",
")",
"for",
"val",
"in",
"data",
".",
"values",
"(",
")",
")",
"term_width",
"=",
"get_terminal_size",
"(",
")",
"[",
"0",
"]",
"max_bar_width",
"=",
"term_width",
"-",
"MARGIN",
"-",
"(",
"max_length",
"+",
"3",
"+",
"max_val_length",
"+",
"3",
")",
"template",
"=",
"u\"{key:{key_width}} [ {value:{val_width}} ] {bar}\"",
"for",
"key",
",",
"value",
"in",
"data",
".",
"items",
"(",
")",
":",
"try",
":",
"bar",
"=",
"int",
"(",
"math",
".",
"ceil",
"(",
"max_bar_width",
"*",
"value",
"/",
"max_val",
")",
")",
"*",
"TICK",
"except",
"ZeroDivisionError",
":",
"bar",
"=",
"''",
"line",
"=",
"template",
".",
"format",
"(",
"key",
"=",
"key",
"[",
":",
"max_length",
"]",
",",
"value",
"=",
"_style_value",
"(",
"value",
")",
",",
"bar",
"=",
"bar",
",",
"key_width",
"=",
"max_length",
",",
"val_width",
"=",
"max_val_length",
")",
"lines",
".",
"append",
"(",
"line",
")",
"return",
"'\\n'",
".",
"join",
"(",
"lines",
")"
] | Return a bar graph as a string, given a dictionary of data. | [
"Return",
"a",
"bar",
"graph",
"as",
"a",
"string",
"given",
"a",
"dictionary",
"of",
"data",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L328-L352 | train |
sloria/pypi-cli | pypi_cli.py | Package.max_version | def max_version(self):
"""Version with the most downloads.
:return: A tuple of the form (version, n_downloads)
"""
data = self.version_downloads
if not data:
return None, 0
return max(data.items(), key=lambda item: item[1]) | python | def max_version(self):
"""Version with the most downloads.
:return: A tuple of the form (version, n_downloads)
"""
data = self.version_downloads
if not data:
return None, 0
return max(data.items(), key=lambda item: item[1]) | [
"def",
"max_version",
"(",
"self",
")",
":",
"data",
"=",
"self",
".",
"version_downloads",
"if",
"not",
"data",
":",
"return",
"None",
",",
"0",
"return",
"max",
"(",
"data",
".",
"items",
"(",
")",
",",
"key",
"=",
"lambda",
"item",
":",
"item",
"[",
"1",
"]",
")"
] | Version with the most downloads.
:return: A tuple of the form (version, n_downloads) | [
"Version",
"with",
"the",
"most",
"downloads",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L440-L448 | train |
sloria/pypi-cli | pypi_cli.py | Package.min_version | def min_version(self):
"""Version with the fewest downloads."""
data = self.version_downloads
if not data:
return (None, 0)
return min(data.items(), key=lambda item: item[1]) | python | def min_version(self):
"""Version with the fewest downloads."""
data = self.version_downloads
if not data:
return (None, 0)
return min(data.items(), key=lambda item: item[1]) | [
"def",
"min_version",
"(",
"self",
")",
":",
"data",
"=",
"self",
".",
"version_downloads",
"if",
"not",
"data",
":",
"return",
"(",
"None",
",",
"0",
")",
"return",
"min",
"(",
"data",
".",
"items",
"(",
")",
",",
"key",
"=",
"lambda",
"item",
":",
"item",
"[",
"1",
"]",
")"
] | Version with the fewest downloads. | [
"Version",
"with",
"the",
"fewest",
"downloads",
"."
] | beb007bf2bdd285209876ce2758982b5d8b54d5d | https://github.com/sloria/pypi-cli/blob/beb007bf2bdd285209876ce2758982b5d8b54d5d/pypi_cli.py#L451-L456 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | ripping_of_cds | def ripping_of_cds():
'''Install the tools ripit and burnit in order to rip and burn audio cds.
More info: http://forums.debian.net/viewtopic.php?f=16&t=36826
'''
# install and configure ripit
install_package('ripit')
install_file_legacy(path='~/.ripit/config', username=env.user)
# install burnit
run('mkdir -p ~/bin')
install_file_legacy('~/bin/burnit')
run('chmod 755 ~/bin/burnit') | python | def ripping_of_cds():
'''Install the tools ripit and burnit in order to rip and burn audio cds.
More info: http://forums.debian.net/viewtopic.php?f=16&t=36826
'''
# install and configure ripit
install_package('ripit')
install_file_legacy(path='~/.ripit/config', username=env.user)
# install burnit
run('mkdir -p ~/bin')
install_file_legacy('~/bin/burnit')
run('chmod 755 ~/bin/burnit') | [
"def",
"ripping_of_cds",
"(",
")",
":",
"# install and configure ripit",
"install_package",
"(",
"'ripit'",
")",
"install_file_legacy",
"(",
"path",
"=",
"'~/.ripit/config'",
",",
"username",
"=",
"env",
".",
"user",
")",
"# install burnit",
"run",
"(",
"'mkdir -p ~/bin'",
")",
"install_file_legacy",
"(",
"'~/bin/burnit'",
")",
"run",
"(",
"'chmod 755 ~/bin/burnit'",
")"
] | Install the tools ripit and burnit in order to rip and burn audio cds.
More info: http://forums.debian.net/viewtopic.php?f=16&t=36826 | [
"Install",
"the",
"tools",
"ripit",
"and",
"burnit",
"in",
"order",
"to",
"rip",
"and",
"burn",
"audio",
"cds",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L34-L45 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | i3 | def i3():
'''Install and customize the tiling window manager i3.'''
install_package('i3')
install_file_legacy(path='~/.i3/config', username=env.user, repos_dir='repos')
# setup: hide the mouse if not in use
# in ~/.i3/config: 'exec /home/<USERNAME>/repos/hhpc/hhpc -i 10 &'
install_packages(['make', 'pkg-config', 'gcc', 'libc6-dev', 'libx11-dev'])
checkup_git_repo_legacy(url='https://github.com/aktau/hhpc.git')
run('cd ~/repos/hhpc && make') | python | def i3():
'''Install and customize the tiling window manager i3.'''
install_package('i3')
install_file_legacy(path='~/.i3/config', username=env.user, repos_dir='repos')
# setup: hide the mouse if not in use
# in ~/.i3/config: 'exec /home/<USERNAME>/repos/hhpc/hhpc -i 10 &'
install_packages(['make', 'pkg-config', 'gcc', 'libc6-dev', 'libx11-dev'])
checkup_git_repo_legacy(url='https://github.com/aktau/hhpc.git')
run('cd ~/repos/hhpc && make') | [
"def",
"i3",
"(",
")",
":",
"install_package",
"(",
"'i3'",
")",
"install_file_legacy",
"(",
"path",
"=",
"'~/.i3/config'",
",",
"username",
"=",
"env",
".",
"user",
",",
"repos_dir",
"=",
"'repos'",
")",
"# setup: hide the mouse if not in use",
"# in ~/.i3/config: 'exec /home/<USERNAME>/repos/hhpc/hhpc -i 10 &'",
"install_packages",
"(",
"[",
"'make'",
",",
"'pkg-config'",
",",
"'gcc'",
",",
"'libc6-dev'",
",",
"'libx11-dev'",
"]",
")",
"checkup_git_repo_legacy",
"(",
"url",
"=",
"'https://github.com/aktau/hhpc.git'",
")",
"run",
"(",
"'cd ~/repos/hhpc && make'",
")"
] | Install and customize the tiling window manager i3. | [
"Install",
"and",
"customize",
"the",
"tiling",
"window",
"manager",
"i3",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L78-L87 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | solarized | def solarized():
'''Set solarized colors in urxvt, tmux, and vim.
More Infos:
* Getting solarized colors right with urxvt, st, tmux and vim:
https://bbs.archlinux.org/viewtopic.php?id=164108
* Creating ~/.Xresources:
https://wiki.archlinux.org/index.php/Rxvt-unicode#Creating_.7E.2F.Xresources
* Select a good font on Ubuntu:
https://michaelheap.com/getting-solarized-working-on-ubuntu/
* tmux and 256 colors:
http://unix.stackexchange.com/a/118903
'''
install_packages(['rxvt-unicode', 'tmux', 'vim'])
install_file_legacy('~/.Xresources')
if env.host_string == 'localhost':
run('xrdb ~/.Xresources')
# install and call term_colors
run('mkdir -p ~/bin')
install_file_legacy('~/bin/term_colors')
run('chmod 755 ~/bin/term_colors')
run('~/bin/term_colors') | python | def solarized():
'''Set solarized colors in urxvt, tmux, and vim.
More Infos:
* Getting solarized colors right with urxvt, st, tmux and vim:
https://bbs.archlinux.org/viewtopic.php?id=164108
* Creating ~/.Xresources:
https://wiki.archlinux.org/index.php/Rxvt-unicode#Creating_.7E.2F.Xresources
* Select a good font on Ubuntu:
https://michaelheap.com/getting-solarized-working-on-ubuntu/
* tmux and 256 colors:
http://unix.stackexchange.com/a/118903
'''
install_packages(['rxvt-unicode', 'tmux', 'vim'])
install_file_legacy('~/.Xresources')
if env.host_string == 'localhost':
run('xrdb ~/.Xresources')
# install and call term_colors
run('mkdir -p ~/bin')
install_file_legacy('~/bin/term_colors')
run('chmod 755 ~/bin/term_colors')
run('~/bin/term_colors') | [
"def",
"solarized",
"(",
")",
":",
"install_packages",
"(",
"[",
"'rxvt-unicode'",
",",
"'tmux'",
",",
"'vim'",
"]",
")",
"install_file_legacy",
"(",
"'~/.Xresources'",
")",
"if",
"env",
".",
"host_string",
"==",
"'localhost'",
":",
"run",
"(",
"'xrdb ~/.Xresources'",
")",
"# install and call term_colors",
"run",
"(",
"'mkdir -p ~/bin'",
")",
"install_file_legacy",
"(",
"'~/bin/term_colors'",
")",
"run",
"(",
"'chmod 755 ~/bin/term_colors'",
")",
"run",
"(",
"'~/bin/term_colors'",
")"
] | Set solarized colors in urxvt, tmux, and vim.
More Infos:
* Getting solarized colors right with urxvt, st, tmux and vim:
https://bbs.archlinux.org/viewtopic.php?id=164108
* Creating ~/.Xresources:
https://wiki.archlinux.org/index.php/Rxvt-unicode#Creating_.7E.2F.Xresources
* Select a good font on Ubuntu:
https://michaelheap.com/getting-solarized-working-on-ubuntu/
* tmux and 256 colors:
http://unix.stackexchange.com/a/118903 | [
"Set",
"solarized",
"colors",
"in",
"urxvt",
"tmux",
"and",
"vim",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L92-L117 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | vim | def vim():
'''Customize vim, install package manager pathogen and some vim-packages.
pathogen will be installed as a git repo at ~/repos/vim-pathogen and
activated in vim by a symbolic link at ~/.vim/autoload/pathogen.vim
A ~/.vimrc will be installed which loads the package manager within of vim.
The vim packages vim-colors-solarized, nerdtree, and tagbar are installed
as git repos placed at dir ~/.vim/bundle/
If you want to install more vim packages also place them at this dir, cf.
https://logicalfriday.com/2011/07/18/using-vim-with-pathogen/
'''
install_package('vim')
print_msg('## install ~/.vimrc\n')
install_file_legacy('~/.vimrc')
print_msg('\n## set up pathogen\n')
run('mkdir -p ~/.vim/autoload ~/.vim/bundle')
checkup_git_repo_legacy(url='https://github.com/tpope/vim-pathogen.git')
run('ln -snf ~/repos/vim-pathogen/autoload/pathogen.vim '
'~/.vim/autoload/pathogen.vim')
print_msg('\n## install vim packages\n')
install_package('ctags') # required by package tagbar
repos = [
{
'name': 'vim-colors-solarized',
'url': 'git://github.com/altercation/vim-colors-solarized.git',
},
{
'name': 'nerdtree',
'url': 'https://github.com/scrooloose/nerdtree.git',
},
{
'name': 'vim-nerdtree-tabs',
'url': 'https://github.com/jistr/vim-nerdtree-tabs.git',
},
{
'name': 'tagbar',
'url': 'https://github.com/majutsushi/tagbar.git',
},
]
checkup_git_repos_legacy(repos, base_dir='~/.vim/bundle') | python | def vim():
'''Customize vim, install package manager pathogen and some vim-packages.
pathogen will be installed as a git repo at ~/repos/vim-pathogen and
activated in vim by a symbolic link at ~/.vim/autoload/pathogen.vim
A ~/.vimrc will be installed which loads the package manager within of vim.
The vim packages vim-colors-solarized, nerdtree, and tagbar are installed
as git repos placed at dir ~/.vim/bundle/
If you want to install more vim packages also place them at this dir, cf.
https://logicalfriday.com/2011/07/18/using-vim-with-pathogen/
'''
install_package('vim')
print_msg('## install ~/.vimrc\n')
install_file_legacy('~/.vimrc')
print_msg('\n## set up pathogen\n')
run('mkdir -p ~/.vim/autoload ~/.vim/bundle')
checkup_git_repo_legacy(url='https://github.com/tpope/vim-pathogen.git')
run('ln -snf ~/repos/vim-pathogen/autoload/pathogen.vim '
'~/.vim/autoload/pathogen.vim')
print_msg('\n## install vim packages\n')
install_package('ctags') # required by package tagbar
repos = [
{
'name': 'vim-colors-solarized',
'url': 'git://github.com/altercation/vim-colors-solarized.git',
},
{
'name': 'nerdtree',
'url': 'https://github.com/scrooloose/nerdtree.git',
},
{
'name': 'vim-nerdtree-tabs',
'url': 'https://github.com/jistr/vim-nerdtree-tabs.git',
},
{
'name': 'tagbar',
'url': 'https://github.com/majutsushi/tagbar.git',
},
]
checkup_git_repos_legacy(repos, base_dir='~/.vim/bundle') | [
"def",
"vim",
"(",
")",
":",
"install_package",
"(",
"'vim'",
")",
"print_msg",
"(",
"'## install ~/.vimrc\\n'",
")",
"install_file_legacy",
"(",
"'~/.vimrc'",
")",
"print_msg",
"(",
"'\\n## set up pathogen\\n'",
")",
"run",
"(",
"'mkdir -p ~/.vim/autoload ~/.vim/bundle'",
")",
"checkup_git_repo_legacy",
"(",
"url",
"=",
"'https://github.com/tpope/vim-pathogen.git'",
")",
"run",
"(",
"'ln -snf ~/repos/vim-pathogen/autoload/pathogen.vim '",
"'~/.vim/autoload/pathogen.vim'",
")",
"print_msg",
"(",
"'\\n## install vim packages\\n'",
")",
"install_package",
"(",
"'ctags'",
")",
"# required by package tagbar",
"repos",
"=",
"[",
"{",
"'name'",
":",
"'vim-colors-solarized'",
",",
"'url'",
":",
"'git://github.com/altercation/vim-colors-solarized.git'",
",",
"}",
",",
"{",
"'name'",
":",
"'nerdtree'",
",",
"'url'",
":",
"'https://github.com/scrooloose/nerdtree.git'",
",",
"}",
",",
"{",
"'name'",
":",
"'vim-nerdtree-tabs'",
",",
"'url'",
":",
"'https://github.com/jistr/vim-nerdtree-tabs.git'",
",",
"}",
",",
"{",
"'name'",
":",
"'tagbar'",
",",
"'url'",
":",
"'https://github.com/majutsushi/tagbar.git'",
",",
"}",
",",
"]",
"checkup_git_repos_legacy",
"(",
"repos",
",",
"base_dir",
"=",
"'~/.vim/bundle'",
")"
] | Customize vim, install package manager pathogen and some vim-packages.
pathogen will be installed as a git repo at ~/repos/vim-pathogen and
activated in vim by a symbolic link at ~/.vim/autoload/pathogen.vim
A ~/.vimrc will be installed which loads the package manager within of vim.
The vim packages vim-colors-solarized, nerdtree, and tagbar are installed
as git repos placed at dir ~/.vim/bundle/
If you want to install more vim packages also place them at this dir, cf.
https://logicalfriday.com/2011/07/18/using-vim-with-pathogen/ | [
"Customize",
"vim",
"install",
"package",
"manager",
"pathogen",
"and",
"some",
"vim",
"-",
"packages",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L122-L167 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | pyenv | def pyenv():
'''Install or update the pyenv python environment.
Checkout or update the pyenv repo at ~/.pyenv and enable the pyenv.
Pyenv wird also als Github-Repo "installiert" unter ~/.pyenv
More info:
* https://github.com/yyuu/pyenv
* https://github.com/yyuu/pyenv/wiki/Common-build-problems#requirements
Tutorial:
* http://amaral-lab.org/resources/guides/pyenv-tutorial
'''
install_packages([
'make',
'build-essential',
'libssl-dev',
'zlib1g-dev',
'libbz2-dev',
'libreadline-dev',
'libsqlite3-dev',
'wget',
'curl',
'llvm',
'libncurses5-dev',
'libncursesw5-dev',
])
if exists('~/.pyenv'):
run('cd ~/.pyenv && git pull')
run('~/.pyenv/bin/pyenv update')
else:
run('curl -L https://raw.githubusercontent.com/yyuu/pyenv-installer/'
'master/bin/pyenv-installer | bash')
# add pyenv to $PATH and set up pyenv init
bash_snippet = '~/.bashrc_pyenv'
install_file_legacy(path=bash_snippet)
prefix = flo('if [ -f {bash_snippet} ]; ')
enabler = flo('if [ -f {bash_snippet} ]; then source {bash_snippet}; fi')
if env.host == 'localhost':
# FIXME: next function currently only works for localhost
uncomment_or_update_or_append_line(filename='~/.bashrc', prefix=prefix,
new_line=enabler)
else:
print(cyan('\nappend to ~/.bashrc:\n\n ') + enabler) | python | def pyenv():
'''Install or update the pyenv python environment.
Checkout or update the pyenv repo at ~/.pyenv and enable the pyenv.
Pyenv wird also als Github-Repo "installiert" unter ~/.pyenv
More info:
* https://github.com/yyuu/pyenv
* https://github.com/yyuu/pyenv/wiki/Common-build-problems#requirements
Tutorial:
* http://amaral-lab.org/resources/guides/pyenv-tutorial
'''
install_packages([
'make',
'build-essential',
'libssl-dev',
'zlib1g-dev',
'libbz2-dev',
'libreadline-dev',
'libsqlite3-dev',
'wget',
'curl',
'llvm',
'libncurses5-dev',
'libncursesw5-dev',
])
if exists('~/.pyenv'):
run('cd ~/.pyenv && git pull')
run('~/.pyenv/bin/pyenv update')
else:
run('curl -L https://raw.githubusercontent.com/yyuu/pyenv-installer/'
'master/bin/pyenv-installer | bash')
# add pyenv to $PATH and set up pyenv init
bash_snippet = '~/.bashrc_pyenv'
install_file_legacy(path=bash_snippet)
prefix = flo('if [ -f {bash_snippet} ]; ')
enabler = flo('if [ -f {bash_snippet} ]; then source {bash_snippet}; fi')
if env.host == 'localhost':
# FIXME: next function currently only works for localhost
uncomment_or_update_or_append_line(filename='~/.bashrc', prefix=prefix,
new_line=enabler)
else:
print(cyan('\nappend to ~/.bashrc:\n\n ') + enabler) | [
"def",
"pyenv",
"(",
")",
":",
"install_packages",
"(",
"[",
"'make'",
",",
"'build-essential'",
",",
"'libssl-dev'",
",",
"'zlib1g-dev'",
",",
"'libbz2-dev'",
",",
"'libreadline-dev'",
",",
"'libsqlite3-dev'",
",",
"'wget'",
",",
"'curl'",
",",
"'llvm'",
",",
"'libncurses5-dev'",
",",
"'libncursesw5-dev'",
",",
"]",
")",
"if",
"exists",
"(",
"'~/.pyenv'",
")",
":",
"run",
"(",
"'cd ~/.pyenv && git pull'",
")",
"run",
"(",
"'~/.pyenv/bin/pyenv update'",
")",
"else",
":",
"run",
"(",
"'curl -L https://raw.githubusercontent.com/yyuu/pyenv-installer/'",
"'master/bin/pyenv-installer | bash'",
")",
"# add pyenv to $PATH and set up pyenv init",
"bash_snippet",
"=",
"'~/.bashrc_pyenv'",
"install_file_legacy",
"(",
"path",
"=",
"bash_snippet",
")",
"prefix",
"=",
"flo",
"(",
"'if [ -f {bash_snippet} ]; '",
")",
"enabler",
"=",
"flo",
"(",
"'if [ -f {bash_snippet} ]; then source {bash_snippet}; fi'",
")",
"if",
"env",
".",
"host",
"==",
"'localhost'",
":",
"# FIXME: next function currently only works for localhost",
"uncomment_or_update_or_append_line",
"(",
"filename",
"=",
"'~/.bashrc'",
",",
"prefix",
"=",
"prefix",
",",
"new_line",
"=",
"enabler",
")",
"else",
":",
"print",
"(",
"cyan",
"(",
"'\\nappend to ~/.bashrc:\\n\\n '",
")",
"+",
"enabler",
")"
] | Install or update the pyenv python environment.
Checkout or update the pyenv repo at ~/.pyenv and enable the pyenv.
Pyenv wird also als Github-Repo "installiert" unter ~/.pyenv
More info:
* https://github.com/yyuu/pyenv
* https://github.com/yyuu/pyenv/wiki/Common-build-problems#requirements
Tutorial:
* http://amaral-lab.org/resources/guides/pyenv-tutorial | [
"Install",
"or",
"update",
"the",
"pyenv",
"python",
"environment",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L172-L215 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | virtualbox_host | def virtualbox_host():
'''Install a VirtualBox host system.
More Infos:
* overview: https://wiki.ubuntuusers.de/VirtualBox/
* installation: https://wiki.ubuntuusers.de/VirtualBox/Installation/
'''
if query_yes_no(question='Uninstall virtualbox-dkms?', default='yes'):
run('sudo apt-get remove virtualbox-dkms')
install_packages([
'virtualbox',
'virtualbox-qt',
'virtualbox-dkms',
'virtualbox-guest-dkms',
'virtualbox-guest-additions-iso',
])
users = [env.user]
for username in users:
run(flo('sudo adduser {username} vboxusers')) | python | def virtualbox_host():
'''Install a VirtualBox host system.
More Infos:
* overview: https://wiki.ubuntuusers.de/VirtualBox/
* installation: https://wiki.ubuntuusers.de/VirtualBox/Installation/
'''
if query_yes_no(question='Uninstall virtualbox-dkms?', default='yes'):
run('sudo apt-get remove virtualbox-dkms')
install_packages([
'virtualbox',
'virtualbox-qt',
'virtualbox-dkms',
'virtualbox-guest-dkms',
'virtualbox-guest-additions-iso',
])
users = [env.user]
for username in users:
run(flo('sudo adduser {username} vboxusers')) | [
"def",
"virtualbox_host",
"(",
")",
":",
"if",
"query_yes_no",
"(",
"question",
"=",
"'Uninstall virtualbox-dkms?'",
",",
"default",
"=",
"'yes'",
")",
":",
"run",
"(",
"'sudo apt-get remove virtualbox-dkms'",
")",
"install_packages",
"(",
"[",
"'virtualbox'",
",",
"'virtualbox-qt'",
",",
"'virtualbox-dkms'",
",",
"'virtualbox-guest-dkms'",
",",
"'virtualbox-guest-additions-iso'",
",",
"]",
")",
"users",
"=",
"[",
"env",
".",
"user",
"]",
"for",
"username",
"in",
"users",
":",
"run",
"(",
"flo",
"(",
"'sudo adduser {username} vboxusers'",
")",
")"
] | Install a VirtualBox host system.
More Infos:
* overview: https://wiki.ubuntuusers.de/VirtualBox/
* installation: https://wiki.ubuntuusers.de/VirtualBox/Installation/ | [
"Install",
"a",
"VirtualBox",
"host",
"system",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L220-L238 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | pencil2 | def pencil2():
'''Install or update latest Pencil version 2, a GUI prototyping tool.
Tip: For svg exports displayed proper in other programs (eg. inkscape,
okular, reveal.js presentations) only use the 'Common Shapes' and
'Desktop - Sketchy GUI' elements.
More info:
github repo (forked version 2): https://github.com/prikhi/pencil
'''
repo_name = 'pencil2'
repo_dir = flo('~/repos/{repo_name}')
print_msg('## fetch latest pencil\n')
checkup_git_repo_legacy(url='https://github.com/prikhi/pencil.git',
name=repo_name)
print_msg('\n## build properties\n')
update_or_append_line(flo('{repo_dir}/build/properties.sh'),
prefix='export MAX_VERSION=',
new_line="export MAX_VERSION='100.*'")
run(flo('cat {repo_dir}/build/properties.sh'))
run(flo('cd {repo_dir}/build && ./build.sh linux'),
msg='\n## build pencil\n')
install_user_command_legacy('pencil2', pencil2_repodir=repo_dir)
print_msg('\nNow You can start pencil version 2 with this command:\n\n'
' pencil2') | python | def pencil2():
'''Install or update latest Pencil version 2, a GUI prototyping tool.
Tip: For svg exports displayed proper in other programs (eg. inkscape,
okular, reveal.js presentations) only use the 'Common Shapes' and
'Desktop - Sketchy GUI' elements.
More info:
github repo (forked version 2): https://github.com/prikhi/pencil
'''
repo_name = 'pencil2'
repo_dir = flo('~/repos/{repo_name}')
print_msg('## fetch latest pencil\n')
checkup_git_repo_legacy(url='https://github.com/prikhi/pencil.git',
name=repo_name)
print_msg('\n## build properties\n')
update_or_append_line(flo('{repo_dir}/build/properties.sh'),
prefix='export MAX_VERSION=',
new_line="export MAX_VERSION='100.*'")
run(flo('cat {repo_dir}/build/properties.sh'))
run(flo('cd {repo_dir}/build && ./build.sh linux'),
msg='\n## build pencil\n')
install_user_command_legacy('pencil2', pencil2_repodir=repo_dir)
print_msg('\nNow You can start pencil version 2 with this command:\n\n'
' pencil2') | [
"def",
"pencil2",
"(",
")",
":",
"repo_name",
"=",
"'pencil2'",
"repo_dir",
"=",
"flo",
"(",
"'~/repos/{repo_name}'",
")",
"print_msg",
"(",
"'## fetch latest pencil\\n'",
")",
"checkup_git_repo_legacy",
"(",
"url",
"=",
"'https://github.com/prikhi/pencil.git'",
",",
"name",
"=",
"repo_name",
")",
"print_msg",
"(",
"'\\n## build properties\\n'",
")",
"update_or_append_line",
"(",
"flo",
"(",
"'{repo_dir}/build/properties.sh'",
")",
",",
"prefix",
"=",
"'export MAX_VERSION='",
",",
"new_line",
"=",
"\"export MAX_VERSION='100.*'\"",
")",
"run",
"(",
"flo",
"(",
"'cat {repo_dir}/build/properties.sh'",
")",
")",
"run",
"(",
"flo",
"(",
"'cd {repo_dir}/build && ./build.sh linux'",
")",
",",
"msg",
"=",
"'\\n## build pencil\\n'",
")",
"install_user_command_legacy",
"(",
"'pencil2'",
",",
"pencil2_repodir",
"=",
"repo_dir",
")",
"print_msg",
"(",
"'\\nNow You can start pencil version 2 with this command:\\n\\n'",
"' pencil2'",
")"
] | Install or update latest Pencil version 2, a GUI prototyping tool.
Tip: For svg exports displayed proper in other programs (eg. inkscape,
okular, reveal.js presentations) only use the 'Common Shapes' and
'Desktop - Sketchy GUI' elements.
More info:
github repo (forked version 2): https://github.com/prikhi/pencil | [
"Install",
"or",
"update",
"latest",
"Pencil",
"version",
"2",
"a",
"GUI",
"prototyping",
"tool",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L277-L304 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | pencil3 | def pencil3():
'''Install or update latest Pencil version 3, a GUI prototyping tool.
While it is the newer one and the GUI is more fancy, it is the "more beta"
version of pencil. For exmaple, to display a svg export may fail from
within a reveal.js presentation.
More info:
Homepage: http://pencil.evolus.vn/Next.html
github repo: https://github.com/evolus/pencil
'''
repo_name = 'pencil3'
repo_dir = flo('~/repos/{repo_name}')
print_msg('## fetch latest pencil\n')
checkup_git_repo_legacy(url='https://github.com/evolus/pencil.git',
name=repo_name)
run(flo('cd {repo_dir} && npm install'), msg='\n## install npms\n')
install_user_command_legacy('pencil3', pencil3_repodir=repo_dir)
print_msg('\nNow You can start pencil version 3 with this command:\n\n'
' pencil3') | python | def pencil3():
'''Install or update latest Pencil version 3, a GUI prototyping tool.
While it is the newer one and the GUI is more fancy, it is the "more beta"
version of pencil. For exmaple, to display a svg export may fail from
within a reveal.js presentation.
More info:
Homepage: http://pencil.evolus.vn/Next.html
github repo: https://github.com/evolus/pencil
'''
repo_name = 'pencil3'
repo_dir = flo('~/repos/{repo_name}')
print_msg('## fetch latest pencil\n')
checkup_git_repo_legacy(url='https://github.com/evolus/pencil.git',
name=repo_name)
run(flo('cd {repo_dir} && npm install'), msg='\n## install npms\n')
install_user_command_legacy('pencil3', pencil3_repodir=repo_dir)
print_msg('\nNow You can start pencil version 3 with this command:\n\n'
' pencil3') | [
"def",
"pencil3",
"(",
")",
":",
"repo_name",
"=",
"'pencil3'",
"repo_dir",
"=",
"flo",
"(",
"'~/repos/{repo_name}'",
")",
"print_msg",
"(",
"'## fetch latest pencil\\n'",
")",
"checkup_git_repo_legacy",
"(",
"url",
"=",
"'https://github.com/evolus/pencil.git'",
",",
"name",
"=",
"repo_name",
")",
"run",
"(",
"flo",
"(",
"'cd {repo_dir} && npm install'",
")",
",",
"msg",
"=",
"'\\n## install npms\\n'",
")",
"install_user_command_legacy",
"(",
"'pencil3'",
",",
"pencil3_repodir",
"=",
"repo_dir",
")",
"print_msg",
"(",
"'\\nNow You can start pencil version 3 with this command:\\n\\n'",
"' pencil3'",
")"
] | Install or update latest Pencil version 3, a GUI prototyping tool.
While it is the newer one and the GUI is more fancy, it is the "more beta"
version of pencil. For exmaple, to display a svg export may fail from
within a reveal.js presentation.
More info:
Homepage: http://pencil.evolus.vn/Next.html
github repo: https://github.com/evolus/pencil | [
"Install",
"or",
"update",
"latest",
"Pencil",
"version",
"3",
"a",
"GUI",
"prototyping",
"tool",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L309-L328 | train |
theno/fabsetup | fabsetup/fabfile/setup/__init__.py | powerline_shell | def powerline_shell():
'''Install and set up powerline-shell prompt.
More infos:
* https://github.com/banga/powerline-shell
* https://github.com/ohnonot/powerline-shell
* https://askubuntu.com/questions/283908/how-can-i-install-and-use-powerline-plugin
'''
assert env.host == 'localhost', 'This task cannot run on a remote host'
# set up fonts for powerline
checkup_git_repo_legacy('https://github.com/powerline/fonts.git',
name='powerline-fonts')
run('cd ~/repos/powerline-fonts && ./install.sh')
# run('fc-cache -vf ~/.local/share/fonts')
prefix = 'URxvt*font: '
from config import fontlist
line = prefix + fontlist
update_or_append_line(filename='~/.Xresources', prefix=prefix,
new_line=line)
if env.host_string == 'localhost':
run('xrdb ~/.Xresources')
# set up powerline-shell
checkup_git_repo_legacy('https://github.com/banga/powerline-shell.git')
# checkup_git_repo_legacy('https://github.com/ohnonot/powerline-shell.git')
install_file_legacy(path='~/repos/powerline-shell/config.py')
run('cd ~/repos/powerline-shell && ./install.py')
question = 'Use normal question mark (u003F) for untracked files instead '\
'of fancy "black question mark ornament" (u2753, which may not work)?'
if query_yes_no(question, default='yes'):
filename = '~/repos/powerline-shell/powerline-shell.py'
update_or_append_line(filename, keep_backup=False,
prefix=" 'untracked': u'\u2753',",
new_line=" 'untracked': u'\u003F',")
run(flo('chmod u+x {filename}'))
bash_snippet = '~/.bashrc_powerline_shell'
install_file_legacy(path=bash_snippet)
prefix = flo('if [ -f {bash_snippet} ]; ')
enabler = flo('if [ -f {bash_snippet} ]; then source {bash_snippet}; fi')
uncomment_or_update_or_append_line(filename='~/.bashrc', prefix=prefix,
new_line=enabler) | python | def powerline_shell():
'''Install and set up powerline-shell prompt.
More infos:
* https://github.com/banga/powerline-shell
* https://github.com/ohnonot/powerline-shell
* https://askubuntu.com/questions/283908/how-can-i-install-and-use-powerline-plugin
'''
assert env.host == 'localhost', 'This task cannot run on a remote host'
# set up fonts for powerline
checkup_git_repo_legacy('https://github.com/powerline/fonts.git',
name='powerline-fonts')
run('cd ~/repos/powerline-fonts && ./install.sh')
# run('fc-cache -vf ~/.local/share/fonts')
prefix = 'URxvt*font: '
from config import fontlist
line = prefix + fontlist
update_or_append_line(filename='~/.Xresources', prefix=prefix,
new_line=line)
if env.host_string == 'localhost':
run('xrdb ~/.Xresources')
# set up powerline-shell
checkup_git_repo_legacy('https://github.com/banga/powerline-shell.git')
# checkup_git_repo_legacy('https://github.com/ohnonot/powerline-shell.git')
install_file_legacy(path='~/repos/powerline-shell/config.py')
run('cd ~/repos/powerline-shell && ./install.py')
question = 'Use normal question mark (u003F) for untracked files instead '\
'of fancy "black question mark ornament" (u2753, which may not work)?'
if query_yes_no(question, default='yes'):
filename = '~/repos/powerline-shell/powerline-shell.py'
update_or_append_line(filename, keep_backup=False,
prefix=" 'untracked': u'\u2753',",
new_line=" 'untracked': u'\u003F',")
run(flo('chmod u+x {filename}'))
bash_snippet = '~/.bashrc_powerline_shell'
install_file_legacy(path=bash_snippet)
prefix = flo('if [ -f {bash_snippet} ]; ')
enabler = flo('if [ -f {bash_snippet} ]; then source {bash_snippet}; fi')
uncomment_or_update_or_append_line(filename='~/.bashrc', prefix=prefix,
new_line=enabler) | [
"def",
"powerline_shell",
"(",
")",
":",
"assert",
"env",
".",
"host",
"==",
"'localhost'",
",",
"'This task cannot run on a remote host'",
"# set up fonts for powerline",
"checkup_git_repo_legacy",
"(",
"'https://github.com/powerline/fonts.git'",
",",
"name",
"=",
"'powerline-fonts'",
")",
"run",
"(",
"'cd ~/repos/powerline-fonts && ./install.sh'",
")",
"# run('fc-cache -vf ~/.local/share/fonts')",
"prefix",
"=",
"'URxvt*font: '",
"from",
"config",
"import",
"fontlist",
"line",
"=",
"prefix",
"+",
"fontlist",
"update_or_append_line",
"(",
"filename",
"=",
"'~/.Xresources'",
",",
"prefix",
"=",
"prefix",
",",
"new_line",
"=",
"line",
")",
"if",
"env",
".",
"host_string",
"==",
"'localhost'",
":",
"run",
"(",
"'xrdb ~/.Xresources'",
")",
"# set up powerline-shell",
"checkup_git_repo_legacy",
"(",
"'https://github.com/banga/powerline-shell.git'",
")",
"# checkup_git_repo_legacy('https://github.com/ohnonot/powerline-shell.git')",
"install_file_legacy",
"(",
"path",
"=",
"'~/repos/powerline-shell/config.py'",
")",
"run",
"(",
"'cd ~/repos/powerline-shell && ./install.py'",
")",
"question",
"=",
"'Use normal question mark (u003F) for untracked files instead '",
"'of fancy \"black question mark ornament\" (u2753, which may not work)?'",
"if",
"query_yes_no",
"(",
"question",
",",
"default",
"=",
"'yes'",
")",
":",
"filename",
"=",
"'~/repos/powerline-shell/powerline-shell.py'",
"update_or_append_line",
"(",
"filename",
",",
"keep_backup",
"=",
"False",
",",
"prefix",
"=",
"\" 'untracked': u'\\u2753',\"",
",",
"new_line",
"=",
"\" 'untracked': u'\\u003F',\"",
")",
"run",
"(",
"flo",
"(",
"'chmod u+x {filename}'",
")",
")",
"bash_snippet",
"=",
"'~/.bashrc_powerline_shell'",
"install_file_legacy",
"(",
"path",
"=",
"bash_snippet",
")",
"prefix",
"=",
"flo",
"(",
"'if [ -f {bash_snippet} ]; '",
")",
"enabler",
"=",
"flo",
"(",
"'if [ -f {bash_snippet} ]; then source {bash_snippet}; fi'",
")",
"uncomment_or_update_or_append_line",
"(",
"filename",
"=",
"'~/.bashrc'",
",",
"prefix",
"=",
"prefix",
",",
"new_line",
"=",
"enabler",
")"
] | Install and set up powerline-shell prompt.
More infos:
* https://github.com/banga/powerline-shell
* https://github.com/ohnonot/powerline-shell
* https://askubuntu.com/questions/283908/how-can-i-install-and-use-powerline-plugin | [
"Install",
"and",
"set",
"up",
"powerline",
"-",
"shell",
"prompt",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/__init__.py#L378-L423 | train |
muckamuck/stackility | stackility/drift.py | DriftTool._init_boto3_clients | def _init_boto3_clients(self, profile, region):
"""
The utililty requires boto3 clients to CloudFormation.
Args:
None
Returns:
Good or Bad; True or False
"""
try:
session = None
if profile and region:
session = boto3.session.Session(profile_name=profile, region_name=region)
elif profile:
session = boto3.session.Session(profile_name=profile)
elif region:
session = boto3.session.Session(region_name=region)
else:
session = boto3.session.Session()
self._cloud_formation = session.client('cloudformation')
return True
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False | python | def _init_boto3_clients(self, profile, region):
"""
The utililty requires boto3 clients to CloudFormation.
Args:
None
Returns:
Good or Bad; True or False
"""
try:
session = None
if profile and region:
session = boto3.session.Session(profile_name=profile, region_name=region)
elif profile:
session = boto3.session.Session(profile_name=profile)
elif region:
session = boto3.session.Session(region_name=region)
else:
session = boto3.session.Session()
self._cloud_formation = session.client('cloudformation')
return True
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False | [
"def",
"_init_boto3_clients",
"(",
"self",
",",
"profile",
",",
"region",
")",
":",
"try",
":",
"session",
"=",
"None",
"if",
"profile",
"and",
"region",
":",
"session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
"profile_name",
"=",
"profile",
",",
"region_name",
"=",
"region",
")",
"elif",
"profile",
":",
"session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
"profile_name",
"=",
"profile",
")",
"elif",
"region",
":",
"session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
"region_name",
"=",
"region",
")",
"else",
":",
"session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
")",
"self",
".",
"_cloud_formation",
"=",
"session",
".",
"client",
"(",
"'cloudformation'",
")",
"return",
"True",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"error",
"(",
"wtf",
",",
"exc_info",
"=",
"True",
")",
"return",
"False"
] | The utililty requires boto3 clients to CloudFormation.
Args:
None
Returns:
Good or Bad; True or False | [
"The",
"utililty",
"requires",
"boto3",
"clients",
"to",
"CloudFormation",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/drift.py#L54-L79 | train |
muckamuck/stackility | stackility/drift.py | DriftTool.determine_drift | def determine_drift(self):
"""
Determine the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False
"""
try:
response = self._cloud_formation.detect_stack_drift(StackName=self._stack_name)
drift_request_id = response.get('StackDriftDetectionId', None)
if drift_request_id:
logging.info('drift_request_id: %s - polling', drift_request_id)
drift_calc_done = False
while not drift_calc_done:
time.sleep(self.nap_time)
response = self._cloud_formation.describe_stack_drift_detection_status(
StackDriftDetectionId=drift_request_id
)
current_state = response.get('DetectionStatus', None)
logging.info(
'describe_stack_drift_detection_status(): {}'.format(current_state)
)
drift_calc_done = current_state in CALC_DONE_STATES
drift_answer = response.get('StackDriftStatus', 'UNKNOWN')
logging.info('drift of {}: {}'.format(
self._stack_name,
drift_answer
))
if drift_answer == 'DRIFTED':
if self._verbose:
self._print_drift_report()
return False
else:
return True
else:
logging.warning('drift_request_id is None')
return False
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False | python | def determine_drift(self):
"""
Determine the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False
"""
try:
response = self._cloud_formation.detect_stack_drift(StackName=self._stack_name)
drift_request_id = response.get('StackDriftDetectionId', None)
if drift_request_id:
logging.info('drift_request_id: %s - polling', drift_request_id)
drift_calc_done = False
while not drift_calc_done:
time.sleep(self.nap_time)
response = self._cloud_formation.describe_stack_drift_detection_status(
StackDriftDetectionId=drift_request_id
)
current_state = response.get('DetectionStatus', None)
logging.info(
'describe_stack_drift_detection_status(): {}'.format(current_state)
)
drift_calc_done = current_state in CALC_DONE_STATES
drift_answer = response.get('StackDriftStatus', 'UNKNOWN')
logging.info('drift of {}: {}'.format(
self._stack_name,
drift_answer
))
if drift_answer == 'DRIFTED':
if self._verbose:
self._print_drift_report()
return False
else:
return True
else:
logging.warning('drift_request_id is None')
return False
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False | [
"def",
"determine_drift",
"(",
"self",
")",
":",
"try",
":",
"response",
"=",
"self",
".",
"_cloud_formation",
".",
"detect_stack_drift",
"(",
"StackName",
"=",
"self",
".",
"_stack_name",
")",
"drift_request_id",
"=",
"response",
".",
"get",
"(",
"'StackDriftDetectionId'",
",",
"None",
")",
"if",
"drift_request_id",
":",
"logging",
".",
"info",
"(",
"'drift_request_id: %s - polling'",
",",
"drift_request_id",
")",
"drift_calc_done",
"=",
"False",
"while",
"not",
"drift_calc_done",
":",
"time",
".",
"sleep",
"(",
"self",
".",
"nap_time",
")",
"response",
"=",
"self",
".",
"_cloud_formation",
".",
"describe_stack_drift_detection_status",
"(",
"StackDriftDetectionId",
"=",
"drift_request_id",
")",
"current_state",
"=",
"response",
".",
"get",
"(",
"'DetectionStatus'",
",",
"None",
")",
"logging",
".",
"info",
"(",
"'describe_stack_drift_detection_status(): {}'",
".",
"format",
"(",
"current_state",
")",
")",
"drift_calc_done",
"=",
"current_state",
"in",
"CALC_DONE_STATES",
"drift_answer",
"=",
"response",
".",
"get",
"(",
"'StackDriftStatus'",
",",
"'UNKNOWN'",
")",
"logging",
".",
"info",
"(",
"'drift of {}: {}'",
".",
"format",
"(",
"self",
".",
"_stack_name",
",",
"drift_answer",
")",
")",
"if",
"drift_answer",
"==",
"'DRIFTED'",
":",
"if",
"self",
".",
"_verbose",
":",
"self",
".",
"_print_drift_report",
"(",
")",
"return",
"False",
"else",
":",
"return",
"True",
"else",
":",
"logging",
".",
"warning",
"(",
"'drift_request_id is None'",
")",
"return",
"False",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"error",
"(",
"wtf",
",",
"exc_info",
"=",
"True",
")",
"return",
"False"
] | Determine the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False | [
"Determine",
"the",
"drift",
"of",
"the",
"stack",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/drift.py#L81-L126 | train |
muckamuck/stackility | stackility/drift.py | DriftTool._print_drift_report | def _print_drift_report(self):
"""
Report the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False
Note: not yet implemented
"""
try:
response = self._cloud_formation.describe_stack_resources(StackName=self._stack_name)
rows = []
for resource in response.get('StackResources', []):
row = []
row.append(resource.get('LogicalResourceId', 'unknown'))
row.append(resource.get('PhysicalResourceId', 'unknown'))
row.append(resource.get('ResourceStatus', 'unknown'))
row.append(resource.get('DriftInformation', {}).get('StackResourceDriftStatus', 'unknown'))
rows.append(row)
print('Drift Report:')
print(tabulate(rows, headers=[
'Logical ID',
'Physical ID',
'Resource Status',
'Drift Info'
]))
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False
return True | python | def _print_drift_report(self):
"""
Report the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False
Note: not yet implemented
"""
try:
response = self._cloud_formation.describe_stack_resources(StackName=self._stack_name)
rows = []
for resource in response.get('StackResources', []):
row = []
row.append(resource.get('LogicalResourceId', 'unknown'))
row.append(resource.get('PhysicalResourceId', 'unknown'))
row.append(resource.get('ResourceStatus', 'unknown'))
row.append(resource.get('DriftInformation', {}).get('StackResourceDriftStatus', 'unknown'))
rows.append(row)
print('Drift Report:')
print(tabulate(rows, headers=[
'Logical ID',
'Physical ID',
'Resource Status',
'Drift Info'
]))
except Exception as wtf:
logging.error(wtf, exc_info=True)
return False
return True | [
"def",
"_print_drift_report",
"(",
"self",
")",
":",
"try",
":",
"response",
"=",
"self",
".",
"_cloud_formation",
".",
"describe_stack_resources",
"(",
"StackName",
"=",
"self",
".",
"_stack_name",
")",
"rows",
"=",
"[",
"]",
"for",
"resource",
"in",
"response",
".",
"get",
"(",
"'StackResources'",
",",
"[",
"]",
")",
":",
"row",
"=",
"[",
"]",
"row",
".",
"append",
"(",
"resource",
".",
"get",
"(",
"'LogicalResourceId'",
",",
"'unknown'",
")",
")",
"row",
".",
"append",
"(",
"resource",
".",
"get",
"(",
"'PhysicalResourceId'",
",",
"'unknown'",
")",
")",
"row",
".",
"append",
"(",
"resource",
".",
"get",
"(",
"'ResourceStatus'",
",",
"'unknown'",
")",
")",
"row",
".",
"append",
"(",
"resource",
".",
"get",
"(",
"'DriftInformation'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'StackResourceDriftStatus'",
",",
"'unknown'",
")",
")",
"rows",
".",
"append",
"(",
"row",
")",
"print",
"(",
"'Drift Report:'",
")",
"print",
"(",
"tabulate",
"(",
"rows",
",",
"headers",
"=",
"[",
"'Logical ID'",
",",
"'Physical ID'",
",",
"'Resource Status'",
",",
"'Drift Info'",
"]",
")",
")",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"error",
"(",
"wtf",
",",
"exc_info",
"=",
"True",
")",
"return",
"False",
"return",
"True"
] | Report the drift of the stack.
Args:
None
Returns:
Good or Bad; True or False
Note: not yet implemented | [
"Report",
"the",
"drift",
"of",
"the",
"stack",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/drift.py#L128-L162 | train |
jaraco/jaraco.windows | jaraco/windows/dpapi.py | DATA_BLOB.set_data | def set_data(self, data):
"Use this method to set the data for this blob"
if data is None:
self.data_size = 0
self.data = None
return
self.data_size = len(data)
# create a string buffer so that null bytes aren't interpreted
# as the end of the string
self.data = ctypes.cast(ctypes.create_string_buffer(data), ctypes.c_void_p) | python | def set_data(self, data):
"Use this method to set the data for this blob"
if data is None:
self.data_size = 0
self.data = None
return
self.data_size = len(data)
# create a string buffer so that null bytes aren't interpreted
# as the end of the string
self.data = ctypes.cast(ctypes.create_string_buffer(data), ctypes.c_void_p) | [
"def",
"set_data",
"(",
"self",
",",
"data",
")",
":",
"if",
"data",
"is",
"None",
":",
"self",
".",
"data_size",
"=",
"0",
"self",
".",
"data",
"=",
"None",
"return",
"self",
".",
"data_size",
"=",
"len",
"(",
"data",
")",
"# create a string buffer so that null bytes aren't interpreted",
"# as the end of the string",
"self",
".",
"data",
"=",
"ctypes",
".",
"cast",
"(",
"ctypes",
".",
"create_string_buffer",
"(",
"data",
")",
",",
"ctypes",
".",
"c_void_p",
")"
] | Use this method to set the data for this blob | [
"Use",
"this",
"method",
"to",
"set",
"the",
"data",
"for",
"this",
"blob"
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/dpapi.py#L43-L52 | train |
jaraco/jaraco.windows | jaraco/windows/dpapi.py | DATA_BLOB.get_data | def get_data(self):
"Get the data for this blob"
array = ctypes.POINTER(ctypes.c_char * len(self))
return ctypes.cast(self.data, array).contents.raw | python | def get_data(self):
"Get the data for this blob"
array = ctypes.POINTER(ctypes.c_char * len(self))
return ctypes.cast(self.data, array).contents.raw | [
"def",
"get_data",
"(",
"self",
")",
":",
"array",
"=",
"ctypes",
".",
"POINTER",
"(",
"ctypes",
".",
"c_char",
"*",
"len",
"(",
"self",
")",
")",
"return",
"ctypes",
".",
"cast",
"(",
"self",
".",
"data",
",",
"array",
")",
".",
"contents",
".",
"raw"
] | Get the data for this blob | [
"Get",
"the",
"data",
"for",
"this",
"blob"
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/dpapi.py#L54-L57 | train |
sbmlteam/libCombine | examples/python/printExample.py | printMetaDataFor | def printMetaDataFor(archive, location):
""" Prints metadata for given location.
:param archive: CombineArchive instance
:param location:
:return:
"""
desc = archive.getMetadataForLocation(location)
if desc.isEmpty():
print(" no metadata for '{0}'".format(location))
return None
print(" metadata for '{0}':".format(location))
print(" Created : {0}".format(desc.getCreated().getDateAsString()))
for i in range(desc.getNumModified()):
print(" Modified : {0}".format(desc.getModified(i).getDateAsString()))
print(" # Creators: {0}".format(desc.getNumCreators()))
for i in range(desc.getNumCreators()):
creator = desc.getCreator(i)
print(" {0} {1}".format(creator.getGivenName(), creator.getFamilyName())) | python | def printMetaDataFor(archive, location):
""" Prints metadata for given location.
:param archive: CombineArchive instance
:param location:
:return:
"""
desc = archive.getMetadataForLocation(location)
if desc.isEmpty():
print(" no metadata for '{0}'".format(location))
return None
print(" metadata for '{0}':".format(location))
print(" Created : {0}".format(desc.getCreated().getDateAsString()))
for i in range(desc.getNumModified()):
print(" Modified : {0}".format(desc.getModified(i).getDateAsString()))
print(" # Creators: {0}".format(desc.getNumCreators()))
for i in range(desc.getNumCreators()):
creator = desc.getCreator(i)
print(" {0} {1}".format(creator.getGivenName(), creator.getFamilyName())) | [
"def",
"printMetaDataFor",
"(",
"archive",
",",
"location",
")",
":",
"desc",
"=",
"archive",
".",
"getMetadataForLocation",
"(",
"location",
")",
"if",
"desc",
".",
"isEmpty",
"(",
")",
":",
"print",
"(",
"\" no metadata for '{0}'\"",
".",
"format",
"(",
"location",
")",
")",
"return",
"None",
"print",
"(",
"\" metadata for '{0}':\"",
".",
"format",
"(",
"location",
")",
")",
"print",
"(",
"\" Created : {0}\"",
".",
"format",
"(",
"desc",
".",
"getCreated",
"(",
")",
".",
"getDateAsString",
"(",
")",
")",
")",
"for",
"i",
"in",
"range",
"(",
"desc",
".",
"getNumModified",
"(",
")",
")",
":",
"print",
"(",
"\" Modified : {0}\"",
".",
"format",
"(",
"desc",
".",
"getModified",
"(",
"i",
")",
".",
"getDateAsString",
"(",
")",
")",
")",
"print",
"(",
"\" # Creators: {0}\"",
".",
"format",
"(",
"desc",
".",
"getNumCreators",
"(",
")",
")",
")",
"for",
"i",
"in",
"range",
"(",
"desc",
".",
"getNumCreators",
"(",
")",
")",
":",
"creator",
"=",
"desc",
".",
"getCreator",
"(",
"i",
")",
"print",
"(",
"\" {0} {1}\"",
".",
"format",
"(",
"creator",
".",
"getGivenName",
"(",
")",
",",
"creator",
".",
"getFamilyName",
"(",
")",
")",
")"
] | Prints metadata for given location.
:param archive: CombineArchive instance
:param location:
:return: | [
"Prints",
"metadata",
"for",
"given",
"location",
"."
] | d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4 | https://github.com/sbmlteam/libCombine/blob/d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4/examples/python/printExample.py#L11-L31 | train |
sbmlteam/libCombine | examples/python/printExample.py | printArchive | def printArchive(fileName):
""" Prints content of combine archive
:param fileName: path of archive
:return: None
"""
archive = CombineArchive()
if archive.initializeFromArchive(fileName) is None:
print("Invalid Combine Archive")
return None
print('*'*80)
print('Print archive:', fileName)
print('*' * 80)
printMetaDataFor(archive, ".")
print("Num Entries: {0}".format(archive.getNumEntries()))
for i in range(archive.getNumEntries()):
entry = archive.getEntry(i)
print(" {0}: location: {1} format: {2}".format(i, entry.getLocation(), entry.getFormat()))
printMetaDataFor(archive, entry.getLocation())
for j in range(entry.getNumCrossRefs()):
print(" {0}: crossRef location {1}".format(j, entry.getCrossRef(j).getLocation()))
# the entry could now be extracted via
# archive.extractEntry(entry.getLocation(), <filename or folder>)
# or used as string
# content = archive.extractEntryToString(entry.getLocation());
archive.cleanUp() | python | def printArchive(fileName):
""" Prints content of combine archive
:param fileName: path of archive
:return: None
"""
archive = CombineArchive()
if archive.initializeFromArchive(fileName) is None:
print("Invalid Combine Archive")
return None
print('*'*80)
print('Print archive:', fileName)
print('*' * 80)
printMetaDataFor(archive, ".")
print("Num Entries: {0}".format(archive.getNumEntries()))
for i in range(archive.getNumEntries()):
entry = archive.getEntry(i)
print(" {0}: location: {1} format: {2}".format(i, entry.getLocation(), entry.getFormat()))
printMetaDataFor(archive, entry.getLocation())
for j in range(entry.getNumCrossRefs()):
print(" {0}: crossRef location {1}".format(j, entry.getCrossRef(j).getLocation()))
# the entry could now be extracted via
# archive.extractEntry(entry.getLocation(), <filename or folder>)
# or used as string
# content = archive.extractEntryToString(entry.getLocation());
archive.cleanUp() | [
"def",
"printArchive",
"(",
"fileName",
")",
":",
"archive",
"=",
"CombineArchive",
"(",
")",
"if",
"archive",
".",
"initializeFromArchive",
"(",
"fileName",
")",
"is",
"None",
":",
"print",
"(",
"\"Invalid Combine Archive\"",
")",
"return",
"None",
"print",
"(",
"'*'",
"*",
"80",
")",
"print",
"(",
"'Print archive:'",
",",
"fileName",
")",
"print",
"(",
"'*'",
"*",
"80",
")",
"printMetaDataFor",
"(",
"archive",
",",
"\".\"",
")",
"print",
"(",
"\"Num Entries: {0}\"",
".",
"format",
"(",
"archive",
".",
"getNumEntries",
"(",
")",
")",
")",
"for",
"i",
"in",
"range",
"(",
"archive",
".",
"getNumEntries",
"(",
")",
")",
":",
"entry",
"=",
"archive",
".",
"getEntry",
"(",
"i",
")",
"print",
"(",
"\" {0}: location: {1} format: {2}\"",
".",
"format",
"(",
"i",
",",
"entry",
".",
"getLocation",
"(",
")",
",",
"entry",
".",
"getFormat",
"(",
")",
")",
")",
"printMetaDataFor",
"(",
"archive",
",",
"entry",
".",
"getLocation",
"(",
")",
")",
"for",
"j",
"in",
"range",
"(",
"entry",
".",
"getNumCrossRefs",
"(",
")",
")",
":",
"print",
"(",
"\" {0}: crossRef location {1}\"",
".",
"format",
"(",
"j",
",",
"entry",
".",
"getCrossRef",
"(",
"j",
")",
".",
"getLocation",
"(",
")",
")",
")",
"# the entry could now be extracted via",
"# archive.extractEntry(entry.getLocation(), <filename or folder>)",
"# or used as string",
"# content = archive.extractEntryToString(entry.getLocation());",
"archive",
".",
"cleanUp",
"(",
")"
] | Prints content of combine archive
:param fileName: path of archive
:return: None | [
"Prints",
"content",
"of",
"combine",
"archive"
] | d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4 | https://github.com/sbmlteam/libCombine/blob/d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4/examples/python/printExample.py#L34-L65 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | mklink | def mklink():
"""
Like cmd.exe's mklink except it will infer directory status of the
target.
"""
from optparse import OptionParser
parser = OptionParser(usage="usage: %prog [options] link target")
parser.add_option(
'-d', '--directory',
help="Target is a directory (only necessary if not present)",
action="store_true")
options, args = parser.parse_args()
try:
link, target = args
except ValueError:
parser.error("incorrect number of arguments")
symlink(target, link, options.directory)
sys.stdout.write("Symbolic link created: %(link)s --> %(target)s\n" % vars()) | python | def mklink():
"""
Like cmd.exe's mklink except it will infer directory status of the
target.
"""
from optparse import OptionParser
parser = OptionParser(usage="usage: %prog [options] link target")
parser.add_option(
'-d', '--directory',
help="Target is a directory (only necessary if not present)",
action="store_true")
options, args = parser.parse_args()
try:
link, target = args
except ValueError:
parser.error("incorrect number of arguments")
symlink(target, link, options.directory)
sys.stdout.write("Symbolic link created: %(link)s --> %(target)s\n" % vars()) | [
"def",
"mklink",
"(",
")",
":",
"from",
"optparse",
"import",
"OptionParser",
"parser",
"=",
"OptionParser",
"(",
"usage",
"=",
"\"usage: %prog [options] link target\"",
")",
"parser",
".",
"add_option",
"(",
"'-d'",
",",
"'--directory'",
",",
"help",
"=",
"\"Target is a directory (only necessary if not present)\"",
",",
"action",
"=",
"\"store_true\"",
")",
"options",
",",
"args",
"=",
"parser",
".",
"parse_args",
"(",
")",
"try",
":",
"link",
",",
"target",
"=",
"args",
"except",
"ValueError",
":",
"parser",
".",
"error",
"(",
"\"incorrect number of arguments\"",
")",
"symlink",
"(",
"target",
",",
"link",
",",
"options",
".",
"directory",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"\"Symbolic link created: %(link)s --> %(target)s\\n\"",
"%",
"vars",
"(",
")",
")"
] | Like cmd.exe's mklink except it will infer directory status of the
target. | [
"Like",
"cmd",
".",
"exe",
"s",
"mklink",
"except",
"it",
"will",
"infer",
"directory",
"status",
"of",
"the",
"target",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L28-L45 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | is_reparse_point | def is_reparse_point(path):
"""
Determine if the given path is a reparse point.
Return False if the file does not exist or the file attributes cannot
be determined.
"""
res = api.GetFileAttributes(path)
return (
res != api.INVALID_FILE_ATTRIBUTES
and bool(res & api.FILE_ATTRIBUTE_REPARSE_POINT)
) | python | def is_reparse_point(path):
"""
Determine if the given path is a reparse point.
Return False if the file does not exist or the file attributes cannot
be determined.
"""
res = api.GetFileAttributes(path)
return (
res != api.INVALID_FILE_ATTRIBUTES
and bool(res & api.FILE_ATTRIBUTE_REPARSE_POINT)
) | [
"def",
"is_reparse_point",
"(",
"path",
")",
":",
"res",
"=",
"api",
".",
"GetFileAttributes",
"(",
"path",
")",
"return",
"(",
"res",
"!=",
"api",
".",
"INVALID_FILE_ATTRIBUTES",
"and",
"bool",
"(",
"res",
"&",
"api",
".",
"FILE_ATTRIBUTE_REPARSE_POINT",
")",
")"
] | Determine if the given path is a reparse point.
Return False if the file does not exist or the file attributes cannot
be determined. | [
"Determine",
"if",
"the",
"given",
"path",
"is",
"a",
"reparse",
"point",
".",
"Return",
"False",
"if",
"the",
"file",
"does",
"not",
"exist",
"or",
"the",
"file",
"attributes",
"cannot",
"be",
"determined",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L78-L88 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | is_symlink | def is_symlink(path):
"""
Assuming path is a reparse point, determine if it's a symlink.
"""
path = _patch_path(path)
try:
return _is_symlink(next(find_files(path)))
# comment below workaround for PyCQA/pyflakes#376
except WindowsError as orig_error: # noqa: F841
tmpl = "Error accessing {path}: {orig_error.message}"
raise builtins.WindowsError(tmpl.format(**locals())) | python | def is_symlink(path):
"""
Assuming path is a reparse point, determine if it's a symlink.
"""
path = _patch_path(path)
try:
return _is_symlink(next(find_files(path)))
# comment below workaround for PyCQA/pyflakes#376
except WindowsError as orig_error: # noqa: F841
tmpl = "Error accessing {path}: {orig_error.message}"
raise builtins.WindowsError(tmpl.format(**locals())) | [
"def",
"is_symlink",
"(",
"path",
")",
":",
"path",
"=",
"_patch_path",
"(",
"path",
")",
"try",
":",
"return",
"_is_symlink",
"(",
"next",
"(",
"find_files",
"(",
"path",
")",
")",
")",
"# comment below workaround for PyCQA/pyflakes#376",
"except",
"WindowsError",
"as",
"orig_error",
":",
"# noqa: F841",
"tmpl",
"=",
"\"Error accessing {path}: {orig_error.message}\"",
"raise",
"builtins",
".",
"WindowsError",
"(",
"tmpl",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")"
] | Assuming path is a reparse point, determine if it's a symlink. | [
"Assuming",
"path",
"is",
"a",
"reparse",
"point",
"determine",
"if",
"it",
"s",
"a",
"symlink",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L113-L123 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | get_final_path | def get_final_path(path):
r"""
For a given path, determine the ultimate location of that path.
Useful for resolving symlink targets.
This functions wraps the GetFinalPathNameByHandle from the Windows
SDK.
Note, this function fails if a handle cannot be obtained (such as
for C:\Pagefile.sys on a stock windows system). Consider using
trace_symlink_target instead.
"""
desired_access = api.NULL
share_mode = (
api.FILE_SHARE_READ | api.FILE_SHARE_WRITE | api.FILE_SHARE_DELETE
)
security_attributes = api.LPSECURITY_ATTRIBUTES() # NULL pointer
hFile = api.CreateFile(
path,
desired_access,
share_mode,
security_attributes,
api.OPEN_EXISTING,
api.FILE_FLAG_BACKUP_SEMANTICS,
api.NULL,
)
if hFile == api.INVALID_HANDLE_VALUE:
raise WindowsError()
buf_size = api.GetFinalPathNameByHandle(
hFile, LPWSTR(), 0, api.VOLUME_NAME_DOS)
handle_nonzero_success(buf_size)
buf = create_unicode_buffer(buf_size)
result_length = api.GetFinalPathNameByHandle(
hFile, buf, len(buf), api.VOLUME_NAME_DOS)
assert result_length < len(buf)
handle_nonzero_success(result_length)
handle_nonzero_success(api.CloseHandle(hFile))
return buf[:result_length] | python | def get_final_path(path):
r"""
For a given path, determine the ultimate location of that path.
Useful for resolving symlink targets.
This functions wraps the GetFinalPathNameByHandle from the Windows
SDK.
Note, this function fails if a handle cannot be obtained (such as
for C:\Pagefile.sys on a stock windows system). Consider using
trace_symlink_target instead.
"""
desired_access = api.NULL
share_mode = (
api.FILE_SHARE_READ | api.FILE_SHARE_WRITE | api.FILE_SHARE_DELETE
)
security_attributes = api.LPSECURITY_ATTRIBUTES() # NULL pointer
hFile = api.CreateFile(
path,
desired_access,
share_mode,
security_attributes,
api.OPEN_EXISTING,
api.FILE_FLAG_BACKUP_SEMANTICS,
api.NULL,
)
if hFile == api.INVALID_HANDLE_VALUE:
raise WindowsError()
buf_size = api.GetFinalPathNameByHandle(
hFile, LPWSTR(), 0, api.VOLUME_NAME_DOS)
handle_nonzero_success(buf_size)
buf = create_unicode_buffer(buf_size)
result_length = api.GetFinalPathNameByHandle(
hFile, buf, len(buf), api.VOLUME_NAME_DOS)
assert result_length < len(buf)
handle_nonzero_success(result_length)
handle_nonzero_success(api.CloseHandle(hFile))
return buf[:result_length] | [
"def",
"get_final_path",
"(",
"path",
")",
":",
"desired_access",
"=",
"api",
".",
"NULL",
"share_mode",
"=",
"(",
"api",
".",
"FILE_SHARE_READ",
"|",
"api",
".",
"FILE_SHARE_WRITE",
"|",
"api",
".",
"FILE_SHARE_DELETE",
")",
"security_attributes",
"=",
"api",
".",
"LPSECURITY_ATTRIBUTES",
"(",
")",
"# NULL pointer",
"hFile",
"=",
"api",
".",
"CreateFile",
"(",
"path",
",",
"desired_access",
",",
"share_mode",
",",
"security_attributes",
",",
"api",
".",
"OPEN_EXISTING",
",",
"api",
".",
"FILE_FLAG_BACKUP_SEMANTICS",
",",
"api",
".",
"NULL",
",",
")",
"if",
"hFile",
"==",
"api",
".",
"INVALID_HANDLE_VALUE",
":",
"raise",
"WindowsError",
"(",
")",
"buf_size",
"=",
"api",
".",
"GetFinalPathNameByHandle",
"(",
"hFile",
",",
"LPWSTR",
"(",
")",
",",
"0",
",",
"api",
".",
"VOLUME_NAME_DOS",
")",
"handle_nonzero_success",
"(",
"buf_size",
")",
"buf",
"=",
"create_unicode_buffer",
"(",
"buf_size",
")",
"result_length",
"=",
"api",
".",
"GetFinalPathNameByHandle",
"(",
"hFile",
",",
"buf",
",",
"len",
"(",
"buf",
")",
",",
"api",
".",
"VOLUME_NAME_DOS",
")",
"assert",
"result_length",
"<",
"len",
"(",
"buf",
")",
"handle_nonzero_success",
"(",
"result_length",
")",
"handle_nonzero_success",
"(",
"api",
".",
"CloseHandle",
"(",
"hFile",
")",
")",
"return",
"buf",
"[",
":",
"result_length",
"]"
] | r"""
For a given path, determine the ultimate location of that path.
Useful for resolving symlink targets.
This functions wraps the GetFinalPathNameByHandle from the Windows
SDK.
Note, this function fails if a handle cannot be obtained (such as
for C:\Pagefile.sys on a stock windows system). Consider using
trace_symlink_target instead. | [
"r",
"For",
"a",
"given",
"path",
"determine",
"the",
"ultimate",
"location",
"of",
"that",
"path",
".",
"Useful",
"for",
"resolving",
"symlink",
"targets",
".",
"This",
"functions",
"wraps",
"the",
"GetFinalPathNameByHandle",
"from",
"the",
"Windows",
"SDK",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L163-L203 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | join | def join(*paths):
r"""
Wrapper around os.path.join that works with Windows drive letters.
>>> join('d:\\foo', '\\bar')
'd:\\bar'
"""
paths_with_drives = map(os.path.splitdrive, paths)
drives, paths = zip(*paths_with_drives)
# the drive we care about is the last one in the list
drive = next(filter(None, reversed(drives)), '')
return os.path.join(drive, os.path.join(*paths)) | python | def join(*paths):
r"""
Wrapper around os.path.join that works with Windows drive letters.
>>> join('d:\\foo', '\\bar')
'd:\\bar'
"""
paths_with_drives = map(os.path.splitdrive, paths)
drives, paths = zip(*paths_with_drives)
# the drive we care about is the last one in the list
drive = next(filter(None, reversed(drives)), '')
return os.path.join(drive, os.path.join(*paths)) | [
"def",
"join",
"(",
"*",
"paths",
")",
":",
"paths_with_drives",
"=",
"map",
"(",
"os",
".",
"path",
".",
"splitdrive",
",",
"paths",
")",
"drives",
",",
"paths",
"=",
"zip",
"(",
"*",
"paths_with_drives",
")",
"# the drive we care about is the last one in the list",
"drive",
"=",
"next",
"(",
"filter",
"(",
"None",
",",
"reversed",
"(",
"drives",
")",
")",
",",
"''",
")",
"return",
"os",
".",
"path",
".",
"join",
"(",
"drive",
",",
"os",
".",
"path",
".",
"join",
"(",
"*",
"paths",
")",
")"
] | r"""
Wrapper around os.path.join that works with Windows drive letters.
>>> join('d:\\foo', '\\bar')
'd:\\bar' | [
"r",
"Wrapper",
"around",
"os",
".",
"path",
".",
"join",
"that",
"works",
"with",
"Windows",
"drive",
"letters",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L292-L303 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | resolve_path | def resolve_path(target, start=os.path.curdir):
r"""
Find a path from start to target where target is relative to start.
>>> tmp = str(getfixture('tmpdir_as_cwd'))
>>> findpath('d:\\')
'd:\\'
>>> findpath('d:\\', tmp)
'd:\\'
>>> findpath('\\bar', 'd:\\')
'd:\\bar'
>>> findpath('\\bar', 'd:\\foo') # fails with '\\bar'
'd:\\bar'
>>> findpath('bar', 'd:\\foo')
'd:\\foo\\bar'
>>> findpath('\\baz', 'd:\\foo\\bar') # fails with '\\baz'
'd:\\baz'
>>> os.path.abspath(findpath('\\bar')).lower()
'c:\\bar'
>>> os.path.abspath(findpath('bar'))
'...\\bar'
>>> findpath('..', 'd:\\foo\\bar')
'd:\\foo'
The parent of the root directory is the root directory.
>>> findpath('..', 'd:\\')
'd:\\'
"""
return os.path.normpath(join(start, target)) | python | def resolve_path(target, start=os.path.curdir):
r"""
Find a path from start to target where target is relative to start.
>>> tmp = str(getfixture('tmpdir_as_cwd'))
>>> findpath('d:\\')
'd:\\'
>>> findpath('d:\\', tmp)
'd:\\'
>>> findpath('\\bar', 'd:\\')
'd:\\bar'
>>> findpath('\\bar', 'd:\\foo') # fails with '\\bar'
'd:\\bar'
>>> findpath('bar', 'd:\\foo')
'd:\\foo\\bar'
>>> findpath('\\baz', 'd:\\foo\\bar') # fails with '\\baz'
'd:\\baz'
>>> os.path.abspath(findpath('\\bar')).lower()
'c:\\bar'
>>> os.path.abspath(findpath('bar'))
'...\\bar'
>>> findpath('..', 'd:\\foo\\bar')
'd:\\foo'
The parent of the root directory is the root directory.
>>> findpath('..', 'd:\\')
'd:\\'
"""
return os.path.normpath(join(start, target)) | [
"def",
"resolve_path",
"(",
"target",
",",
"start",
"=",
"os",
".",
"path",
".",
"curdir",
")",
":",
"return",
"os",
".",
"path",
".",
"normpath",
"(",
"join",
"(",
"start",
",",
"target",
")",
")"
] | r"""
Find a path from start to target where target is relative to start.
>>> tmp = str(getfixture('tmpdir_as_cwd'))
>>> findpath('d:\\')
'd:\\'
>>> findpath('d:\\', tmp)
'd:\\'
>>> findpath('\\bar', 'd:\\')
'd:\\bar'
>>> findpath('\\bar', 'd:\\foo') # fails with '\\bar'
'd:\\bar'
>>> findpath('bar', 'd:\\foo')
'd:\\foo\\bar'
>>> findpath('\\baz', 'd:\\foo\\bar') # fails with '\\baz'
'd:\\baz'
>>> os.path.abspath(findpath('\\bar')).lower()
'c:\\bar'
>>> os.path.abspath(findpath('bar'))
'...\\bar'
>>> findpath('..', 'd:\\foo\\bar')
'd:\\foo'
The parent of the root directory is the root directory.
>>> findpath('..', 'd:\\')
'd:\\' | [
"r",
"Find",
"a",
"path",
"from",
"start",
"to",
"target",
"where",
"target",
"is",
"relative",
"to",
"start",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L306-L343 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | trace_symlink_target | def trace_symlink_target(link):
"""
Given a file that is known to be a symlink, trace it to its ultimate
target.
Raises TargetNotPresent when the target cannot be determined.
Raises ValueError when the specified link is not a symlink.
"""
if not is_symlink(link):
raise ValueError("link must point to a symlink on the system")
while is_symlink(link):
orig = os.path.dirname(link)
link = readlink(link)
link = resolve_path(link, orig)
return link | python | def trace_symlink_target(link):
"""
Given a file that is known to be a symlink, trace it to its ultimate
target.
Raises TargetNotPresent when the target cannot be determined.
Raises ValueError when the specified link is not a symlink.
"""
if not is_symlink(link):
raise ValueError("link must point to a symlink on the system")
while is_symlink(link):
orig = os.path.dirname(link)
link = readlink(link)
link = resolve_path(link, orig)
return link | [
"def",
"trace_symlink_target",
"(",
"link",
")",
":",
"if",
"not",
"is_symlink",
"(",
"link",
")",
":",
"raise",
"ValueError",
"(",
"\"link must point to a symlink on the system\"",
")",
"while",
"is_symlink",
"(",
"link",
")",
":",
"orig",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"link",
")",
"link",
"=",
"readlink",
"(",
"link",
")",
"link",
"=",
"resolve_path",
"(",
"link",
",",
"orig",
")",
"return",
"link"
] | Given a file that is known to be a symlink, trace it to its ultimate
target.
Raises TargetNotPresent when the target cannot be determined.
Raises ValueError when the specified link is not a symlink. | [
"Given",
"a",
"file",
"that",
"is",
"known",
"to",
"be",
"a",
"symlink",
"trace",
"it",
"to",
"its",
"ultimate",
"target",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L349-L364 | train |
jaraco/jaraco.windows | jaraco/windows/filesystem/__init__.py | patch_os_module | def patch_os_module():
"""
jaraco.windows provides the os.symlink and os.readlink functions.
Monkey-patch the os module to include them if not present.
"""
if not hasattr(os, 'symlink'):
os.symlink = symlink
os.path.islink = islink
if not hasattr(os, 'readlink'):
os.readlink = readlink | python | def patch_os_module():
"""
jaraco.windows provides the os.symlink and os.readlink functions.
Monkey-patch the os module to include them if not present.
"""
if not hasattr(os, 'symlink'):
os.symlink = symlink
os.path.islink = islink
if not hasattr(os, 'readlink'):
os.readlink = readlink | [
"def",
"patch_os_module",
"(",
")",
":",
"if",
"not",
"hasattr",
"(",
"os",
",",
"'symlink'",
")",
":",
"os",
".",
"symlink",
"=",
"symlink",
"os",
".",
"path",
".",
"islink",
"=",
"islink",
"if",
"not",
"hasattr",
"(",
"os",
",",
"'readlink'",
")",
":",
"os",
".",
"readlink",
"=",
"readlink"
] | jaraco.windows provides the os.symlink and os.readlink functions.
Monkey-patch the os module to include them if not present. | [
"jaraco",
".",
"windows",
"provides",
"the",
"os",
".",
"symlink",
"and",
"os",
".",
"readlink",
"functions",
".",
"Monkey",
"-",
"patch",
"the",
"os",
"module",
"to",
"include",
"them",
"if",
"not",
"present",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/filesystem/__init__.py#L398-L407 | train |
theno/fabsetup | fabsetup/fabutils.py | task | def task(func, *args, **kwargs):
'''Composition of decorator functions for inherent self-documentation on
task execution.
On execution, each task prints out its name and its first docstring line.
'''
prefix = '\n# '
tail = '\n'
return fabric.api.task(
print_full_name(color=magenta,
prefix=prefix,
tail=tail)(print_doc1(func)),
*args,
**kwargs) | python | def task(func, *args, **kwargs):
'''Composition of decorator functions for inherent self-documentation on
task execution.
On execution, each task prints out its name and its first docstring line.
'''
prefix = '\n# '
tail = '\n'
return fabric.api.task(
print_full_name(color=magenta,
prefix=prefix,
tail=tail)(print_doc1(func)),
*args,
**kwargs) | [
"def",
"task",
"(",
"func",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"prefix",
"=",
"'\\n# '",
"tail",
"=",
"'\\n'",
"return",
"fabric",
".",
"api",
".",
"task",
"(",
"print_full_name",
"(",
"color",
"=",
"magenta",
",",
"prefix",
"=",
"prefix",
",",
"tail",
"=",
"tail",
")",
"(",
"print_doc1",
"(",
"func",
")",
")",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | Composition of decorator functions for inherent self-documentation on
task execution.
On execution, each task prints out its name and its first docstring line. | [
"Composition",
"of",
"decorator",
"functions",
"for",
"inherent",
"self",
"-",
"documentation",
"on",
"task",
"execution",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L191-L204 | train |
theno/fabsetup | fabsetup/fabutils.py | subtask | def subtask(*args, **kwargs):
'''Decorator which prints out the name of the decorated function on
execution.
'''
depth = kwargs.get('depth', 2)
prefix = kwargs.get('prefix', '\n' + '#' * depth + ' ')
tail = kwargs.get('tail', '\n')
doc1 = kwargs.get('doc1', False)
color = kwargs.get('color', cyan)
def real_decorator(func):
if doc1:
return print_full_name(color=color, prefix=prefix,
tail=tail)(print_doc1(func))
return print_full_name(color=color, prefix=prefix, tail=tail)(func)
invoked = bool(not args or kwargs)
if not invoked:
# invoke decorator function which returns the wrapper function
return real_decorator(func=args[0])
return real_decorator | python | def subtask(*args, **kwargs):
'''Decorator which prints out the name of the decorated function on
execution.
'''
depth = kwargs.get('depth', 2)
prefix = kwargs.get('prefix', '\n' + '#' * depth + ' ')
tail = kwargs.get('tail', '\n')
doc1 = kwargs.get('doc1', False)
color = kwargs.get('color', cyan)
def real_decorator(func):
if doc1:
return print_full_name(color=color, prefix=prefix,
tail=tail)(print_doc1(func))
return print_full_name(color=color, prefix=prefix, tail=tail)(func)
invoked = bool(not args or kwargs)
if not invoked:
# invoke decorator function which returns the wrapper function
return real_decorator(func=args[0])
return real_decorator | [
"def",
"subtask",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"depth",
"=",
"kwargs",
".",
"get",
"(",
"'depth'",
",",
"2",
")",
"prefix",
"=",
"kwargs",
".",
"get",
"(",
"'prefix'",
",",
"'\\n'",
"+",
"'#'",
"*",
"depth",
"+",
"' '",
")",
"tail",
"=",
"kwargs",
".",
"get",
"(",
"'tail'",
",",
"'\\n'",
")",
"doc1",
"=",
"kwargs",
".",
"get",
"(",
"'doc1'",
",",
"False",
")",
"color",
"=",
"kwargs",
".",
"get",
"(",
"'color'",
",",
"cyan",
")",
"def",
"real_decorator",
"(",
"func",
")",
":",
"if",
"doc1",
":",
"return",
"print_full_name",
"(",
"color",
"=",
"color",
",",
"prefix",
"=",
"prefix",
",",
"tail",
"=",
"tail",
")",
"(",
"print_doc1",
"(",
"func",
")",
")",
"return",
"print_full_name",
"(",
"color",
"=",
"color",
",",
"prefix",
"=",
"prefix",
",",
"tail",
"=",
"tail",
")",
"(",
"func",
")",
"invoked",
"=",
"bool",
"(",
"not",
"args",
"or",
"kwargs",
")",
"if",
"not",
"invoked",
":",
"# invoke decorator function which returns the wrapper function",
"return",
"real_decorator",
"(",
"func",
"=",
"args",
"[",
"0",
"]",
")",
"return",
"real_decorator"
] | Decorator which prints out the name of the decorated function on
execution. | [
"Decorator",
"which",
"prints",
"out",
"the",
"name",
"of",
"the",
"decorated",
"function",
"on",
"execution",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L213-L233 | train |
theno/fabsetup | fabsetup/fabutils.py | _is_sudoer | def _is_sudoer(what_for=''):
'''Return True if current user is a sudoer, else False.
Should be called non-eager if sudo is wanted only.
'''
if env.get('nosudo', None) is None:
if what_for:
print(yellow(what_for))
with quiet():
# possible outputs:
# en: "Sorry, user winhost-tester may not run sudo on <hostname>"
# en: "sudo: a password is required" (=> is sudoer)
# de: "sudo: Ein Passwort ist notwendig" (=> is sudoer)
output = run('sudo -nv', capture=True)
env.nosudo = not (output.startswith('sudo: ') or output == '')
if env.nosudo:
print('Cannot execute sudo-commands')
return not env.nosudo | python | def _is_sudoer(what_for=''):
'''Return True if current user is a sudoer, else False.
Should be called non-eager if sudo is wanted only.
'''
if env.get('nosudo', None) is None:
if what_for:
print(yellow(what_for))
with quiet():
# possible outputs:
# en: "Sorry, user winhost-tester may not run sudo on <hostname>"
# en: "sudo: a password is required" (=> is sudoer)
# de: "sudo: Ein Passwort ist notwendig" (=> is sudoer)
output = run('sudo -nv', capture=True)
env.nosudo = not (output.startswith('sudo: ') or output == '')
if env.nosudo:
print('Cannot execute sudo-commands')
return not env.nosudo | [
"def",
"_is_sudoer",
"(",
"what_for",
"=",
"''",
")",
":",
"if",
"env",
".",
"get",
"(",
"'nosudo'",
",",
"None",
")",
"is",
"None",
":",
"if",
"what_for",
":",
"print",
"(",
"yellow",
"(",
"what_for",
")",
")",
"with",
"quiet",
"(",
")",
":",
"# possible outputs:",
"# en: \"Sorry, user winhost-tester may not run sudo on <hostname>\"",
"# en: \"sudo: a password is required\" (=> is sudoer)",
"# de: \"sudo: Ein Passwort ist notwendig\" (=> is sudoer)",
"output",
"=",
"run",
"(",
"'sudo -nv'",
",",
"capture",
"=",
"True",
")",
"env",
".",
"nosudo",
"=",
"not",
"(",
"output",
".",
"startswith",
"(",
"'sudo: '",
")",
"or",
"output",
"==",
"''",
")",
"if",
"env",
".",
"nosudo",
":",
"print",
"(",
"'Cannot execute sudo-commands'",
")",
"return",
"not",
"env",
".",
"nosudo"
] | Return True if current user is a sudoer, else False.
Should be called non-eager if sudo is wanted only. | [
"Return",
"True",
"if",
"current",
"user",
"is",
"a",
"sudoer",
"else",
"False",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L248-L265 | train |
theno/fabsetup | fabsetup/fabutils.py | install_packages | def install_packages(packages,
what_for='for a complete setup to work properly'):
'''Try to install .deb packages given by list.
Return True, if packages could be installed or are installed already, or if
they cannot be installed but the user gives feedback to continue.
Else return False.
'''
res = True
non_installed_packages = _non_installed(packages)
packages_str = ' '.join(non_installed_packages)
if non_installed_packages:
with quiet():
dpkg = _has_dpkg()
hint = ' (You may have to install them manually)'
do_install = False
go_on = True
if dpkg:
if _is_sudoer('Want to install dpkg packages'):
do_install = True
else:
do_install is False # cannot install anything
info = yellow(' '.join([
'This deb packages are missing to be installed',
flo("{what_for}: "), ', '.join(non_installed_packages),
]))
question = ' Continue anyway?'
go_on = query_yes_no(info + hint + question, default='no')
else:
# dpkg == False, unable to determine if packages are installed
do_install = False # cannot install anything
info = yellow(' '.join([
flo('Required {what_for}: '),
', '.join(non_installed_packages),
]))
go_on = query_yes_no(info + hint + ' Continue?', default='yes')
if not go_on:
sys.exit('Abort')
if do_install:
command = flo('sudo apt-get install {packages_str}')
res = run(command).return_code == 0
return res | python | def install_packages(packages,
what_for='for a complete setup to work properly'):
'''Try to install .deb packages given by list.
Return True, if packages could be installed or are installed already, or if
they cannot be installed but the user gives feedback to continue.
Else return False.
'''
res = True
non_installed_packages = _non_installed(packages)
packages_str = ' '.join(non_installed_packages)
if non_installed_packages:
with quiet():
dpkg = _has_dpkg()
hint = ' (You may have to install them manually)'
do_install = False
go_on = True
if dpkg:
if _is_sudoer('Want to install dpkg packages'):
do_install = True
else:
do_install is False # cannot install anything
info = yellow(' '.join([
'This deb packages are missing to be installed',
flo("{what_for}: "), ', '.join(non_installed_packages),
]))
question = ' Continue anyway?'
go_on = query_yes_no(info + hint + question, default='no')
else:
# dpkg == False, unable to determine if packages are installed
do_install = False # cannot install anything
info = yellow(' '.join([
flo('Required {what_for}: '),
', '.join(non_installed_packages),
]))
go_on = query_yes_no(info + hint + ' Continue?', default='yes')
if not go_on:
sys.exit('Abort')
if do_install:
command = flo('sudo apt-get install {packages_str}')
res = run(command).return_code == 0
return res | [
"def",
"install_packages",
"(",
"packages",
",",
"what_for",
"=",
"'for a complete setup to work properly'",
")",
":",
"res",
"=",
"True",
"non_installed_packages",
"=",
"_non_installed",
"(",
"packages",
")",
"packages_str",
"=",
"' '",
".",
"join",
"(",
"non_installed_packages",
")",
"if",
"non_installed_packages",
":",
"with",
"quiet",
"(",
")",
":",
"dpkg",
"=",
"_has_dpkg",
"(",
")",
"hint",
"=",
"' (You may have to install them manually)'",
"do_install",
"=",
"False",
"go_on",
"=",
"True",
"if",
"dpkg",
":",
"if",
"_is_sudoer",
"(",
"'Want to install dpkg packages'",
")",
":",
"do_install",
"=",
"True",
"else",
":",
"do_install",
"is",
"False",
"# cannot install anything",
"info",
"=",
"yellow",
"(",
"' '",
".",
"join",
"(",
"[",
"'This deb packages are missing to be installed'",
",",
"flo",
"(",
"\"{what_for}: \"",
")",
",",
"', '",
".",
"join",
"(",
"non_installed_packages",
")",
",",
"]",
")",
")",
"question",
"=",
"' Continue anyway?'",
"go_on",
"=",
"query_yes_no",
"(",
"info",
"+",
"hint",
"+",
"question",
",",
"default",
"=",
"'no'",
")",
"else",
":",
"# dpkg == False, unable to determine if packages are installed",
"do_install",
"=",
"False",
"# cannot install anything",
"info",
"=",
"yellow",
"(",
"' '",
".",
"join",
"(",
"[",
"flo",
"(",
"'Required {what_for}: '",
")",
",",
"', '",
".",
"join",
"(",
"non_installed_packages",
")",
",",
"]",
")",
")",
"go_on",
"=",
"query_yes_no",
"(",
"info",
"+",
"hint",
"+",
"' Continue?'",
",",
"default",
"=",
"'yes'",
")",
"if",
"not",
"go_on",
":",
"sys",
".",
"exit",
"(",
"'Abort'",
")",
"if",
"do_install",
":",
"command",
"=",
"flo",
"(",
"'sudo apt-get install {packages_str}'",
")",
"res",
"=",
"run",
"(",
"command",
")",
".",
"return_code",
"==",
"0",
"return",
"res"
] | Try to install .deb packages given by list.
Return True, if packages could be installed or are installed already, or if
they cannot be installed but the user gives feedback to continue.
Else return False. | [
"Try",
"to",
"install",
".",
"deb",
"packages",
"given",
"by",
"list",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L273-L315 | train |
theno/fabsetup | fabsetup/fabutils.py | checkup_git_repos_legacy | def checkup_git_repos_legacy(repos, base_dir='~/repos',
verbose=False, prefix='', postfix=''):
'''Checkout or update git repos.
repos must be a list of dicts each with an url and optional with a name
value.
'''
run(flo('mkdir -p {base_dir}'))
for repo in repos:
cur_base_dir = repo.get('base_dir', base_dir)
checkup_git_repo_legacy(url=repo['url'], name=repo.get('name', None),
base_dir=cur_base_dir, verbose=verbose,
prefix=prefix, postfix=postfix) | python | def checkup_git_repos_legacy(repos, base_dir='~/repos',
verbose=False, prefix='', postfix=''):
'''Checkout or update git repos.
repos must be a list of dicts each with an url and optional with a name
value.
'''
run(flo('mkdir -p {base_dir}'))
for repo in repos:
cur_base_dir = repo.get('base_dir', base_dir)
checkup_git_repo_legacy(url=repo['url'], name=repo.get('name', None),
base_dir=cur_base_dir, verbose=verbose,
prefix=prefix, postfix=postfix) | [
"def",
"checkup_git_repos_legacy",
"(",
"repos",
",",
"base_dir",
"=",
"'~/repos'",
",",
"verbose",
"=",
"False",
",",
"prefix",
"=",
"''",
",",
"postfix",
"=",
"''",
")",
":",
"run",
"(",
"flo",
"(",
"'mkdir -p {base_dir}'",
")",
")",
"for",
"repo",
"in",
"repos",
":",
"cur_base_dir",
"=",
"repo",
".",
"get",
"(",
"'base_dir'",
",",
"base_dir",
")",
"checkup_git_repo_legacy",
"(",
"url",
"=",
"repo",
"[",
"'url'",
"]",
",",
"name",
"=",
"repo",
".",
"get",
"(",
"'name'",
",",
"None",
")",
",",
"base_dir",
"=",
"cur_base_dir",
",",
"verbose",
"=",
"verbose",
",",
"prefix",
"=",
"prefix",
",",
"postfix",
"=",
"postfix",
")"
] | Checkout or update git repos.
repos must be a list of dicts each with an url and optional with a name
value. | [
"Checkout",
"or",
"update",
"git",
"repos",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L324-L336 | train |
theno/fabsetup | fabsetup/fabutils.py | checkup_git_repo_legacy | def checkup_git_repo_legacy(url, name=None, base_dir='~/repos',
verbose=False, prefix='', postfix=''):
'''Checkout or update a git repo.'''
if not name:
match = re.match(r'.*/(.+)\.git', url)
assert match, flo("Unable to extract repo name from '{url}'")
name = match.group(1)
assert name is not None, flo('Cannot extract repo name from repo: {url}')
assert name != '', flo('Cannot extract repo name from repo: {url} (empty)')
if verbose:
name_blue = blue(name)
print_msg(flo('{prefix}Checkout or update {name_blue}{postfix}'))
if not exists(base_dir):
run(flo('mkdir -p {base_dir}'))
if not exists(flo('{base_dir}/{name}/.git')):
run(flo(' && '.join([
'cd {base_dir}',
'git clone {url} {name}'])),
msg='clone repo')
else:
if verbose:
print_msg('update: pull from origin')
run(flo('cd {base_dir}/{name} && git pull'))
return name | python | def checkup_git_repo_legacy(url, name=None, base_dir='~/repos',
verbose=False, prefix='', postfix=''):
'''Checkout or update a git repo.'''
if not name:
match = re.match(r'.*/(.+)\.git', url)
assert match, flo("Unable to extract repo name from '{url}'")
name = match.group(1)
assert name is not None, flo('Cannot extract repo name from repo: {url}')
assert name != '', flo('Cannot extract repo name from repo: {url} (empty)')
if verbose:
name_blue = blue(name)
print_msg(flo('{prefix}Checkout or update {name_blue}{postfix}'))
if not exists(base_dir):
run(flo('mkdir -p {base_dir}'))
if not exists(flo('{base_dir}/{name}/.git')):
run(flo(' && '.join([
'cd {base_dir}',
'git clone {url} {name}'])),
msg='clone repo')
else:
if verbose:
print_msg('update: pull from origin')
run(flo('cd {base_dir}/{name} && git pull'))
return name | [
"def",
"checkup_git_repo_legacy",
"(",
"url",
",",
"name",
"=",
"None",
",",
"base_dir",
"=",
"'~/repos'",
",",
"verbose",
"=",
"False",
",",
"prefix",
"=",
"''",
",",
"postfix",
"=",
"''",
")",
":",
"if",
"not",
"name",
":",
"match",
"=",
"re",
".",
"match",
"(",
"r'.*/(.+)\\.git'",
",",
"url",
")",
"assert",
"match",
",",
"flo",
"(",
"\"Unable to extract repo name from '{url}'\"",
")",
"name",
"=",
"match",
".",
"group",
"(",
"1",
")",
"assert",
"name",
"is",
"not",
"None",
",",
"flo",
"(",
"'Cannot extract repo name from repo: {url}'",
")",
"assert",
"name",
"!=",
"''",
",",
"flo",
"(",
"'Cannot extract repo name from repo: {url} (empty)'",
")",
"if",
"verbose",
":",
"name_blue",
"=",
"blue",
"(",
"name",
")",
"print_msg",
"(",
"flo",
"(",
"'{prefix}Checkout or update {name_blue}{postfix}'",
")",
")",
"if",
"not",
"exists",
"(",
"base_dir",
")",
":",
"run",
"(",
"flo",
"(",
"'mkdir -p {base_dir}'",
")",
")",
"if",
"not",
"exists",
"(",
"flo",
"(",
"'{base_dir}/{name}/.git'",
")",
")",
":",
"run",
"(",
"flo",
"(",
"' && '",
".",
"join",
"(",
"[",
"'cd {base_dir}'",
",",
"'git clone {url} {name}'",
"]",
")",
")",
",",
"msg",
"=",
"'clone repo'",
")",
"else",
":",
"if",
"verbose",
":",
"print_msg",
"(",
"'update: pull from origin'",
")",
"run",
"(",
"flo",
"(",
"'cd {base_dir}/{name} && git pull'",
")",
")",
"return",
"name"
] | Checkout or update a git repo. | [
"Checkout",
"or",
"update",
"a",
"git",
"repo",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L339-L362 | train |
theno/fabsetup | fabsetup/fabutils.py | install_file_legacy | def install_file_legacy(path, sudo=False, from_path=None, **substitutions):
'''Install file with path on the host target.
The from file is the first of this list which exists:
* custom file
* custom file.template
* common file
* common file.template
'''
# source paths 'from_custom' and 'from_common'
from_path = from_path or path
# remove beginning '/' (if any), eg '/foo/bar' -> 'foo/bar'
from_tail = join('files', from_path.lstrip(os.sep))
if from_path.startswith('~/'):
from_tail = join('files', 'home', 'USERNAME',
from_path[2:]) # without beginning '~/'
from_common = join(FABFILE_DATA_DIR, from_tail)
from_custom = join(FABSETUP_CUSTOM_DIR, from_tail)
# target path 'to_' (path or tempfile)
for subst in ['SITENAME', 'USER', 'ADDON', 'TASK']:
sitename = substitutions.get(subst, False)
if sitename:
path = path.replace(subst, sitename)
to_ = path
if sudo:
to_ = join(os.sep, 'tmp', 'fabsetup_' + os.path.basename(path))
path_dir = dirname(path)
# copy file
if isfile(from_custom):
run(flo('mkdir -p {path_dir}'))
put(from_custom, to_)
elif isfile(from_custom + '.template'):
_install_file_from_template_legacy(from_custom + '.template', to_=to_,
**substitutions)
elif isfile(from_common):
run(flo('mkdir -p {path_dir}'))
put(from_common, to_)
else:
_install_file_from_template_legacy(from_common + '.template', to_=to_,
**substitutions)
if sudo:
run(flo('sudo mv --force {to_} {path}')) | python | def install_file_legacy(path, sudo=False, from_path=None, **substitutions):
'''Install file with path on the host target.
The from file is the first of this list which exists:
* custom file
* custom file.template
* common file
* common file.template
'''
# source paths 'from_custom' and 'from_common'
from_path = from_path or path
# remove beginning '/' (if any), eg '/foo/bar' -> 'foo/bar'
from_tail = join('files', from_path.lstrip(os.sep))
if from_path.startswith('~/'):
from_tail = join('files', 'home', 'USERNAME',
from_path[2:]) # without beginning '~/'
from_common = join(FABFILE_DATA_DIR, from_tail)
from_custom = join(FABSETUP_CUSTOM_DIR, from_tail)
# target path 'to_' (path or tempfile)
for subst in ['SITENAME', 'USER', 'ADDON', 'TASK']:
sitename = substitutions.get(subst, False)
if sitename:
path = path.replace(subst, sitename)
to_ = path
if sudo:
to_ = join(os.sep, 'tmp', 'fabsetup_' + os.path.basename(path))
path_dir = dirname(path)
# copy file
if isfile(from_custom):
run(flo('mkdir -p {path_dir}'))
put(from_custom, to_)
elif isfile(from_custom + '.template'):
_install_file_from_template_legacy(from_custom + '.template', to_=to_,
**substitutions)
elif isfile(from_common):
run(flo('mkdir -p {path_dir}'))
put(from_common, to_)
else:
_install_file_from_template_legacy(from_common + '.template', to_=to_,
**substitutions)
if sudo:
run(flo('sudo mv --force {to_} {path}')) | [
"def",
"install_file_legacy",
"(",
"path",
",",
"sudo",
"=",
"False",
",",
"from_path",
"=",
"None",
",",
"*",
"*",
"substitutions",
")",
":",
"# source paths 'from_custom' and 'from_common'",
"from_path",
"=",
"from_path",
"or",
"path",
"# remove beginning '/' (if any), eg '/foo/bar' -> 'foo/bar'",
"from_tail",
"=",
"join",
"(",
"'files'",
",",
"from_path",
".",
"lstrip",
"(",
"os",
".",
"sep",
")",
")",
"if",
"from_path",
".",
"startswith",
"(",
"'~/'",
")",
":",
"from_tail",
"=",
"join",
"(",
"'files'",
",",
"'home'",
",",
"'USERNAME'",
",",
"from_path",
"[",
"2",
":",
"]",
")",
"# without beginning '~/'",
"from_common",
"=",
"join",
"(",
"FABFILE_DATA_DIR",
",",
"from_tail",
")",
"from_custom",
"=",
"join",
"(",
"FABSETUP_CUSTOM_DIR",
",",
"from_tail",
")",
"# target path 'to_' (path or tempfile)",
"for",
"subst",
"in",
"[",
"'SITENAME'",
",",
"'USER'",
",",
"'ADDON'",
",",
"'TASK'",
"]",
":",
"sitename",
"=",
"substitutions",
".",
"get",
"(",
"subst",
",",
"False",
")",
"if",
"sitename",
":",
"path",
"=",
"path",
".",
"replace",
"(",
"subst",
",",
"sitename",
")",
"to_",
"=",
"path",
"if",
"sudo",
":",
"to_",
"=",
"join",
"(",
"os",
".",
"sep",
",",
"'tmp'",
",",
"'fabsetup_'",
"+",
"os",
".",
"path",
".",
"basename",
"(",
"path",
")",
")",
"path_dir",
"=",
"dirname",
"(",
"path",
")",
"# copy file",
"if",
"isfile",
"(",
"from_custom",
")",
":",
"run",
"(",
"flo",
"(",
"'mkdir -p {path_dir}'",
")",
")",
"put",
"(",
"from_custom",
",",
"to_",
")",
"elif",
"isfile",
"(",
"from_custom",
"+",
"'.template'",
")",
":",
"_install_file_from_template_legacy",
"(",
"from_custom",
"+",
"'.template'",
",",
"to_",
"=",
"to_",
",",
"*",
"*",
"substitutions",
")",
"elif",
"isfile",
"(",
"from_common",
")",
":",
"run",
"(",
"flo",
"(",
"'mkdir -p {path_dir}'",
")",
")",
"put",
"(",
"from_common",
",",
"to_",
")",
"else",
":",
"_install_file_from_template_legacy",
"(",
"from_common",
"+",
"'.template'",
",",
"to_",
"=",
"to_",
",",
"*",
"*",
"substitutions",
")",
"if",
"sudo",
":",
"run",
"(",
"flo",
"(",
"'sudo mv --force {to_} {path}'",
")",
")"
] | Install file with path on the host target.
The from file is the first of this list which exists:
* custom file
* custom file.template
* common file
* common file.template | [
"Install",
"file",
"with",
"path",
"on",
"the",
"host",
"target",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L481-L524 | train |
theno/fabsetup | fabsetup/fabutils.py | install_user_command_legacy | def install_user_command_legacy(command, **substitutions):
'''Install command executable file into users bin dir.
If a custom executable exists it would be installed instead of the "normal"
one. The executable also could exist as a <command>.template file.
'''
path = flo('~/bin/{command}')
install_file_legacy(path, **substitutions)
run(flo('chmod 755 {path}')) | python | def install_user_command_legacy(command, **substitutions):
'''Install command executable file into users bin dir.
If a custom executable exists it would be installed instead of the "normal"
one. The executable also could exist as a <command>.template file.
'''
path = flo('~/bin/{command}')
install_file_legacy(path, **substitutions)
run(flo('chmod 755 {path}')) | [
"def",
"install_user_command_legacy",
"(",
"command",
",",
"*",
"*",
"substitutions",
")",
":",
"path",
"=",
"flo",
"(",
"'~/bin/{command}'",
")",
"install_file_legacy",
"(",
"path",
",",
"*",
"*",
"substitutions",
")",
"run",
"(",
"flo",
"(",
"'chmod 755 {path}'",
")",
")"
] | Install command executable file into users bin dir.
If a custom executable exists it would be installed instead of the "normal"
one. The executable also could exist as a <command>.template file. | [
"Install",
"command",
"executable",
"file",
"into",
"users",
"bin",
"dir",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L527-L535 | train |
theno/fabsetup | fabsetup/fabutils.py | _line_2_pair | def _line_2_pair(line):
'''Return bash variable declaration as name-value pair.
Name as lower case str. Value itself only without surrounding '"' (if any).
For example, _line_2_pair('NAME="Ubuntu"') will return ('name', 'Ubuntu')
'''
key, val = line.split('=')
return key.lower(), val.strip('"') | python | def _line_2_pair(line):
'''Return bash variable declaration as name-value pair.
Name as lower case str. Value itself only without surrounding '"' (if any).
For example, _line_2_pair('NAME="Ubuntu"') will return ('name', 'Ubuntu')
'''
key, val = line.split('=')
return key.lower(), val.strip('"') | [
"def",
"_line_2_pair",
"(",
"line",
")",
":",
"key",
",",
"val",
"=",
"line",
".",
"split",
"(",
"'='",
")",
"return",
"key",
".",
"lower",
"(",
")",
",",
"val",
".",
"strip",
"(",
"'\"'",
")"
] | Return bash variable declaration as name-value pair.
Name as lower case str. Value itself only without surrounding '"' (if any).
For example, _line_2_pair('NAME="Ubuntu"') will return ('name', 'Ubuntu') | [
"Return",
"bash",
"variable",
"declaration",
"as",
"name",
"-",
"value",
"pair",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L735-L743 | train |
theno/fabsetup | fabsetup/fabutils.py | extract_minors_from_setup_py | def extract_minors_from_setup_py(filename_setup_py):
'''Extract supported python minor versions from setup.py and return them
as a list of str.
Return example:
['2.6', '2.7', '3.3', '3.4', '3.5', '3.6']
'''
# eg: minors_str = '2.6\n2.7\n3.3\n3.4\n3.5\n3.6'
minors_str = fabric.api.local(
flo('grep --perl-regexp --only-matching '
'"(?<=Programming Language :: Python :: )\\d+\\.\\d+" '
'{filename_setup_py}'),
capture=True)
# eg: minors = ['2.6', '2.7', '3.3', '3.4', '3.5', '3.6']
minors = minors_str.split()
return minors | python | def extract_minors_from_setup_py(filename_setup_py):
'''Extract supported python minor versions from setup.py and return them
as a list of str.
Return example:
['2.6', '2.7', '3.3', '3.4', '3.5', '3.6']
'''
# eg: minors_str = '2.6\n2.7\n3.3\n3.4\n3.5\n3.6'
minors_str = fabric.api.local(
flo('grep --perl-regexp --only-matching '
'"(?<=Programming Language :: Python :: )\\d+\\.\\d+" '
'{filename_setup_py}'),
capture=True)
# eg: minors = ['2.6', '2.7', '3.3', '3.4', '3.5', '3.6']
minors = minors_str.split()
return minors | [
"def",
"extract_minors_from_setup_py",
"(",
"filename_setup_py",
")",
":",
"# eg: minors_str = '2.6\\n2.7\\n3.3\\n3.4\\n3.5\\n3.6'",
"minors_str",
"=",
"fabric",
".",
"api",
".",
"local",
"(",
"flo",
"(",
"'grep --perl-regexp --only-matching '",
"'\"(?<=Programming Language :: Python :: )\\\\d+\\\\.\\\\d+\" '",
"'{filename_setup_py}'",
")",
",",
"capture",
"=",
"True",
")",
"# eg: minors = ['2.6', '2.7', '3.3', '3.4', '3.5', '3.6']",
"minors",
"=",
"minors_str",
".",
"split",
"(",
")",
"return",
"minors"
] | Extract supported python minor versions from setup.py and return them
as a list of str.
Return example:
['2.6', '2.7', '3.3', '3.4', '3.5', '3.6'] | [
"Extract",
"supported",
"python",
"minor",
"versions",
"from",
"setup",
".",
"py",
"and",
"return",
"them",
"as",
"a",
"list",
"of",
"str",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabutils.py#L834-L850 | train |
theno/fabsetup | fabsetup/fabfile/setup/vim_janus.py | vim_janus | def vim_janus(uninstall=None):
'''Install or update Janus, a distribution of addons and mappings for vim.
More info:
https://github.com/carlhuda/janus
Customization: https://github.com/carlhuda/janus/wiki/Customization
Args:
uninstall: If not None, Uninstall janus and restore old vim config
'''
if uninstall is not None:
uninstall_janus()
else:
if not exists('~/.vim/janus'):
print_msg('not installed => install')
install_janus()
else:
print_msg('already installed => update')
update_janus()
customize_janus()
show_files_used_by_vim_and_janus() | python | def vim_janus(uninstall=None):
'''Install or update Janus, a distribution of addons and mappings for vim.
More info:
https://github.com/carlhuda/janus
Customization: https://github.com/carlhuda/janus/wiki/Customization
Args:
uninstall: If not None, Uninstall janus and restore old vim config
'''
if uninstall is not None:
uninstall_janus()
else:
if not exists('~/.vim/janus'):
print_msg('not installed => install')
install_janus()
else:
print_msg('already installed => update')
update_janus()
customize_janus()
show_files_used_by_vim_and_janus() | [
"def",
"vim_janus",
"(",
"uninstall",
"=",
"None",
")",
":",
"if",
"uninstall",
"is",
"not",
"None",
":",
"uninstall_janus",
"(",
")",
"else",
":",
"if",
"not",
"exists",
"(",
"'~/.vim/janus'",
")",
":",
"print_msg",
"(",
"'not installed => install'",
")",
"install_janus",
"(",
")",
"else",
":",
"print_msg",
"(",
"'already installed => update'",
")",
"update_janus",
"(",
")",
"customize_janus",
"(",
")",
"show_files_used_by_vim_and_janus",
"(",
")"
] | Install or update Janus, a distribution of addons and mappings for vim.
More info:
https://github.com/carlhuda/janus
Customization: https://github.com/carlhuda/janus/wiki/Customization
Args:
uninstall: If not None, Uninstall janus and restore old vim config | [
"Install",
"or",
"update",
"Janus",
"a",
"distribution",
"of",
"addons",
"and",
"mappings",
"for",
"vim",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/vim_janus.py#L18-L38 | train |
yahoo/serviceping | serviceping/network.py | scan | def scan(host, port=80, url=None, https=False, timeout=1, max_size=65535):
"""
Scan a network port
Parameters
----------
host : str
Host or ip address to scan
port : int, optional
Port to scan, default=80
url : str, optional
URL to perform get request to on the host and port specified
https : bool, optional
Perform ssl connection on the socket, default=False
timeout : float
Timeout for network operations, default=1
Returns
-------
dict
Result dictionary that contains the following keys:
host - The host or IP address that was scanned
port - The port number that was scanned
state - The state of the port, will be either "open" or "closed"
durations - An ordered dictionary with floating point value of the
time elapsed for each connection operation
Raises
------
ScanFailed - The scan operation failed
"""
starts = OrderedDict()
ends = OrderedDict()
port = int(port)
result = dict(
host=host, port=port, state='closed', durations=OrderedDict()
)
if url:
timeout = 1
result['code'] = None
starts['all'] = starts['dns'] = datetime.datetime.now()
# DNS Lookup
try:
hostip = socket.gethostbyname(host)
result['ip'] = hostip
ends['dns'] = datetime.datetime.now()
except socket.gaierror:
raise ScanFailed('DNS Lookup failed', result=result)
# TCP Connect
starts['connect'] = datetime.datetime.now()
network_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
network_socket.settimeout(timeout)
result_connection = network_socket.connect_ex((hostip, port))
ends['connect'] = datetime.datetime.now()
# SSL
if https:
starts['ssl'] = datetime.datetime.now()
try:
network_socket = ssl.wrap_socket(network_socket)
except socket.timeout:
raise ScanFailed('SSL socket timeout', result=result)
ends['ssl'] = datetime.datetime.now()
# Get request
if result_connection == 0 and url:
starts['request'] = datetime.datetime.now()
network_socket.send(
"GET {0} HTTP/1.0\r\nHost: {1}\r\n\r\n".format(
url, host
).encode('ascii'))
if max_size:
data = network_socket.recv(max_size)
else:
data = network_socket.recv()
result['length'] = len(data)
data = data.decode('ascii', errors='ignore')
result['response'] = (data)
try:
result['code'] = int(data.split('\n')[0].split()[1])
except IndexError:
pass
ends['request'] = datetime.datetime.now()
network_socket.close()
# Calculate durations
ends['all'] = datetime.datetime.now()
for duration in starts.keys():
if duration in ends.keys():
result['durations'][duration] = ends[duration] - starts[duration]
if result_connection == 0:
result['state'] = 'open'
return result | python | def scan(host, port=80, url=None, https=False, timeout=1, max_size=65535):
"""
Scan a network port
Parameters
----------
host : str
Host or ip address to scan
port : int, optional
Port to scan, default=80
url : str, optional
URL to perform get request to on the host and port specified
https : bool, optional
Perform ssl connection on the socket, default=False
timeout : float
Timeout for network operations, default=1
Returns
-------
dict
Result dictionary that contains the following keys:
host - The host or IP address that was scanned
port - The port number that was scanned
state - The state of the port, will be either "open" or "closed"
durations - An ordered dictionary with floating point value of the
time elapsed for each connection operation
Raises
------
ScanFailed - The scan operation failed
"""
starts = OrderedDict()
ends = OrderedDict()
port = int(port)
result = dict(
host=host, port=port, state='closed', durations=OrderedDict()
)
if url:
timeout = 1
result['code'] = None
starts['all'] = starts['dns'] = datetime.datetime.now()
# DNS Lookup
try:
hostip = socket.gethostbyname(host)
result['ip'] = hostip
ends['dns'] = datetime.datetime.now()
except socket.gaierror:
raise ScanFailed('DNS Lookup failed', result=result)
# TCP Connect
starts['connect'] = datetime.datetime.now()
network_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
network_socket.settimeout(timeout)
result_connection = network_socket.connect_ex((hostip, port))
ends['connect'] = datetime.datetime.now()
# SSL
if https:
starts['ssl'] = datetime.datetime.now()
try:
network_socket = ssl.wrap_socket(network_socket)
except socket.timeout:
raise ScanFailed('SSL socket timeout', result=result)
ends['ssl'] = datetime.datetime.now()
# Get request
if result_connection == 0 and url:
starts['request'] = datetime.datetime.now()
network_socket.send(
"GET {0} HTTP/1.0\r\nHost: {1}\r\n\r\n".format(
url, host
).encode('ascii'))
if max_size:
data = network_socket.recv(max_size)
else:
data = network_socket.recv()
result['length'] = len(data)
data = data.decode('ascii', errors='ignore')
result['response'] = (data)
try:
result['code'] = int(data.split('\n')[0].split()[1])
except IndexError:
pass
ends['request'] = datetime.datetime.now()
network_socket.close()
# Calculate durations
ends['all'] = datetime.datetime.now()
for duration in starts.keys():
if duration in ends.keys():
result['durations'][duration] = ends[duration] - starts[duration]
if result_connection == 0:
result['state'] = 'open'
return result | [
"def",
"scan",
"(",
"host",
",",
"port",
"=",
"80",
",",
"url",
"=",
"None",
",",
"https",
"=",
"False",
",",
"timeout",
"=",
"1",
",",
"max_size",
"=",
"65535",
")",
":",
"starts",
"=",
"OrderedDict",
"(",
")",
"ends",
"=",
"OrderedDict",
"(",
")",
"port",
"=",
"int",
"(",
"port",
")",
"result",
"=",
"dict",
"(",
"host",
"=",
"host",
",",
"port",
"=",
"port",
",",
"state",
"=",
"'closed'",
",",
"durations",
"=",
"OrderedDict",
"(",
")",
")",
"if",
"url",
":",
"timeout",
"=",
"1",
"result",
"[",
"'code'",
"]",
"=",
"None",
"starts",
"[",
"'all'",
"]",
"=",
"starts",
"[",
"'dns'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"# DNS Lookup",
"try",
":",
"hostip",
"=",
"socket",
".",
"gethostbyname",
"(",
"host",
")",
"result",
"[",
"'ip'",
"]",
"=",
"hostip",
"ends",
"[",
"'dns'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"except",
"socket",
".",
"gaierror",
":",
"raise",
"ScanFailed",
"(",
"'DNS Lookup failed'",
",",
"result",
"=",
"result",
")",
"# TCP Connect",
"starts",
"[",
"'connect'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"network_socket",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_INET",
",",
"socket",
".",
"SOCK_STREAM",
")",
"network_socket",
".",
"settimeout",
"(",
"timeout",
")",
"result_connection",
"=",
"network_socket",
".",
"connect_ex",
"(",
"(",
"hostip",
",",
"port",
")",
")",
"ends",
"[",
"'connect'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"# SSL",
"if",
"https",
":",
"starts",
"[",
"'ssl'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"try",
":",
"network_socket",
"=",
"ssl",
".",
"wrap_socket",
"(",
"network_socket",
")",
"except",
"socket",
".",
"timeout",
":",
"raise",
"ScanFailed",
"(",
"'SSL socket timeout'",
",",
"result",
"=",
"result",
")",
"ends",
"[",
"'ssl'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"# Get request",
"if",
"result_connection",
"==",
"0",
"and",
"url",
":",
"starts",
"[",
"'request'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"network_socket",
".",
"send",
"(",
"\"GET {0} HTTP/1.0\\r\\nHost: {1}\\r\\n\\r\\n\"",
".",
"format",
"(",
"url",
",",
"host",
")",
".",
"encode",
"(",
"'ascii'",
")",
")",
"if",
"max_size",
":",
"data",
"=",
"network_socket",
".",
"recv",
"(",
"max_size",
")",
"else",
":",
"data",
"=",
"network_socket",
".",
"recv",
"(",
")",
"result",
"[",
"'length'",
"]",
"=",
"len",
"(",
"data",
")",
"data",
"=",
"data",
".",
"decode",
"(",
"'ascii'",
",",
"errors",
"=",
"'ignore'",
")",
"result",
"[",
"'response'",
"]",
"=",
"(",
"data",
")",
"try",
":",
"result",
"[",
"'code'",
"]",
"=",
"int",
"(",
"data",
".",
"split",
"(",
"'\\n'",
")",
"[",
"0",
"]",
".",
"split",
"(",
")",
"[",
"1",
"]",
")",
"except",
"IndexError",
":",
"pass",
"ends",
"[",
"'request'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"network_socket",
".",
"close",
"(",
")",
"# Calculate durations",
"ends",
"[",
"'all'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"for",
"duration",
"in",
"starts",
".",
"keys",
"(",
")",
":",
"if",
"duration",
"in",
"ends",
".",
"keys",
"(",
")",
":",
"result",
"[",
"'durations'",
"]",
"[",
"duration",
"]",
"=",
"ends",
"[",
"duration",
"]",
"-",
"starts",
"[",
"duration",
"]",
"if",
"result_connection",
"==",
"0",
":",
"result",
"[",
"'state'",
"]",
"=",
"'open'",
"return",
"result"
] | Scan a network port
Parameters
----------
host : str
Host or ip address to scan
port : int, optional
Port to scan, default=80
url : str, optional
URL to perform get request to on the host and port specified
https : bool, optional
Perform ssl connection on the socket, default=False
timeout : float
Timeout for network operations, default=1
Returns
-------
dict
Result dictionary that contains the following keys:
host - The host or IP address that was scanned
port - The port number that was scanned
state - The state of the port, will be either "open" or "closed"
durations - An ordered dictionary with floating point value of the
time elapsed for each connection operation
Raises
------
ScanFailed - The scan operation failed | [
"Scan",
"a",
"network",
"port"
] | 1f9df5ee5b3cba466426b1164262278472ba4977 | https://github.com/yahoo/serviceping/blob/1f9df5ee5b3cba466426b1164262278472ba4977/serviceping/network.py#L25-L124 | train |
yahoo/serviceping | serviceping/network.py | ping | def ping(host, port=80, url=None, https=False, timeout=1, max_size=65535, sequence=0):
"""
Ping a host
Parameters
----------
host: str
The host or ip address to ping
port: int, optional
The port to ping, default=80
url: str, optional
URL to ping, will do a host/port ping if not provided
https: bool, optional
Connect via ssl, default=False
timeout: int, optional
Number of seconds to wait for a response before timing out, default=1 second
max_size: int, optional
The max size of response that can be retrieved. This should be a power of 2
default=65535.
sequence: int, optional
Sequence number for the ping request
Returns
-------
PingResponse:
The ping response object
"""
try:
result = scan(host=host, port=port, url=url, https=https, timeout=timeout, max_size=max_size)
except ScanFailed as failure:
result = failure.result
result['error'] = True
result['error_message'] = str(failure)
result_obj = PingResponse(
host=host, port=port, ip=result.get('ip', None), sequence=sequence,
durations=result.get('durations', None),
code=result.get('code', None),
state=result.get('state', 'unknown'),
length=result.get('length', 0),
response=result.get('response', None),
error=result.get('error', False),
error_message=result.get('error_message', None),
responding=True if result.get('state', 'unknown') in ['open'] else False,
start=datetime.datetime.now(),
end=datetime.datetime.now() + result['durations'].get('all', datetime.timedelta(0)) if result.get('durations', None) else None
)
return result_obj | python | def ping(host, port=80, url=None, https=False, timeout=1, max_size=65535, sequence=0):
"""
Ping a host
Parameters
----------
host: str
The host or ip address to ping
port: int, optional
The port to ping, default=80
url: str, optional
URL to ping, will do a host/port ping if not provided
https: bool, optional
Connect via ssl, default=False
timeout: int, optional
Number of seconds to wait for a response before timing out, default=1 second
max_size: int, optional
The max size of response that can be retrieved. This should be a power of 2
default=65535.
sequence: int, optional
Sequence number for the ping request
Returns
-------
PingResponse:
The ping response object
"""
try:
result = scan(host=host, port=port, url=url, https=https, timeout=timeout, max_size=max_size)
except ScanFailed as failure:
result = failure.result
result['error'] = True
result['error_message'] = str(failure)
result_obj = PingResponse(
host=host, port=port, ip=result.get('ip', None), sequence=sequence,
durations=result.get('durations', None),
code=result.get('code', None),
state=result.get('state', 'unknown'),
length=result.get('length', 0),
response=result.get('response', None),
error=result.get('error', False),
error_message=result.get('error_message', None),
responding=True if result.get('state', 'unknown') in ['open'] else False,
start=datetime.datetime.now(),
end=datetime.datetime.now() + result['durations'].get('all', datetime.timedelta(0)) if result.get('durations', None) else None
)
return result_obj | [
"def",
"ping",
"(",
"host",
",",
"port",
"=",
"80",
",",
"url",
"=",
"None",
",",
"https",
"=",
"False",
",",
"timeout",
"=",
"1",
",",
"max_size",
"=",
"65535",
",",
"sequence",
"=",
"0",
")",
":",
"try",
":",
"result",
"=",
"scan",
"(",
"host",
"=",
"host",
",",
"port",
"=",
"port",
",",
"url",
"=",
"url",
",",
"https",
"=",
"https",
",",
"timeout",
"=",
"timeout",
",",
"max_size",
"=",
"max_size",
")",
"except",
"ScanFailed",
"as",
"failure",
":",
"result",
"=",
"failure",
".",
"result",
"result",
"[",
"'error'",
"]",
"=",
"True",
"result",
"[",
"'error_message'",
"]",
"=",
"str",
"(",
"failure",
")",
"result_obj",
"=",
"PingResponse",
"(",
"host",
"=",
"host",
",",
"port",
"=",
"port",
",",
"ip",
"=",
"result",
".",
"get",
"(",
"'ip'",
",",
"None",
")",
",",
"sequence",
"=",
"sequence",
",",
"durations",
"=",
"result",
".",
"get",
"(",
"'durations'",
",",
"None",
")",
",",
"code",
"=",
"result",
".",
"get",
"(",
"'code'",
",",
"None",
")",
",",
"state",
"=",
"result",
".",
"get",
"(",
"'state'",
",",
"'unknown'",
")",
",",
"length",
"=",
"result",
".",
"get",
"(",
"'length'",
",",
"0",
")",
",",
"response",
"=",
"result",
".",
"get",
"(",
"'response'",
",",
"None",
")",
",",
"error",
"=",
"result",
".",
"get",
"(",
"'error'",
",",
"False",
")",
",",
"error_message",
"=",
"result",
".",
"get",
"(",
"'error_message'",
",",
"None",
")",
",",
"responding",
"=",
"True",
"if",
"result",
".",
"get",
"(",
"'state'",
",",
"'unknown'",
")",
"in",
"[",
"'open'",
"]",
"else",
"False",
",",
"start",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
",",
"end",
"=",
"datetime",
".",
"datetime",
".",
"now",
"(",
")",
"+",
"result",
"[",
"'durations'",
"]",
".",
"get",
"(",
"'all'",
",",
"datetime",
".",
"timedelta",
"(",
"0",
")",
")",
"if",
"result",
".",
"get",
"(",
"'durations'",
",",
"None",
")",
"else",
"None",
")",
"return",
"result_obj"
] | Ping a host
Parameters
----------
host: str
The host or ip address to ping
port: int, optional
The port to ping, default=80
url: str, optional
URL to ping, will do a host/port ping if not provided
https: bool, optional
Connect via ssl, default=False
timeout: int, optional
Number of seconds to wait for a response before timing out, default=1 second
max_size: int, optional
The max size of response that can be retrieved. This should be a power of 2
default=65535.
sequence: int, optional
Sequence number for the ping request
Returns
-------
PingResponse:
The ping response object | [
"Ping",
"a",
"host"
] | 1f9df5ee5b3cba466426b1164262278472ba4977 | https://github.com/yahoo/serviceping/blob/1f9df5ee5b3cba466426b1164262278472ba4977/serviceping/network.py#L213-L265 | train |
muckamuck/stackility | stackility/command.py | delete | def delete(stack, region, profile):
"""
Delete the given CloudFormation stack.
"""
ini_data = {}
environment = {}
environment['stack_name'] = stack
if region:
environment['region'] = region
else:
environment['region'] = find_myself()
if profile:
environment['profile'] = profile
ini_data['environment'] = environment
if start_smash(ini_data):
sys.exit(0)
else:
sys.exit(1) | python | def delete(stack, region, profile):
"""
Delete the given CloudFormation stack.
"""
ini_data = {}
environment = {}
environment['stack_name'] = stack
if region:
environment['region'] = region
else:
environment['region'] = find_myself()
if profile:
environment['profile'] = profile
ini_data['environment'] = environment
if start_smash(ini_data):
sys.exit(0)
else:
sys.exit(1) | [
"def",
"delete",
"(",
"stack",
",",
"region",
",",
"profile",
")",
":",
"ini_data",
"=",
"{",
"}",
"environment",
"=",
"{",
"}",
"environment",
"[",
"'stack_name'",
"]",
"=",
"stack",
"if",
"region",
":",
"environment",
"[",
"'region'",
"]",
"=",
"region",
"else",
":",
"environment",
"[",
"'region'",
"]",
"=",
"find_myself",
"(",
")",
"if",
"profile",
":",
"environment",
"[",
"'profile'",
"]",
"=",
"profile",
"ini_data",
"[",
"'environment'",
"]",
"=",
"environment",
"if",
"start_smash",
"(",
"ini_data",
")",
":",
"sys",
".",
"exit",
"(",
"0",
")",
"else",
":",
"sys",
".",
"exit",
"(",
"1",
")"
] | Delete the given CloudFormation stack. | [
"Delete",
"the",
"given",
"CloudFormation",
"stack",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/command.py#L81-L102 | train |
muckamuck/stackility | stackility/command.py | list | def list(region, profile):
"""
List all the CloudFormation stacks in the given region.
"""
ini_data = {}
environment = {}
if region:
environment['region'] = region
else:
environment['region'] = find_myself()
if profile:
environment['profile'] = profile
ini_data['environment'] = environment
if start_list(ini_data):
sys.exit(0)
else:
sys.exit(1) | python | def list(region, profile):
"""
List all the CloudFormation stacks in the given region.
"""
ini_data = {}
environment = {}
if region:
environment['region'] = region
else:
environment['region'] = find_myself()
if profile:
environment['profile'] = profile
ini_data['environment'] = environment
if start_list(ini_data):
sys.exit(0)
else:
sys.exit(1) | [
"def",
"list",
"(",
"region",
",",
"profile",
")",
":",
"ini_data",
"=",
"{",
"}",
"environment",
"=",
"{",
"}",
"if",
"region",
":",
"environment",
"[",
"'region'",
"]",
"=",
"region",
"else",
":",
"environment",
"[",
"'region'",
"]",
"=",
"find_myself",
"(",
")",
"if",
"profile",
":",
"environment",
"[",
"'profile'",
"]",
"=",
"profile",
"ini_data",
"[",
"'environment'",
"]",
"=",
"environment",
"if",
"start_list",
"(",
"ini_data",
")",
":",
"sys",
".",
"exit",
"(",
"0",
")",
"else",
":",
"sys",
".",
"exit",
"(",
"1",
")"
] | List all the CloudFormation stacks in the given region. | [
"List",
"all",
"the",
"CloudFormation",
"stacks",
"in",
"the",
"given",
"region",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/command.py#L108-L127 | train |
muckamuck/stackility | stackility/command.py | drift | def drift(stack, region, profile):
"""
Produce a CloudFormation drift report for the given stack.
"""
logging.debug('finding drift - stack: {}'.format(stack))
logging.debug('region: {}'.format(region))
logging.debug('profile: {}'.format(profile))
tool = DriftTool(
Stack=stack,
Region=region,
Profile=profile,
Verbose=True
)
if tool.determine_drift():
sys.exit(0)
else:
sys.exit(1) | python | def drift(stack, region, profile):
"""
Produce a CloudFormation drift report for the given stack.
"""
logging.debug('finding drift - stack: {}'.format(stack))
logging.debug('region: {}'.format(region))
logging.debug('profile: {}'.format(profile))
tool = DriftTool(
Stack=stack,
Region=region,
Profile=profile,
Verbose=True
)
if tool.determine_drift():
sys.exit(0)
else:
sys.exit(1) | [
"def",
"drift",
"(",
"stack",
",",
"region",
",",
"profile",
")",
":",
"logging",
".",
"debug",
"(",
"'finding drift - stack: {}'",
".",
"format",
"(",
"stack",
")",
")",
"logging",
".",
"debug",
"(",
"'region: {}'",
".",
"format",
"(",
"region",
")",
")",
"logging",
".",
"debug",
"(",
"'profile: {}'",
".",
"format",
"(",
"profile",
")",
")",
"tool",
"=",
"DriftTool",
"(",
"Stack",
"=",
"stack",
",",
"Region",
"=",
"region",
",",
"Profile",
"=",
"profile",
",",
"Verbose",
"=",
"True",
")",
"if",
"tool",
".",
"determine_drift",
"(",
")",
":",
"sys",
".",
"exit",
"(",
"0",
")",
"else",
":",
"sys",
".",
"exit",
"(",
"1",
")"
] | Produce a CloudFormation drift report for the given stack. | [
"Produce",
"a",
"CloudFormation",
"drift",
"report",
"for",
"the",
"given",
"stack",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/command.py#L134-L151 | train |
muckamuck/stackility | stackility/command.py | start_upsert | def start_upsert(ini_data):
"""
Helper function to facilitate upsert.
Args:
ini_date - the dictionary of info to run upsert
Exit:
0 - good
1 - bad
"""
stack_driver = CloudStackUtility(ini_data)
poll_stack = not ini_data.get('no_poll', False)
if stack_driver.upsert():
logging.info('stack create/update was started successfully.')
if poll_stack:
stack_tool = None
try:
profile = ini_data.get('environment', {}).get('profile')
if profile:
boto3_session = boto3.session.Session(profile_name=profile)
else:
boto3_session = boto3.session.Session()
region = ini_data['environment']['region']
stack_name = ini_data['environment']['stack_name']
cf_client = stack_driver.get_cloud_formation_client()
if not cf_client:
cf_client = boto3_session.client('cloudformation', region_name=region)
stack_tool = stack_tool = StackTool(
stack_name,
region,
cf_client
)
except Exception as wtf:
logging.warning('there was a problems creating stack tool: {}'.format(wtf))
if stack_driver.poll_stack():
try:
logging.info('stack create/update was finished successfully.')
stack_tool.print_stack_info()
except Exception as wtf:
logging.warning('there was a problems printing stack info: {}'.format(wtf))
sys.exit(0)
else:
try:
logging.error('stack create/update was did not go well.')
stack_tool.print_stack_events()
except Exception as wtf:
logging.warning('there was a problems printing stack events: {}'.format(wtf))
sys.exit(1)
else:
logging.error('start of stack create/update did not go well.')
sys.exit(1) | python | def start_upsert(ini_data):
"""
Helper function to facilitate upsert.
Args:
ini_date - the dictionary of info to run upsert
Exit:
0 - good
1 - bad
"""
stack_driver = CloudStackUtility(ini_data)
poll_stack = not ini_data.get('no_poll', False)
if stack_driver.upsert():
logging.info('stack create/update was started successfully.')
if poll_stack:
stack_tool = None
try:
profile = ini_data.get('environment', {}).get('profile')
if profile:
boto3_session = boto3.session.Session(profile_name=profile)
else:
boto3_session = boto3.session.Session()
region = ini_data['environment']['region']
stack_name = ini_data['environment']['stack_name']
cf_client = stack_driver.get_cloud_formation_client()
if not cf_client:
cf_client = boto3_session.client('cloudformation', region_name=region)
stack_tool = stack_tool = StackTool(
stack_name,
region,
cf_client
)
except Exception as wtf:
logging.warning('there was a problems creating stack tool: {}'.format(wtf))
if stack_driver.poll_stack():
try:
logging.info('stack create/update was finished successfully.')
stack_tool.print_stack_info()
except Exception as wtf:
logging.warning('there was a problems printing stack info: {}'.format(wtf))
sys.exit(0)
else:
try:
logging.error('stack create/update was did not go well.')
stack_tool.print_stack_events()
except Exception as wtf:
logging.warning('there was a problems printing stack events: {}'.format(wtf))
sys.exit(1)
else:
logging.error('start of stack create/update did not go well.')
sys.exit(1) | [
"def",
"start_upsert",
"(",
"ini_data",
")",
":",
"stack_driver",
"=",
"CloudStackUtility",
"(",
"ini_data",
")",
"poll_stack",
"=",
"not",
"ini_data",
".",
"get",
"(",
"'no_poll'",
",",
"False",
")",
"if",
"stack_driver",
".",
"upsert",
"(",
")",
":",
"logging",
".",
"info",
"(",
"'stack create/update was started successfully.'",
")",
"if",
"poll_stack",
":",
"stack_tool",
"=",
"None",
"try",
":",
"profile",
"=",
"ini_data",
".",
"get",
"(",
"'environment'",
",",
"{",
"}",
")",
".",
"get",
"(",
"'profile'",
")",
"if",
"profile",
":",
"boto3_session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
"profile_name",
"=",
"profile",
")",
"else",
":",
"boto3_session",
"=",
"boto3",
".",
"session",
".",
"Session",
"(",
")",
"region",
"=",
"ini_data",
"[",
"'environment'",
"]",
"[",
"'region'",
"]",
"stack_name",
"=",
"ini_data",
"[",
"'environment'",
"]",
"[",
"'stack_name'",
"]",
"cf_client",
"=",
"stack_driver",
".",
"get_cloud_formation_client",
"(",
")",
"if",
"not",
"cf_client",
":",
"cf_client",
"=",
"boto3_session",
".",
"client",
"(",
"'cloudformation'",
",",
"region_name",
"=",
"region",
")",
"stack_tool",
"=",
"stack_tool",
"=",
"StackTool",
"(",
"stack_name",
",",
"region",
",",
"cf_client",
")",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"warning",
"(",
"'there was a problems creating stack tool: {}'",
".",
"format",
"(",
"wtf",
")",
")",
"if",
"stack_driver",
".",
"poll_stack",
"(",
")",
":",
"try",
":",
"logging",
".",
"info",
"(",
"'stack create/update was finished successfully.'",
")",
"stack_tool",
".",
"print_stack_info",
"(",
")",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"warning",
"(",
"'there was a problems printing stack info: {}'",
".",
"format",
"(",
"wtf",
")",
")",
"sys",
".",
"exit",
"(",
"0",
")",
"else",
":",
"try",
":",
"logging",
".",
"error",
"(",
"'stack create/update was did not go well.'",
")",
"stack_tool",
".",
"print_stack_events",
"(",
")",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"warning",
"(",
"'there was a problems printing stack events: {}'",
".",
"format",
"(",
"wtf",
")",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"else",
":",
"logging",
".",
"error",
"(",
"'start of stack create/update did not go well.'",
")",
"sys",
".",
"exit",
"(",
"1",
")"
] | Helper function to facilitate upsert.
Args:
ini_date - the dictionary of info to run upsert
Exit:
0 - good
1 - bad | [
"Helper",
"function",
"to",
"facilitate",
"upsert",
"."
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/command.py#L154-L212 | train |
muckamuck/stackility | stackility/command.py | read_config_info | def read_config_info(ini_file):
"""
Read the INI file
Args:
ini_file - path to the file
Returns:
A dictionary of stuff from the INI file
Exits:
1 - if problems are encountered
"""
try:
config = RawConfigParser()
config.optionxform = lambda option: option
config.read(ini_file)
the_stuff = {}
for section in config.sections():
the_stuff[section] = {}
for option in config.options(section):
the_stuff[section][option] = config.get(section, option)
return the_stuff
except Exception as wtf:
logging.error('Exception caught in read_config_info(): {}'.format(wtf))
traceback.print_exc(file=sys.stdout)
return sys.exit(1) | python | def read_config_info(ini_file):
"""
Read the INI file
Args:
ini_file - path to the file
Returns:
A dictionary of stuff from the INI file
Exits:
1 - if problems are encountered
"""
try:
config = RawConfigParser()
config.optionxform = lambda option: option
config.read(ini_file)
the_stuff = {}
for section in config.sections():
the_stuff[section] = {}
for option in config.options(section):
the_stuff[section][option] = config.get(section, option)
return the_stuff
except Exception as wtf:
logging.error('Exception caught in read_config_info(): {}'.format(wtf))
traceback.print_exc(file=sys.stdout)
return sys.exit(1) | [
"def",
"read_config_info",
"(",
"ini_file",
")",
":",
"try",
":",
"config",
"=",
"RawConfigParser",
"(",
")",
"config",
".",
"optionxform",
"=",
"lambda",
"option",
":",
"option",
"config",
".",
"read",
"(",
"ini_file",
")",
"the_stuff",
"=",
"{",
"}",
"for",
"section",
"in",
"config",
".",
"sections",
"(",
")",
":",
"the_stuff",
"[",
"section",
"]",
"=",
"{",
"}",
"for",
"option",
"in",
"config",
".",
"options",
"(",
"section",
")",
":",
"the_stuff",
"[",
"section",
"]",
"[",
"option",
"]",
"=",
"config",
".",
"get",
"(",
"section",
",",
"option",
")",
"return",
"the_stuff",
"except",
"Exception",
"as",
"wtf",
":",
"logging",
".",
"error",
"(",
"'Exception caught in read_config_info(): {}'",
".",
"format",
"(",
"wtf",
")",
")",
"traceback",
".",
"print_exc",
"(",
"file",
"=",
"sys",
".",
"stdout",
")",
"return",
"sys",
".",
"exit",
"(",
"1",
")"
] | Read the INI file
Args:
ini_file - path to the file
Returns:
A dictionary of stuff from the INI file
Exits:
1 - if problems are encountered | [
"Read",
"the",
"INI",
"file"
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/command.py#L257-L284 | train |
muckamuck/stackility | stackility/stack_tool.py | StackTool.print_stack_info | def print_stack_info(self):
'''
List resources from the given stack
Args:
None
Returns:
A dictionary filled resources or None if things went sideways
'''
try:
rest_api_id = None
deployment_found = False
response = self._cf_client.describe_stack_resources(
StackName=self._stack_name
)
print('\nThe following resources were created:')
rows = []
for resource in response['StackResources']:
if resource['ResourceType'] == 'AWS::ApiGateway::RestApi':
rest_api_id = resource['PhysicalResourceId']
elif resource['ResourceType'] == 'AWS::ApiGateway::Deployment':
deployment_found = True
row = []
row.append(resource['ResourceType'])
row.append(resource['LogicalResourceId'])
row.append(resource['PhysicalResourceId'])
rows.append(row)
'''
print('\t{}\t{}\t{}'.format(
resource['ResourceType'],
resource['LogicalResourceId'],
resource['PhysicalResourceId']
)
)
'''
print(tabulate(rows, headers=['Resource Type', 'Logical ID', 'Physical ID']))
if rest_api_id and deployment_found:
url = 'https://{}.execute-api.{}.amazonaws.com/{}'.format(
rest_api_id,
self._region,
'<stage>'
)
print('\nThe deployed service can be found at this URL:')
print('\t{}\n'.format(url))
return response
except Exception as wtf:
print(wtf)
return None | python | def print_stack_info(self):
'''
List resources from the given stack
Args:
None
Returns:
A dictionary filled resources or None if things went sideways
'''
try:
rest_api_id = None
deployment_found = False
response = self._cf_client.describe_stack_resources(
StackName=self._stack_name
)
print('\nThe following resources were created:')
rows = []
for resource in response['StackResources']:
if resource['ResourceType'] == 'AWS::ApiGateway::RestApi':
rest_api_id = resource['PhysicalResourceId']
elif resource['ResourceType'] == 'AWS::ApiGateway::Deployment':
deployment_found = True
row = []
row.append(resource['ResourceType'])
row.append(resource['LogicalResourceId'])
row.append(resource['PhysicalResourceId'])
rows.append(row)
'''
print('\t{}\t{}\t{}'.format(
resource['ResourceType'],
resource['LogicalResourceId'],
resource['PhysicalResourceId']
)
)
'''
print(tabulate(rows, headers=['Resource Type', 'Logical ID', 'Physical ID']))
if rest_api_id and deployment_found:
url = 'https://{}.execute-api.{}.amazonaws.com/{}'.format(
rest_api_id,
self._region,
'<stage>'
)
print('\nThe deployed service can be found at this URL:')
print('\t{}\n'.format(url))
return response
except Exception as wtf:
print(wtf)
return None | [
"def",
"print_stack_info",
"(",
"self",
")",
":",
"try",
":",
"rest_api_id",
"=",
"None",
"deployment_found",
"=",
"False",
"response",
"=",
"self",
".",
"_cf_client",
".",
"describe_stack_resources",
"(",
"StackName",
"=",
"self",
".",
"_stack_name",
")",
"print",
"(",
"'\\nThe following resources were created:'",
")",
"rows",
"=",
"[",
"]",
"for",
"resource",
"in",
"response",
"[",
"'StackResources'",
"]",
":",
"if",
"resource",
"[",
"'ResourceType'",
"]",
"==",
"'AWS::ApiGateway::RestApi'",
":",
"rest_api_id",
"=",
"resource",
"[",
"'PhysicalResourceId'",
"]",
"elif",
"resource",
"[",
"'ResourceType'",
"]",
"==",
"'AWS::ApiGateway::Deployment'",
":",
"deployment_found",
"=",
"True",
"row",
"=",
"[",
"]",
"row",
".",
"append",
"(",
"resource",
"[",
"'ResourceType'",
"]",
")",
"row",
".",
"append",
"(",
"resource",
"[",
"'LogicalResourceId'",
"]",
")",
"row",
".",
"append",
"(",
"resource",
"[",
"'PhysicalResourceId'",
"]",
")",
"rows",
".",
"append",
"(",
"row",
")",
"'''\n print('\\t{}\\t{}\\t{}'.format(\n resource['ResourceType'],\n resource['LogicalResourceId'],\n resource['PhysicalResourceId']\n )\n )\n '''",
"print",
"(",
"tabulate",
"(",
"rows",
",",
"headers",
"=",
"[",
"'Resource Type'",
",",
"'Logical ID'",
",",
"'Physical ID'",
"]",
")",
")",
"if",
"rest_api_id",
"and",
"deployment_found",
":",
"url",
"=",
"'https://{}.execute-api.{}.amazonaws.com/{}'",
".",
"format",
"(",
"rest_api_id",
",",
"self",
".",
"_region",
",",
"'<stage>'",
")",
"print",
"(",
"'\\nThe deployed service can be found at this URL:'",
")",
"print",
"(",
"'\\t{}\\n'",
".",
"format",
"(",
"url",
")",
")",
"return",
"response",
"except",
"Exception",
"as",
"wtf",
":",
"print",
"(",
"wtf",
")",
"return",
"None"
] | List resources from the given stack
Args:
None
Returns:
A dictionary filled resources or None if things went sideways | [
"List",
"resources",
"from",
"the",
"given",
"stack"
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/stack_tool.py#L39-L92 | train |
muckamuck/stackility | stackility/stack_tool.py | StackTool.print_stack_events | def print_stack_events(self):
'''
List events from the given stack
Args:
None
Returns:
None
'''
first_token = '7be7981bd6287dd8112305e8f3822a6f'
keep_going = True
next_token = first_token
current_request_token = None
rows = []
try:
while keep_going and next_token:
if next_token == first_token:
response = self._cf_client.describe_stack_events(
StackName=self._stack_name
)
else:
response = self._cf_client.describe_stack_events(
StackName=self._stack_name,
NextToken=next_token
)
next_token = response.get('NextToken', None)
for event in response['StackEvents']:
row = []
event_time = event.get('Timestamp')
request_token = event.get('ClientRequestToken', 'unknown')
if current_request_token is None:
current_request_token = request_token
elif current_request_token != request_token:
keep_going = False
break
row.append(event_time.strftime('%x %X'))
row.append(event.get('LogicalResourceId'))
row.append(event.get('ResourceStatus'))
row.append(event.get('ResourceStatusReason', ''))
rows.append(row)
if len(rows) > 0:
print('\nEvents for the current upsert:')
print(tabulate(rows, headers=['Time', 'Logical ID', 'Status', 'Message']))
return True
else:
print('\nNo stack events found\n')
except Exception as wtf:
print(wtf)
return False | python | def print_stack_events(self):
'''
List events from the given stack
Args:
None
Returns:
None
'''
first_token = '7be7981bd6287dd8112305e8f3822a6f'
keep_going = True
next_token = first_token
current_request_token = None
rows = []
try:
while keep_going and next_token:
if next_token == first_token:
response = self._cf_client.describe_stack_events(
StackName=self._stack_name
)
else:
response = self._cf_client.describe_stack_events(
StackName=self._stack_name,
NextToken=next_token
)
next_token = response.get('NextToken', None)
for event in response['StackEvents']:
row = []
event_time = event.get('Timestamp')
request_token = event.get('ClientRequestToken', 'unknown')
if current_request_token is None:
current_request_token = request_token
elif current_request_token != request_token:
keep_going = False
break
row.append(event_time.strftime('%x %X'))
row.append(event.get('LogicalResourceId'))
row.append(event.get('ResourceStatus'))
row.append(event.get('ResourceStatusReason', ''))
rows.append(row)
if len(rows) > 0:
print('\nEvents for the current upsert:')
print(tabulate(rows, headers=['Time', 'Logical ID', 'Status', 'Message']))
return True
else:
print('\nNo stack events found\n')
except Exception as wtf:
print(wtf)
return False | [
"def",
"print_stack_events",
"(",
"self",
")",
":",
"first_token",
"=",
"'7be7981bd6287dd8112305e8f3822a6f'",
"keep_going",
"=",
"True",
"next_token",
"=",
"first_token",
"current_request_token",
"=",
"None",
"rows",
"=",
"[",
"]",
"try",
":",
"while",
"keep_going",
"and",
"next_token",
":",
"if",
"next_token",
"==",
"first_token",
":",
"response",
"=",
"self",
".",
"_cf_client",
".",
"describe_stack_events",
"(",
"StackName",
"=",
"self",
".",
"_stack_name",
")",
"else",
":",
"response",
"=",
"self",
".",
"_cf_client",
".",
"describe_stack_events",
"(",
"StackName",
"=",
"self",
".",
"_stack_name",
",",
"NextToken",
"=",
"next_token",
")",
"next_token",
"=",
"response",
".",
"get",
"(",
"'NextToken'",
",",
"None",
")",
"for",
"event",
"in",
"response",
"[",
"'StackEvents'",
"]",
":",
"row",
"=",
"[",
"]",
"event_time",
"=",
"event",
".",
"get",
"(",
"'Timestamp'",
")",
"request_token",
"=",
"event",
".",
"get",
"(",
"'ClientRequestToken'",
",",
"'unknown'",
")",
"if",
"current_request_token",
"is",
"None",
":",
"current_request_token",
"=",
"request_token",
"elif",
"current_request_token",
"!=",
"request_token",
":",
"keep_going",
"=",
"False",
"break",
"row",
".",
"append",
"(",
"event_time",
".",
"strftime",
"(",
"'%x %X'",
")",
")",
"row",
".",
"append",
"(",
"event",
".",
"get",
"(",
"'LogicalResourceId'",
")",
")",
"row",
".",
"append",
"(",
"event",
".",
"get",
"(",
"'ResourceStatus'",
")",
")",
"row",
".",
"append",
"(",
"event",
".",
"get",
"(",
"'ResourceStatusReason'",
",",
"''",
")",
")",
"rows",
".",
"append",
"(",
"row",
")",
"if",
"len",
"(",
"rows",
")",
">",
"0",
":",
"print",
"(",
"'\\nEvents for the current upsert:'",
")",
"print",
"(",
"tabulate",
"(",
"rows",
",",
"headers",
"=",
"[",
"'Time'",
",",
"'Logical ID'",
",",
"'Status'",
",",
"'Message'",
"]",
")",
")",
"return",
"True",
"else",
":",
"print",
"(",
"'\\nNo stack events found\\n'",
")",
"except",
"Exception",
"as",
"wtf",
":",
"print",
"(",
"wtf",
")",
"return",
"False"
] | List events from the given stack
Args:
None
Returns:
None | [
"List",
"events",
"from",
"the",
"given",
"stack"
] | b1696f02661134d31b99b4dea7c0d21d09482d33 | https://github.com/muckamuck/stackility/blob/b1696f02661134d31b99b4dea7c0d21d09482d33/stackility/stack_tool.py#L94-L147 | train |
theno/fabsetup | fabsetup/fabfile/setup/service/trac.py | trac | def trac():
'''Set up or update a trac project.
This trac installation uses python2, git, sqlite (trac-default), gunicorn,
and nginx.
The connection is https-only and secured by a letsencrypt certificate. This
certificate must be created separately with task setup.server_letsencrypt.
This task installes or updates to the latest trac version hosted at
https://pypi.python.org/pypi/Trac
Created and modified files and dirs of this task:
```
~/sites/<sitename> <--- example: sitename = trac.example.com
│
├── backup.sh <--- create a local backup (deletes dir backup/
├── backup | before it creates
│ └── <sitename>_tracenv_hotcopy.tar.gz <--´ the tarball)
├── run
│ └── trac.sock <--- file-socket for binding to nginx
├── scripts
│ └── tracwsgi.py
├── tracenv
│ ├── conf
│ │ ├── trac.htpasswd <--- trac user password hashes
│ │ └── trac.ini <--- trac config file
│ ├── db
│ │ └── trac.db <--- sqlite database
│ ├── files
│ ├── git
│ ├── htdocs
│ ├── log
│ ├── plugins
│ ├── README
│ ├── templates
│ └── VERSION
└── virtualenv
├── bin
├── include
├── lib
├── local
└── pip-selfcheck.json
```
How to create a backup tarball "manually":
`~/sites/<sitename>/backup/tracenv_hotcopy_<yyyy-mm-dd>.tar.gz`:
```
cd ~/sites/<sitename> && rm -rf ./backup
./virtualenv/bin/trac-admin ./tracenv hotcopy ./backup/tracenv_hotcopy
mkdir -p ./backup && cd ./backup
tar czf <sitename>_tracenv_hotcopy_$(date +%F).tar.gz tracenv_hotcopy/
rm -rf tracenv_hotcopy; ls -hl
```
More infos:
https://trac.edgewall.org/wiki/TracInstall
https://trac.edgewall.org/wiki/TracFastCgi#NginxConfiguration
https://trac.edgewall.org/wiki/TracNginxRecipe
https://trac.edgewall.org/wiki/Gunicorn
http://www.obeythetestinggoat.com/book/chapter_08.html#_getting_to_a_production_ready_deployment
Setting REMOTE_USER for Trac in Gunicorn behind Nginx:
http://serverfault.com/a/392096
https://trac.edgewall.org/wiki/TracBackup
'''
hostname = re.sub(r'^[^@]+@', '', env.host) # without username if any
sitename = query_input(
question='\nEnter site-name of Your trac web service',
default=flo('trac.{hostname}'))
username = env.user
site_dir = flo('/home/{username}/sites/{sitename}')
bin_dir = flo('{site_dir}/virtualenv/bin')
# provisioning steps
install_or_upgrade_virtualenv_pip_package()
create_directory_structure(site_dir)
update_virtualenv(site_dir, sitename)
set_up_trac_plugins(sitename, site_dir, bin_dir)
set_up_gunicorn(site_dir, sitename)
configure_nginx(username, sitename, hostname)
if query_yes_no('\nRestore trac environment from backup tarball?',
default='no'):
restore_tracenv_from_backup_tarball(site_dir, bin_dir)
elif not tracenv_exists(site_dir):
init_tracenv(site_dir, bin_dir, username)
upgrade_tracenv(site_dir, bin_dir)
set_up_upstart_for_gunicorn(sitename, username, site_dir) | python | def trac():
'''Set up or update a trac project.
This trac installation uses python2, git, sqlite (trac-default), gunicorn,
and nginx.
The connection is https-only and secured by a letsencrypt certificate. This
certificate must be created separately with task setup.server_letsencrypt.
This task installes or updates to the latest trac version hosted at
https://pypi.python.org/pypi/Trac
Created and modified files and dirs of this task:
```
~/sites/<sitename> <--- example: sitename = trac.example.com
│
├── backup.sh <--- create a local backup (deletes dir backup/
├── backup | before it creates
│ └── <sitename>_tracenv_hotcopy.tar.gz <--´ the tarball)
├── run
│ └── trac.sock <--- file-socket for binding to nginx
├── scripts
│ └── tracwsgi.py
├── tracenv
│ ├── conf
│ │ ├── trac.htpasswd <--- trac user password hashes
│ │ └── trac.ini <--- trac config file
│ ├── db
│ │ └── trac.db <--- sqlite database
│ ├── files
│ ├── git
│ ├── htdocs
│ ├── log
│ ├── plugins
│ ├── README
│ ├── templates
│ └── VERSION
└── virtualenv
├── bin
├── include
├── lib
├── local
└── pip-selfcheck.json
```
How to create a backup tarball "manually":
`~/sites/<sitename>/backup/tracenv_hotcopy_<yyyy-mm-dd>.tar.gz`:
```
cd ~/sites/<sitename> && rm -rf ./backup
./virtualenv/bin/trac-admin ./tracenv hotcopy ./backup/tracenv_hotcopy
mkdir -p ./backup && cd ./backup
tar czf <sitename>_tracenv_hotcopy_$(date +%F).tar.gz tracenv_hotcopy/
rm -rf tracenv_hotcopy; ls -hl
```
More infos:
https://trac.edgewall.org/wiki/TracInstall
https://trac.edgewall.org/wiki/TracFastCgi#NginxConfiguration
https://trac.edgewall.org/wiki/TracNginxRecipe
https://trac.edgewall.org/wiki/Gunicorn
http://www.obeythetestinggoat.com/book/chapter_08.html#_getting_to_a_production_ready_deployment
Setting REMOTE_USER for Trac in Gunicorn behind Nginx:
http://serverfault.com/a/392096
https://trac.edgewall.org/wiki/TracBackup
'''
hostname = re.sub(r'^[^@]+@', '', env.host) # without username if any
sitename = query_input(
question='\nEnter site-name of Your trac web service',
default=flo('trac.{hostname}'))
username = env.user
site_dir = flo('/home/{username}/sites/{sitename}')
bin_dir = flo('{site_dir}/virtualenv/bin')
# provisioning steps
install_or_upgrade_virtualenv_pip_package()
create_directory_structure(site_dir)
update_virtualenv(site_dir, sitename)
set_up_trac_plugins(sitename, site_dir, bin_dir)
set_up_gunicorn(site_dir, sitename)
configure_nginx(username, sitename, hostname)
if query_yes_no('\nRestore trac environment from backup tarball?',
default='no'):
restore_tracenv_from_backup_tarball(site_dir, bin_dir)
elif not tracenv_exists(site_dir):
init_tracenv(site_dir, bin_dir, username)
upgrade_tracenv(site_dir, bin_dir)
set_up_upstart_for_gunicorn(sitename, username, site_dir) | [
"def",
"trac",
"(",
")",
":",
"hostname",
"=",
"re",
".",
"sub",
"(",
"r'^[^@]+@'",
",",
"''",
",",
"env",
".",
"host",
")",
"# without username if any",
"sitename",
"=",
"query_input",
"(",
"question",
"=",
"'\\nEnter site-name of Your trac web service'",
",",
"default",
"=",
"flo",
"(",
"'trac.{hostname}'",
")",
")",
"username",
"=",
"env",
".",
"user",
"site_dir",
"=",
"flo",
"(",
"'/home/{username}/sites/{sitename}'",
")",
"bin_dir",
"=",
"flo",
"(",
"'{site_dir}/virtualenv/bin'",
")",
"# provisioning steps",
"install_or_upgrade_virtualenv_pip_package",
"(",
")",
"create_directory_structure",
"(",
"site_dir",
")",
"update_virtualenv",
"(",
"site_dir",
",",
"sitename",
")",
"set_up_trac_plugins",
"(",
"sitename",
",",
"site_dir",
",",
"bin_dir",
")",
"set_up_gunicorn",
"(",
"site_dir",
",",
"sitename",
")",
"configure_nginx",
"(",
"username",
",",
"sitename",
",",
"hostname",
")",
"if",
"query_yes_no",
"(",
"'\\nRestore trac environment from backup tarball?'",
",",
"default",
"=",
"'no'",
")",
":",
"restore_tracenv_from_backup_tarball",
"(",
"site_dir",
",",
"bin_dir",
")",
"elif",
"not",
"tracenv_exists",
"(",
"site_dir",
")",
":",
"init_tracenv",
"(",
"site_dir",
",",
"bin_dir",
",",
"username",
")",
"upgrade_tracenv",
"(",
"site_dir",
",",
"bin_dir",
")",
"set_up_upstart_for_gunicorn",
"(",
"sitename",
",",
"username",
",",
"site_dir",
")"
] | Set up or update a trac project.
This trac installation uses python2, git, sqlite (trac-default), gunicorn,
and nginx.
The connection is https-only and secured by a letsencrypt certificate. This
certificate must be created separately with task setup.server_letsencrypt.
This task installes or updates to the latest trac version hosted at
https://pypi.python.org/pypi/Trac
Created and modified files and dirs of this task:
```
~/sites/<sitename> <--- example: sitename = trac.example.com
│
├── backup.sh <--- create a local backup (deletes dir backup/
├── backup | before it creates
│ └── <sitename>_tracenv_hotcopy.tar.gz <--´ the tarball)
├── run
│ └── trac.sock <--- file-socket for binding to nginx
├── scripts
│ └── tracwsgi.py
├── tracenv
│ ├── conf
│ │ ├── trac.htpasswd <--- trac user password hashes
│ │ └── trac.ini <--- trac config file
│ ├── db
│ │ └── trac.db <--- sqlite database
│ ├── files
│ ├── git
│ ├── htdocs
│ ├── log
│ ├── plugins
│ ├── README
│ ├── templates
│ └── VERSION
└── virtualenv
├── bin
├── include
├── lib
├── local
└── pip-selfcheck.json
```
How to create a backup tarball "manually":
`~/sites/<sitename>/backup/tracenv_hotcopy_<yyyy-mm-dd>.tar.gz`:
```
cd ~/sites/<sitename> && rm -rf ./backup
./virtualenv/bin/trac-admin ./tracenv hotcopy ./backup/tracenv_hotcopy
mkdir -p ./backup && cd ./backup
tar czf <sitename>_tracenv_hotcopy_$(date +%F).tar.gz tracenv_hotcopy/
rm -rf tracenv_hotcopy; ls -hl
```
More infos:
https://trac.edgewall.org/wiki/TracInstall
https://trac.edgewall.org/wiki/TracFastCgi#NginxConfiguration
https://trac.edgewall.org/wiki/TracNginxRecipe
https://trac.edgewall.org/wiki/Gunicorn
http://www.obeythetestinggoat.com/book/chapter_08.html#_getting_to_a_production_ready_deployment
Setting REMOTE_USER for Trac in Gunicorn behind Nginx:
http://serverfault.com/a/392096
https://trac.edgewall.org/wiki/TracBackup | [
"Set",
"up",
"or",
"update",
"a",
"trac",
"project",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/service/trac.py#L28-L120 | train |
push-things/wallabag_api | wallabag_api/wallabag.py | Wallabag.query | async def query(self, path, method='get', **params):
"""
Do a query to the System API
:param path: url to the API
:param method: the kind of query to do
:param params: a dict with all the
necessary things to query the API
:return json data
"""
if method in ('get', 'post', 'patch', 'delete', 'put'):
full_path = self.host + path
if method == 'get':
resp = await self.aio_sess.get(full_path, params=params)
elif method == 'post':
resp = await self.aio_sess.post(full_path, data=params)
elif method == 'patch':
resp = await self.aio_sess.patch(full_path, data=params)
elif method == 'delete':
resp = await self.aio_sess.delete(full_path, params=params, headers=params)
elif method == 'put':
resp = await self.aio_sess.put(full_path, data=params)
async with resp:
# return the content if its a binary one
if resp.content_type.startswith('application/pdf') or \
resp.content_type.startswith('application/epub'):
return await resp.read()
return await self.handle_json_response(resp)
else:
raise ValueError('method expected: get, post, patch, delete, put') | python | async def query(self, path, method='get', **params):
"""
Do a query to the System API
:param path: url to the API
:param method: the kind of query to do
:param params: a dict with all the
necessary things to query the API
:return json data
"""
if method in ('get', 'post', 'patch', 'delete', 'put'):
full_path = self.host + path
if method == 'get':
resp = await self.aio_sess.get(full_path, params=params)
elif method == 'post':
resp = await self.aio_sess.post(full_path, data=params)
elif method == 'patch':
resp = await self.aio_sess.patch(full_path, data=params)
elif method == 'delete':
resp = await self.aio_sess.delete(full_path, params=params, headers=params)
elif method == 'put':
resp = await self.aio_sess.put(full_path, data=params)
async with resp:
# return the content if its a binary one
if resp.content_type.startswith('application/pdf') or \
resp.content_type.startswith('application/epub'):
return await resp.read()
return await self.handle_json_response(resp)
else:
raise ValueError('method expected: get, post, patch, delete, put') | [
"async",
"def",
"query",
"(",
"self",
",",
"path",
",",
"method",
"=",
"'get'",
",",
"*",
"*",
"params",
")",
":",
"if",
"method",
"in",
"(",
"'get'",
",",
"'post'",
",",
"'patch'",
",",
"'delete'",
",",
"'put'",
")",
":",
"full_path",
"=",
"self",
".",
"host",
"+",
"path",
"if",
"method",
"==",
"'get'",
":",
"resp",
"=",
"await",
"self",
".",
"aio_sess",
".",
"get",
"(",
"full_path",
",",
"params",
"=",
"params",
")",
"elif",
"method",
"==",
"'post'",
":",
"resp",
"=",
"await",
"self",
".",
"aio_sess",
".",
"post",
"(",
"full_path",
",",
"data",
"=",
"params",
")",
"elif",
"method",
"==",
"'patch'",
":",
"resp",
"=",
"await",
"self",
".",
"aio_sess",
".",
"patch",
"(",
"full_path",
",",
"data",
"=",
"params",
")",
"elif",
"method",
"==",
"'delete'",
":",
"resp",
"=",
"await",
"self",
".",
"aio_sess",
".",
"delete",
"(",
"full_path",
",",
"params",
"=",
"params",
",",
"headers",
"=",
"params",
")",
"elif",
"method",
"==",
"'put'",
":",
"resp",
"=",
"await",
"self",
".",
"aio_sess",
".",
"put",
"(",
"full_path",
",",
"data",
"=",
"params",
")",
"async",
"with",
"resp",
":",
"# return the content if its a binary one",
"if",
"resp",
".",
"content_type",
".",
"startswith",
"(",
"'application/pdf'",
")",
"or",
"resp",
".",
"content_type",
".",
"startswith",
"(",
"'application/epub'",
")",
":",
"return",
"await",
"resp",
".",
"read",
"(",
")",
"return",
"await",
"self",
".",
"handle_json_response",
"(",
"resp",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'method expected: get, post, patch, delete, put'",
")"
] | Do a query to the System API
:param path: url to the API
:param method: the kind of query to do
:param params: a dict with all the
necessary things to query the API
:return json data | [
"Do",
"a",
"query",
"to",
"the",
"System",
"API"
] | 8d1e10a6ebc03d1ac9af2b38b57eb69f29b4216e | https://github.com/push-things/wallabag_api/blob/8d1e10a6ebc03d1ac9af2b38b57eb69f29b4216e/wallabag_api/wallabag.py#L60-L91 | train |
theno/fabsetup | fabsetup/fabfile/setup/revealjs.py | revealjs | def revealjs(basedir=None, title=None, subtitle=None, description=None,
github_user=None, github_repo=None):
'''Set up or update a reveals.js presentation with slides written in markdown.
Several reveal.js plugins will be set up, too.
More info:
Demo: https://theno.github.io/revealjs_template
http://lab.hakim.se/reveal-js/
https://github.com/hakimel/reveal.js
plugins:
https://github.com/hakimel/reveal.js/wiki/Plugins,-Tools-and-Hardware
https://github.com/rajgoel/reveal.js-plugins/
https://github.com/e-gor/Reveal.js-TOC-Progress
https://github.com/e-gor/Reveal.js-Title-Footer
'''
basedir = basedir or query_input('Base dir of the presentation?',
default='~/repos/my_presi')
revealjs_repo_name = 'reveal.js'
revealjs_dir = flo('{basedir}/{revealjs_repo_name}')
_lazy_dict['presi_title'] = title
_lazy_dict['presi_subtitle'] = subtitle
_lazy_dict['presi_description'] = description
_lazy_dict['github_user'] = github_user
_lazy_dict['github_repo'] = github_repo
question = flo("Base dir already contains a sub dir '{revealjs_repo_name}'."
' Reset (and re-download) reveal.js codebase?')
if not exists(revealjs_dir) or query_yes_no(question, default='no'):
run(flo('mkdir -p {basedir}'))
set_up_revealjs_codebase(basedir, revealjs_repo_name)
install_plugins(revealjs_dir)
apply_customizations(repo_dir=revealjs_dir)
if exists(revealjs_dir):
install_files_in_basedir(basedir, repo_dir=revealjs_dir)
init_git_repo(basedir)
create_github_remote_repo(basedir)
setup_npm(revealjs_dir)
else:
print('abort') | python | def revealjs(basedir=None, title=None, subtitle=None, description=None,
github_user=None, github_repo=None):
'''Set up or update a reveals.js presentation with slides written in markdown.
Several reveal.js plugins will be set up, too.
More info:
Demo: https://theno.github.io/revealjs_template
http://lab.hakim.se/reveal-js/
https://github.com/hakimel/reveal.js
plugins:
https://github.com/hakimel/reveal.js/wiki/Plugins,-Tools-and-Hardware
https://github.com/rajgoel/reveal.js-plugins/
https://github.com/e-gor/Reveal.js-TOC-Progress
https://github.com/e-gor/Reveal.js-Title-Footer
'''
basedir = basedir or query_input('Base dir of the presentation?',
default='~/repos/my_presi')
revealjs_repo_name = 'reveal.js'
revealjs_dir = flo('{basedir}/{revealjs_repo_name}')
_lazy_dict['presi_title'] = title
_lazy_dict['presi_subtitle'] = subtitle
_lazy_dict['presi_description'] = description
_lazy_dict['github_user'] = github_user
_lazy_dict['github_repo'] = github_repo
question = flo("Base dir already contains a sub dir '{revealjs_repo_name}'."
' Reset (and re-download) reveal.js codebase?')
if not exists(revealjs_dir) or query_yes_no(question, default='no'):
run(flo('mkdir -p {basedir}'))
set_up_revealjs_codebase(basedir, revealjs_repo_name)
install_plugins(revealjs_dir)
apply_customizations(repo_dir=revealjs_dir)
if exists(revealjs_dir):
install_files_in_basedir(basedir, repo_dir=revealjs_dir)
init_git_repo(basedir)
create_github_remote_repo(basedir)
setup_npm(revealjs_dir)
else:
print('abort') | [
"def",
"revealjs",
"(",
"basedir",
"=",
"None",
",",
"title",
"=",
"None",
",",
"subtitle",
"=",
"None",
",",
"description",
"=",
"None",
",",
"github_user",
"=",
"None",
",",
"github_repo",
"=",
"None",
")",
":",
"basedir",
"=",
"basedir",
"or",
"query_input",
"(",
"'Base dir of the presentation?'",
",",
"default",
"=",
"'~/repos/my_presi'",
")",
"revealjs_repo_name",
"=",
"'reveal.js'",
"revealjs_dir",
"=",
"flo",
"(",
"'{basedir}/{revealjs_repo_name}'",
")",
"_lazy_dict",
"[",
"'presi_title'",
"]",
"=",
"title",
"_lazy_dict",
"[",
"'presi_subtitle'",
"]",
"=",
"subtitle",
"_lazy_dict",
"[",
"'presi_description'",
"]",
"=",
"description",
"_lazy_dict",
"[",
"'github_user'",
"]",
"=",
"github_user",
"_lazy_dict",
"[",
"'github_repo'",
"]",
"=",
"github_repo",
"question",
"=",
"flo",
"(",
"\"Base dir already contains a sub dir '{revealjs_repo_name}'.\"",
"' Reset (and re-download) reveal.js codebase?'",
")",
"if",
"not",
"exists",
"(",
"revealjs_dir",
")",
"or",
"query_yes_no",
"(",
"question",
",",
"default",
"=",
"'no'",
")",
":",
"run",
"(",
"flo",
"(",
"'mkdir -p {basedir}'",
")",
")",
"set_up_revealjs_codebase",
"(",
"basedir",
",",
"revealjs_repo_name",
")",
"install_plugins",
"(",
"revealjs_dir",
")",
"apply_customizations",
"(",
"repo_dir",
"=",
"revealjs_dir",
")",
"if",
"exists",
"(",
"revealjs_dir",
")",
":",
"install_files_in_basedir",
"(",
"basedir",
",",
"repo_dir",
"=",
"revealjs_dir",
")",
"init_git_repo",
"(",
"basedir",
")",
"create_github_remote_repo",
"(",
"basedir",
")",
"setup_npm",
"(",
"revealjs_dir",
")",
"else",
":",
"print",
"(",
"'abort'",
")"
] | Set up or update a reveals.js presentation with slides written in markdown.
Several reveal.js plugins will be set up, too.
More info:
Demo: https://theno.github.io/revealjs_template
http://lab.hakim.se/reveal-js/
https://github.com/hakimel/reveal.js
plugins:
https://github.com/hakimel/reveal.js/wiki/Plugins,-Tools-and-Hardware
https://github.com/rajgoel/reveal.js-plugins/
https://github.com/e-gor/Reveal.js-TOC-Progress
https://github.com/e-gor/Reveal.js-Title-Footer | [
"Set",
"up",
"or",
"update",
"a",
"reveals",
".",
"js",
"presentation",
"with",
"slides",
"written",
"in",
"markdown",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/revealjs.py#L50-L90 | train |
theno/fabsetup | fabsetup/fabfile/setup/revealjs.py | tweak_css | def tweak_css(repo_dir):
'''Comment out some css settings.'''
print_msg("* don't capitalize titles (no uppercase headings)")
files = [
'beige.css', 'black.css', 'blood.css', 'league.css', 'moon.css',
'night.css', 'serif.css', 'simple.css', 'sky.css', 'solarized.css',
'white.css',
]
line = ' text-transform: uppercase;'
for file_ in files:
update_or_append_line(filename=flo('{repo_dir}/css/theme/{file_}'),
prefix=line, new_line=flo('/*{line}*/'))
print_msg('* images without border')
data = [
{'file': 'beige.css', 'line': ' border: 4px solid #333;'},
{'file': 'black.css', 'line': ' border: 4px solid #fff;'},
{'file': 'blood.css', 'line': ' border: 4px solid #eee;'},
{'file': 'league.css', 'line': ' border: 4px solid #eee;'},
{'file': 'moon.css', 'line': ' border: 4px solid #93a1a1;'},
{'file': 'night.css', 'line': ' border: 4px solid #eee;'},
{'file': 'serif.css', 'line': ' border: 4px solid #000;'},
{'file': 'simple.css', 'line': ' border: 4px solid #000;'},
{'file': 'sky.css', 'line': ' border: 4px solid #333;'},
{'file': 'solarized.css', 'line': ' border: 4px solid #657b83;'},
{'file': 'white.css', 'line': ' border: 4px solid #222;'},
]
for item in data:
file_ = item['file']
lines = [item['line'], ]
lines.extend([' box-shadow: 0 0 10px rgba(0, 0, 0, 0.15); }',
' box-shadow: 0 0 20px rgba(0, 0, 0, 0.55); }'])
for line in lines:
update_or_append_line(filename=flo('{repo_dir}/css/theme/{file_}'),
prefix=line, new_line=flo('/*{line}*/')) | python | def tweak_css(repo_dir):
'''Comment out some css settings.'''
print_msg("* don't capitalize titles (no uppercase headings)")
files = [
'beige.css', 'black.css', 'blood.css', 'league.css', 'moon.css',
'night.css', 'serif.css', 'simple.css', 'sky.css', 'solarized.css',
'white.css',
]
line = ' text-transform: uppercase;'
for file_ in files:
update_or_append_line(filename=flo('{repo_dir}/css/theme/{file_}'),
prefix=line, new_line=flo('/*{line}*/'))
print_msg('* images without border')
data = [
{'file': 'beige.css', 'line': ' border: 4px solid #333;'},
{'file': 'black.css', 'line': ' border: 4px solid #fff;'},
{'file': 'blood.css', 'line': ' border: 4px solid #eee;'},
{'file': 'league.css', 'line': ' border: 4px solid #eee;'},
{'file': 'moon.css', 'line': ' border: 4px solid #93a1a1;'},
{'file': 'night.css', 'line': ' border: 4px solid #eee;'},
{'file': 'serif.css', 'line': ' border: 4px solid #000;'},
{'file': 'simple.css', 'line': ' border: 4px solid #000;'},
{'file': 'sky.css', 'line': ' border: 4px solid #333;'},
{'file': 'solarized.css', 'line': ' border: 4px solid #657b83;'},
{'file': 'white.css', 'line': ' border: 4px solid #222;'},
]
for item in data:
file_ = item['file']
lines = [item['line'], ]
lines.extend([' box-shadow: 0 0 10px rgba(0, 0, 0, 0.15); }',
' box-shadow: 0 0 20px rgba(0, 0, 0, 0.55); }'])
for line in lines:
update_or_append_line(filename=flo('{repo_dir}/css/theme/{file_}'),
prefix=line, new_line=flo('/*{line}*/')) | [
"def",
"tweak_css",
"(",
"repo_dir",
")",
":",
"print_msg",
"(",
"\"* don't capitalize titles (no uppercase headings)\"",
")",
"files",
"=",
"[",
"'beige.css'",
",",
"'black.css'",
",",
"'blood.css'",
",",
"'league.css'",
",",
"'moon.css'",
",",
"'night.css'",
",",
"'serif.css'",
",",
"'simple.css'",
",",
"'sky.css'",
",",
"'solarized.css'",
",",
"'white.css'",
",",
"]",
"line",
"=",
"' text-transform: uppercase;'",
"for",
"file_",
"in",
"files",
":",
"update_or_append_line",
"(",
"filename",
"=",
"flo",
"(",
"'{repo_dir}/css/theme/{file_}'",
")",
",",
"prefix",
"=",
"line",
",",
"new_line",
"=",
"flo",
"(",
"'/*{line}*/'",
")",
")",
"print_msg",
"(",
"'* images without border'",
")",
"data",
"=",
"[",
"{",
"'file'",
":",
"'beige.css'",
",",
"'line'",
":",
"' border: 4px solid #333;'",
"}",
",",
"{",
"'file'",
":",
"'black.css'",
",",
"'line'",
":",
"' border: 4px solid #fff;'",
"}",
",",
"{",
"'file'",
":",
"'blood.css'",
",",
"'line'",
":",
"' border: 4px solid #eee;'",
"}",
",",
"{",
"'file'",
":",
"'league.css'",
",",
"'line'",
":",
"' border: 4px solid #eee;'",
"}",
",",
"{",
"'file'",
":",
"'moon.css'",
",",
"'line'",
":",
"' border: 4px solid #93a1a1;'",
"}",
",",
"{",
"'file'",
":",
"'night.css'",
",",
"'line'",
":",
"' border: 4px solid #eee;'",
"}",
",",
"{",
"'file'",
":",
"'serif.css'",
",",
"'line'",
":",
"' border: 4px solid #000;'",
"}",
",",
"{",
"'file'",
":",
"'simple.css'",
",",
"'line'",
":",
"' border: 4px solid #000;'",
"}",
",",
"{",
"'file'",
":",
"'sky.css'",
",",
"'line'",
":",
"' border: 4px solid #333;'",
"}",
",",
"{",
"'file'",
":",
"'solarized.css'",
",",
"'line'",
":",
"' border: 4px solid #657b83;'",
"}",
",",
"{",
"'file'",
":",
"'white.css'",
",",
"'line'",
":",
"' border: 4px solid #222;'",
"}",
",",
"]",
"for",
"item",
"in",
"data",
":",
"file_",
"=",
"item",
"[",
"'file'",
"]",
"lines",
"=",
"[",
"item",
"[",
"'line'",
"]",
",",
"]",
"lines",
".",
"extend",
"(",
"[",
"' box-shadow: 0 0 10px rgba(0, 0, 0, 0.15); }'",
",",
"' box-shadow: 0 0 20px rgba(0, 0, 0, 0.55); }'",
"]",
")",
"for",
"line",
"in",
"lines",
":",
"update_or_append_line",
"(",
"filename",
"=",
"flo",
"(",
"'{repo_dir}/css/theme/{file_}'",
")",
",",
"prefix",
"=",
"line",
",",
"new_line",
"=",
"flo",
"(",
"'/*{line}*/'",
")",
")"
] | Comment out some css settings. | [
"Comment",
"out",
"some",
"css",
"settings",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/revealjs.py#L184-L218 | train |
theno/fabsetup | fabsetup/fabfile/setup/revealjs.py | decktape | def decktape():
'''Install DeckTape.
DeckTape is a "high-quality PDF exporter for HTML5 presentation
frameworks". It can be used to create PDFs from reveal.js presentations.
More info:
https://github.com/astefanutti/decktape
https://github.com/hakimel/reveal.js/issues/1252#issuecomment-198270915
'''
run('mkdir -p ~/bin/decktape')
if not exists('~/bin/decktape/decktape-1.0.0'):
print_msg('\n## download decktape 1.0.0\n')
run('cd ~/bin/decktape && '
'curl -L https://github.com/astefanutti/decktape/archive/'
'v1.0.0.tar.gz | tar -xz --exclude phantomjs')
run('cd ~/bin/decktape/decktape-1.0.0 && '
'curl -L https://github.com/astefanutti/decktape/releases/'
'download/v1.0.0/phantomjs-linux-x86-64 -o phantomjs')
run('cd ~/bin/decktape/decktape-1.0.0 && '
'chmod +x phantomjs')
run('ln -snf ~/bin/decktape/decktape-1.0.0 ~/bin/decktape/active',
msg='\n## link installed decktape version as active')
print_msg('\nCreate PDF from reveal.js presentation:\n\n '
'# serve presentation:\n '
'cd ~/repos/my_presi/reveal.js/ && npm start\n\n '
'# create pdf in another shell:\n '
'cd ~/bin/decktape/active && \\\n '
'./phantomjs decktape.js --size 1280x800 localhost:8000 '
'~/repos/my_presi/my_presi.pdf') | python | def decktape():
'''Install DeckTape.
DeckTape is a "high-quality PDF exporter for HTML5 presentation
frameworks". It can be used to create PDFs from reveal.js presentations.
More info:
https://github.com/astefanutti/decktape
https://github.com/hakimel/reveal.js/issues/1252#issuecomment-198270915
'''
run('mkdir -p ~/bin/decktape')
if not exists('~/bin/decktape/decktape-1.0.0'):
print_msg('\n## download decktape 1.0.0\n')
run('cd ~/bin/decktape && '
'curl -L https://github.com/astefanutti/decktape/archive/'
'v1.0.0.tar.gz | tar -xz --exclude phantomjs')
run('cd ~/bin/decktape/decktape-1.0.0 && '
'curl -L https://github.com/astefanutti/decktape/releases/'
'download/v1.0.0/phantomjs-linux-x86-64 -o phantomjs')
run('cd ~/bin/decktape/decktape-1.0.0 && '
'chmod +x phantomjs')
run('ln -snf ~/bin/decktape/decktape-1.0.0 ~/bin/decktape/active',
msg='\n## link installed decktape version as active')
print_msg('\nCreate PDF from reveal.js presentation:\n\n '
'# serve presentation:\n '
'cd ~/repos/my_presi/reveal.js/ && npm start\n\n '
'# create pdf in another shell:\n '
'cd ~/bin/decktape/active && \\\n '
'./phantomjs decktape.js --size 1280x800 localhost:8000 '
'~/repos/my_presi/my_presi.pdf') | [
"def",
"decktape",
"(",
")",
":",
"run",
"(",
"'mkdir -p ~/bin/decktape'",
")",
"if",
"not",
"exists",
"(",
"'~/bin/decktape/decktape-1.0.0'",
")",
":",
"print_msg",
"(",
"'\\n## download decktape 1.0.0\\n'",
")",
"run",
"(",
"'cd ~/bin/decktape && '",
"'curl -L https://github.com/astefanutti/decktape/archive/'",
"'v1.0.0.tar.gz | tar -xz --exclude phantomjs'",
")",
"run",
"(",
"'cd ~/bin/decktape/decktape-1.0.0 && '",
"'curl -L https://github.com/astefanutti/decktape/releases/'",
"'download/v1.0.0/phantomjs-linux-x86-64 -o phantomjs'",
")",
"run",
"(",
"'cd ~/bin/decktape/decktape-1.0.0 && '",
"'chmod +x phantomjs'",
")",
"run",
"(",
"'ln -snf ~/bin/decktape/decktape-1.0.0 ~/bin/decktape/active'",
",",
"msg",
"=",
"'\\n## link installed decktape version as active'",
")",
"print_msg",
"(",
"'\\nCreate PDF from reveal.js presentation:\\n\\n '",
"'# serve presentation:\\n '",
"'cd ~/repos/my_presi/reveal.js/ && npm start\\n\\n '",
"'# create pdf in another shell:\\n '",
"'cd ~/bin/decktape/active && \\\\\\n '",
"'./phantomjs decktape.js --size 1280x800 localhost:8000 '",
"'~/repos/my_presi/my_presi.pdf'",
")"
] | Install DeckTape.
DeckTape is a "high-quality PDF exporter for HTML5 presentation
frameworks". It can be used to create PDFs from reveal.js presentations.
More info:
https://github.com/astefanutti/decktape
https://github.com/hakimel/reveal.js/issues/1252#issuecomment-198270915 | [
"Install",
"DeckTape",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/revealjs.py#L366-L395 | train |
theno/fabsetup | fabsetup/fabfile/setup/revealjs.py | revealjs_template | def revealjs_template():
'''Create or update the template presentation demo using task `revealjs`.
'''
from config import basedir, github_user, github_repo
run(flo('rm -f {basedir}/index.html'))
run(flo('rm -f {basedir}/slides.md'))
run(flo('rm -f {basedir}/README.md'))
run(flo('rm -rf {basedir}/img/'))
title = 'reveal.js template'
subtitle = '[reveal.js][3] presentation written ' \
'in [markdown][4] set up with [fabric][5] & [fabsetup][6]'
description = '''\
This presentation shows how to create a reveal.js presentation which will be
set up with the fabric task `setup.revealjs` of fabsetup.
Also, you can use this presentation source as a reveal.js template:
* Checkout this repo
* Then set the title in the `index.html` and edit the
`slides.md`.'''
execute(revealjs, basedir, title, subtitle, description,
github_user, github_repo)
# (README.md was removed, but not the github remote repo)
print_msg('\n## Re-add github repo infos into README.md')
basename = os.path.basename(basedir)
_insert_repo_infos_into_readme(basedir, github_user=_lazy('github_user'),
github_repo=_lazy('github_repo',
default=basename))
print_msg('\n## Assure symbolic link not tracked by git exists\n')
run(flo('ln -snf ../reveal.js {basedir}/reveal.js/reveal.js')) | python | def revealjs_template():
'''Create or update the template presentation demo using task `revealjs`.
'''
from config import basedir, github_user, github_repo
run(flo('rm -f {basedir}/index.html'))
run(flo('rm -f {basedir}/slides.md'))
run(flo('rm -f {basedir}/README.md'))
run(flo('rm -rf {basedir}/img/'))
title = 'reveal.js template'
subtitle = '[reveal.js][3] presentation written ' \
'in [markdown][4] set up with [fabric][5] & [fabsetup][6]'
description = '''\
This presentation shows how to create a reveal.js presentation which will be
set up with the fabric task `setup.revealjs` of fabsetup.
Also, you can use this presentation source as a reveal.js template:
* Checkout this repo
* Then set the title in the `index.html` and edit the
`slides.md`.'''
execute(revealjs, basedir, title, subtitle, description,
github_user, github_repo)
# (README.md was removed, but not the github remote repo)
print_msg('\n## Re-add github repo infos into README.md')
basename = os.path.basename(basedir)
_insert_repo_infos_into_readme(basedir, github_user=_lazy('github_user'),
github_repo=_lazy('github_repo',
default=basename))
print_msg('\n## Assure symbolic link not tracked by git exists\n')
run(flo('ln -snf ../reveal.js {basedir}/reveal.js/reveal.js')) | [
"def",
"revealjs_template",
"(",
")",
":",
"from",
"config",
"import",
"basedir",
",",
"github_user",
",",
"github_repo",
"run",
"(",
"flo",
"(",
"'rm -f {basedir}/index.html'",
")",
")",
"run",
"(",
"flo",
"(",
"'rm -f {basedir}/slides.md'",
")",
")",
"run",
"(",
"flo",
"(",
"'rm -f {basedir}/README.md'",
")",
")",
"run",
"(",
"flo",
"(",
"'rm -rf {basedir}/img/'",
")",
")",
"title",
"=",
"'reveal.js template'",
"subtitle",
"=",
"'[reveal.js][3] presentation written '",
"'in [markdown][4] set up with [fabric][5] & [fabsetup][6]'",
"description",
"=",
"'''\\\nThis presentation shows how to create a reveal.js presentation which will be\nset up with the fabric task `setup.revealjs` of fabsetup.\n\nAlso, you can use this presentation source as a reveal.js template:\n* Checkout this repo\n* Then set the title in the `index.html` and edit the\n `slides.md`.'''",
"execute",
"(",
"revealjs",
",",
"basedir",
",",
"title",
",",
"subtitle",
",",
"description",
",",
"github_user",
",",
"github_repo",
")",
"# (README.md was removed, but not the github remote repo)",
"print_msg",
"(",
"'\\n## Re-add github repo infos into README.md'",
")",
"basename",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"basedir",
")",
"_insert_repo_infos_into_readme",
"(",
"basedir",
",",
"github_user",
"=",
"_lazy",
"(",
"'github_user'",
")",
",",
"github_repo",
"=",
"_lazy",
"(",
"'github_repo'",
",",
"default",
"=",
"basename",
")",
")",
"print_msg",
"(",
"'\\n## Assure symbolic link not tracked by git exists\\n'",
")",
"run",
"(",
"flo",
"(",
"'ln -snf ../reveal.js {basedir}/reveal.js/reveal.js'",
")",
")"
] | Create or update the template presentation demo using task `revealjs`. | [
"Create",
"or",
"update",
"the",
"template",
"presentation",
"demo",
"using",
"task",
"revealjs",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/revealjs.py#L401-L434 | train |
awickert/gFlex | gflex/f1d.py | F1D.spatialDomainNoGrid | def spatialDomainNoGrid(self):
"""
Superposition of analytical solutions without a gridded domain
"""
self.w = np.zeros(self.xw.shape)
if self.Debug:
print("w = ")
print(self.w.shape)
for i in range(len(self.q)):
# More efficient if we have created some 0-load points
# (e.g., for where we want output)
if self.q[i] != 0:
dist = np.abs(self.xw - self.x[i])
self.w -= self.q[i] * self.coeff * np.exp(-dist/self.alpha) * \
( np.cos(dist/self.alpha) + np.sin(dist/self.alpha) ) | python | def spatialDomainNoGrid(self):
"""
Superposition of analytical solutions without a gridded domain
"""
self.w = np.zeros(self.xw.shape)
if self.Debug:
print("w = ")
print(self.w.shape)
for i in range(len(self.q)):
# More efficient if we have created some 0-load points
# (e.g., for where we want output)
if self.q[i] != 0:
dist = np.abs(self.xw - self.x[i])
self.w -= self.q[i] * self.coeff * np.exp(-dist/self.alpha) * \
( np.cos(dist/self.alpha) + np.sin(dist/self.alpha) ) | [
"def",
"spatialDomainNoGrid",
"(",
"self",
")",
":",
"self",
".",
"w",
"=",
"np",
".",
"zeros",
"(",
"self",
".",
"xw",
".",
"shape",
")",
"if",
"self",
".",
"Debug",
":",
"print",
"(",
"\"w = \"",
")",
"print",
"(",
"self",
".",
"w",
".",
"shape",
")",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"self",
".",
"q",
")",
")",
":",
"# More efficient if we have created some 0-load points\r",
"# (e.g., for where we want output)\r",
"if",
"self",
".",
"q",
"[",
"i",
"]",
"!=",
"0",
":",
"dist",
"=",
"np",
".",
"abs",
"(",
"self",
".",
"xw",
"-",
"self",
".",
"x",
"[",
"i",
"]",
")",
"self",
".",
"w",
"-=",
"self",
".",
"q",
"[",
"i",
"]",
"*",
"self",
".",
"coeff",
"*",
"np",
".",
"exp",
"(",
"-",
"dist",
"/",
"self",
".",
"alpha",
")",
"*",
"(",
"np",
".",
"cos",
"(",
"dist",
"/",
"self",
".",
"alpha",
")",
"+",
"np",
".",
"sin",
"(",
"dist",
"/",
"self",
".",
"alpha",
")",
")"
] | Superposition of analytical solutions without a gridded domain | [
"Superposition",
"of",
"analytical",
"solutions",
"without",
"a",
"gridded",
"domain"
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/f1d.py#L143-L159 | train |
awickert/gFlex | gflex/f1d.py | F1D.build_diagonals | def build_diagonals(self):
"""
Builds the diagonals for the coefficient array
"""
##########################################################
# INCORPORATE BOUNDARY CONDITIONS INTO COEFFICIENT ARRAY #
##########################################################
# Roll to keep the proper coefficients at the proper places in the
# arrays: Python will naturally just do vertical shifts instead of
# diagonal shifts, so this takes into account the horizontal compoent
# to ensure that boundary values are at the right place.
self.l2 = np.roll(self.l2, -2)
self.l1 = np.roll(self.l1, -1)
self.r1 = np.roll(self.r1, 1)
self.r2 = np.roll(self.r2, 2)
# Then assemble these rows: this is where the periodic boundary condition
# can matter.
if self.coeff_matrix is not None:
pass
elif self.BC_E == 'Periodic' and self.BC_W == 'Periodic':
# In this case, the boundary-condition-related stacking has already
# happened inside b.c.-handling function. This is because periodic
# boundary conditions require extra diagonals to exist on the edges of
# the solution array
pass
else:
self.diags = np.vstack((self.l2,self.l1,self.c0,self.r1,self.r2))
self.offsets = np.array([-2,-1,0,1,2])
# Everybody now (including periodic b.c. cases)
self.coeff_matrix = spdiags(self.diags, self.offsets, self.nx, self.nx, format='csr') | python | def build_diagonals(self):
"""
Builds the diagonals for the coefficient array
"""
##########################################################
# INCORPORATE BOUNDARY CONDITIONS INTO COEFFICIENT ARRAY #
##########################################################
# Roll to keep the proper coefficients at the proper places in the
# arrays: Python will naturally just do vertical shifts instead of
# diagonal shifts, so this takes into account the horizontal compoent
# to ensure that boundary values are at the right place.
self.l2 = np.roll(self.l2, -2)
self.l1 = np.roll(self.l1, -1)
self.r1 = np.roll(self.r1, 1)
self.r2 = np.roll(self.r2, 2)
# Then assemble these rows: this is where the periodic boundary condition
# can matter.
if self.coeff_matrix is not None:
pass
elif self.BC_E == 'Periodic' and self.BC_W == 'Periodic':
# In this case, the boundary-condition-related stacking has already
# happened inside b.c.-handling function. This is because periodic
# boundary conditions require extra diagonals to exist on the edges of
# the solution array
pass
else:
self.diags = np.vstack((self.l2,self.l1,self.c0,self.r1,self.r2))
self.offsets = np.array([-2,-1,0,1,2])
# Everybody now (including periodic b.c. cases)
self.coeff_matrix = spdiags(self.diags, self.offsets, self.nx, self.nx, format='csr') | [
"def",
"build_diagonals",
"(",
"self",
")",
":",
"##########################################################\r",
"# INCORPORATE BOUNDARY CONDITIONS INTO COEFFICIENT ARRAY #\r",
"##########################################################\r",
"# Roll to keep the proper coefficients at the proper places in the\r",
"# arrays: Python will naturally just do vertical shifts instead of \r",
"# diagonal shifts, so this takes into account the horizontal compoent \r",
"# to ensure that boundary values are at the right place.\r",
"self",
".",
"l2",
"=",
"np",
".",
"roll",
"(",
"self",
".",
"l2",
",",
"-",
"2",
")",
"self",
".",
"l1",
"=",
"np",
".",
"roll",
"(",
"self",
".",
"l1",
",",
"-",
"1",
")",
"self",
".",
"r1",
"=",
"np",
".",
"roll",
"(",
"self",
".",
"r1",
",",
"1",
")",
"self",
".",
"r2",
"=",
"np",
".",
"roll",
"(",
"self",
".",
"r2",
",",
"2",
")",
"# Then assemble these rows: this is where the periodic boundary condition \r",
"# can matter.\r",
"if",
"self",
".",
"coeff_matrix",
"is",
"not",
"None",
":",
"pass",
"elif",
"self",
".",
"BC_E",
"==",
"'Periodic'",
"and",
"self",
".",
"BC_W",
"==",
"'Periodic'",
":",
"# In this case, the boundary-condition-related stacking has already \r",
"# happened inside b.c.-handling function. This is because periodic\r",
"# boundary conditions require extra diagonals to exist on the edges of \r",
"# the solution array\r",
"pass",
"else",
":",
"self",
".",
"diags",
"=",
"np",
".",
"vstack",
"(",
"(",
"self",
".",
"l2",
",",
"self",
".",
"l1",
",",
"self",
".",
"c0",
",",
"self",
".",
"r1",
",",
"self",
".",
"r2",
")",
")",
"self",
".",
"offsets",
"=",
"np",
".",
"array",
"(",
"[",
"-",
"2",
",",
"-",
"1",
",",
"0",
",",
"1",
",",
"2",
"]",
")",
"# Everybody now (including periodic b.c. cases)\r",
"self",
".",
"coeff_matrix",
"=",
"spdiags",
"(",
"self",
".",
"diags",
",",
"self",
".",
"offsets",
",",
"self",
".",
"nx",
",",
"self",
".",
"nx",
",",
"format",
"=",
"'csr'",
")"
] | Builds the diagonals for the coefficient array | [
"Builds",
"the",
"diagonals",
"for",
"the",
"coefficient",
"array"
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/f1d.py#L344-L377 | train |
sbmlteam/libCombine | examples/python/createArchiveExample.py | createArchiveExample | def createArchiveExample(fileName):
""" Creates Combine Archive containing the given file.
:param fileName: file to include in the archive
:return: None
"""
print('*' * 80)
print('Create archive')
print('*' * 80)
archive = CombineArchive()
archive.addFile(
fileName, # filename
"./models/model.xml", # target file name
KnownFormats.lookupFormat("sbml"), # look up identifier for SBML models
True # mark file as master
)
# add metadata to the archive itself
description = OmexDescription()
description.setAbout(".")
description.setDescription("Simple test archive including one SBML model")
description.setCreated(OmexDescription.getCurrentDateAndTime())
creator = VCard()
creator.setFamilyName("Bergmann")
creator.setGivenName("Frank")
creator.setEmail("[email protected]")
creator.setOrganization("Caltech")
description.addCreator(creator)
archive.addMetadata(".", description)
# add metadata to the added file
location = "./models/model.xml"
description = OmexDescription()
description.setAbout(location)
description.setDescription("SBML model")
description.setCreated(OmexDescription.getCurrentDateAndTime())
archive.addMetadata(location, description)
# write the archive
out_file = "out.omex"
archive.writeToFile(out_file)
print('Archive created:', out_file) | python | def createArchiveExample(fileName):
""" Creates Combine Archive containing the given file.
:param fileName: file to include in the archive
:return: None
"""
print('*' * 80)
print('Create archive')
print('*' * 80)
archive = CombineArchive()
archive.addFile(
fileName, # filename
"./models/model.xml", # target file name
KnownFormats.lookupFormat("sbml"), # look up identifier for SBML models
True # mark file as master
)
# add metadata to the archive itself
description = OmexDescription()
description.setAbout(".")
description.setDescription("Simple test archive including one SBML model")
description.setCreated(OmexDescription.getCurrentDateAndTime())
creator = VCard()
creator.setFamilyName("Bergmann")
creator.setGivenName("Frank")
creator.setEmail("[email protected]")
creator.setOrganization("Caltech")
description.addCreator(creator)
archive.addMetadata(".", description)
# add metadata to the added file
location = "./models/model.xml"
description = OmexDescription()
description.setAbout(location)
description.setDescription("SBML model")
description.setCreated(OmexDescription.getCurrentDateAndTime())
archive.addMetadata(location, description)
# write the archive
out_file = "out.omex"
archive.writeToFile(out_file)
print('Archive created:', out_file) | [
"def",
"createArchiveExample",
"(",
"fileName",
")",
":",
"print",
"(",
"'*'",
"*",
"80",
")",
"print",
"(",
"'Create archive'",
")",
"print",
"(",
"'*'",
"*",
"80",
")",
"archive",
"=",
"CombineArchive",
"(",
")",
"archive",
".",
"addFile",
"(",
"fileName",
",",
"# filename",
"\"./models/model.xml\"",
",",
"# target file name",
"KnownFormats",
".",
"lookupFormat",
"(",
"\"sbml\"",
")",
",",
"# look up identifier for SBML models",
"True",
"# mark file as master",
")",
"# add metadata to the archive itself",
"description",
"=",
"OmexDescription",
"(",
")",
"description",
".",
"setAbout",
"(",
"\".\"",
")",
"description",
".",
"setDescription",
"(",
"\"Simple test archive including one SBML model\"",
")",
"description",
".",
"setCreated",
"(",
"OmexDescription",
".",
"getCurrentDateAndTime",
"(",
")",
")",
"creator",
"=",
"VCard",
"(",
")",
"creator",
".",
"setFamilyName",
"(",
"\"Bergmann\"",
")",
"creator",
".",
"setGivenName",
"(",
"\"Frank\"",
")",
"creator",
".",
"setEmail",
"(",
"\"[email protected]\"",
")",
"creator",
".",
"setOrganization",
"(",
"\"Caltech\"",
")",
"description",
".",
"addCreator",
"(",
"creator",
")",
"archive",
".",
"addMetadata",
"(",
"\".\"",
",",
"description",
")",
"# add metadata to the added file",
"location",
"=",
"\"./models/model.xml\"",
"description",
"=",
"OmexDescription",
"(",
")",
"description",
".",
"setAbout",
"(",
"location",
")",
"description",
".",
"setDescription",
"(",
"\"SBML model\"",
")",
"description",
".",
"setCreated",
"(",
"OmexDescription",
".",
"getCurrentDateAndTime",
"(",
")",
")",
"archive",
".",
"addMetadata",
"(",
"location",
",",
"description",
")",
"# write the archive",
"out_file",
"=",
"\"out.omex\"",
"archive",
".",
"writeToFile",
"(",
"out_file",
")",
"print",
"(",
"'Archive created:'",
",",
"out_file",
")"
] | Creates Combine Archive containing the given file.
:param fileName: file to include in the archive
:return: None | [
"Creates",
"Combine",
"Archive",
"containing",
"the",
"given",
"file",
"."
] | d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4 | https://github.com/sbmlteam/libCombine/blob/d7c11a90129dedbcc8bdba8d204be03f1dd0c3e4/examples/python/createArchiveExample.py#L11-L57 | train |
yahoo/serviceping | serviceping/serviceping.py | calc_deviation | def calc_deviation(values, average):
"""
Calculate the standard deviation of a list of values
@param values: list(float)
@param average:
@return:
"""
size = len(values)
if size < 2:
return 0
calc_sum = 0.0
for number in range(0, size):
calc_sum += math.sqrt((values[number] - average) ** 2)
return math.sqrt((1.0 / (size - 1)) * (calc_sum / size)) | python | def calc_deviation(values, average):
"""
Calculate the standard deviation of a list of values
@param values: list(float)
@param average:
@return:
"""
size = len(values)
if size < 2:
return 0
calc_sum = 0.0
for number in range(0, size):
calc_sum += math.sqrt((values[number] - average) ** 2)
return math.sqrt((1.0 / (size - 1)) * (calc_sum / size)) | [
"def",
"calc_deviation",
"(",
"values",
",",
"average",
")",
":",
"size",
"=",
"len",
"(",
"values",
")",
"if",
"size",
"<",
"2",
":",
"return",
"0",
"calc_sum",
"=",
"0.0",
"for",
"number",
"in",
"range",
"(",
"0",
",",
"size",
")",
":",
"calc_sum",
"+=",
"math",
".",
"sqrt",
"(",
"(",
"values",
"[",
"number",
"]",
"-",
"average",
")",
"**",
"2",
")",
"return",
"math",
".",
"sqrt",
"(",
"(",
"1.0",
"/",
"(",
"size",
"-",
"1",
")",
")",
"*",
"(",
"calc_sum",
"/",
"size",
")",
")"
] | Calculate the standard deviation of a list of values
@param values: list(float)
@param average:
@return: | [
"Calculate",
"the",
"standard",
"deviation",
"of",
"a",
"list",
"of",
"values"
] | 1f9df5ee5b3cba466426b1164262278472ba4977 | https://github.com/yahoo/serviceping/blob/1f9df5ee5b3cba466426b1164262278472ba4977/serviceping/serviceping.py#L106-L120 | train |
yahoo/serviceping | serviceping/serviceping.py | StatsList.append | def append(self, value):
"""
Append a value to the stats list
Parameters
----------
value : float
The value to add
"""
self.count += 1
if self.count == 1:
self.old_m = self.new_m = value
self.old_s = 0
else:
self.new_m = self.old_m + (value - self.old_m) / self.count
self.new_s = self.old_s + (value - self.old_m) * (value - self.new_m)
self.old_m = self.new_m
self.old_s = self.new_s | python | def append(self, value):
"""
Append a value to the stats list
Parameters
----------
value : float
The value to add
"""
self.count += 1
if self.count == 1:
self.old_m = self.new_m = value
self.old_s = 0
else:
self.new_m = self.old_m + (value - self.old_m) / self.count
self.new_s = self.old_s + (value - self.old_m) * (value - self.new_m)
self.old_m = self.new_m
self.old_s = self.new_s | [
"def",
"append",
"(",
"self",
",",
"value",
")",
":",
"self",
".",
"count",
"+=",
"1",
"if",
"self",
".",
"count",
"==",
"1",
":",
"self",
".",
"old_m",
"=",
"self",
".",
"new_m",
"=",
"value",
"self",
".",
"old_s",
"=",
"0",
"else",
":",
"self",
".",
"new_m",
"=",
"self",
".",
"old_m",
"+",
"(",
"value",
"-",
"self",
".",
"old_m",
")",
"/",
"self",
".",
"count",
"self",
".",
"new_s",
"=",
"self",
".",
"old_s",
"+",
"(",
"value",
"-",
"self",
".",
"old_m",
")",
"*",
"(",
"value",
"-",
"self",
".",
"new_m",
")",
"self",
".",
"old_m",
"=",
"self",
".",
"new_m",
"self",
".",
"old_s",
"=",
"self",
".",
"new_s"
] | Append a value to the stats list
Parameters
----------
value : float
The value to add | [
"Append",
"a",
"value",
"to",
"the",
"stats",
"list"
] | 1f9df5ee5b3cba466426b1164262278472ba4977 | https://github.com/yahoo/serviceping/blob/1f9df5ee5b3cba466426b1164262278472ba4977/serviceping/serviceping.py#L35-L54 | train |
bionikspoon/pureyaml | docs/github_docs.py | pipeline | def pipeline(steps, initial=None):
"""
Chain results from a list of functions. Inverted reduce.
:param (function) steps: List of function callbacks
:param initial: Starting value for pipeline.
"""
def apply(result, step):
return step(result)
return reduce(apply, steps, initial) | python | def pipeline(steps, initial=None):
"""
Chain results from a list of functions. Inverted reduce.
:param (function) steps: List of function callbacks
:param initial: Starting value for pipeline.
"""
def apply(result, step):
return step(result)
return reduce(apply, steps, initial) | [
"def",
"pipeline",
"(",
"steps",
",",
"initial",
"=",
"None",
")",
":",
"def",
"apply",
"(",
"result",
",",
"step",
")",
":",
"return",
"step",
"(",
"result",
")",
"return",
"reduce",
"(",
"apply",
",",
"steps",
",",
"initial",
")"
] | Chain results from a list of functions. Inverted reduce.
:param (function) steps: List of function callbacks
:param initial: Starting value for pipeline. | [
"Chain",
"results",
"from",
"a",
"list",
"of",
"functions",
".",
"Inverted",
"reduce",
"."
] | 784830b907ca14525c4cecdb6ae35306f6f8a877 | https://github.com/bionikspoon/pureyaml/blob/784830b907ca14525c4cecdb6ae35306f6f8a877/docs/github_docs.py#L186-L197 | train |
jaraco/jaraco.windows | jaraco/windows/environ.py | RegisteredEnvironment.add | def add(class_, name, value, sep=';'):
"""
Add a value to a delimited variable, but only when the value isn't
already present.
"""
values = class_.get_values_list(name, sep)
if value in values:
return
new_value = sep.join(values + [value])
winreg.SetValueEx(
class_.key, name, 0, winreg.REG_EXPAND_SZ, new_value)
class_.notify() | python | def add(class_, name, value, sep=';'):
"""
Add a value to a delimited variable, but only when the value isn't
already present.
"""
values = class_.get_values_list(name, sep)
if value in values:
return
new_value = sep.join(values + [value])
winreg.SetValueEx(
class_.key, name, 0, winreg.REG_EXPAND_SZ, new_value)
class_.notify() | [
"def",
"add",
"(",
"class_",
",",
"name",
",",
"value",
",",
"sep",
"=",
"';'",
")",
":",
"values",
"=",
"class_",
".",
"get_values_list",
"(",
"name",
",",
"sep",
")",
"if",
"value",
"in",
"values",
":",
"return",
"new_value",
"=",
"sep",
".",
"join",
"(",
"values",
"+",
"[",
"value",
"]",
")",
"winreg",
".",
"SetValueEx",
"(",
"class_",
".",
"key",
",",
"name",
",",
"0",
",",
"winreg",
".",
"REG_EXPAND_SZ",
",",
"new_value",
")",
"class_",
".",
"notify",
"(",
")"
] | Add a value to a delimited variable, but only when the value isn't
already present. | [
"Add",
"a",
"value",
"to",
"a",
"delimited",
"variable",
"but",
"only",
"when",
"the",
"value",
"isn",
"t",
"already",
"present",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/environ.py#L84-L95 | train |
jaraco/jaraco.windows | jaraco/windows/timezone.py | Info.current | def current(class_):
"Windows Platform SDK GetTimeZoneInformation"
tzi = class_()
kernel32 = ctypes.windll.kernel32
getter = kernel32.GetTimeZoneInformation
getter = getattr(kernel32, 'GetDynamicTimeZoneInformation', getter)
code = getter(ctypes.byref(tzi))
return code, tzi | python | def current(class_):
"Windows Platform SDK GetTimeZoneInformation"
tzi = class_()
kernel32 = ctypes.windll.kernel32
getter = kernel32.GetTimeZoneInformation
getter = getattr(kernel32, 'GetDynamicTimeZoneInformation', getter)
code = getter(ctypes.byref(tzi))
return code, tzi | [
"def",
"current",
"(",
"class_",
")",
":",
"tzi",
"=",
"class_",
"(",
")",
"kernel32",
"=",
"ctypes",
".",
"windll",
".",
"kernel32",
"getter",
"=",
"kernel32",
".",
"GetTimeZoneInformation",
"getter",
"=",
"getattr",
"(",
"kernel32",
",",
"'GetDynamicTimeZoneInformation'",
",",
"getter",
")",
"code",
"=",
"getter",
"(",
"ctypes",
".",
"byref",
"(",
"tzi",
")",
")",
"return",
"code",
",",
"tzi"
] | Windows Platform SDK GetTimeZoneInformation | [
"Windows",
"Platform",
"SDK",
"GetTimeZoneInformation"
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/timezone.py#L168-L175 | train |
jaraco/jaraco.windows | jaraco/windows/timezone.py | Info.dynamic_info | def dynamic_info(self):
"Return a map that for a given year will return the correct Info"
if self.key_name:
dyn_key = self.get_key().subkey('Dynamic DST')
del dyn_key['FirstEntry']
del dyn_key['LastEntry']
years = map(int, dyn_key.keys())
values = map(Info, dyn_key.values())
# create a range mapping that searches by descending year and matches
# if the target year is greater or equal.
return RangeMap(zip(years, values), RangeMap.descending, operator.ge)
else:
return AnyDict(self) | python | def dynamic_info(self):
"Return a map that for a given year will return the correct Info"
if self.key_name:
dyn_key = self.get_key().subkey('Dynamic DST')
del dyn_key['FirstEntry']
del dyn_key['LastEntry']
years = map(int, dyn_key.keys())
values = map(Info, dyn_key.values())
# create a range mapping that searches by descending year and matches
# if the target year is greater or equal.
return RangeMap(zip(years, values), RangeMap.descending, operator.ge)
else:
return AnyDict(self) | [
"def",
"dynamic_info",
"(",
"self",
")",
":",
"if",
"self",
".",
"key_name",
":",
"dyn_key",
"=",
"self",
".",
"get_key",
"(",
")",
".",
"subkey",
"(",
"'Dynamic DST'",
")",
"del",
"dyn_key",
"[",
"'FirstEntry'",
"]",
"del",
"dyn_key",
"[",
"'LastEntry'",
"]",
"years",
"=",
"map",
"(",
"int",
",",
"dyn_key",
".",
"keys",
"(",
")",
")",
"values",
"=",
"map",
"(",
"Info",
",",
"dyn_key",
".",
"values",
"(",
")",
")",
"# create a range mapping that searches by descending year and matches",
"# if the target year is greater or equal.",
"return",
"RangeMap",
"(",
"zip",
"(",
"years",
",",
"values",
")",
",",
"RangeMap",
".",
"descending",
",",
"operator",
".",
"ge",
")",
"else",
":",
"return",
"AnyDict",
"(",
"self",
")"
] | Return a map that for a given year will return the correct Info | [
"Return",
"a",
"map",
"that",
"for",
"a",
"given",
"year",
"will",
"return",
"the",
"correct",
"Info"
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/timezone.py#L198-L210 | train |
jaraco/jaraco.windows | jaraco/windows/timezone.py | Info._locate_day | def _locate_day(year, cutoff):
"""
Takes a SYSTEMTIME object, such as retrieved from a TIME_ZONE_INFORMATION
structure or call to GetTimeZoneInformation and interprets
it based on the given
year to identify the actual day.
This method is necessary because the SYSTEMTIME structure
refers to a day by its
day of the week and week of the month (e.g. 4th saturday in March).
>>> SATURDAY = 6
>>> MARCH = 3
>>> st = SYSTEMTIME(2000, MARCH, SATURDAY, 4, 0, 0, 0, 0)
# according to my calendar, the 4th Saturday in March in 2009 was the 28th
>>> expected_date = datetime.datetime(2009, 3, 28)
>>> Info._locate_day(2009, st) == expected_date
True
"""
# MS stores Sunday as 0, Python datetime stores Monday as zero
target_weekday = (cutoff.day_of_week + 6) % 7
# For SYSTEMTIMEs relating to time zone inforamtion, cutoff.day
# is the week of the month
week_of_month = cutoff.day
# so the following is the first day of that week
day = (week_of_month - 1) * 7 + 1
result = datetime.datetime(
year, cutoff.month, day,
cutoff.hour, cutoff.minute, cutoff.second, cutoff.millisecond)
# now the result is the correct week, but not necessarily
# the correct day of the week
days_to_go = (target_weekday - result.weekday()) % 7
result += datetime.timedelta(days_to_go)
# if we selected a day in the month following the target month,
# move back a week or two.
# This is necessary because Microsoft defines the fifth week in a month
# to be the last week in a month and adding the time delta might have
# pushed the result into the next month.
while result.month == cutoff.month + 1:
result -= datetime.timedelta(weeks=1)
return result | python | def _locate_day(year, cutoff):
"""
Takes a SYSTEMTIME object, such as retrieved from a TIME_ZONE_INFORMATION
structure or call to GetTimeZoneInformation and interprets
it based on the given
year to identify the actual day.
This method is necessary because the SYSTEMTIME structure
refers to a day by its
day of the week and week of the month (e.g. 4th saturday in March).
>>> SATURDAY = 6
>>> MARCH = 3
>>> st = SYSTEMTIME(2000, MARCH, SATURDAY, 4, 0, 0, 0, 0)
# according to my calendar, the 4th Saturday in March in 2009 was the 28th
>>> expected_date = datetime.datetime(2009, 3, 28)
>>> Info._locate_day(2009, st) == expected_date
True
"""
# MS stores Sunday as 0, Python datetime stores Monday as zero
target_weekday = (cutoff.day_of_week + 6) % 7
# For SYSTEMTIMEs relating to time zone inforamtion, cutoff.day
# is the week of the month
week_of_month = cutoff.day
# so the following is the first day of that week
day = (week_of_month - 1) * 7 + 1
result = datetime.datetime(
year, cutoff.month, day,
cutoff.hour, cutoff.minute, cutoff.second, cutoff.millisecond)
# now the result is the correct week, but not necessarily
# the correct day of the week
days_to_go = (target_weekday - result.weekday()) % 7
result += datetime.timedelta(days_to_go)
# if we selected a day in the month following the target month,
# move back a week or two.
# This is necessary because Microsoft defines the fifth week in a month
# to be the last week in a month and adding the time delta might have
# pushed the result into the next month.
while result.month == cutoff.month + 1:
result -= datetime.timedelta(weeks=1)
return result | [
"def",
"_locate_day",
"(",
"year",
",",
"cutoff",
")",
":",
"# MS stores Sunday as 0, Python datetime stores Monday as zero",
"target_weekday",
"=",
"(",
"cutoff",
".",
"day_of_week",
"+",
"6",
")",
"%",
"7",
"# For SYSTEMTIMEs relating to time zone inforamtion, cutoff.day",
"# is the week of the month",
"week_of_month",
"=",
"cutoff",
".",
"day",
"# so the following is the first day of that week",
"day",
"=",
"(",
"week_of_month",
"-",
"1",
")",
"*",
"7",
"+",
"1",
"result",
"=",
"datetime",
".",
"datetime",
"(",
"year",
",",
"cutoff",
".",
"month",
",",
"day",
",",
"cutoff",
".",
"hour",
",",
"cutoff",
".",
"minute",
",",
"cutoff",
".",
"second",
",",
"cutoff",
".",
"millisecond",
")",
"# now the result is the correct week, but not necessarily",
"# the correct day of the week",
"days_to_go",
"=",
"(",
"target_weekday",
"-",
"result",
".",
"weekday",
"(",
")",
")",
"%",
"7",
"result",
"+=",
"datetime",
".",
"timedelta",
"(",
"days_to_go",
")",
"# if we selected a day in the month following the target month,",
"# move back a week or two.",
"# This is necessary because Microsoft defines the fifth week in a month",
"# to be the last week in a month and adding the time delta might have",
"# pushed the result into the next month.",
"while",
"result",
".",
"month",
"==",
"cutoff",
".",
"month",
"+",
"1",
":",
"result",
"-=",
"datetime",
".",
"timedelta",
"(",
"weeks",
"=",
"1",
")",
"return",
"result"
] | Takes a SYSTEMTIME object, such as retrieved from a TIME_ZONE_INFORMATION
structure or call to GetTimeZoneInformation and interprets
it based on the given
year to identify the actual day.
This method is necessary because the SYSTEMTIME structure
refers to a day by its
day of the week and week of the month (e.g. 4th saturday in March).
>>> SATURDAY = 6
>>> MARCH = 3
>>> st = SYSTEMTIME(2000, MARCH, SATURDAY, 4, 0, 0, 0, 0)
# according to my calendar, the 4th Saturday in March in 2009 was the 28th
>>> expected_date = datetime.datetime(2009, 3, 28)
>>> Info._locate_day(2009, st) == expected_date
True | [
"Takes",
"a",
"SYSTEMTIME",
"object",
"such",
"as",
"retrieved",
"from",
"a",
"TIME_ZONE_INFORMATION",
"structure",
"or",
"call",
"to",
"GetTimeZoneInformation",
"and",
"interprets",
"it",
"based",
"on",
"the",
"given",
"year",
"to",
"identify",
"the",
"actual",
"day",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/timezone.py#L213-L254 | train |
pmac/django-redirect-urls | redirect_urls/utils.py | redirect | def redirect(pattern, to, permanent=True, locale_prefix=True, anchor=None, name=None,
query=None, vary=None, cache_timeout=12, decorators=None, re_flags=None,
to_args=None, to_kwargs=None, prepend_locale=True, merge_query=False):
"""
Return a url matcher suited for urlpatterns.
pattern: the regex against which to match the requested URL.
to: either a url name that `reverse` will find, a url that will simply be returned,
or a function that will be given the request and url captures, and return the
destination.
permanent: boolean whether to send a 301 or 302 response.
locale_prefix: automatically prepend `pattern` with a regex for an optional locale
in the url. This locale (or None) will show up in captured kwargs as 'locale'.
anchor: if set it will be appended to the destination url after a '#'.
name: if used in a `urls.py` the redirect URL will be available as the name
for use in calls to `reverse()`. Does _NOT_ work if used in a `redirects.py` file.
query: a dict of query params to add to the destination url.
vary: if you used an HTTP header to decide where to send users you should include that
header's name in the `vary` arg.
cache_timeout: number of hours to cache this redirect. just sets the proper `cache-control`
and `expires` headers.
decorators: a callable (or list of callables) that will wrap the view used to redirect
the user. equivalent to adding a decorator to any other view.
re_flags: a string of any of the characters: "iLmsux". Will modify the `pattern` regex
based on the documented meaning of the flags (see python re module docs).
to_args: a tuple or list of args to pass to reverse if `to` is a url name.
to_kwargs: a dict of keyword args to pass to reverse if `to` is a url name.
prepend_locale: if true the redirect URL will be prepended with the locale from the
requested URL.
merge_query: merge the requested query params from the `query` arg with any query params
from the request.
Usage:
urlpatterns = [
redirect(r'projects/$', 'mozorg.product'),
redirect(r'^projects/seamonkey$', 'mozorg.product', locale_prefix=False),
redirect(r'apps/$', 'https://marketplace.firefox.com'),
redirect(r'firefox/$', 'firefox.new', name='firefox'),
redirect(r'the/dude$', 'abides', query={'aggression': 'not_stand'}),
]
"""
if permanent:
redirect_class = HttpResponsePermanentRedirect
else:
redirect_class = HttpResponseRedirect
if locale_prefix:
pattern = pattern.lstrip('^/')
pattern = LOCALE_RE + pattern
if re_flags:
pattern = '(?{})'.format(re_flags) + pattern
view_decorators = []
if cache_timeout is not None:
view_decorators.append(cache_control_expires(cache_timeout))
if vary:
if isinstance(vary, basestring):
vary = [vary]
view_decorators.append(vary_on_headers(*vary))
if decorators:
if callable(decorators):
view_decorators.append(decorators)
else:
view_decorators.extend(decorators)
def _view(request, *args, **kwargs):
# don't want to have 'None' in substitutions
kwargs = {k: v or '' for k, v in kwargs.items()}
args = [x or '' for x in args]
# If it's a callable, call it and get the url out.
if callable(to):
to_value = to(request, *args, **kwargs)
else:
to_value = to
if to_value.startswith('/') or HTTP_RE.match(to_value):
redirect_url = to_value
else:
try:
redirect_url = reverse(to_value, args=to_args, kwargs=to_kwargs)
except NoReverseMatch:
# Assume it's a URL
redirect_url = to_value
if prepend_locale and redirect_url.startswith('/') and kwargs.get('locale'):
redirect_url = '/{locale}' + redirect_url.lstrip('/')
# use info from url captures.
if args or kwargs:
redirect_url = strip_tags(force_text(redirect_url).format(*args, **kwargs))
if query:
if merge_query:
req_query = parse_qs(request.META.get('QUERY_STRING'))
req_query.update(query)
querystring = urlencode(req_query, doseq=True)
else:
querystring = urlencode(query, doseq=True)
elif query is None:
querystring = request.META.get('QUERY_STRING')
else:
querystring = ''
if querystring:
redirect_url = '?'.join([redirect_url, querystring])
if anchor:
redirect_url = '#'.join([redirect_url, anchor])
if PROTOCOL_RELATIVE_RE.match(redirect_url):
redirect_url = '/' + redirect_url.lstrip('/')
return redirect_class(redirect_url)
# Apply decorators
try:
# Decorators should be applied in reverse order so that input
# can be sent in the order your would write nested decorators
# e.g. dec1(dec2(_view)) -> [dec1, dec2]
for decorator in reversed(view_decorators):
_view = decorator(_view)
except TypeError:
log.exception('decorators not iterable or does not contain '
'callable items')
return url(pattern, _view, name=name) | python | def redirect(pattern, to, permanent=True, locale_prefix=True, anchor=None, name=None,
query=None, vary=None, cache_timeout=12, decorators=None, re_flags=None,
to_args=None, to_kwargs=None, prepend_locale=True, merge_query=False):
"""
Return a url matcher suited for urlpatterns.
pattern: the regex against which to match the requested URL.
to: either a url name that `reverse` will find, a url that will simply be returned,
or a function that will be given the request and url captures, and return the
destination.
permanent: boolean whether to send a 301 or 302 response.
locale_prefix: automatically prepend `pattern` with a regex for an optional locale
in the url. This locale (or None) will show up in captured kwargs as 'locale'.
anchor: if set it will be appended to the destination url after a '#'.
name: if used in a `urls.py` the redirect URL will be available as the name
for use in calls to `reverse()`. Does _NOT_ work if used in a `redirects.py` file.
query: a dict of query params to add to the destination url.
vary: if you used an HTTP header to decide where to send users you should include that
header's name in the `vary` arg.
cache_timeout: number of hours to cache this redirect. just sets the proper `cache-control`
and `expires` headers.
decorators: a callable (or list of callables) that will wrap the view used to redirect
the user. equivalent to adding a decorator to any other view.
re_flags: a string of any of the characters: "iLmsux". Will modify the `pattern` regex
based on the documented meaning of the flags (see python re module docs).
to_args: a tuple or list of args to pass to reverse if `to` is a url name.
to_kwargs: a dict of keyword args to pass to reverse if `to` is a url name.
prepend_locale: if true the redirect URL will be prepended with the locale from the
requested URL.
merge_query: merge the requested query params from the `query` arg with any query params
from the request.
Usage:
urlpatterns = [
redirect(r'projects/$', 'mozorg.product'),
redirect(r'^projects/seamonkey$', 'mozorg.product', locale_prefix=False),
redirect(r'apps/$', 'https://marketplace.firefox.com'),
redirect(r'firefox/$', 'firefox.new', name='firefox'),
redirect(r'the/dude$', 'abides', query={'aggression': 'not_stand'}),
]
"""
if permanent:
redirect_class = HttpResponsePermanentRedirect
else:
redirect_class = HttpResponseRedirect
if locale_prefix:
pattern = pattern.lstrip('^/')
pattern = LOCALE_RE + pattern
if re_flags:
pattern = '(?{})'.format(re_flags) + pattern
view_decorators = []
if cache_timeout is not None:
view_decorators.append(cache_control_expires(cache_timeout))
if vary:
if isinstance(vary, basestring):
vary = [vary]
view_decorators.append(vary_on_headers(*vary))
if decorators:
if callable(decorators):
view_decorators.append(decorators)
else:
view_decorators.extend(decorators)
def _view(request, *args, **kwargs):
# don't want to have 'None' in substitutions
kwargs = {k: v or '' for k, v in kwargs.items()}
args = [x or '' for x in args]
# If it's a callable, call it and get the url out.
if callable(to):
to_value = to(request, *args, **kwargs)
else:
to_value = to
if to_value.startswith('/') or HTTP_RE.match(to_value):
redirect_url = to_value
else:
try:
redirect_url = reverse(to_value, args=to_args, kwargs=to_kwargs)
except NoReverseMatch:
# Assume it's a URL
redirect_url = to_value
if prepend_locale and redirect_url.startswith('/') and kwargs.get('locale'):
redirect_url = '/{locale}' + redirect_url.lstrip('/')
# use info from url captures.
if args or kwargs:
redirect_url = strip_tags(force_text(redirect_url).format(*args, **kwargs))
if query:
if merge_query:
req_query = parse_qs(request.META.get('QUERY_STRING'))
req_query.update(query)
querystring = urlencode(req_query, doseq=True)
else:
querystring = urlencode(query, doseq=True)
elif query is None:
querystring = request.META.get('QUERY_STRING')
else:
querystring = ''
if querystring:
redirect_url = '?'.join([redirect_url, querystring])
if anchor:
redirect_url = '#'.join([redirect_url, anchor])
if PROTOCOL_RELATIVE_RE.match(redirect_url):
redirect_url = '/' + redirect_url.lstrip('/')
return redirect_class(redirect_url)
# Apply decorators
try:
# Decorators should be applied in reverse order so that input
# can be sent in the order your would write nested decorators
# e.g. dec1(dec2(_view)) -> [dec1, dec2]
for decorator in reversed(view_decorators):
_view = decorator(_view)
except TypeError:
log.exception('decorators not iterable or does not contain '
'callable items')
return url(pattern, _view, name=name) | [
"def",
"redirect",
"(",
"pattern",
",",
"to",
",",
"permanent",
"=",
"True",
",",
"locale_prefix",
"=",
"True",
",",
"anchor",
"=",
"None",
",",
"name",
"=",
"None",
",",
"query",
"=",
"None",
",",
"vary",
"=",
"None",
",",
"cache_timeout",
"=",
"12",
",",
"decorators",
"=",
"None",
",",
"re_flags",
"=",
"None",
",",
"to_args",
"=",
"None",
",",
"to_kwargs",
"=",
"None",
",",
"prepend_locale",
"=",
"True",
",",
"merge_query",
"=",
"False",
")",
":",
"if",
"permanent",
":",
"redirect_class",
"=",
"HttpResponsePermanentRedirect",
"else",
":",
"redirect_class",
"=",
"HttpResponseRedirect",
"if",
"locale_prefix",
":",
"pattern",
"=",
"pattern",
".",
"lstrip",
"(",
"'^/'",
")",
"pattern",
"=",
"LOCALE_RE",
"+",
"pattern",
"if",
"re_flags",
":",
"pattern",
"=",
"'(?{})'",
".",
"format",
"(",
"re_flags",
")",
"+",
"pattern",
"view_decorators",
"=",
"[",
"]",
"if",
"cache_timeout",
"is",
"not",
"None",
":",
"view_decorators",
".",
"append",
"(",
"cache_control_expires",
"(",
"cache_timeout",
")",
")",
"if",
"vary",
":",
"if",
"isinstance",
"(",
"vary",
",",
"basestring",
")",
":",
"vary",
"=",
"[",
"vary",
"]",
"view_decorators",
".",
"append",
"(",
"vary_on_headers",
"(",
"*",
"vary",
")",
")",
"if",
"decorators",
":",
"if",
"callable",
"(",
"decorators",
")",
":",
"view_decorators",
".",
"append",
"(",
"decorators",
")",
"else",
":",
"view_decorators",
".",
"extend",
"(",
"decorators",
")",
"def",
"_view",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"# don't want to have 'None' in substitutions",
"kwargs",
"=",
"{",
"k",
":",
"v",
"or",
"''",
"for",
"k",
",",
"v",
"in",
"kwargs",
".",
"items",
"(",
")",
"}",
"args",
"=",
"[",
"x",
"or",
"''",
"for",
"x",
"in",
"args",
"]",
"# If it's a callable, call it and get the url out.",
"if",
"callable",
"(",
"to",
")",
":",
"to_value",
"=",
"to",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"else",
":",
"to_value",
"=",
"to",
"if",
"to_value",
".",
"startswith",
"(",
"'/'",
")",
"or",
"HTTP_RE",
".",
"match",
"(",
"to_value",
")",
":",
"redirect_url",
"=",
"to_value",
"else",
":",
"try",
":",
"redirect_url",
"=",
"reverse",
"(",
"to_value",
",",
"args",
"=",
"to_args",
",",
"kwargs",
"=",
"to_kwargs",
")",
"except",
"NoReverseMatch",
":",
"# Assume it's a URL",
"redirect_url",
"=",
"to_value",
"if",
"prepend_locale",
"and",
"redirect_url",
".",
"startswith",
"(",
"'/'",
")",
"and",
"kwargs",
".",
"get",
"(",
"'locale'",
")",
":",
"redirect_url",
"=",
"'/{locale}'",
"+",
"redirect_url",
".",
"lstrip",
"(",
"'/'",
")",
"# use info from url captures.",
"if",
"args",
"or",
"kwargs",
":",
"redirect_url",
"=",
"strip_tags",
"(",
"force_text",
"(",
"redirect_url",
")",
".",
"format",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
")",
"if",
"query",
":",
"if",
"merge_query",
":",
"req_query",
"=",
"parse_qs",
"(",
"request",
".",
"META",
".",
"get",
"(",
"'QUERY_STRING'",
")",
")",
"req_query",
".",
"update",
"(",
"query",
")",
"querystring",
"=",
"urlencode",
"(",
"req_query",
",",
"doseq",
"=",
"True",
")",
"else",
":",
"querystring",
"=",
"urlencode",
"(",
"query",
",",
"doseq",
"=",
"True",
")",
"elif",
"query",
"is",
"None",
":",
"querystring",
"=",
"request",
".",
"META",
".",
"get",
"(",
"'QUERY_STRING'",
")",
"else",
":",
"querystring",
"=",
"''",
"if",
"querystring",
":",
"redirect_url",
"=",
"'?'",
".",
"join",
"(",
"[",
"redirect_url",
",",
"querystring",
"]",
")",
"if",
"anchor",
":",
"redirect_url",
"=",
"'#'",
".",
"join",
"(",
"[",
"redirect_url",
",",
"anchor",
"]",
")",
"if",
"PROTOCOL_RELATIVE_RE",
".",
"match",
"(",
"redirect_url",
")",
":",
"redirect_url",
"=",
"'/'",
"+",
"redirect_url",
".",
"lstrip",
"(",
"'/'",
")",
"return",
"redirect_class",
"(",
"redirect_url",
")",
"# Apply decorators",
"try",
":",
"# Decorators should be applied in reverse order so that input",
"# can be sent in the order your would write nested decorators",
"# e.g. dec1(dec2(_view)) -> [dec1, dec2]",
"for",
"decorator",
"in",
"reversed",
"(",
"view_decorators",
")",
":",
"_view",
"=",
"decorator",
"(",
"_view",
")",
"except",
"TypeError",
":",
"log",
".",
"exception",
"(",
"'decorators not iterable or does not contain '",
"'callable items'",
")",
"return",
"url",
"(",
"pattern",
",",
"_view",
",",
"name",
"=",
"name",
")"
] | Return a url matcher suited for urlpatterns.
pattern: the regex against which to match the requested URL.
to: either a url name that `reverse` will find, a url that will simply be returned,
or a function that will be given the request and url captures, and return the
destination.
permanent: boolean whether to send a 301 or 302 response.
locale_prefix: automatically prepend `pattern` with a regex for an optional locale
in the url. This locale (or None) will show up in captured kwargs as 'locale'.
anchor: if set it will be appended to the destination url after a '#'.
name: if used in a `urls.py` the redirect URL will be available as the name
for use in calls to `reverse()`. Does _NOT_ work if used in a `redirects.py` file.
query: a dict of query params to add to the destination url.
vary: if you used an HTTP header to decide where to send users you should include that
header's name in the `vary` arg.
cache_timeout: number of hours to cache this redirect. just sets the proper `cache-control`
and `expires` headers.
decorators: a callable (or list of callables) that will wrap the view used to redirect
the user. equivalent to adding a decorator to any other view.
re_flags: a string of any of the characters: "iLmsux". Will modify the `pattern` regex
based on the documented meaning of the flags (see python re module docs).
to_args: a tuple or list of args to pass to reverse if `to` is a url name.
to_kwargs: a dict of keyword args to pass to reverse if `to` is a url name.
prepend_locale: if true the redirect URL will be prepended with the locale from the
requested URL.
merge_query: merge the requested query params from the `query` arg with any query params
from the request.
Usage:
urlpatterns = [
redirect(r'projects/$', 'mozorg.product'),
redirect(r'^projects/seamonkey$', 'mozorg.product', locale_prefix=False),
redirect(r'apps/$', 'https://marketplace.firefox.com'),
redirect(r'firefox/$', 'firefox.new', name='firefox'),
redirect(r'the/dude$', 'abides', query={'aggression': 'not_stand'}),
] | [
"Return",
"a",
"url",
"matcher",
"suited",
"for",
"urlpatterns",
"."
] | 21495194b0b2a2bdd1013e13ec0d54d34dd7f750 | https://github.com/pmac/django-redirect-urls/blob/21495194b0b2a2bdd1013e13ec0d54d34dd7f750/redirect_urls/utils.py#L128-L257 | train |
jaraco/jaraco.windows | jaraco/windows/inet.py | AllocatedTable.__get_table_size | def __get_table_size(self):
"""
Retrieve the size of the buffer needed by calling the method
with a null pointer and length of zero. This should trigger an
insufficient buffer error and return the size needed for the
buffer.
"""
length = ctypes.wintypes.DWORD()
res = self.method(None, length, False)
if res != errors.ERROR_INSUFFICIENT_BUFFER:
raise RuntimeError("Error getting table length (%d)" % res)
return length.value | python | def __get_table_size(self):
"""
Retrieve the size of the buffer needed by calling the method
with a null pointer and length of zero. This should trigger an
insufficient buffer error and return the size needed for the
buffer.
"""
length = ctypes.wintypes.DWORD()
res = self.method(None, length, False)
if res != errors.ERROR_INSUFFICIENT_BUFFER:
raise RuntimeError("Error getting table length (%d)" % res)
return length.value | [
"def",
"__get_table_size",
"(",
"self",
")",
":",
"length",
"=",
"ctypes",
".",
"wintypes",
".",
"DWORD",
"(",
")",
"res",
"=",
"self",
".",
"method",
"(",
"None",
",",
"length",
",",
"False",
")",
"if",
"res",
"!=",
"errors",
".",
"ERROR_INSUFFICIENT_BUFFER",
":",
"raise",
"RuntimeError",
"(",
"\"Error getting table length (%d)\"",
"%",
"res",
")",
"return",
"length",
".",
"value"
] | Retrieve the size of the buffer needed by calling the method
with a null pointer and length of zero. This should trigger an
insufficient buffer error and return the size needed for the
buffer. | [
"Retrieve",
"the",
"size",
"of",
"the",
"buffer",
"needed",
"by",
"calling",
"the",
"method",
"with",
"a",
"null",
"pointer",
"and",
"length",
"of",
"zero",
".",
"This",
"should",
"trigger",
"an",
"insufficient",
"buffer",
"error",
"and",
"return",
"the",
"size",
"needed",
"for",
"the",
"buffer",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/inet.py#L49-L60 | train |
jaraco/jaraco.windows | jaraco/windows/inet.py | AllocatedTable.get_table | def get_table(self):
"""
Get the table
"""
buffer_length = self.__get_table_size()
returned_buffer_length = ctypes.wintypes.DWORD(buffer_length)
buffer = ctypes.create_string_buffer(buffer_length)
pointer_type = ctypes.POINTER(self.structure)
table_p = ctypes.cast(buffer, pointer_type)
res = self.method(table_p, returned_buffer_length, False)
if res != errors.NO_ERROR:
raise RuntimeError("Error retrieving table (%d)" % res)
return table_p.contents | python | def get_table(self):
"""
Get the table
"""
buffer_length = self.__get_table_size()
returned_buffer_length = ctypes.wintypes.DWORD(buffer_length)
buffer = ctypes.create_string_buffer(buffer_length)
pointer_type = ctypes.POINTER(self.structure)
table_p = ctypes.cast(buffer, pointer_type)
res = self.method(table_p, returned_buffer_length, False)
if res != errors.NO_ERROR:
raise RuntimeError("Error retrieving table (%d)" % res)
return table_p.contents | [
"def",
"get_table",
"(",
"self",
")",
":",
"buffer_length",
"=",
"self",
".",
"__get_table_size",
"(",
")",
"returned_buffer_length",
"=",
"ctypes",
".",
"wintypes",
".",
"DWORD",
"(",
"buffer_length",
")",
"buffer",
"=",
"ctypes",
".",
"create_string_buffer",
"(",
"buffer_length",
")",
"pointer_type",
"=",
"ctypes",
".",
"POINTER",
"(",
"self",
".",
"structure",
")",
"table_p",
"=",
"ctypes",
".",
"cast",
"(",
"buffer",
",",
"pointer_type",
")",
"res",
"=",
"self",
".",
"method",
"(",
"table_p",
",",
"returned_buffer_length",
",",
"False",
")",
"if",
"res",
"!=",
"errors",
".",
"NO_ERROR",
":",
"raise",
"RuntimeError",
"(",
"\"Error retrieving table (%d)\"",
"%",
"res",
")",
"return",
"table_p",
".",
"contents"
] | Get the table | [
"Get",
"the",
"table"
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/inet.py#L62-L74 | train |
jaraco/jaraco.windows | jaraco/windows/inet.py | AllocatedTable.entries | def entries(self):
"""
Using the table structure, return the array of entries based
on the table size.
"""
table = self.get_table()
entries_array = self.row_structure * table.num_entries
pointer_type = ctypes.POINTER(entries_array)
return ctypes.cast(table.entries, pointer_type).contents | python | def entries(self):
"""
Using the table structure, return the array of entries based
on the table size.
"""
table = self.get_table()
entries_array = self.row_structure * table.num_entries
pointer_type = ctypes.POINTER(entries_array)
return ctypes.cast(table.entries, pointer_type).contents | [
"def",
"entries",
"(",
"self",
")",
":",
"table",
"=",
"self",
".",
"get_table",
"(",
")",
"entries_array",
"=",
"self",
".",
"row_structure",
"*",
"table",
".",
"num_entries",
"pointer_type",
"=",
"ctypes",
".",
"POINTER",
"(",
"entries_array",
")",
"return",
"ctypes",
".",
"cast",
"(",
"table",
".",
"entries",
",",
"pointer_type",
")",
".",
"contents"
] | Using the table structure, return the array of entries based
on the table size. | [
"Using",
"the",
"table",
"structure",
"return",
"the",
"array",
"of",
"entries",
"based",
"on",
"the",
"table",
"size",
"."
] | 51811efed50b46ad08daa25408a1cc806bc8d519 | https://github.com/jaraco/jaraco.windows/blob/51811efed50b46ad08daa25408a1cc806bc8d519/jaraco/windows/inet.py#L77-L85 | train |
theno/fabsetup | fabsetup/fabfile/setup/service/__init__.py | owncloud | def owncloud():
'''Set up owncloud.
Package 'owncloud' pulls package 'mysql' which asks for a password.
'''
hostname = re.sub(r'^[^@]+@', '', env.host) # without username if any
sitename = query_input(
question='\nEnter site-name of Your Owncloud web service',
default=flo('owncloud.{hostname}'), color=cyan)
username = env.user
fabfile_data_dir = FABFILE_DATA_DIR
print(magenta(' install owncloud'))
repository = ''.join([
'http://download.opensuse.org/repositories/',
'isv:/ownCloud:/community/Debian_7.0/',
])
with hide('output'):
sudo(flo('wget -O - {repository}Release.key | apt-key add -'))
filename = '/etc/apt/sources.list.d/owncloud.list'
sudo(flo("echo 'deb {repository} /' > {filename}"))
sudo('apt-get update')
install_packages([
'owncloud',
'php5-fpm',
'php-apc',
'memcached',
'php5-memcache',
])
# This server uses nginx. owncloud pulls apache2 => Disable apache2
print(magenta(' disable apache'))
with hide('output'):
sudo('service apache2 stop')
sudo('update-rc.d apache2 disable')
print(magenta(' nginx setup for owncloud'))
filename = 'owncloud_site_config.template'
path = flo('{fabfile_data_dir}/files/etc/nginx/sites-available/{filename}')
from_str = filled_out_template(path, username=username, sitename=sitename,
hostname=hostname)
with tempfile.NamedTemporaryFile(prefix=filename) as tmp_file:
with open(tmp_file.name, 'w') as fp:
fp.write(from_str)
put(tmp_file.name, flo('/tmp/{filename}'))
to = flo('/etc/nginx/sites-available/{sitename}')
sudo(flo('mv /tmp/{filename} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
sudo(flo(' '.join([
'ln -snf ../sites-available/{sitename}',
'/etc/nginx/sites-enabled/{sitename}',
])))
# php5 fpm fast-cgi config
template = 'www.conf'
to = flo('/etc/php5/fpm/pool.d/{template}')
from_ = flo('{fabfile_data_dir}/files{to}')
put(from_, '/tmp/')
sudo(flo('mv /tmp/{template} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
template = 'php.ini'
to = flo('/etc/php5/fpm/{template}')
from_ = flo('{fabfile_data_dir}/files{to}')
put(from_, '/tmp/')
sudo(flo('mv /tmp/{template} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
sudo('service php5-fpm restart')
sudo('service nginx reload') | python | def owncloud():
'''Set up owncloud.
Package 'owncloud' pulls package 'mysql' which asks for a password.
'''
hostname = re.sub(r'^[^@]+@', '', env.host) # without username if any
sitename = query_input(
question='\nEnter site-name of Your Owncloud web service',
default=flo('owncloud.{hostname}'), color=cyan)
username = env.user
fabfile_data_dir = FABFILE_DATA_DIR
print(magenta(' install owncloud'))
repository = ''.join([
'http://download.opensuse.org/repositories/',
'isv:/ownCloud:/community/Debian_7.0/',
])
with hide('output'):
sudo(flo('wget -O - {repository}Release.key | apt-key add -'))
filename = '/etc/apt/sources.list.d/owncloud.list'
sudo(flo("echo 'deb {repository} /' > {filename}"))
sudo('apt-get update')
install_packages([
'owncloud',
'php5-fpm',
'php-apc',
'memcached',
'php5-memcache',
])
# This server uses nginx. owncloud pulls apache2 => Disable apache2
print(magenta(' disable apache'))
with hide('output'):
sudo('service apache2 stop')
sudo('update-rc.d apache2 disable')
print(magenta(' nginx setup for owncloud'))
filename = 'owncloud_site_config.template'
path = flo('{fabfile_data_dir}/files/etc/nginx/sites-available/{filename}')
from_str = filled_out_template(path, username=username, sitename=sitename,
hostname=hostname)
with tempfile.NamedTemporaryFile(prefix=filename) as tmp_file:
with open(tmp_file.name, 'w') as fp:
fp.write(from_str)
put(tmp_file.name, flo('/tmp/{filename}'))
to = flo('/etc/nginx/sites-available/{sitename}')
sudo(flo('mv /tmp/{filename} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
sudo(flo(' '.join([
'ln -snf ../sites-available/{sitename}',
'/etc/nginx/sites-enabled/{sitename}',
])))
# php5 fpm fast-cgi config
template = 'www.conf'
to = flo('/etc/php5/fpm/pool.d/{template}')
from_ = flo('{fabfile_data_dir}/files{to}')
put(from_, '/tmp/')
sudo(flo('mv /tmp/{template} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
template = 'php.ini'
to = flo('/etc/php5/fpm/{template}')
from_ = flo('{fabfile_data_dir}/files{to}')
put(from_, '/tmp/')
sudo(flo('mv /tmp/{template} {to}'))
sudo(flo('chown root.root {to}'))
sudo(flo('chmod 644 {to}'))
sudo('service php5-fpm restart')
sudo('service nginx reload') | [
"def",
"owncloud",
"(",
")",
":",
"hostname",
"=",
"re",
".",
"sub",
"(",
"r'^[^@]+@'",
",",
"''",
",",
"env",
".",
"host",
")",
"# without username if any",
"sitename",
"=",
"query_input",
"(",
"question",
"=",
"'\\nEnter site-name of Your Owncloud web service'",
",",
"default",
"=",
"flo",
"(",
"'owncloud.{hostname}'",
")",
",",
"color",
"=",
"cyan",
")",
"username",
"=",
"env",
".",
"user",
"fabfile_data_dir",
"=",
"FABFILE_DATA_DIR",
"print",
"(",
"magenta",
"(",
"' install owncloud'",
")",
")",
"repository",
"=",
"''",
".",
"join",
"(",
"[",
"'http://download.opensuse.org/repositories/'",
",",
"'isv:/ownCloud:/community/Debian_7.0/'",
",",
"]",
")",
"with",
"hide",
"(",
"'output'",
")",
":",
"sudo",
"(",
"flo",
"(",
"'wget -O - {repository}Release.key | apt-key add -'",
")",
")",
"filename",
"=",
"'/etc/apt/sources.list.d/owncloud.list'",
"sudo",
"(",
"flo",
"(",
"\"echo 'deb {repository} /' > {filename}\"",
")",
")",
"sudo",
"(",
"'apt-get update'",
")",
"install_packages",
"(",
"[",
"'owncloud'",
",",
"'php5-fpm'",
",",
"'php-apc'",
",",
"'memcached'",
",",
"'php5-memcache'",
",",
"]",
")",
"# This server uses nginx. owncloud pulls apache2 => Disable apache2",
"print",
"(",
"magenta",
"(",
"' disable apache'",
")",
")",
"with",
"hide",
"(",
"'output'",
")",
":",
"sudo",
"(",
"'service apache2 stop'",
")",
"sudo",
"(",
"'update-rc.d apache2 disable'",
")",
"print",
"(",
"magenta",
"(",
"' nginx setup for owncloud'",
")",
")",
"filename",
"=",
"'owncloud_site_config.template'",
"path",
"=",
"flo",
"(",
"'{fabfile_data_dir}/files/etc/nginx/sites-available/{filename}'",
")",
"from_str",
"=",
"filled_out_template",
"(",
"path",
",",
"username",
"=",
"username",
",",
"sitename",
"=",
"sitename",
",",
"hostname",
"=",
"hostname",
")",
"with",
"tempfile",
".",
"NamedTemporaryFile",
"(",
"prefix",
"=",
"filename",
")",
"as",
"tmp_file",
":",
"with",
"open",
"(",
"tmp_file",
".",
"name",
",",
"'w'",
")",
"as",
"fp",
":",
"fp",
".",
"write",
"(",
"from_str",
")",
"put",
"(",
"tmp_file",
".",
"name",
",",
"flo",
"(",
"'/tmp/{filename}'",
")",
")",
"to",
"=",
"flo",
"(",
"'/etc/nginx/sites-available/{sitename}'",
")",
"sudo",
"(",
"flo",
"(",
"'mv /tmp/{filename} {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chown root.root {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chmod 644 {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"' '",
".",
"join",
"(",
"[",
"'ln -snf ../sites-available/{sitename}'",
",",
"'/etc/nginx/sites-enabled/{sitename}'",
",",
"]",
")",
")",
")",
"# php5 fpm fast-cgi config",
"template",
"=",
"'www.conf'",
"to",
"=",
"flo",
"(",
"'/etc/php5/fpm/pool.d/{template}'",
")",
"from_",
"=",
"flo",
"(",
"'{fabfile_data_dir}/files{to}'",
")",
"put",
"(",
"from_",
",",
"'/tmp/'",
")",
"sudo",
"(",
"flo",
"(",
"'mv /tmp/{template} {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chown root.root {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chmod 644 {to}'",
")",
")",
"template",
"=",
"'php.ini'",
"to",
"=",
"flo",
"(",
"'/etc/php5/fpm/{template}'",
")",
"from_",
"=",
"flo",
"(",
"'{fabfile_data_dir}/files{to}'",
")",
"put",
"(",
"from_",
",",
"'/tmp/'",
")",
"sudo",
"(",
"flo",
"(",
"'mv /tmp/{template} {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chown root.root {to}'",
")",
")",
"sudo",
"(",
"flo",
"(",
"'chmod 644 {to}'",
")",
")",
"sudo",
"(",
"'service php5-fpm restart'",
")",
"sudo",
"(",
"'service nginx reload'",
")"
] | Set up owncloud.
Package 'owncloud' pulls package 'mysql' which asks for a password. | [
"Set",
"up",
"owncloud",
"."
] | ced728abff93551ba5677e63bc1bdc0ef5ca5777 | https://github.com/theno/fabsetup/blob/ced728abff93551ba5677e63bc1bdc0ef5ca5777/fabsetup/fabfile/setup/service/__init__.py#L23-L98 | train |
bionikspoon/pureyaml | docs/source/conf.py | doctree_read_handler | def doctree_read_handler(app, doctree):
"""
Add 'orphan' to metadata for partials
:type app: sphinx.application.Sphinx
:type doctree: docutils.nodes.document
"""
# noinspection PyProtectedMember
docname = sys._getframe(2).f_locals['docname']
if docname.startswith('_partial'):
app.env.metadata[docname]['orphan'] = True | python | def doctree_read_handler(app, doctree):
"""
Add 'orphan' to metadata for partials
:type app: sphinx.application.Sphinx
:type doctree: docutils.nodes.document
"""
# noinspection PyProtectedMember
docname = sys._getframe(2).f_locals['docname']
if docname.startswith('_partial'):
app.env.metadata[docname]['orphan'] = True | [
"def",
"doctree_read_handler",
"(",
"app",
",",
"doctree",
")",
":",
"# noinspection PyProtectedMember",
"docname",
"=",
"sys",
".",
"_getframe",
"(",
"2",
")",
".",
"f_locals",
"[",
"'docname'",
"]",
"if",
"docname",
".",
"startswith",
"(",
"'_partial'",
")",
":",
"app",
".",
"env",
".",
"metadata",
"[",
"docname",
"]",
"[",
"'orphan'",
"]",
"=",
"True"
] | Add 'orphan' to metadata for partials
:type app: sphinx.application.Sphinx
:type doctree: docutils.nodes.document | [
"Add",
"orphan",
"to",
"metadata",
"for",
"partials"
] | 784830b907ca14525c4cecdb6ae35306f6f8a877 | https://github.com/bionikspoon/pureyaml/blob/784830b907ca14525c4cecdb6ae35306f6f8a877/docs/source/conf.py#L26-L36 | train |
bionikspoon/pureyaml | docs/source/conf.py | autodoc_skip_member_handler | def autodoc_skip_member_handler(app, what, name, obj, skip, options):
"""
Skip un parseable functions.
:type app: sphinx.application.Sphinx
:param str what: the type of the object which the docstring belongs to
(one of "module", "class", "exception", "function", "method", "attribute")
:param str name: the fully qualified name of the object
:param type obj: the object itself
:param bool skip: a boolean indicating if autodoc will skip this member
:param sphinx.ext.autodoc.Options options: the options given to the directive
:rtype: bool
"""
if 'YAMLTokens' in name:
return True
return False | python | def autodoc_skip_member_handler(app, what, name, obj, skip, options):
"""
Skip un parseable functions.
:type app: sphinx.application.Sphinx
:param str what: the type of the object which the docstring belongs to
(one of "module", "class", "exception", "function", "method", "attribute")
:param str name: the fully qualified name of the object
:param type obj: the object itself
:param bool skip: a boolean indicating if autodoc will skip this member
:param sphinx.ext.autodoc.Options options: the options given to the directive
:rtype: bool
"""
if 'YAMLTokens' in name:
return True
return False | [
"def",
"autodoc_skip_member_handler",
"(",
"app",
",",
"what",
",",
"name",
",",
"obj",
",",
"skip",
",",
"options",
")",
":",
"if",
"'YAMLTokens'",
"in",
"name",
":",
"return",
"True",
"return",
"False"
] | Skip un parseable functions.
:type app: sphinx.application.Sphinx
:param str what: the type of the object which the docstring belongs to
(one of "module", "class", "exception", "function", "method", "attribute")
:param str name: the fully qualified name of the object
:param type obj: the object itself
:param bool skip: a boolean indicating if autodoc will skip this member
:param sphinx.ext.autodoc.Options options: the options given to the directive
:rtype: bool | [
"Skip",
"un",
"parseable",
"functions",
"."
] | 784830b907ca14525c4cecdb6ae35306f6f8a877 | https://github.com/bionikspoon/pureyaml/blob/784830b907ca14525c4cecdb6ae35306f6f8a877/docs/source/conf.py#L39-L54 | train |
awickert/gFlex | gflex/base.py | Plotting.surfplot | def surfplot(self, z, titletext):
"""
Plot if you want to - for troubleshooting - 1 figure
"""
if self.latlon:
plt.imshow(z, extent=(0, self.dx*z.shape[0], self.dy*z.shape[1], 0)) #,interpolation='nearest'
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(z, extent=(0, self.dx/1000.*z.shape[0], self.dy/1000.*z.shape[1], 0)) #,interpolation='nearest'
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar()
plt.title(titletext,fontsize=16) | python | def surfplot(self, z, titletext):
"""
Plot if you want to - for troubleshooting - 1 figure
"""
if self.latlon:
plt.imshow(z, extent=(0, self.dx*z.shape[0], self.dy*z.shape[1], 0)) #,interpolation='nearest'
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(z, extent=(0, self.dx/1000.*z.shape[0], self.dy/1000.*z.shape[1], 0)) #,interpolation='nearest'
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar()
plt.title(titletext,fontsize=16) | [
"def",
"surfplot",
"(",
"self",
",",
"z",
",",
"titletext",
")",
":",
"if",
"self",
".",
"latlon",
":",
"plt",
".",
"imshow",
"(",
"z",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"*",
"z",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"*",
"z",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"#,interpolation='nearest'",
"plt",
".",
"xlabel",
"(",
"'longitude [deg E]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'latitude [deg N]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"else",
":",
"plt",
".",
"imshow",
"(",
"z",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"/",
"1000.",
"*",
"z",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"/",
"1000.",
"*",
"z",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"#,interpolation='nearest'",
"plt",
".",
"xlabel",
"(",
"'x [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'y [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"colorbar",
"(",
")",
"plt",
".",
"title",
"(",
"titletext",
",",
"fontsize",
"=",
"16",
")"
] | Plot if you want to - for troubleshooting - 1 figure | [
"Plot",
"if",
"you",
"want",
"to",
"-",
"for",
"troubleshooting",
"-",
"1",
"figure"
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/base.py#L376-L390 | train |
awickert/gFlex | gflex/base.py | Plotting.twoSurfplots | def twoSurfplots(self):
"""
Plot multiple subplot figure for 2D array
"""
# Could more elegantly just call surfplot twice
# And also could include xyzinterp as an option inside surfplot.
# Noted here in case anyone wants to take that on in the future...
plt.subplot(211)
plt.title('Load thickness, mantle equivalent [m]',fontsize=16)
if self.latlon:
plt.imshow(self.qs/(self.rho_m*self.g), extent=(0, self.dx*self.qs.shape[0], self.dy*self.qs.shape[1], 0))
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(self.qs/(self.rho_m*self.g), extent=(0, self.dx/1000.*self.qs.shape[0], self.dy/1000.*self.qs.shape[1], 0))
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar()
plt.subplot(212)
plt.title('Deflection [m]')
if self.latlon:
plt.imshow(self.w, extent=(0, self.dx*self.w.shape[0], self.dy*self.w.shape[1], 0))
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(self.w, extent=(0, self.dx/1000.*self.w.shape[0], self.dy/1000.*self.w.shape[1], 0))
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar() | python | def twoSurfplots(self):
"""
Plot multiple subplot figure for 2D array
"""
# Could more elegantly just call surfplot twice
# And also could include xyzinterp as an option inside surfplot.
# Noted here in case anyone wants to take that on in the future...
plt.subplot(211)
plt.title('Load thickness, mantle equivalent [m]',fontsize=16)
if self.latlon:
plt.imshow(self.qs/(self.rho_m*self.g), extent=(0, self.dx*self.qs.shape[0], self.dy*self.qs.shape[1], 0))
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(self.qs/(self.rho_m*self.g), extent=(0, self.dx/1000.*self.qs.shape[0], self.dy/1000.*self.qs.shape[1], 0))
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar()
plt.subplot(212)
plt.title('Deflection [m]')
if self.latlon:
plt.imshow(self.w, extent=(0, self.dx*self.w.shape[0], self.dy*self.w.shape[1], 0))
plt.xlabel('longitude [deg E]', fontsize=12, fontweight='bold')
plt.ylabel('latitude [deg N]', fontsize=12, fontweight='bold')
else:
plt.imshow(self.w, extent=(0, self.dx/1000.*self.w.shape[0], self.dy/1000.*self.w.shape[1], 0))
plt.xlabel('x [km]', fontsize=12, fontweight='bold')
plt.ylabel('y [km]', fontsize=12, fontweight='bold')
plt.colorbar() | [
"def",
"twoSurfplots",
"(",
"self",
")",
":",
"# Could more elegantly just call surfplot twice",
"# And also could include xyzinterp as an option inside surfplot.",
"# Noted here in case anyone wants to take that on in the future...",
"plt",
".",
"subplot",
"(",
"211",
")",
"plt",
".",
"title",
"(",
"'Load thickness, mantle equivalent [m]'",
",",
"fontsize",
"=",
"16",
")",
"if",
"self",
".",
"latlon",
":",
"plt",
".",
"imshow",
"(",
"self",
".",
"qs",
"/",
"(",
"self",
".",
"rho_m",
"*",
"self",
".",
"g",
")",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"plt",
".",
"xlabel",
"(",
"'longitude [deg E]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'latitude [deg N]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"else",
":",
"plt",
".",
"imshow",
"(",
"self",
".",
"qs",
"/",
"(",
"self",
".",
"rho_m",
"*",
"self",
".",
"g",
")",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"/",
"1000.",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"/",
"1000.",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"plt",
".",
"xlabel",
"(",
"'x [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'y [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"colorbar",
"(",
")",
"plt",
".",
"subplot",
"(",
"212",
")",
"plt",
".",
"title",
"(",
"'Deflection [m]'",
")",
"if",
"self",
".",
"latlon",
":",
"plt",
".",
"imshow",
"(",
"self",
".",
"w",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"*",
"self",
".",
"w",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"*",
"self",
".",
"w",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"plt",
".",
"xlabel",
"(",
"'longitude [deg E]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'latitude [deg N]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"else",
":",
"plt",
".",
"imshow",
"(",
"self",
".",
"w",
",",
"extent",
"=",
"(",
"0",
",",
"self",
".",
"dx",
"/",
"1000.",
"*",
"self",
".",
"w",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dy",
"/",
"1000.",
"*",
"self",
".",
"w",
".",
"shape",
"[",
"1",
"]",
",",
"0",
")",
")",
"plt",
".",
"xlabel",
"(",
"'x [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"ylabel",
"(",
"'y [km]'",
",",
"fontsize",
"=",
"12",
",",
"fontweight",
"=",
"'bold'",
")",
"plt",
".",
"colorbar",
"(",
")"
] | Plot multiple subplot figure for 2D array | [
"Plot",
"multiple",
"subplot",
"figure",
"for",
"2D",
"array"
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/base.py#L392-L422 | train |
awickert/gFlex | gflex/base.py | Flexure.outputDeflections | def outputDeflections(self):
"""
Outputs a grid of deflections if an output directory is defined in the
configuration file
If the filename given in the configuration file ends in ".npy", then a binary
numpy grid will be exported.
Otherwise, an ASCII grid will be exported.
"""
try:
# If wOutFile exists, has already been set by a setter
self.wOutFile
if self.Verbose:
print("Output filename provided.")
# Otherwise, it needs to be set by an configuration file
except:
try:
self.wOutFile = self.configGet("string", "output", "DeflectionOut", optional=True)
except:
# if there is no parsable output string, do not generate output;
# this allows the user to leave the line blank and produce no output
if self.Debug:
print("No output filename provided:")
print(" not writing any deflection output to file")
if self.wOutFile:
if self.wOutFile[-4:] == '.npy':
from numpy import save
save(self.wOutFile,self.w)
else:
from numpy import savetxt
# Shouldn't need more than mm precision, at very most
savetxt(self.wOutFile,self.w,fmt='%.3f')
if self.Verbose:
print("Saving deflections --> " + self.wOutFile) | python | def outputDeflections(self):
"""
Outputs a grid of deflections if an output directory is defined in the
configuration file
If the filename given in the configuration file ends in ".npy", then a binary
numpy grid will be exported.
Otherwise, an ASCII grid will be exported.
"""
try:
# If wOutFile exists, has already been set by a setter
self.wOutFile
if self.Verbose:
print("Output filename provided.")
# Otherwise, it needs to be set by an configuration file
except:
try:
self.wOutFile = self.configGet("string", "output", "DeflectionOut", optional=True)
except:
# if there is no parsable output string, do not generate output;
# this allows the user to leave the line blank and produce no output
if self.Debug:
print("No output filename provided:")
print(" not writing any deflection output to file")
if self.wOutFile:
if self.wOutFile[-4:] == '.npy':
from numpy import save
save(self.wOutFile,self.w)
else:
from numpy import savetxt
# Shouldn't need more than mm precision, at very most
savetxt(self.wOutFile,self.w,fmt='%.3f')
if self.Verbose:
print("Saving deflections --> " + self.wOutFile) | [
"def",
"outputDeflections",
"(",
"self",
")",
":",
"try",
":",
"# If wOutFile exists, has already been set by a setter",
"self",
".",
"wOutFile",
"if",
"self",
".",
"Verbose",
":",
"print",
"(",
"\"Output filename provided.\"",
")",
"# Otherwise, it needs to be set by an configuration file",
"except",
":",
"try",
":",
"self",
".",
"wOutFile",
"=",
"self",
".",
"configGet",
"(",
"\"string\"",
",",
"\"output\"",
",",
"\"DeflectionOut\"",
",",
"optional",
"=",
"True",
")",
"except",
":",
"# if there is no parsable output string, do not generate output;",
"# this allows the user to leave the line blank and produce no output",
"if",
"self",
".",
"Debug",
":",
"print",
"(",
"\"No output filename provided:\"",
")",
"print",
"(",
"\" not writing any deflection output to file\"",
")",
"if",
"self",
".",
"wOutFile",
":",
"if",
"self",
".",
"wOutFile",
"[",
"-",
"4",
":",
"]",
"==",
"'.npy'",
":",
"from",
"numpy",
"import",
"save",
"save",
"(",
"self",
".",
"wOutFile",
",",
"self",
".",
"w",
")",
"else",
":",
"from",
"numpy",
"import",
"savetxt",
"# Shouldn't need more than mm precision, at very most",
"savetxt",
"(",
"self",
".",
"wOutFile",
",",
"self",
".",
"w",
",",
"fmt",
"=",
"'%.3f'",
")",
"if",
"self",
".",
"Verbose",
":",
"print",
"(",
"\"Saving deflections --> \"",
"+",
"self",
".",
"wOutFile",
")"
] | Outputs a grid of deflections if an output directory is defined in the
configuration file
If the filename given in the configuration file ends in ".npy", then a binary
numpy grid will be exported.
Otherwise, an ASCII grid will be exported. | [
"Outputs",
"a",
"grid",
"of",
"deflections",
"if",
"an",
"output",
"directory",
"is",
"defined",
"in",
"the",
"configuration",
"file",
"If",
"the",
"filename",
"given",
"in",
"the",
"configuration",
"file",
"ends",
"in",
".",
"npy",
"then",
"a",
"binary",
"numpy",
"grid",
"will",
"be",
"exported",
".",
"Otherwise",
"an",
"ASCII",
"grid",
"will",
"be",
"exported",
"."
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/base.py#L762-L796 | train |
awickert/gFlex | gflex/base.py | Flexure.TeArraySizeCheck | def TeArraySizeCheck(self):
"""
Checks that Te and q0 array sizes are compatible
For finite difference solution.
"""
# Only if they are both defined and are arrays
# Both being arrays is a possible bug in this check routine that I have
# intentionally introduced
if type(self.Te) == np.ndarray and type(self.qs) == np.ndarray:
# Doesn't touch non-arrays or 1D arrays
if type(self.Te) is np.ndarray:
if (np.array(self.Te.shape) != np.array(self.qs.shape)).any():
sys.exit("q0 and Te arrays have incompatible shapes. Exiting.")
else:
if self.Debug: print("Te and qs array sizes pass consistency check") | python | def TeArraySizeCheck(self):
"""
Checks that Te and q0 array sizes are compatible
For finite difference solution.
"""
# Only if they are both defined and are arrays
# Both being arrays is a possible bug in this check routine that I have
# intentionally introduced
if type(self.Te) == np.ndarray and type(self.qs) == np.ndarray:
# Doesn't touch non-arrays or 1D arrays
if type(self.Te) is np.ndarray:
if (np.array(self.Te.shape) != np.array(self.qs.shape)).any():
sys.exit("q0 and Te arrays have incompatible shapes. Exiting.")
else:
if self.Debug: print("Te and qs array sizes pass consistency check") | [
"def",
"TeArraySizeCheck",
"(",
"self",
")",
":",
"# Only if they are both defined and are arrays",
"# Both being arrays is a possible bug in this check routine that I have ",
"# intentionally introduced",
"if",
"type",
"(",
"self",
".",
"Te",
")",
"==",
"np",
".",
"ndarray",
"and",
"type",
"(",
"self",
".",
"qs",
")",
"==",
"np",
".",
"ndarray",
":",
"# Doesn't touch non-arrays or 1D arrays",
"if",
"type",
"(",
"self",
".",
"Te",
")",
"is",
"np",
".",
"ndarray",
":",
"if",
"(",
"np",
".",
"array",
"(",
"self",
".",
"Te",
".",
"shape",
")",
"!=",
"np",
".",
"array",
"(",
"self",
".",
"qs",
".",
"shape",
")",
")",
".",
"any",
"(",
")",
":",
"sys",
".",
"exit",
"(",
"\"q0 and Te arrays have incompatible shapes. Exiting.\"",
")",
"else",
":",
"if",
"self",
".",
"Debug",
":",
"print",
"(",
"\"Te and qs array sizes pass consistency check\"",
")"
] | Checks that Te and q0 array sizes are compatible
For finite difference solution. | [
"Checks",
"that",
"Te",
"and",
"q0",
"array",
"sizes",
"are",
"compatible",
"For",
"finite",
"difference",
"solution",
"."
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/base.py#L907-L921 | train |
awickert/gFlex | gflex/base.py | Flexure.FD | def FD(self):
"""
Set-up for the finite difference solution method
"""
if self.Verbose:
print("Finite Difference Solution Technique")
# Used to check for coeff_matrix here, but now doing so in self.bc_check()
# called by f1d and f2d at the start
#
# Define a stress-based qs = q0
# But only if the latter has not already been defined
# (e.g., by the getters and setters)
try:
self.qs
except:
self.qs = self.q0.copy()
# Remove self.q0 to avoid issues with multiply-defined inputs
# q0 is the parsable input to either a qs grid or contains (x,(y),q)
del self.q0
# Give it x and y dimensions for help with plotting tools
# (not implemented internally, but a help with external methods)
self.x = np.arange(self.dx/2., self.dx * self.qs.shape[0], self.dx)
if self.dimension == 2:
self.y = np.arange(self.dy/2., self.dy * self.qs.shape[1], self.dy)
# Is there a solver defined
try:
self.Solver # See if it exists already
except:
# Well, will fail if it doesn't see this, maybe not the most reasonable
# error message.
if self.filename:
self.Solver = self.configGet("string", "numerical", "Solver")
else:
sys.exit("No solver defined!")
# Check consistency of size if coeff array was loaded
if self.filename:
# In the case that it is iterative, find the convergence criterion
self.iterative_ConvergenceTolerance = self.configGet("float", "numerical", "ConvergenceTolerance")
# Try to import Te grid or scalar for the finite difference solution
try:
self.Te = self.configGet("float", "input", "ElasticThickness", optional=False)
if self.Te is None:
Tepath = self.configGet("string", "input", "ElasticThickness", optional=False)
self.Te = Tepath
else:
Tepath = None
except:
Tepath = self.configGet("string", "input", "ElasticThickness", optional=False)
self.Te = Tepath
if self.Te is None:
if self.coeff_matrix is not None:
pass
else:
# Have to bring this out here in case it was discovered in the
# try statement that there is no value given
sys.exit("No input elastic thickness or coefficient matrix supplied.")
# or if getter/setter
if type(self.Te) == str:
# Try to import Te grid or scalar for the finite difference solution
Tepath = self.Te
else:
Tepath = None # in case no self.filename present (like for GRASS GIS)
# If there is a Tepath, import Te
# Assume that even if a coeff_matrix is defined
# That the user wants Te if they gave the path
if Tepath:
self.Te = self.loadFile(self.Te, close_on_fail = False)
if self.Te is None:
print("Requested Te file is provided but cannot be located.")
print("No scalar elastic thickness is provided in configuration file")
print("(Typo in path to input Te grid?)")
if self.coeff_matrix is not None:
print("But a coefficient matrix has been found.")
print("Calculations will be carried forward using it.")
else:
print("Exiting.")
sys.exit()
# Check that Te is the proper size if it was loaded
# Will be array if it was loaded
if self.Te.any():
self.TeArraySizeCheck() | python | def FD(self):
"""
Set-up for the finite difference solution method
"""
if self.Verbose:
print("Finite Difference Solution Technique")
# Used to check for coeff_matrix here, but now doing so in self.bc_check()
# called by f1d and f2d at the start
#
# Define a stress-based qs = q0
# But only if the latter has not already been defined
# (e.g., by the getters and setters)
try:
self.qs
except:
self.qs = self.q0.copy()
# Remove self.q0 to avoid issues with multiply-defined inputs
# q0 is the parsable input to either a qs grid or contains (x,(y),q)
del self.q0
# Give it x and y dimensions for help with plotting tools
# (not implemented internally, but a help with external methods)
self.x = np.arange(self.dx/2., self.dx * self.qs.shape[0], self.dx)
if self.dimension == 2:
self.y = np.arange(self.dy/2., self.dy * self.qs.shape[1], self.dy)
# Is there a solver defined
try:
self.Solver # See if it exists already
except:
# Well, will fail if it doesn't see this, maybe not the most reasonable
# error message.
if self.filename:
self.Solver = self.configGet("string", "numerical", "Solver")
else:
sys.exit("No solver defined!")
# Check consistency of size if coeff array was loaded
if self.filename:
# In the case that it is iterative, find the convergence criterion
self.iterative_ConvergenceTolerance = self.configGet("float", "numerical", "ConvergenceTolerance")
# Try to import Te grid or scalar for the finite difference solution
try:
self.Te = self.configGet("float", "input", "ElasticThickness", optional=False)
if self.Te is None:
Tepath = self.configGet("string", "input", "ElasticThickness", optional=False)
self.Te = Tepath
else:
Tepath = None
except:
Tepath = self.configGet("string", "input", "ElasticThickness", optional=False)
self.Te = Tepath
if self.Te is None:
if self.coeff_matrix is not None:
pass
else:
# Have to bring this out here in case it was discovered in the
# try statement that there is no value given
sys.exit("No input elastic thickness or coefficient matrix supplied.")
# or if getter/setter
if type(self.Te) == str:
# Try to import Te grid or scalar for the finite difference solution
Tepath = self.Te
else:
Tepath = None # in case no self.filename present (like for GRASS GIS)
# If there is a Tepath, import Te
# Assume that even if a coeff_matrix is defined
# That the user wants Te if they gave the path
if Tepath:
self.Te = self.loadFile(self.Te, close_on_fail = False)
if self.Te is None:
print("Requested Te file is provided but cannot be located.")
print("No scalar elastic thickness is provided in configuration file")
print("(Typo in path to input Te grid?)")
if self.coeff_matrix is not None:
print("But a coefficient matrix has been found.")
print("Calculations will be carried forward using it.")
else:
print("Exiting.")
sys.exit()
# Check that Te is the proper size if it was loaded
# Will be array if it was loaded
if self.Te.any():
self.TeArraySizeCheck() | [
"def",
"FD",
"(",
"self",
")",
":",
"if",
"self",
".",
"Verbose",
":",
"print",
"(",
"\"Finite Difference Solution Technique\"",
")",
"# Used to check for coeff_matrix here, but now doing so in self.bc_check()",
"# called by f1d and f2d at the start",
"# ",
"# Define a stress-based qs = q0",
"# But only if the latter has not already been defined",
"# (e.g., by the getters and setters)",
"try",
":",
"self",
".",
"qs",
"except",
":",
"self",
".",
"qs",
"=",
"self",
".",
"q0",
".",
"copy",
"(",
")",
"# Remove self.q0 to avoid issues with multiply-defined inputs",
"# q0 is the parsable input to either a qs grid or contains (x,(y),q)",
"del",
"self",
".",
"q0",
"# Give it x and y dimensions for help with plotting tools",
"# (not implemented internally, but a help with external methods)",
"self",
".",
"x",
"=",
"np",
".",
"arange",
"(",
"self",
".",
"dx",
"/",
"2.",
",",
"self",
".",
"dx",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"0",
"]",
",",
"self",
".",
"dx",
")",
"if",
"self",
".",
"dimension",
"==",
"2",
":",
"self",
".",
"y",
"=",
"np",
".",
"arange",
"(",
"self",
".",
"dy",
"/",
"2.",
",",
"self",
".",
"dy",
"*",
"self",
".",
"qs",
".",
"shape",
"[",
"1",
"]",
",",
"self",
".",
"dy",
")",
"# Is there a solver defined",
"try",
":",
"self",
".",
"Solver",
"# See if it exists already",
"except",
":",
"# Well, will fail if it doesn't see this, maybe not the most reasonable",
"# error message.",
"if",
"self",
".",
"filename",
":",
"self",
".",
"Solver",
"=",
"self",
".",
"configGet",
"(",
"\"string\"",
",",
"\"numerical\"",
",",
"\"Solver\"",
")",
"else",
":",
"sys",
".",
"exit",
"(",
"\"No solver defined!\"",
")",
"# Check consistency of size if coeff array was loaded",
"if",
"self",
".",
"filename",
":",
"# In the case that it is iterative, find the convergence criterion",
"self",
".",
"iterative_ConvergenceTolerance",
"=",
"self",
".",
"configGet",
"(",
"\"float\"",
",",
"\"numerical\"",
",",
"\"ConvergenceTolerance\"",
")",
"# Try to import Te grid or scalar for the finite difference solution",
"try",
":",
"self",
".",
"Te",
"=",
"self",
".",
"configGet",
"(",
"\"float\"",
",",
"\"input\"",
",",
"\"ElasticThickness\"",
",",
"optional",
"=",
"False",
")",
"if",
"self",
".",
"Te",
"is",
"None",
":",
"Tepath",
"=",
"self",
".",
"configGet",
"(",
"\"string\"",
",",
"\"input\"",
",",
"\"ElasticThickness\"",
",",
"optional",
"=",
"False",
")",
"self",
".",
"Te",
"=",
"Tepath",
"else",
":",
"Tepath",
"=",
"None",
"except",
":",
"Tepath",
"=",
"self",
".",
"configGet",
"(",
"\"string\"",
",",
"\"input\"",
",",
"\"ElasticThickness\"",
",",
"optional",
"=",
"False",
")",
"self",
".",
"Te",
"=",
"Tepath",
"if",
"self",
".",
"Te",
"is",
"None",
":",
"if",
"self",
".",
"coeff_matrix",
"is",
"not",
"None",
":",
"pass",
"else",
":",
"# Have to bring this out here in case it was discovered in the ",
"# try statement that there is no value given",
"sys",
".",
"exit",
"(",
"\"No input elastic thickness or coefficient matrix supplied.\"",
")",
"# or if getter/setter",
"if",
"type",
"(",
"self",
".",
"Te",
")",
"==",
"str",
":",
"# Try to import Te grid or scalar for the finite difference solution",
"Tepath",
"=",
"self",
".",
"Te",
"else",
":",
"Tepath",
"=",
"None",
"# in case no self.filename present (like for GRASS GIS)",
"# If there is a Tepath, import Te",
"# Assume that even if a coeff_matrix is defined",
"# That the user wants Te if they gave the path",
"if",
"Tepath",
":",
"self",
".",
"Te",
"=",
"self",
".",
"loadFile",
"(",
"self",
".",
"Te",
",",
"close_on_fail",
"=",
"False",
")",
"if",
"self",
".",
"Te",
"is",
"None",
":",
"print",
"(",
"\"Requested Te file is provided but cannot be located.\"",
")",
"print",
"(",
"\"No scalar elastic thickness is provided in configuration file\"",
")",
"print",
"(",
"\"(Typo in path to input Te grid?)\"",
")",
"if",
"self",
".",
"coeff_matrix",
"is",
"not",
"None",
":",
"print",
"(",
"\"But a coefficient matrix has been found.\"",
")",
"print",
"(",
"\"Calculations will be carried forward using it.\"",
")",
"else",
":",
"print",
"(",
"\"Exiting.\"",
")",
"sys",
".",
"exit",
"(",
")",
"# Check that Te is the proper size if it was loaded",
"# Will be array if it was loaded",
"if",
"self",
".",
"Te",
".",
"any",
"(",
")",
":",
"self",
".",
"TeArraySizeCheck",
"(",
")"
] | Set-up for the finite difference solution method | [
"Set",
"-",
"up",
"for",
"the",
"finite",
"difference",
"solution",
"method"
] | 3ac32249375b0f8d342a142585d86ea4d905a5a0 | https://github.com/awickert/gFlex/blob/3ac32249375b0f8d342a142585d86ea4d905a5a0/gflex/base.py#L927-L1008 | train |