Unnamed: 0 int64 0 10k | function stringlengths 79 138k | label stringclasses 20
values | info stringlengths 42 261 |
|---|---|---|---|
2,000 | def logSaveToFile(self, *ignore):
filename = self.uiFileDialog("save",
title=_("arelle - Save Messages Log"),
initialdir=".",
filetypes=[(_("Txt file"), "*.txt")],
defaultextension=".txt")
if not filename:
return False
t... | IOError | dataset/ETHPy150Open Arelle/Arelle/arelle/CntlrWinMain.py/CntlrWinMain.logSaveToFile |
2,001 | def __call__(self, *args):
"""Apply first function SUBST to arguments, than FUNC."""
try:
if self.subst:
args = self.subst(*args)
return self.func(*args)
except __HOLE__ as msg:
raise SystemExit(msg)
except Exception:
# this... | SystemExit | dataset/ETHPy150Open Arelle/Arelle/arelle/CntlrWinMain.py/TkinterCallWrapper.__call__ |
2,002 | def detect_encoding(readline):
"""
The detect_encoding() function is used to detect the encoding that should
be used to decode a Python source file. It requires one argment, readline,
in the same way as the tokenize() generator.
It will call readline a maximum of twice, and return the encoding used... | StopIteration | dataset/ETHPy150Open inducer/pudb/pudb/lowlevel.py/detect_encoding |
2,003 | def list_post_default(self, request, **kwargs):
data = self.get_request_data()
serialiser = self.get_serialiser()
serialiser_kwargs = self.get_serialiser_kwargs()
try:
with transaction.atomic():
obj = serialiser.object_inflate(data, **serialiser_kwargs)
... | ValueError | dataset/ETHPy150Open funkybob/django-nap/nap/rest/models.py/ModelPublisher.list_post_default |
2,004 | def clear_history(self):
try:
readline.clear_history()
except __HOLE__:
len = self.get_max_length()
readline.set_history_length(0)
readline.set_history_length(len) | AttributeError | dataset/ETHPy150Open AppScale/appscale/AppServer/lib/grizzled/grizzled/history.py/ReadlineHistory.clear_history |
2,005 | def main():
paths = sys.argv[1:] or ['.']
print('Importing nltk...')
try:
import nltk
except __HOLE__:
print('Unable to import nltk -- check your PYTHONPATH.')
sys.exit(-1)
print('Finding definitions of deprecated funtions & classes in nltk...')
find_deprecated_defs(nlt... | ImportError | dataset/ETHPy150Open nltk/nltk/tools/find_deprecated.py/main |
2,006 | def get_compute_capability(device_id=None, verbose=False):
"""
Query compute capability through PyCuda and check it's 5.0 (Maxwell) or
greater.
5.0 (GTX750 Ti) only fp32 support
5.2 (GTX9xx series) required for fp16
By default, check all devices and return the highest compute capability.
Ar... | ImportError | dataset/ETHPy150Open NervanaSystems/neon/neon/backends/util/check_gpu.py/get_compute_capability |
2,007 | def get_device_count(verbose=False):
"""
Query device count through PyCuda.
Arguments:
verbose (bool): prints verbose logging if True, default False.
Returns:
int: Number of GPUs available.
"""
try:
import pycuda
import pycuda.driver as drv
except __HOLE__:
... | ImportError | dataset/ETHPy150Open NervanaSystems/neon/neon/backends/util/check_gpu.py/get_device_count |
2,008 | def read_rules(self):
if not exists(self.rules_file):
self.clear()
return
# Only read if the rules file has been modified
try:
mtime = getmtime(self.rules_file)
except __HOLE__:
log.err("Failed to get mtime of %s" % self.rules_file)
return
if mtime <= self.rules_last_r... | OSError | dataset/ETHPy150Open graphite-project/carbon/lib/carbon/aggregator/rules.py/RuleManager.read_rules |
2,009 | def parse_definition(self, line):
try:
left_side, right_side = line.split('=', 1)
output_pattern, frequency = left_side.split()
method, input_pattern = right_side.split()
frequency = int( frequency.lstrip('(').rstrip(')') )
return AggregationRule(input_pattern, output_pattern, method, ... | ValueError | dataset/ETHPy150Open graphite-project/carbon/lib/carbon/aggregator/rules.py/RuleManager.parse_definition |
2,010 | def get_aggregate_metric(self, metric_path):
if metric_path in self.cache:
return self.cache[metric_path]
match = self.regex.match(metric_path)
result = None
if match:
extracted_fields = match.groupdict()
try:
result = self.output_template % extracted_fields
except __HO... | TypeError | dataset/ETHPy150Open graphite-project/carbon/lib/carbon/aggregator/rules.py/AggregationRule.get_aggregate_metric |
2,011 | def refresh_cache(f):
"""Decorator to update the instance_info_cache
Requires context and instance as function args
"""
argspec = inspect.getargspec(f)
@functools.wraps(f)
def wrapper(self, context, *args, **kwargs):
try:
# get the instance from arguments (or raise ValueErr... | ValueError | dataset/ETHPy150Open BU-NU-CLOUD-SP16/Trusted-Platform-Module-nova/nova/network/base_api.py/refresh_cache |
2,012 | def render(self, name, value, attrs=None):
elements = []
try:
values = json.loads(value)
except __HOLE__:
values = {}
# value sometimes come as unicode and we need to treat it
if type(values) == unicode:
values = json.loads(values)
ob... | TypeError | dataset/ETHPy150Open opps/opps/opps/fields/widgets.py/JSONField.render |
2,013 | def update(self):
'''
Update the form in background
'''
# get the information
try:
disk_info = self.statistics['Disk']['text']['/']
swap_info = self.statistics['Memory']['text']['swap_memory']
memory_info = self.statistics['Memory']... | KeyError | dataset/ETHPy150Open black-perl/ptop/ptop/interfaces/GUI.py/PtopGUI.update |
2,014 | def _set_bind_addr(self, value):
if value is None:
self.socket_file = None
self.socket_host = None
self.socket_port = None
elif isinstance(value, basestring):
self.socket_file = value
self.socket_host = None
self.socket_port = None
... | ValueError | dataset/ETHPy150Open AppScale/appscale/AppServer/lib/cherrypy/cherrypy/_cpserver.py/Server._set_bind_addr |
2,015 | def _fore(self, color):
def get(what):
try:
r = getattr(self._modifiers, what)
except __HOLE__:
r = getattr(self._forecolors, what)
return r
args = map(get, color.split(u"_"))
return u"".join(args) | AttributeError | dataset/ETHPy150Open gabrielfalcao/couleur/couleur/__init__.py/Shell._fore |
2,016 | def parse_addr_spec(spec, defhost = None, defport = None):
"""Parse a host:port specification and return a 2-tuple ("host", port) as
understood by the Python socket functions.
>>> parse_addr_spec("192.168.0.1:9999")
('192.168.0.1', 9999)
If defhost or defport are given and not None, the respective... | ValueError | dataset/ETHPy150Open arlolra/flashproxy/flashproxy/util.py/parse_addr_spec |
2,017 | def approximate_current_flow_betweenness_centrality(G, normalized=True,
weight='weight',
dtype=float, solver='full',
epsilon=0.5, kmax=10000):
r"""Compute the a... | ImportError | dataset/ETHPy150Open networkx/networkx/networkx/algorithms/centrality/current_flow_betweenness.py/approximate_current_flow_betweenness_centrality |
2,018 | def current_flow_betweenness_centrality(G, normalized=True, weight='weight',
dtype=float, solver='full'):
r"""Compute current-flow betweenness centrality for nodes.
Current-flow betweenness centrality uses an electrical current
model for information spreading in cont... | ImportError | dataset/ETHPy150Open networkx/networkx/networkx/algorithms/centrality/current_flow_betweenness.py/current_flow_betweenness_centrality |
2,019 | def edge_current_flow_betweenness_centrality(G, normalized=True,
weight='weight',
dtype=float, solver='full'):
"""Compute current-flow betweenness centrality for edges.
Current-flow betweenness centrality uses an electric... | ImportError | dataset/ETHPy150Open networkx/networkx/networkx/algorithms/centrality/current_flow_betweenness.py/edge_current_flow_betweenness_centrality |
2,020 | def get_or_create_iexact(self, **kwargs):
"""
Case insensitive title version of ``get_or_create``. Also
allows for multiple existing results.
"""
lookup = dict(**kwargs)
try:
lookup["title__iexact"] = lookup.pop("title")
except __HOLE__:
pa... | KeyError | dataset/ETHPy150Open stephenmcd/mezzanine/mezzanine/generic/managers.py/KeywordManager.get_or_create_iexact |
2,021 | def run_cl(argv=[]):
logging.basicConfig(level=logging.INFO)
try:
app = MultioptParser(
clsname='floyd',
version=floyd.get_version(),
desc_short="Static website generator",
global_options=[
make_option("-v", "--verbose", action="store_true", dest="verbose"),
make_optio... | KeyboardInterrupt | dataset/ETHPy150Open nikcub/floyd/floyd/core/command_utils.py/run_cl |
2,022 | def integer(x):
try:
int(x)
except (ValueError, __HOLE__):
raise ValueError("%r is not a valid integer" % x)
else:
return x | TypeError | dataset/ETHPy150Open cloudtools/troposphere/troposphere/validators.py/integer |
2,023 | def add_buffer(self, data=''):
if data:
self.__buffer += data
packets = self.__buffer.split(DELIMITER)
self.__buffer = b''
for packet in packets:
try:
self.__process_packet(packet)
except __HOLE__:
self.__buffer = pack... | ValueError | dataset/ETHPy150Open circuits/circuits/circuits/node/protocol.py/Protocol.add_buffer |
2,024 | def _kill(self, f):
# close and remove file
try:
f.close()
except (SystemExit, KeyboardInterrupt):
raise
except:
pass
try:
os.unlink(f.name)
except (SystemExit, __HOLE__):
raise
except:
pass | KeyboardInterrupt | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/test/test_uu.py/UUFileTest._kill |
2,025 | def mutate_value(element,parts,rel_amount=None):
i = random.randint(0,len(element.values)-1)
val = element.values[i]
name = element.spice_name
if rel_amount==None:
try:
val[i] = log_dist(parts[name]['value'][0],parts[name]['value'][1])
except:
return element
e... | KeyError | dataset/ETHPy150Open Ttl/evolutionary-circuits/evolutionary/chromosomes/netlist.py/mutate_value |
2,026 | def mutate(self):
m = random.randint(0,7)
i = random.randint(0,len(self.elements)-1)
if m==0:
#Change value of one component
m = random.randint(0,1)
if m==0:
#New value
self.elements[i] = mutate_value(self.elements[i],self.parts... | IndexError | dataset/ETHPy150Open Ttl/evolutionary-circuits/evolutionary/chromosomes/netlist.py/Chromosome.mutate |
2,027 | def parse_commits(self):
import Commit
events = []
# if this is a cloned repository
if not self.from_feed:
repo_dir = os.path.join(settings.REPO_ROOT, self.project.url_path)
# add the commits
backend = get_backend(self.vcs if self.vcs != 'svn' else 'git')
repository =... | IndexError | dataset/ETHPy150Open rcos/Observatory/observatory/dashboard/models/Repository.py/Repository.parse_commits |
2,028 | def clone_or_fetch(self):
if self.from_feed: return
fresh_clone = True
# ensure that REPO_ROOT already exists
try:
os.makedirs(settings.REPO_ROOT, 0770)
except __HOLE__ as e:
pass
# construct the name of the directory into which to clone the repository
dest_dir = os.pa... | OSError | dataset/ETHPy150Open rcos/Observatory/observatory/dashboard/models/Repository.py/Repository.clone_or_fetch |
2,029 | def clone_svn_repo(clone_url, destination_dir, fresh_clone = False):
if fresh_clone:
# make the repo's directory
try:
os.makedirs(destination_dir, 0770)
except __HOLE__ as e:
pass
clone_cmdline = ["git", "svn", "clone", clone_url, destination_dir]
else:
clone_cmdline = ["git", "... | OSError | dataset/ETHPy150Open rcos/Observatory/observatory/dashboard/models/Repository.py/clone_svn_repo |
2,030 | def getImage(self):
if pil is None or self._pilimage == 'failed':
return None
if self._pilimage:
return self._pilimage
else:
data = self.getData()
if not data:
self._pilimage = 'failed'
return None
try:
... | IOError | dataset/ETHPy150Open pycollada/pycollada/collada/material.py/CImage.getImage |
2,031 | @staticmethod
def load(collada, localscope, node):
localscope = {} # we have our own scope, shadow it
params = []
id = node.get('id')
profilenode = node.find( tag('profile_COMMON') )
if profilenode is None:
raise DaeUnsupportedError('Found effect with profile othe... | ValueError | dataset/ETHPy150Open pycollada/pycollada/collada/material.py/Effect.load |
2,032 | @staticmethod
def _loadShadingParam( collada, localscope, node ):
"""Load from the node a definition for a material property."""
children = node.getchildren()
if not children: raise DaeIncompleteError('Incorrect effect shading parameter '+node.tag)
vnode = children[0]
if vnod... | ValueError | dataset/ETHPy150Open pycollada/pycollada/collada/material.py/Effect._loadShadingParam |
2,033 | def get_current(self):
"""
Returns the current ``Site`` based on the SITE_ID in the
project's settings. The ``Site`` object is cached the first
time it's retrieved from the database.
"""
from django.conf import settings
try:
sid = settings.SITE_ID
... | AttributeError | dataset/ETHPy150Open cloudera/hue/desktop/core/ext-py/Django-1.6.10/django/contrib/sites/models.py/SiteManager.get_current |
2,034 | def clear_site_cache(sender, **kwargs):
"""
Clears the cache (if primed) each time a site is saved or deleted
"""
instance = kwargs['instance']
try:
del SITE_CACHE[instance.pk]
except __HOLE__:
pass | KeyError | dataset/ETHPy150Open cloudera/hue/desktop/core/ext-py/Django-1.6.10/django/contrib/sites/models.py/clear_site_cache |
2,035 | @register.filter
def date(value, arg=None):
"""Formats a date according to the given format."""
if not value:
return u''
if arg is None:
arg = settings.DATE_FORMAT
try:
return formats.date_format(value, arg)
except __HOLE__:
try:
return format(value, arg)
... | AttributeError | dataset/ETHPy150Open crate-archive/crate-site/crateweb/apps/core/helpers.py/date |
2,036 | def ensure_oauth2(provider):
"""
Decorator to ensure a user has been authenticated with the given oauth2 provider.
Usage:
from tethys_sdk.services import ensure_oauth2, get_dataset_engine
@ensure_oauth2('hydroshare-oauth2')
def controller(request):
engine = get_dataset... | AttributeError | dataset/ETHPy150Open tethysplatform/tethys/tethys_services/utilities.py/ensure_oauth2 |
2,037 | def initialize_engine_object(engine, endpoint, apikey=None, username=None, password=None, request=None):
"""
Initialize a DatasetEngine object from a string that points at the engine class.
"""
# Constants
HYDROSHARE_OAUTH_PROVIDER_NAME = 'hydroshare'
# Derive import parts from engine string
... | ObjectDoesNotExist | dataset/ETHPy150Open tethysplatform/tethys/tethys_services/utilities.py/initialize_engine_object |
2,038 | def abstract_is_link(process):
"""
Determine if the process abstract is a link.
Args:
process (owslib.wps.Process): WPS Process object.
Returns:
(bool): True if abstract is a link, False otherwise.
"""
try:
abstract = process.abstract
except __HOLE__:
return Fal... | AttributeError | dataset/ETHPy150Open tethysplatform/tethys/tethys_services/utilities.py/abstract_is_link |
2,039 | def activate_wps(wps, endpoint, name):
"""
Activate a WebProcessingService object by calling getcapabilities() on it and handle errors appropriately.
Args:
wps (owslib.wps.WebProcessingService): A owslib.wps.WebProcessingService object.
Returns:
(owslib.wps.WebProcessingService): Returns a... | HTTPError | dataset/ETHPy150Open tethysplatform/tethys/tethys_services/utilities.py/activate_wps |
2,040 | @register.tag
def settings(parser, token):
var_name = None
try:
tag_name, arg = token.contents.split(None, 1)
except __HOLE__:
raise template.TemplateSyntaxError(
"%r tag requires arguments" % token.contents.split()[0])
m = re.search(r'(.*?) as (\w+)', arg)
if m:
... | ValueError | dataset/ETHPy150Open jqb/django-settings/django_settings/templatetags/settings_tags.py/settings |
2,041 | def _get_templated_url(self, template, input_url, method=None):
url = None
try:
if method=="members":
match = re.match("^https://publons.com/author/(\d+)/.+", input_url)
user_id = match.group(1)
url = template % user_id
else:
... | AttributeError | dataset/ETHPy150Open Impactstory/total-impact-core/totalimpact/providers/publons.py/Publons._get_templated_url |
2,042 | def _extract_biblio(self, page, id=None):
dict_of_keylists = {
'title' : ['title'],
'authors' : ['author', 'last_name'],
'journal' : ['source', 'provider'],
'review_url' : ['source', 'url'],
'review_type' : ['review_type'],
'create_date' : ... | KeyError | dataset/ETHPy150Open Impactstory/total-impact-core/totalimpact/providers/publons.py/Publons._extract_biblio |
2,043 | def serve_forever(host, port, childnum):
# create, bind. listen
listen_sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
# re-use the port
listen_sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
# put listening socket into non-blocking mode
listen_sock.setblocking(0)
listen... | IOError | dataset/ETHPy150Open rspivak/csdesign/server04.py/serve_forever |
2,044 | @login_required
@transaction.atomic
def create_invoice(request):
pk = request.GET.get('project', None)
to_date = request.GET.get('to_date', None)
if not (pk and to_date):
raise Http404
from_date = request.GET.get('from_date', None)
if not request.user.has_perm('crm.generate_project_invoice')... | ValueError | dataset/ETHPy150Open caktus/django-timepiece/timepiece/contracts/views.py/create_invoice |
2,045 | def flush(self, n=4096):
"""
Flush `n` bytes of data from the reader Stream to the writer Stream.
Returns the number of bytes that were actually flushed. A return value
of zero is not an error.
If EOF has been reached, `None` is returned.
"""
try:
r... | OSError | dataset/ETHPy150Open d11wtq/dockerpty/dockerpty/io.py/Pump.flush |
2,046 | def get_available_name(self, name):
"""
Returns a filename that's free on the target storage system, and
available for new content to be written to.
"""
# If the filename already exists, keep adding an underscore to the name
# of the file until the filename doesn't exist.... | ValueError | dataset/ETHPy150Open dcramer/django-compositepks/django/core/files/storage.py/Storage.get_available_name |
2,047 | def _save(self, name, content):
full_path = self.path(name)
directory = os.path.dirname(full_path)
if not os.path.exists(directory):
os.makedirs(directory)
elif not os.path.isdir(directory):
raise IOError("%s exists and is not a directory." % directory)
... | OSError | dataset/ETHPy150Open dcramer/django-compositepks/django/core/files/storage.py/FileSystemStorage._save |
2,048 | def path(self, name):
try:
path = safe_join(self.location, name)
except __HOLE__:
raise SuspiciousOperation("Attempted access to '%s' denied." % name)
return os.path.normpath(path) | ValueError | dataset/ETHPy150Open dcramer/django-compositepks/django/core/files/storage.py/FileSystemStorage.path |
2,049 | def get_storage_class(import_path):
try:
dot = import_path.rindex('.')
except ValueError:
raise ImproperlyConfigured("%s isn't a storage module." % import_path)
module, classname = import_path[:dot], import_path[dot+1:]
try:
mod = __import__(module, {}, {}, [''])
except __HOL... | ImportError | dataset/ETHPy150Open dcramer/django-compositepks/django/core/files/storage.py/get_storage_class |
2,050 | def replace_wikiwords(value):
def replace_wikiword(m):
slug = m.group(1)
try:
page = WikiPage.objects.get(slug=slug)
kwargs = {
'slug': slug,
}
url = reverse('wakawaka_page', kwargs=kwargs)
return r'<a href="%s">%s</a>' % (u... | ObjectDoesNotExist | dataset/ETHPy150Open bartTC/django-wakawaka/wakawaka/templatetags/wakawaka_tags.py/replace_wikiwords |
2,051 | def get_mapper():
"""Returns the mapper."""
global _CACHED_MAPPER
if not _CACHED_MAPPER:
name = getattr(settings, 'MULTITENANT_MAPPER_CLASS', None)
if not name:
raise ImproperlyConfigured("You must specify MULTITENANT_MAPPER_CLASS in settings.")
try:
module_p... | ImportError | dataset/ETHPy150Open mik3y/django-db-multitenant/db_multitenant/utils.py/get_mapper |
2,052 | def trigger_event(self, event, *args, **kwargs):
"""
Trigger that calls all of the specified events associated with this class.
"""
for event_class, event_callbacks in six.iteritems(self._event_callbacks):
if not isinstance(self, event_class):
continue
... | KeyboardInterrupt | dataset/ETHPy150Open spotify/luigi/luigi/task.py/Task.trigger_event |
2,053 | def getpaths(struct):
"""
Maps all Tasks in a structured data object to their .output().
"""
if isinstance(struct, Task):
return struct.output()
elif isinstance(struct, dict):
r = {}
for k, v in six.iteritems(struct):
r[k] = getpaths(v)
return r
else:
... | TypeError | dataset/ETHPy150Open spotify/luigi/luigi/task.py/getpaths |
2,054 | def flatten(struct):
"""
Creates a flat list of all all items in structured output (dicts, lists, items):
.. code-block:: python
>>> sorted(flatten({'a': 'foo', 'b': 'bar'}))
['bar', 'foo']
>>> sorted(flatten(['foo', ['bar', 'troll']]))
['bar', 'foo', 'troll']
>>> f... | TypeError | dataset/ETHPy150Open spotify/luigi/luigi/task.py/flatten |
2,055 | @extension(EXT_D)
def d_hook(self, node):
# create the compilation task: cpp or cc
task = self.create_task(self.generate_headers and 'd_with_header' or 'd')
try: obj_ext = self.obj_ext
except __HOLE__: obj_ext = '_%d.o' % self.idx
task.inputs = [node]
task.outputs = [node.change_ext(obj_ext)]
self.compiled_task... | AttributeError | dataset/ETHPy150Open appcelerator-archive/poc-nodejs-desktop/Resources/nodejs/builds/linux/node/lib/node/wafadmin/Tools/d.py/d_hook |
2,056 | def handle(self, *args, **options):
try:
build_path, version, build_number = args
except __HOLE__:
raise CommandError('Usage: %s\n%s' % (self.args, self.help))
try:
build_number = int(build_number)
except ValueError:
raise CommandError("Bu... | ValueError | dataset/ETHPy150Open dimagi/commcare-hq/corehq/apps/builds/management/commands/add_commcare_build.py/Command.handle |
2,057 | def receive_request(self, request):
"""
:type request: Request
:param request: the incoming request
:rtype : Transaction
"""
logger.debug("receive_request - " + str(request))
try:
host, port = request.source
except __HOLE__:
return... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.receive_request |
2,058 | def receive_response(self, response):
"""
:type response: Response
:param response:
:rtype : Transaction
"""
logger.debug("receive_response - " + str(response))
try:
host, port = response.source
except __HOLE__:
return
key_... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.receive_response |
2,059 | def receive_empty(self, message):
"""
:type message: Message
:param message:
:rtype : Transaction
"""
logger.debug("receive_empty - " + str(message))
try:
host, port = message.source
except __HOLE__:
return
key_mid = hash(s... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.receive_empty |
2,060 | def send_request(self, request):
"""
:type request: Request
:param request:
"""
logger.debug("send_request - " + str(request))
assert isinstance(request, Request)
try:
host, port = request.destination
except __HOLE__:
return
... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.send_request |
2,061 | def send_response(self, transaction):
"""
:type transaction: Transaction
:param transaction:
"""
logger.debug("send_response - " + str(transaction.response))
if transaction.response.type is None:
if transaction.request.type == defines.Types["CON"] and not tra... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.send_response |
2,062 | def send_empty(self, transaction, related, message):
"""
:param transaction:
:type message: Message
:param message:
"""
logger.debug("send_empty - " + str(message))
if transaction is None:
try:
host, port = message.destination
... | AttributeError | dataset/ETHPy150Open Tanganelli/CoAPthon/coapthon/layers/messagelayer.py/MessageLayer.send_empty |
2,063 | def signal_name(signum):
# Hackety-hack-hack: is there really no better way to reverse lookup the
# signal name? If you read this and know a way: please provide a patch :)
try:
return _signames[signum]
except __HOLE__:
return 'SIG_UNKNOWN' | KeyError | dataset/ETHPy150Open bretth/django-pq/pq/worker.py/signal_name |
2,064 | def _install_signal_handlers(self):
"""Installs signal handlers for handling SIGINT and SIGTERM
gracefully.
"""
def request_force_stop(signum, frame):
"""Terminates the application (cold shutdown).
"""
self.log.warning('Cold shut down.')
... | OSError | dataset/ETHPy150Open bretth/django-pq/pq/worker.py/Worker._install_signal_handlers |
2,065 | def fork_and_perform_job(self, job):
"""Spawns a work horse to perform the actual work and passes it a job.
The worker will wait for the work horse and make sure it executes
within the given timeout bounds, or will end the work horse with
SIGALRM.
"""
child_pid = os.fork(... | OSError | dataset/ETHPy150Open bretth/django-pq/pq/worker.py/Worker.fork_and_perform_job |
2,066 | def perform_job(self, job):
"""Performs the actual work of a job. Will/should only be called
inside the work horse's process.
"""
self.procline('Processing %s from %s since %s' % (
job.func_name,
job.origin, time.time()))
# do it this way to avoid... | NameError | dataset/ETHPy150Open bretth/django-pq/pq/worker.py/Worker.perform_job |
2,067 | def get_all_coverage_modules(app_module):
"""
Returns all possible modules to report coverage on, even if they
aren't loaded.
"""
# We start off with the imported models.py, so we need to import
# the parent app package to find the path.
app_path = app_module.__name__.split('.')[:-1]
app... | ImportError | dataset/ETHPy150Open garethr/django-test-extensions/src/test_extensions/testrunners/codecoverage.py/get_all_coverage_modules |
2,068 | def run_tests(test_labels, verbosity=1, interactive=True,
extra_tests=[], nodatabase=False, xml_out=False, callgraph=False, html_only=False):
"""
Test runner which displays a code coverage report at the end of the
run.
"""
cov = coverage.coverage()
cov.erase()
cov.use_cache(0)
t... | ImportError | dataset/ETHPy150Open garethr/django-test-extensions/src/test_extensions/testrunners/codecoverage.py/run_tests |
2,069 | def __init__(self, config):
self._configure(config)
self._log_handlers = []
# Parse command line args
self.config['server_number'] += self.config['args']['server_number']
self.config['procname'] += "_{}".format(self.config['server_number'])
# setup all our log handlers
... | ImportError | dataset/ETHPy150Open simplecrypto/powerpool/powerpool/main.py/PowerPool.__init__ |
2,070 | def handle(self, data, address):
self.logger.info("Recieved new command {}".format(data))
parts = data.split(" ")
try:
component = self.components[parts[0]]
func = getattr(component, parts[1])
kwargs = {}
args = []
for arg in parts[2:]:... | AttributeError | dataset/ETHPy150Open simplecrypto/powerpool/powerpool/main.py/PowerPool.handle |
2,071 | def select(self):
e = xlib.XEvent()
while xlib.XPending(self._display):
xlib.XNextEvent(self._display, e)
# Key events are filtered by the xlib window event
# handler so they get a shot at the prefiltered event.
if e.xany.type not in (xlib.KeyPress, xlib.... | KeyError | dataset/ETHPy150Open ardekantur/pyglet/pyglet/canvas/xlib.py/XlibDisplay.select |
2,072 | def reverse(viewname, urlconf=None, args=None, kwargs=None, current_app=None):
if urlconf is None:
urlconf = get_urlconf()
resolver = get_resolver(urlconf)
args = args or []
kwargs = kwargs or {}
prefix = get_script_prefix()
if not isinstance(viewname, six.string_types):
view =... | KeyError | dataset/ETHPy150Open django/django/django/urls/base.py/reverse |
2,073 | def clear_script_prefix():
"""
Unset the script prefix for the current thread.
"""
try:
del _prefixes.value
except __HOLE__:
pass | AttributeError | dataset/ETHPy150Open django/django/django/urls/base.py/clear_script_prefix |
2,074 | def ajax_csv_importer(request):
save = True if request.POST.get('save', False) else False
raw_csv_data = request.POST.get('csv-data', '')
primary_attr = request.POST.get('primary-attr', 'hostname')
@transaction.commit_manually
def do_csv_import(data):
try:
return csv_import(data... | ValidationError | dataset/ETHPy150Open mozilla/inventory/mcsv/views.py/ajax_csv_importer |
2,075 | def perm(accessing_obj, accessed_obj, *args, **kwargs):
"""
The basic permission-checker. Ignores case.
Usage:
perm(<permission>)
where <permission> is the permission accessing_obj must
have in order to pass the lock.
If the given permission is part of settings.PERMISSION_HIERARCHY,
... | AttributeError | dataset/ETHPy150Open evennia/evennia/evennia/locks/lockfuncs.py/perm |
2,076 | def dbref(accessing_obj, accessed_obj, *args, **kwargs):
"""
Usage:
dbref(3)
This lock type checks if the checking object
has a particular dbref. Note that this only
works for checking objects that are stored
in the database (e.g. not for commands)
"""
if not args:
return ... | ValueError | dataset/ETHPy150Open evennia/evennia/evennia/locks/lockfuncs.py/dbref |
2,077 | def holds(accessing_obj, accessed_obj, *args, **kwargs):
"""
Usage:
holds() checks if accessed_obj or accessed_obj.obj
is held by accessing_obj
holds(key/dbref) checks if accessing_obj holds an object
with given key/dbref
holds(at... | AttributeError | dataset/ETHPy150Open evennia/evennia/evennia/locks/lockfuncs.py/holds |
2,078 | def add_dev(self, dev):
"""
Add a device to the ring. This device dict should have a minimum of the
following keys:
====== ===============================================================
id unique integer identifier amongst devices. Defaults to the next
id ... | ValueError | dataset/ETHPy150Open openstack/swift/swift/common/ring/builder.py/RingBuilder.add_dev |
2,079 | def _build_dispersion_graph(self, old_replica2part2dev=None):
"""
Build a dict of all tiers in the cluster to a list of the number of
parts with a replica count at each index. The values of the dict will
be lists of length the maximum whole replica + 1 so that the
graph[tier][3]... | IndexError | dataset/ETHPy150Open openstack/swift/swift/common/ring/builder.py/RingBuilder._build_dispersion_graph |
2,080 | @classmethod
def load(cls, builder_file, open=open):
"""
Obtain RingBuilder instance of the provided builder file
:param builder_file: path to builder file to load
:return: RingBuilder instance
"""
try:
fp = open(builder_file, 'rb')
except __HOLE_... | IOError | dataset/ETHPy150Open openstack/swift/swift/common/ring/builder.py/RingBuilder.load |
2,081 | def search_devs(self, search_values):
"""Search devices by parameters.
:param search_values: a dictionary with search values to filter
devices, supported parameters are id,
region, zone, ip, port, replication_ip,
... | ValueError | dataset/ETHPy150Open openstack/swift/swift/common/ring/builder.py/RingBuilder.search_devs |
2,082 | def datetime_u(s):
fmt = "%Y-%m-%dT%H:%M:%S"
try:
return _strptime(s, fmt)
except __HOLE__:
try:
# strip utc offset
if s[-3] == ":" and s[-6] in (' ', '-', '+'):
warnings.warn('removing unsupported UTC offset', RuntimeWarning)
s = s[:-6... | ValueError | dataset/ETHPy150Open uwdata/termite-data-server/web2py/gluon/contrib/pysimplesoap/helpers.py/datetime_u |
2,083 | def symmetric_difference(self, other):
"""Return the symmetric difference of two sets as a new set.
(I.e. all elements that are in exactly one of the sets.)
"""
result = self.__class__()
data = result._data
value = True
selfdata = self._data
try:
... | AttributeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/BaseSet.symmetric_difference |
2,084 | def difference(self, other):
"""Return the difference of two sets as a new Set.
(I.e. all elements that are in this set and not in the other.)
"""
result = self.__class__()
data = result._data
try:
otherdata = other._data
except __HOLE__:
... | AttributeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/BaseSet.difference |
2,085 | def __contains__(self, element):
"""Report whether an element is a member of a set.
(Called in response to the expression `element in self'.)
"""
try:
return element in self._data
except __HOLE__:
transform = getattr(element, "__as_temporarily_immutable__... | TypeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/BaseSet.__contains__ |
2,086 | def _update(self, iterable):
# The main loop for update() and the subclass __init__() methods.
data = self._data
# Use the fast update() method when a dictionary is available.
if isinstance(iterable, BaseSet):
data.update(iterable._data)
return
value = T... | TypeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/BaseSet._update |
2,087 | def add(self, element):
"""Add an element to a set.
This has no effect if the element is already present.
"""
try:
self._data[element] = True
except __HOLE__:
transform = getattr(element, "__as_immutable__", None)
if transform is None:
... | TypeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/Set.add |
2,088 | def remove(self, element):
"""Remove an element from a set; it must be a member.
If the element is not a member, raise a KeyError.
"""
try:
del self._data[element]
except __HOLE__:
transform = getattr(element, "__as_temporarily_immutable__", None)
... | TypeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/Set.remove |
2,089 | def discard(self, element):
"""Remove an element from a set if it is a member.
If the element is not a member, do nothing.
"""
try:
self.remove(element)
except __HOLE__:
pass | KeyError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/sets.py/Set.discard |
2,090 | def add_edge(self, u, v, key=None, attr_dict=None, **attr):
"""Add an edge between u and v.
The nodes u and v will be automatically added if they are
not already in the graph.
Edge attributes can be specified with keywords or by providing
a dictionary with key/value pairs. See... | AttributeError | dataset/ETHPy150Open gkno/gkno_launcher/src/networkx/classes/multidigraph.py/MultiDiGraph.add_edge |
2,091 | def get_available_name(self, name):
"""
In order to prevent file overwriting one another this will generate
a new filename with the format `YYMMDD.uniquehash.filename.extension`
"""
# Set the format of our filename
filename_format = '{path}/{date}.{filename}'
# I... | AttributeError | dataset/ETHPy150Open ofa/connect/open_connect/connect_core/utils/storages.py/AttachmentStorage.get_available_name |
2,092 | def _convert_colors(colors):
"""Convert either a list of colors or nested lists of colors to RGB."""
to_rgb = mpl.colors.colorConverter.to_rgb
if isinstance(colors, pd.DataFrame):
# Convert dataframe
return pd.DataFrame({col: colors[col].map(to_rgb)
for col in co... | ValueError | dataset/ETHPy150Open mwaskom/seaborn/seaborn/matrix.py/_convert_colors |
2,093 | def __init__(self, data, vmin, vmax, cmap, center, robust, annot, fmt,
annot_kws, cbar, cbar_kws,
xticklabels=True, yticklabels=True, mask=None):
"""Initialize the plotting object."""
# We always want to have a DataFrame with semantic information
# and an ndarra... | AttributeError | dataset/ETHPy150Open mwaskom/seaborn/seaborn/matrix.py/_HeatMapper.__init__ |
2,094 | @property
def calculated_linkage(self):
try:
return self._calculate_linkage_fastcluster()
except __HOLE__:
return self._calculate_linkage_scipy() | ImportError | dataset/ETHPy150Open mwaskom/seaborn/seaborn/matrix.py/_DendrogramPlotter.calculated_linkage |
2,095 | def plot_matrix(self, colorbar_kws, xind, yind, **kws):
self.data2d = self.data2d.iloc[yind, xind]
self.mask = self.mask.iloc[yind, xind]
# Try to reorganize specified tick labels, if provided
xtl = kws.pop("xticklabels", True)
try:
xtl = np.asarray(xtl)[xind]
... | IndexError | dataset/ETHPy150Open mwaskom/seaborn/seaborn/matrix.py/ClusterGrid.plot_matrix |
2,096 | def plot(self, metric, method, colorbar_kws, row_cluster, col_cluster,
row_linkage, col_linkage, **kws):
colorbar_kws = {} if colorbar_kws is None else colorbar_kws
self.plot_dendrograms(row_cluster, col_cluster, metric, method,
row_linkage=row_linkage, col_lin... | AttributeError | dataset/ETHPy150Open mwaskom/seaborn/seaborn/matrix.py/ClusterGrid.plot |
2,097 | def _to_node(self, element, groups=None):
try:
state = self.NODE_STATE_MAP[
self._findattr(element, "instanceState/name")
]
except __HOLE__:
state = NodeState.UNKNOWN
n = Node(
id=self._findtext(element, 'instanceId'),
... | KeyError | dataset/ETHPy150Open secondstory/dewpoint/libcloud/drivers/ec2.py/EC2NodeDriver._to_node |
2,098 | def _run_command(args, stdout=_PIPE, stderr=_PIPE, encoding=None, stream=0):
#regarding the shell argument, see: http://bugs.python.org/issue8557
try:
proc = _Popen(args, stdout=stdout, stderr=stderr,
shell=(sys.platform == 'win32'))
data = proc.communicate()[stream]
e... | OSError | dataset/ETHPy150Open GeekTrainer/Flask/Work/Trivia - Module 5/env/Lib/site-packages/setuptools/svn_utils.py/_run_command |
2,099 | @classmethod
def load(cls, dirname=''):
normdir = os.path.normpath(dirname)
code, data = _run_command(['svn', 'info', normdir])
# Must check for some contents, as some use empty directories
# in testcases
svn_dir = os.path.join(normdir, '.svn')
has_svn = (os.path.isfi... | ValueError | dataset/ETHPy150Open GeekTrainer/Flask/Work/Trivia - Module 5/env/Lib/site-packages/setuptools/svn_utils.py/SvnInfo.load |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.