repository_name stringlengths 7 55 | func_path_in_repository stringlengths 4 223 | func_name stringlengths 1 134 | whole_func_string stringlengths 75 104k | language stringclasses 1
value | func_code_string stringlengths 75 104k | func_code_tokens listlengths 19 28.4k | func_documentation_string stringlengths 1 46.9k | func_documentation_tokens listlengths 1 1.97k | split_name stringclasses 1
value | func_code_url stringlengths 87 315 |
|---|---|---|---|---|---|---|---|---|---|---|
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._handle_break | def _handle_break(self, node, scope, ctxt, stream):
"""Handle break node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling break")
raise errors.InterpBreak() | python | def _handle_break(self, node, scope, ctxt, stream):
"""Handle break node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling break")
raise errors.InterpBreak() | [
"def",
"_handle_break",
"(",
"self",
",",
"node",
",",
"scope",
",",
"ctxt",
",",
"stream",
")",
":",
"self",
".",
"_dlog",
"(",
"\"handling break\"",
")",
"raise",
"errors",
".",
"InterpBreak",
"(",
")"
] | Handle break node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO | [
"Handle",
"break",
"node"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2217-L2228 |
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._handle_continue | def _handle_continue(self, node, scope, ctxt, stream):
"""Handle continue node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling continue")
raise errors.InterpContinue() | python | def _handle_continue(self, node, scope, ctxt, stream):
"""Handle continue node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling continue")
raise errors.InterpContinue() | [
"def",
"_handle_continue",
"(",
"self",
",",
"node",
",",
"scope",
",",
"ctxt",
",",
"stream",
")",
":",
"self",
".",
"_dlog",
"(",
"\"handling continue\"",
")",
"raise",
"errors",
".",
"InterpContinue",
"(",
")"
] | Handle continue node
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO | [
"Handle",
"continue",
"node"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2230-L2241 |
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._handle_decl_list | def _handle_decl_list(self, node, scope, ctxt, stream):
"""Handle For nodes
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling decl list")
# just handle each declaration
for decl in node.decls:
... | python | def _handle_decl_list(self, node, scope, ctxt, stream):
"""Handle For nodes
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO
"""
self._dlog("handling decl list")
# just handle each declaration
for decl in node.decls:
... | [
"def",
"_handle_decl_list",
"(",
"self",
",",
"node",
",",
"scope",
",",
"ctxt",
",",
"stream",
")",
":",
"self",
".",
"_dlog",
"(",
"\"handling decl list\"",
")",
"# just handle each declaration",
"for",
"decl",
"in",
"node",
".",
"decls",
":",
"self",
".",... | Handle For nodes
:node: TODO
:scope: TODO
:ctxt: TODO
:stream: TODO
:returns: TODO | [
"Handle",
"For",
"nodes"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2243-L2256 |
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._create_scope | def _create_scope(self):
"""TODO: Docstring for _create_scope.
:returns: TODO
"""
res = Scope(self._log)
for func_name,native_func in six.iteritems(self._natives):
res.add_local(func_name, native_func)
return res | python | def _create_scope(self):
"""TODO: Docstring for _create_scope.
:returns: TODO
"""
res = Scope(self._log)
for func_name,native_func in six.iteritems(self._natives):
res.add_local(func_name, native_func)
return res | [
"def",
"_create_scope",
"(",
"self",
")",
":",
"res",
"=",
"Scope",
"(",
"self",
".",
"_log",
")",
"for",
"func_name",
",",
"native_func",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_natives",
")",
":",
"res",
".",
"add_local",
"(",
"func_name",... | TODO: Docstring for _create_scope.
:returns: TODO | [
"TODO",
":",
"Docstring",
"for",
"_create_scope",
".",
":",
"returns",
":",
"TODO"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2306-L2316 |
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._get_value | def _get_value(self, node, scope, ctxt, stream):
"""Return the value of the node. It is expected to be
either an AST.ID instance or a constant
:node: TODO
:returns: TODO
"""
res = self._handle_node(node, scope, ctxt, stream)
if isinstance(res, fields.Field):
... | python | def _get_value(self, node, scope, ctxt, stream):
"""Return the value of the node. It is expected to be
either an AST.ID instance or a constant
:node: TODO
:returns: TODO
"""
res = self._handle_node(node, scope, ctxt, stream)
if isinstance(res, fields.Field):
... | [
"def",
"_get_value",
"(",
"self",
",",
"node",
",",
"scope",
",",
"ctxt",
",",
"stream",
")",
":",
"res",
"=",
"self",
".",
"_handle_node",
"(",
"node",
",",
"scope",
",",
"ctxt",
",",
"stream",
")",
"if",
"isinstance",
"(",
"res",
",",
"fields",
"... | Return the value of the node. It is expected to be
either an AST.ID instance or a constant
:node: TODO
:returns: TODO | [
"Return",
"the",
"value",
"of",
"the",
"node",
".",
"It",
"is",
"expected",
"to",
"be",
"either",
"an",
"AST",
".",
"ID",
"instance",
"or",
"a",
"constant"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2318-L2334 |
d0c-s4vage/pfp | pfp/interp.py | PfpInterp._resolve_to_field_class | def _resolve_to_field_class(self, names, scope):
"""Resolve the names to a class in fields.py, resolving past
typedefs, etc
:names: TODO
:scope: TODO
:ctxt: TODO
:returns: TODO
"""
switch = {
"char" : "Char",
"int" : "Int",... | python | def _resolve_to_field_class(self, names, scope):
"""Resolve the names to a class in fields.py, resolving past
typedefs, etc
:names: TODO
:scope: TODO
:ctxt: TODO
:returns: TODO
"""
switch = {
"char" : "Char",
"int" : "Int",... | [
"def",
"_resolve_to_field_class",
"(",
"self",
",",
"names",
",",
"scope",
")",
":",
"switch",
"=",
"{",
"\"char\"",
":",
"\"Char\"",
",",
"\"int\"",
":",
"\"Int\"",
",",
"\"long\"",
":",
"\"Int\"",
",",
"\"int64\"",
":",
"\"Int64\"",
",",
"\"uint64\"",
":... | Resolve the names to a class in fields.py, resolving past
typedefs, etc
:names: TODO
:scope: TODO
:ctxt: TODO
:returns: TODO | [
"Resolve",
"the",
"names",
"to",
"a",
"class",
"in",
"fields",
".",
"py",
"resolving",
"past",
"typedefs",
"etc"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/interp.py#L2336-L2385 |
d0c-s4vage/pfp | pfp/bitwrap.py | bits_to_bytes | def bits_to_bytes(bits):
"""Convert the bit list into bytes. (Assumes bits is a list
whose length is a multiple of 8)
"""
if len(bits) % 8 != 0:
raise Exception("num bits must be multiple of 8")
res = ""
for x in six.moves.range(0, len(bits), 8):
byte_bits = bits[x:x+8]
... | python | def bits_to_bytes(bits):
"""Convert the bit list into bytes. (Assumes bits is a list
whose length is a multiple of 8)
"""
if len(bits) % 8 != 0:
raise Exception("num bits must be multiple of 8")
res = ""
for x in six.moves.range(0, len(bits), 8):
byte_bits = bits[x:x+8]
... | [
"def",
"bits_to_bytes",
"(",
"bits",
")",
":",
"if",
"len",
"(",
"bits",
")",
"%",
"8",
"!=",
"0",
":",
"raise",
"Exception",
"(",
"\"num bits must be multiple of 8\"",
")",
"res",
"=",
"\"\"",
"for",
"x",
"in",
"six",
".",
"moves",
".",
"range",
"(",
... | Convert the bit list into bytes. (Assumes bits is a list
whose length is a multiple of 8) | [
"Convert",
"the",
"bit",
"list",
"into",
"bytes",
".",
"(",
"Assumes",
"bits",
"is",
"a",
"list",
"whose",
"length",
"is",
"a",
"multiple",
"of",
"8",
")"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L15-L29 |
d0c-s4vage/pfp | pfp/bitwrap.py | bytes_to_bits | def bytes_to_bits(bytes_):
"""Convert bytes to a list of bits
"""
res = []
for x in bytes_:
if not isinstance(x, int):
x = ord(x)
res += byte_to_bits(x)
return res | python | def bytes_to_bits(bytes_):
"""Convert bytes to a list of bits
"""
res = []
for x in bytes_:
if not isinstance(x, int):
x = ord(x)
res += byte_to_bits(x)
return res | [
"def",
"bytes_to_bits",
"(",
"bytes_",
")",
":",
"res",
"=",
"[",
"]",
"for",
"x",
"in",
"bytes_",
":",
"if",
"not",
"isinstance",
"(",
"x",
",",
"int",
")",
":",
"x",
"=",
"ord",
"(",
"x",
")",
"res",
"+=",
"byte_to_bits",
"(",
"x",
")",
"retu... | Convert bytes to a list of bits | [
"Convert",
"bytes",
"to",
"a",
"list",
"of",
"bits"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L31-L39 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.is_eof | def is_eof(self):
"""Return if the stream has reached EOF or not
without discarding any unflushed bits
:returns: True/False
"""
pos = self._stream.tell()
byte = self._stream.read(1)
self._stream.seek(pos, 0)
return utils.binary(byte) == utils.binary("") | python | def is_eof(self):
"""Return if the stream has reached EOF or not
without discarding any unflushed bits
:returns: True/False
"""
pos = self._stream.tell()
byte = self._stream.read(1)
self._stream.seek(pos, 0)
return utils.binary(byte) == utils.binary("") | [
"def",
"is_eof",
"(",
"self",
")",
":",
"pos",
"=",
"self",
".",
"_stream",
".",
"tell",
"(",
")",
"byte",
"=",
"self",
".",
"_stream",
".",
"read",
"(",
"1",
")",
"self",
".",
"_stream",
".",
"seek",
"(",
"pos",
",",
"0",
")",
"return",
"utils... | Return if the stream has reached EOF or not
without discarding any unflushed bits
:returns: True/False | [
"Return",
"if",
"the",
"stream",
"has",
"reached",
"EOF",
"or",
"not",
"without",
"discarding",
"any",
"unflushed",
"bits"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L70-L80 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.close | def close(self):
"""Close the stream
"""
self.closed = True
self._flush_bits_to_stream()
self._stream.close() | python | def close(self):
"""Close the stream
"""
self.closed = True
self._flush_bits_to_stream()
self._stream.close() | [
"def",
"close",
"(",
"self",
")",
":",
"self",
".",
"closed",
"=",
"True",
"self",
".",
"_flush_bits_to_stream",
"(",
")",
"self",
".",
"_stream",
".",
"close",
"(",
")"
] | Close the stream | [
"Close",
"the",
"stream"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L82-L87 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.read | def read(self, num):
"""Read ``num`` number of bytes from the stream. Note that this will
automatically resets/ends the current bit-reading if it does not
end on an even byte AND ``self.padded`` is True. If ``self.padded`` is
True, then the entire stream is treated as a bitstream.
... | python | def read(self, num):
"""Read ``num`` number of bytes from the stream. Note that this will
automatically resets/ends the current bit-reading if it does not
end on an even byte AND ``self.padded`` is True. If ``self.padded`` is
True, then the entire stream is treated as a bitstream.
... | [
"def",
"read",
"(",
"self",
",",
"num",
")",
":",
"start_pos",
"=",
"self",
".",
"tell",
"(",
")",
"if",
"self",
".",
"padded",
":",
"# we toss out any uneven bytes",
"self",
".",
"_bits",
".",
"clear",
"(",
")",
"res",
"=",
"utils",
".",
"binary",
"... | Read ``num`` number of bytes from the stream. Note that this will
automatically resets/ends the current bit-reading if it does not
end on an even byte AND ``self.padded`` is True. If ``self.padded`` is
True, then the entire stream is treated as a bitstream.
:num: number of bytes to read... | [
"Read",
"num",
"number",
"of",
"bytes",
"from",
"the",
"stream",
".",
"Note",
"that",
"this",
"will",
"automatically",
"resets",
"/",
"ends",
"the",
"current",
"bit",
"-",
"reading",
"if",
"it",
"does",
"not",
"end",
"on",
"an",
"even",
"byte",
"AND",
... | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L100-L123 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.read_bits | def read_bits(self, num):
"""Read ``num`` number of bits from the stream
:num: number of bits to read
:returns: a list of ``num`` bits, or an empty list if EOF has been reached
"""
if num > len(self._bits):
needed = num - len(self._bits)
num_bytes = int(m... | python | def read_bits(self, num):
"""Read ``num`` number of bits from the stream
:num: number of bits to read
:returns: a list of ``num`` bits, or an empty list if EOF has been reached
"""
if num > len(self._bits):
needed = num - len(self._bits)
num_bytes = int(m... | [
"def",
"read_bits",
"(",
"self",
",",
"num",
")",
":",
"if",
"num",
">",
"len",
"(",
"self",
".",
"_bits",
")",
":",
"needed",
"=",
"num",
"-",
"len",
"(",
"self",
".",
"_bits",
")",
"num_bytes",
"=",
"int",
"(",
"math",
".",
"ceil",
"(",
"need... | Read ``num`` number of bits from the stream
:num: number of bits to read
:returns: a list of ``num`` bits, or an empty list if EOF has been reached | [
"Read",
"num",
"number",
"of",
"bits",
"from",
"the",
"stream"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L125-L143 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.write | def write(self, data):
"""Write data to the stream
:data: the data to write to the stream
:returns: None
"""
if self.padded:
# flush out any remaining bits first
if len(self._bits) > 0:
self._flush_bits_to_stream()
self._stream... | python | def write(self, data):
"""Write data to the stream
:data: the data to write to the stream
:returns: None
"""
if self.padded:
# flush out any remaining bits first
if len(self._bits) > 0:
self._flush_bits_to_stream()
self._stream... | [
"def",
"write",
"(",
"self",
",",
"data",
")",
":",
"if",
"self",
".",
"padded",
":",
"# flush out any remaining bits first",
"if",
"len",
"(",
"self",
".",
"_bits",
")",
">",
"0",
":",
"self",
".",
"_flush_bits_to_stream",
"(",
")",
"self",
".",
"_strea... | Write data to the stream
:data: the data to write to the stream
:returns: None | [
"Write",
"data",
"to",
"the",
"stream"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L145-L162 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.write_bits | def write_bits(self, bits):
"""Write the bits to the stream.
Add the bits to the existing unflushed bits and write
complete bytes to the stream.
"""
for bit in bits:
self._bits.append(bit)
while len(self._bits) >= 8:
byte_bits = [self._bits.pople... | python | def write_bits(self, bits):
"""Write the bits to the stream.
Add the bits to the existing unflushed bits and write
complete bytes to the stream.
"""
for bit in bits:
self._bits.append(bit)
while len(self._bits) >= 8:
byte_bits = [self._bits.pople... | [
"def",
"write_bits",
"(",
"self",
",",
"bits",
")",
":",
"for",
"bit",
"in",
"bits",
":",
"self",
".",
"_bits",
".",
"append",
"(",
"bit",
")",
"while",
"len",
"(",
"self",
".",
"_bits",
")",
">=",
"8",
":",
"byte_bits",
"=",
"[",
"self",
".",
... | Write the bits to the stream.
Add the bits to the existing unflushed bits and write
complete bytes to the stream. | [
"Write",
"the",
"bits",
"to",
"the",
"stream",
"."
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L164-L176 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.tell | def tell(self):
"""Return the current position in the stream (ignoring bit
position)
:returns: int for the position in the stream
"""
res = self._stream.tell()
if len(self._bits) > 0:
res -= 1
return res | python | def tell(self):
"""Return the current position in the stream (ignoring bit
position)
:returns: int for the position in the stream
"""
res = self._stream.tell()
if len(self._bits) > 0:
res -= 1
return res | [
"def",
"tell",
"(",
"self",
")",
":",
"res",
"=",
"self",
".",
"_stream",
".",
"tell",
"(",
")",
"if",
"len",
"(",
"self",
".",
"_bits",
")",
">",
"0",
":",
"res",
"-=",
"1",
"return",
"res"
] | Return the current position in the stream (ignoring bit
position)
:returns: int for the position in the stream | [
"Return",
"the",
"current",
"position",
"in",
"the",
"stream",
"(",
"ignoring",
"bit",
"position",
")"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L180-L189 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.seek | def seek(self, pos, seek_type=0):
"""Seek to the specified position in the stream with seek_type.
Unflushed bits will be discarded in the case of a seek.
The stream will also keep track of which bytes have and have
not been consumed so that the dom will capture all of the
bytes ... | python | def seek(self, pos, seek_type=0):
"""Seek to the specified position in the stream with seek_type.
Unflushed bits will be discarded in the case of a seek.
The stream will also keep track of which bytes have and have
not been consumed so that the dom will capture all of the
bytes ... | [
"def",
"seek",
"(",
"self",
",",
"pos",
",",
"seek_type",
"=",
"0",
")",
":",
"self",
".",
"_bits",
".",
"clear",
"(",
")",
"return",
"self",
".",
"_stream",
".",
"seek",
"(",
"pos",
",",
"seek_type",
")"
] | Seek to the specified position in the stream with seek_type.
Unflushed bits will be discarded in the case of a seek.
The stream will also keep track of which bytes have and have
not been consumed so that the dom will capture all of the
bytes in the stream.
:pos: offset
... | [
"Seek",
"to",
"the",
"specified",
"position",
"in",
"the",
"stream",
"with",
"seek_type",
".",
"Unflushed",
"bits",
"will",
"be",
"discarded",
"in",
"the",
"case",
"of",
"a",
"seek",
"."
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L191-L205 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.size | def size(self):
"""Return the size of the stream, or -1 if it cannot
be determined.
"""
pos = self._stream.tell()
# seek to the end of the stream
self._stream.seek(0,2)
size = self._stream.tell()
self._stream.seek(pos, 0)
return size | python | def size(self):
"""Return the size of the stream, or -1 if it cannot
be determined.
"""
pos = self._stream.tell()
# seek to the end of the stream
self._stream.seek(0,2)
size = self._stream.tell()
self._stream.seek(pos, 0)
return size | [
"def",
"size",
"(",
"self",
")",
":",
"pos",
"=",
"self",
".",
"_stream",
".",
"tell",
"(",
")",
"# seek to the end of the stream",
"self",
".",
"_stream",
".",
"seek",
"(",
"0",
",",
"2",
")",
"size",
"=",
"self",
".",
"_stream",
".",
"tell",
"(",
... | Return the size of the stream, or -1 if it cannot
be determined. | [
"Return",
"the",
"size",
"of",
"the",
"stream",
"or",
"-",
"1",
"if",
"it",
"cannot",
"be",
"determined",
"."
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L207-L217 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream.unconsumed_ranges | def unconsumed_ranges(self):
"""Return an IntervalTree of unconsumed ranges, of the format
(start, end] with the end value not being included
"""
res = IntervalTree()
prev = None
# normal iteration is not in a predictable order
ranges = sorted([x for x in self.r... | python | def unconsumed_ranges(self):
"""Return an IntervalTree of unconsumed ranges, of the format
(start, end] with the end value not being included
"""
res = IntervalTree()
prev = None
# normal iteration is not in a predictable order
ranges = sorted([x for x in self.r... | [
"def",
"unconsumed_ranges",
"(",
"self",
")",
":",
"res",
"=",
"IntervalTree",
"(",
")",
"prev",
"=",
"None",
"# normal iteration is not in a predictable order",
"ranges",
"=",
"sorted",
"(",
"[",
"x",
"for",
"x",
"in",
"self",
".",
"range_set",
"]",
",",
"k... | Return an IntervalTree of unconsumed ranges, of the format
(start, end] with the end value not being included | [
"Return",
"an",
"IntervalTree",
"of",
"unconsumed",
"ranges",
"of",
"the",
"format",
"(",
"start",
"end",
"]",
"with",
"the",
"end",
"value",
"not",
"being",
"included"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L219-L241 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream._update_consumed_ranges | def _update_consumed_ranges(self, start_pos, end_pos):
"""Update the ``self.consumed_ranges`` array with which
byte ranges have been consecutively consumed.
"""
self.range_set.add(Interval(start_pos, end_pos+1))
self.range_set.merge_overlaps() | python | def _update_consumed_ranges(self, start_pos, end_pos):
"""Update the ``self.consumed_ranges`` array with which
byte ranges have been consecutively consumed.
"""
self.range_set.add(Interval(start_pos, end_pos+1))
self.range_set.merge_overlaps() | [
"def",
"_update_consumed_ranges",
"(",
"self",
",",
"start_pos",
",",
"end_pos",
")",
":",
"self",
".",
"range_set",
".",
"add",
"(",
"Interval",
"(",
"start_pos",
",",
"end_pos",
"+",
"1",
")",
")",
"self",
".",
"range_set",
".",
"merge_overlaps",
"(",
... | Update the ``self.consumed_ranges`` array with which
byte ranges have been consecutively consumed. | [
"Update",
"the",
"self",
".",
"consumed_ranges",
"array",
"with",
"which",
"byte",
"ranges",
"have",
"been",
"consecutively",
"consumed",
"."
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L247-L252 |
d0c-s4vage/pfp | pfp/bitwrap.py | BitwrappedStream._flush_bits_to_stream | def _flush_bits_to_stream(self):
"""Flush the bits to the stream. This is used when
a few bits have been read and ``self._bits`` contains unconsumed/
flushed bits when data is to be written to the stream
"""
if len(self._bits) == 0:
return 0
bits = list(self.... | python | def _flush_bits_to_stream(self):
"""Flush the bits to the stream. This is used when
a few bits have been read and ``self._bits`` contains unconsumed/
flushed bits when data is to be written to the stream
"""
if len(self._bits) == 0:
return 0
bits = list(self.... | [
"def",
"_flush_bits_to_stream",
"(",
"self",
")",
":",
"if",
"len",
"(",
"self",
".",
"_bits",
")",
"==",
"0",
":",
"return",
"0",
"bits",
"=",
"list",
"(",
"self",
".",
"_bits",
")",
"diff",
"=",
"8",
"-",
"(",
"len",
"(",
"bits",
")",
"%",
"8... | Flush the bits to the stream. This is used when
a few bits have been read and ``self._bits`` contains unconsumed/
flushed bits when data is to be written to the stream | [
"Flush",
"the",
"bits",
"to",
"the",
"stream",
".",
"This",
"is",
"used",
"when",
"a",
"few",
"bits",
"have",
"been",
"read",
"and",
"self",
".",
"_bits",
"contains",
"unconsumed",
"/",
"flushed",
"bits",
"when",
"data",
"is",
"to",
"be",
"written",
"t... | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/bitwrap.py#L254-L271 |
expfactory/expfactory | expfactory/validator/library.py | LibraryValidator._validate_markdown | def _validate_markdown(self, expfile):
'''ensure that fields are present in markdown file'''
try:
import yaml
except:
bot.error('Python yaml is required for testing yml/markdown files.')
sys.exit(1)
self.metadata = {}
uid = os.path.basename(e... | python | def _validate_markdown(self, expfile):
'''ensure that fields are present in markdown file'''
try:
import yaml
except:
bot.error('Python yaml is required for testing yml/markdown files.')
sys.exit(1)
self.metadata = {}
uid = os.path.basename(e... | [
"def",
"_validate_markdown",
"(",
"self",
",",
"expfile",
")",
":",
"try",
":",
"import",
"yaml",
"except",
":",
"bot",
".",
"error",
"(",
"'Python yaml is required for testing yml/markdown files.'",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"self",
".",
"metada... | ensure that fields are present in markdown file | [
"ensure",
"that",
"fields",
"are",
"present",
"in",
"markdown",
"file"
] | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/validator/library.py#L62-L101 |
expfactory/expfactory | expfactory/views/utils.py | perform_checks | def perform_checks(template,
do_redirect=False,
context=None,
next=None,
quiet=False):
'''return all checks for required variables before returning to
desired view
Parameters
==========
template: the html temp... | python | def perform_checks(template,
do_redirect=False,
context=None,
next=None,
quiet=False):
'''return all checks for required variables before returning to
desired view
Parameters
==========
template: the html temp... | [
"def",
"perform_checks",
"(",
"template",
",",
"do_redirect",
"=",
"False",
",",
"context",
"=",
"None",
",",
"next",
"=",
"None",
",",
"quiet",
"=",
"False",
")",
":",
"from",
"expfactory",
".",
"server",
"import",
"app",
"username",
"=",
"session",
"."... | return all checks for required variables before returning to
desired view
Parameters
==========
template: the html template to render
do_redirect: if True, perform a redirect and not render
context: dictionary of context variables to pass to render_template
next: a pre... | [
"return",
"all",
"checks",
"for",
"required",
"variables",
"before",
"returning",
"to",
"desired",
"view"
] | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/views/utils.py#L45-L105 |
d0c-s4vage/pfp | pfp/native/compat_io.py | FSeek | def FSeek(params, ctxt, scope, stream, coord):
"""Returns 0 if successful or -1 if the address is out of range
"""
if len(params) != 1:
raise errors.InvalidArguments(coord, "{} args".format(len(params)), "FSeek accepts only one argument")
pos = PYVAL(params[0])
curr_pos = stream.tell()
... | python | def FSeek(params, ctxt, scope, stream, coord):
"""Returns 0 if successful or -1 if the address is out of range
"""
if len(params) != 1:
raise errors.InvalidArguments(coord, "{} args".format(len(params)), "FSeek accepts only one argument")
pos = PYVAL(params[0])
curr_pos = stream.tell()
... | [
"def",
"FSeek",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"!=",
"1",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
"len",... | Returns 0 if successful or -1 if the address is out of range | [
"Returns",
"0",
"if",
"successful",
"or",
"-",
"1",
"if",
"the",
"address",
"is",
"out",
"of",
"range"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/compat_io.py#L117-L158 |
d0c-s4vage/pfp | pfp/native/compat_io.py | FSkip | def FSkip(params, ctxt, scope, stream, coord):
"""Returns 0 if successful or -1 if the address is out of range
"""
if len(params) != 1:
raise errors.InvalidArguments(coord, "{} args".format(len(params)), "FSkip accepts only one argument")
skip_amt = PYVAL(params[0])
pos = skip_amt + stream.... | python | def FSkip(params, ctxt, scope, stream, coord):
"""Returns 0 if successful or -1 if the address is out of range
"""
if len(params) != 1:
raise errors.InvalidArguments(coord, "{} args".format(len(params)), "FSkip accepts only one argument")
skip_amt = PYVAL(params[0])
pos = skip_amt + stream.... | [
"def",
"FSkip",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"!=",
"1",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
"len",... | Returns 0 if successful or -1 if the address is out of range | [
"Returns",
"0",
"if",
"successful",
"or",
"-",
"1",
"if",
"the",
"address",
"is",
"out",
"of",
"range"
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/compat_io.py#L162-L170 |
d0c-s4vage/pfp | pfp/native/packers.py | packer_gzip | def packer_gzip(params, ctxt, scope, stream, coord):
"""``PackerGZip`` - implements both unpacking and packing. Can be used
as the ``packer`` for a field. When packing, concats the build output
of all params and gzip-compresses the result. When unpacking, concats
the build output of all params and gzip-... | python | def packer_gzip(params, ctxt, scope, stream, coord):
"""``PackerGZip`` - implements both unpacking and packing. Can be used
as the ``packer`` for a field. When packing, concats the build output
of all params and gzip-compresses the result. When unpacking, concats
the build output of all params and gzip-... | [
"def",
"packer_gzip",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"<=",
"1",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
... | ``PackerGZip`` - implements both unpacking and packing. Can be used
as the ``packer`` for a field. When packing, concats the build output
of all params and gzip-compresses the result. When unpacking, concats
the build output of all params and gzip-decompresses the result.
Example:
The code... | [
"PackerGZip",
"-",
"implements",
"both",
"unpacking",
"and",
"packing",
".",
"Can",
"be",
"used",
"as",
"the",
"packer",
"for",
"a",
"field",
".",
"When",
"packing",
"concats",
"the",
"build",
"output",
"of",
"all",
"params",
"and",
"gzip",
"-",
"compresse... | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/packers.py#L14-L40 |
d0c-s4vage/pfp | pfp/native/packers.py | pack_gzip | def pack_gzip(params, ctxt, scope, stream, coord):
"""``PackGZip`` - Concats the build output of all params and gzips the
resulting data, returning a char array.
Example: ::
char data[0x100]<pack=PackGZip, ...>;
"""
if len(params) == 0:
raise errors.InvalidArguments(coord, "{} args... | python | def pack_gzip(params, ctxt, scope, stream, coord):
"""``PackGZip`` - Concats the build output of all params and gzips the
resulting data, returning a char array.
Example: ::
char data[0x100]<pack=PackGZip, ...>;
"""
if len(params) == 0:
raise errors.InvalidArguments(coord, "{} args... | [
"def",
"pack_gzip",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"==",
"0",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
"l... | ``PackGZip`` - Concats the build output of all params and gzips the
resulting data, returning a char array.
Example: ::
char data[0x100]<pack=PackGZip, ...>; | [
"PackGZip",
"-",
"Concats",
"the",
"build",
"output",
"of",
"all",
"params",
"and",
"gzips",
"the",
"resulting",
"data",
"returning",
"a",
"char",
"array",
"."
] | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/packers.py#L43-L61 |
d0c-s4vage/pfp | pfp/native/watchers.py | watch_length | def watch_length(params, ctxt, scope, stream, coord):
"""WatchLength - Watch the total length of each of the params.
Example:
The code below uses the ``WatchLength`` update function to update
the ``length`` field to the length of the ``data`` field ::
int length<watch=data, upd... | python | def watch_length(params, ctxt, scope, stream, coord):
"""WatchLength - Watch the total length of each of the params.
Example:
The code below uses the ``WatchLength`` update function to update
the ``length`` field to the length of the ``data`` field ::
int length<watch=data, upd... | [
"def",
"watch_length",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"<=",
"1",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
... | WatchLength - Watch the total length of each of the params.
Example:
The code below uses the ``WatchLength`` update function to update
the ``length`` field to the length of the ``data`` field ::
int length<watch=data, update=WatchLength>;
char data[length]; | [
"WatchLength",
"-",
"Watch",
"the",
"total",
"length",
"of",
"each",
"of",
"the",
"params",
".",
"Example",
":",
"The",
"code",
"below",
"uses",
"the",
"WatchLength",
"update",
"function",
"to",
"update",
"the",
"length",
"field",
"to",
"the",
"length",
"o... | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/watchers.py#L15-L34 |
d0c-s4vage/pfp | pfp/native/watchers.py | watch_crc | def watch_crc(params, ctxt, scope, stream, coord):
"""WatchCrc32 - Watch the total crc32 of the params.
Example:
The code below uses the ``WatchCrc32`` update function to update
the ``crc`` field to the crc of the ``length`` and ``data`` fields ::
char length;
char ... | python | def watch_crc(params, ctxt, scope, stream, coord):
"""WatchCrc32 - Watch the total crc32 of the params.
Example:
The code below uses the ``WatchCrc32`` update function to update
the ``crc`` field to the crc of the ``length`` and ``data`` fields ::
char length;
char ... | [
"def",
"watch_crc",
"(",
"params",
",",
"ctxt",
",",
"scope",
",",
"stream",
",",
"coord",
")",
":",
"if",
"len",
"(",
"params",
")",
"<=",
"1",
":",
"raise",
"errors",
".",
"InvalidArguments",
"(",
"coord",
",",
"\"{} args\"",
".",
"format",
"(",
"l... | WatchCrc32 - Watch the total crc32 of the params.
Example:
The code below uses the ``WatchCrc32`` update function to update
the ``crc`` field to the crc of the ``length`` and ``data`` fields ::
char length;
char data[length];
int crc<watch=length;data, updat... | [
"WatchCrc32",
"-",
"Watch",
"the",
"total",
"crc32",
"of",
"the",
"params",
".",
"Example",
":",
"The",
"code",
"below",
"uses",
"the",
"WatchCrc32",
"update",
"function",
"to",
"update",
"the",
"crc",
"field",
"to",
"the",
"crc",
"of",
"the",
"length",
... | train | https://github.com/d0c-s4vage/pfp/blob/32f2d34fdec1c70019fa83c7006d5e3be0f92fcd/pfp/native/watchers.py#L37-L57 |
expfactory/expfactory | expfactory/validator/experiments.py | ExperimentValidator._validate_folder | def _validate_folder(self, folder=None):
''' validate folder takes a cloned github repo, ensures
the existence of the config.json, and validates it.
'''
from expfactory.experiment import load_experiment
if folder is None:
folder=os.path.abspath(os.getcwd())
... | python | def _validate_folder(self, folder=None):
''' validate folder takes a cloned github repo, ensures
the existence of the config.json, and validates it.
'''
from expfactory.experiment import load_experiment
if folder is None:
folder=os.path.abspath(os.getcwd())
... | [
"def",
"_validate_folder",
"(",
"self",
",",
"folder",
"=",
"None",
")",
":",
"from",
"expfactory",
".",
"experiment",
"import",
"load_experiment",
"if",
"folder",
"is",
"None",
":",
"folder",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"getc... | validate folder takes a cloned github repo, ensures
the existence of the config.json, and validates it. | [
"validate",
"folder",
"takes",
"a",
"cloned",
"github",
"repo",
"ensures",
"the",
"existence",
"of",
"the",
"config",
".",
"json",
"and",
"validates",
"it",
"."
] | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/validator/experiments.py#L52-L66 |
expfactory/expfactory | expfactory/validator/experiments.py | ExperimentValidator.validate | def validate(self, folder, cleanup=False, validate_folder=True):
''' validate is the entrypoint to all validation, for
a folder, config, or url. If a URL is found, it is
cloned and cleaned up.
:param validate_folder: ensures the folder name (github repo)
... | python | def validate(self, folder, cleanup=False, validate_folder=True):
''' validate is the entrypoint to all validation, for
a folder, config, or url. If a URL is found, it is
cloned and cleaned up.
:param validate_folder: ensures the folder name (github repo)
... | [
"def",
"validate",
"(",
"self",
",",
"folder",
",",
"cleanup",
"=",
"False",
",",
"validate_folder",
"=",
"True",
")",
":",
"# Obtain any repository URL provided",
"if",
"folder",
".",
"startswith",
"(",
"'http'",
")",
"or",
"'github'",
"in",
"folder",
":",
... | validate is the entrypoint to all validation, for
a folder, config, or url. If a URL is found, it is
cloned and cleaned up.
:param validate_folder: ensures the folder name (github repo)
matches. | [
"validate",
"is",
"the",
"entrypoint",
"to",
"all",
"validation",
"for",
"a",
"folder",
"config",
"or",
"url",
".",
"If",
"a",
"URL",
"is",
"found",
"it",
"is",
"cloned",
"and",
"cleaned",
"up",
".",
":",
"param",
"validate_folder",
":",
"ensures",
"the"... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/validator/experiments.py#L69-L90 |
expfactory/expfactory | expfactory/validator/experiments.py | ExperimentValidator._validate_config | def _validate_config(self, folder, validate_folder=True):
''' validate config is the primary validation function that checks
for presence and format of required fields.
Parameters
==========
:folder: full path to folder with config.json
:name: if provided, the folder... | python | def _validate_config(self, folder, validate_folder=True):
''' validate config is the primary validation function that checks
for presence and format of required fields.
Parameters
==========
:folder: full path to folder with config.json
:name: if provided, the folder... | [
"def",
"_validate_config",
"(",
"self",
",",
"folder",
",",
"validate_folder",
"=",
"True",
")",
":",
"config",
"=",
"\"%s/config.json\"",
"%",
"folder",
"name",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"folder",
")",
"if",
"not",
"os",
".",
"path",... | validate config is the primary validation function that checks
for presence and format of required fields.
Parameters
==========
:folder: full path to folder with config.json
:name: if provided, the folder name to check against exp_id | [
"validate",
"config",
"is",
"the",
"primary",
"validation",
"function",
"that",
"checks",
"for",
"presence",
"and",
"format",
"of",
"required",
"fields",
"."
] | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/validator/experiments.py#L93-L146 |
expfactory/expfactory | expfactory/validator/experiments.py | ExperimentValidator.get_validation_fields | def get_validation_fields(self):
'''get_validation_fields returns a list of tuples (each a field)
we only require the exp_id to coincide with the folder name, for the sake
of reproducibility (given that all are served from sample image or Github
organization). All other fields a... | python | def get_validation_fields(self):
'''get_validation_fields returns a list of tuples (each a field)
we only require the exp_id to coincide with the folder name, for the sake
of reproducibility (given that all are served from sample image or Github
organization). All other fields a... | [
"def",
"get_validation_fields",
"(",
"self",
")",
":",
"return",
"[",
"(",
"\"name\"",
",",
"1",
",",
"str",
")",
",",
"# required",
"(",
"\"time\"",
",",
"1",
",",
"int",
")",
",",
"(",
"\"url\"",
",",
"1",
",",
"str",
")",
",",
"(",
"\"descriptio... | get_validation_fields returns a list of tuples (each a field)
we only require the exp_id to coincide with the folder name, for the sake
of reproducibility (given that all are served from sample image or Github
organization). All other fields are optional.
To specify runtime v... | [
"get_validation_fields",
"returns",
"a",
"list",
"of",
"tuples",
"(",
"each",
"a",
"field",
")",
"we",
"only",
"require",
"the",
"exp_id",
"to",
"coincide",
"with",
"the",
"folder",
"name",
"for",
"the",
"sake",
"of",
"reproducibility",
"(",
"given",
"that",... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/validator/experiments.py#L149-L172 |
expfactory/expfactory | expfactory/variables.py | get_runtime_vars | def get_runtime_vars(varset, experiment, token):
'''get_runtime_vars will return the urlparsed string of one or more runtime
variables. If None are present, None is returned.
Parameters
==========
varset: the variable set, a dictionary lookup with exp_id, token, vars
experiment... | python | def get_runtime_vars(varset, experiment, token):
'''get_runtime_vars will return the urlparsed string of one or more runtime
variables. If None are present, None is returned.
Parameters
==========
varset: the variable set, a dictionary lookup with exp_id, token, vars
experiment... | [
"def",
"get_runtime_vars",
"(",
"varset",
",",
"experiment",
",",
"token",
")",
":",
"url",
"=",
"''",
"if",
"experiment",
"in",
"varset",
":",
"variables",
"=",
"dict",
"(",
")",
"# Participant set variables",
"if",
"token",
"in",
"varset",
"[",
"experiment... | get_runtime_vars will return the urlparsed string of one or more runtime
variables. If None are present, None is returned.
Parameters
==========
varset: the variable set, a dictionary lookup with exp_id, token, vars
experiment: the exp_id to look up
token: the participant id... | [
"get_runtime_vars",
"will",
"return",
"the",
"urlparsed",
"string",
"of",
"one",
"or",
"more",
"runtime",
"variables",
".",
"If",
"None",
"are",
"present",
"None",
"is",
"returned",
".",
"Parameters",
"==========",
"varset",
":",
"the",
"variable",
"set",
"a",... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L45-L85 |
expfactory/expfactory | expfactory/variables.py | generate_runtime_vars | def generate_runtime_vars(variable_file=None, sep=','):
'''generate a lookup data structure from a
delimited file. We typically obtain the file name and delimiter from
the environment by way of EXPFACTORY_RUNTIME_VARS, and
EXPFACTORY_RUNTIME_DELIM, respectively, but the user can also parse
... | python | def generate_runtime_vars(variable_file=None, sep=','):
'''generate a lookup data structure from a
delimited file. We typically obtain the file name and delimiter from
the environment by way of EXPFACTORY_RUNTIME_VARS, and
EXPFACTORY_RUNTIME_DELIM, respectively, but the user can also parse
... | [
"def",
"generate_runtime_vars",
"(",
"variable_file",
"=",
"None",
",",
"sep",
"=",
"','",
")",
":",
"# First preference goes to runtime, then environment, then unset",
"if",
"variable_file",
"is",
"None",
":",
"if",
"EXPFACTORY_RUNTIME_VARS",
"is",
"not",
"None",
":",
... | generate a lookup data structure from a
delimited file. We typically obtain the file name and delimiter from
the environment by way of EXPFACTORY_RUNTIME_VARS, and
EXPFACTORY_RUNTIME_DELIM, respectively, but the user can also parse
from a custom variable file by way of specifying it to the ... | [
"generate",
"a",
"lookup",
"data",
"structure",
"from",
"a",
"delimited",
"file",
".",
"We",
"typically",
"obtain",
"the",
"file",
"name",
"and",
"delimiter",
"from",
"the",
"environment",
"by",
"way",
"of",
"EXPFACTORY_RUNTIME_VARS",
"and",
"EXPFACTORY_RUNTIME_DE... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L88-L180 |
expfactory/expfactory | expfactory/variables.py | _read_runtime_vars | def _read_runtime_vars(variable_file, sep=','):
'''read the entire runtime variable file, and return a list of lists,
each corresponding to a row. We also check the header, and exit
if anything is missing or malformed.
Parameters
==========
variable_file: full path to the tabula... | python | def _read_runtime_vars(variable_file, sep=','):
'''read the entire runtime variable file, and return a list of lists,
each corresponding to a row. We also check the header, and exit
if anything is missing or malformed.
Parameters
==========
variable_file: full path to the tabula... | [
"def",
"_read_runtime_vars",
"(",
"variable_file",
",",
"sep",
"=",
"','",
")",
":",
"rows",
"=",
"[",
"x",
"for",
"x",
"in",
"read_file",
"(",
"variable_file",
")",
".",
"split",
"(",
"'\\n'",
")",
"if",
"x",
".",
"strip",
"(",
")",
"]",
"valid_rows... | read the entire runtime variable file, and return a list of lists,
each corresponding to a row. We also check the header, and exit
if anything is missing or malformed.
Parameters
==========
variable_file: full path to the tabular file with token, exp_id, etc.
sep: the default... | [
"read",
"the",
"entire",
"runtime",
"variable",
"file",
"and",
"return",
"a",
"list",
"of",
"lists",
"each",
"corresponding",
"to",
"a",
"row",
".",
"We",
"also",
"check",
"the",
"header",
"and",
"exit",
"if",
"anything",
"is",
"missing",
"or",
"malformed"... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L183-L224 |
expfactory/expfactory | expfactory/variables.py | _validate_row | def _validate_row(row, sep=',', required_length=None):
'''validate_row will ensure that a row has the proper length, and is
not empty and cleaned of extra spaces.
Parameters
==========
row: a single row, not yet parsed.
Returns a valid row, or None if not valid
'''
if ... | python | def _validate_row(row, sep=',', required_length=None):
'''validate_row will ensure that a row has the proper length, and is
not empty and cleaned of extra spaces.
Parameters
==========
row: a single row, not yet parsed.
Returns a valid row, or None if not valid
'''
if ... | [
"def",
"_validate_row",
"(",
"row",
",",
"sep",
"=",
"','",
",",
"required_length",
"=",
"None",
")",
":",
"if",
"not",
"isinstance",
"(",
"row",
",",
"list",
")",
":",
"row",
"=",
"_parse_row",
"(",
"row",
",",
"sep",
")",
"if",
"required_length",
"... | validate_row will ensure that a row has the proper length, and is
not empty and cleaned of extra spaces.
Parameters
==========
row: a single row, not yet parsed.
Returns a valid row, or None if not valid | [
"validate_row",
"will",
"ensure",
"that",
"a",
"row",
"has",
"the",
"proper",
"length",
"and",
"is",
"not",
"empty",
"and",
"cleaned",
"of",
"extra",
"spaces",
".",
"Parameters",
"==========",
"row",
":",
"a",
"single",
"row",
"not",
"yet",
"parsed",
"."
] | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L227-L249 |
expfactory/expfactory | expfactory/variables.py | _parse_row | def _parse_row(row, sep=','):
'''parse row is a helper function to simply clean up a string, and parse
into a row based on a delimiter. If a required length is provided,
we check for this too.
'''
parsed = row.split(sep)
parsed = [x for x in parsed if x.strip()]
return parsed | python | def _parse_row(row, sep=','):
'''parse row is a helper function to simply clean up a string, and parse
into a row based on a delimiter. If a required length is provided,
we check for this too.
'''
parsed = row.split(sep)
parsed = [x for x in parsed if x.strip()]
return parsed | [
"def",
"_parse_row",
"(",
"row",
",",
"sep",
"=",
"','",
")",
":",
"parsed",
"=",
"row",
".",
"split",
"(",
"sep",
")",
"parsed",
"=",
"[",
"x",
"for",
"x",
"in",
"parsed",
"if",
"x",
".",
"strip",
"(",
")",
"]",
"return",
"parsed"
] | parse row is a helper function to simply clean up a string, and parse
into a row based on a delimiter. If a required length is provided,
we check for this too. | [
"parse",
"row",
"is",
"a",
"helper",
"function",
"to",
"simply",
"clean",
"up",
"a",
"string",
"and",
"parse",
"into",
"a",
"row",
"based",
"on",
"a",
"delimiter",
".",
"If",
"a",
"required",
"length",
"is",
"provided",
"we",
"check",
"for",
"this",
"t... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L252-L260 |
expfactory/expfactory | expfactory/variables.py | validate_header | def validate_header(header, required_fields=None):
'''validate_header ensures that the first row contains the exp_id,
var_name, var_value, and token. Capitalization isn't important, but
ordering is. This criteria is very strict, but it's reasonable
to require.
Parameters
=======... | python | def validate_header(header, required_fields=None):
'''validate_header ensures that the first row contains the exp_id,
var_name, var_value, and token. Capitalization isn't important, but
ordering is. This criteria is very strict, but it's reasonable
to require.
Parameters
=======... | [
"def",
"validate_header",
"(",
"header",
",",
"required_fields",
"=",
"None",
")",
":",
"if",
"required_fields",
"is",
"None",
":",
"required_fields",
"=",
"[",
"'exp_id'",
",",
"'var_name'",
",",
"'var_value'",
",",
"'token'",
"]",
"# The required length of the h... | validate_header ensures that the first row contains the exp_id,
var_name, var_value, and token. Capitalization isn't important, but
ordering is. This criteria is very strict, but it's reasonable
to require.
Parameters
==========
header: the header row, as a list
requir... | [
"validate_header",
"ensures",
"that",
"the",
"first",
"row",
"contains",
"the",
"exp_id",
"var_name",
"var_value",
"and",
"token",
".",
"Capitalization",
"isn",
"t",
"important",
"but",
"ordering",
"is",
".",
"This",
"criteria",
"is",
"very",
"strict",
"but",
... | train | https://github.com/expfactory/expfactory/blob/27ce6cc93e17231df8a8024f18e631336afd3501/expfactory/variables.py#L263-L294 |
littlepea/django-docs | docs/views.py | superuser_required | def superuser_required(view_func):
"""
Decorator for views that checks that the user is logged in and is a staff
member, displaying the login page if necessary.
"""
@wraps(view_func)
def _checklogin(request, *args, **kwargs):
if request.user.is_active and request.user.is_superuser:
... | python | def superuser_required(view_func):
"""
Decorator for views that checks that the user is logged in and is a staff
member, displaying the login page if necessary.
"""
@wraps(view_func)
def _checklogin(request, *args, **kwargs):
if request.user.is_active and request.user.is_superuser:
... | [
"def",
"superuser_required",
"(",
"view_func",
")",
":",
"@",
"wraps",
"(",
"view_func",
")",
"def",
"_checklogin",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"request",
".",
"user",
".",
"is_active",
"and",
"request",
"... | Decorator for views that checks that the user is logged in and is a staff
member, displaying the login page if necessary. | [
"Decorator",
"for",
"views",
"that",
"checks",
"that",
"the",
"user",
"is",
"logged",
"in",
"and",
"is",
"a",
"staff",
"member",
"displaying",
"the",
"login",
"page",
"if",
"necessary",
"."
] | train | https://github.com/littlepea/django-docs/blob/8e6c867dab51746c6482bd52c4f1270684af3870/docs/views.py#L18-L40 |
cpburnz/python-path-specification | pathspec/pathspec.py | PathSpec.from_lines | def from_lines(cls, pattern_factory, lines):
"""
Compiles the pattern lines.
*pattern_factory* can be either the name of a registered pattern
factory (:class:`str`), or a :class:`~collections.abc.Callable` used
to compile patterns. It must accept an uncompiled pattern (:class:`str`)
and return the compiled... | python | def from_lines(cls, pattern_factory, lines):
"""
Compiles the pattern lines.
*pattern_factory* can be either the name of a registered pattern
factory (:class:`str`), or a :class:`~collections.abc.Callable` used
to compile patterns. It must accept an uncompiled pattern (:class:`str`)
and return the compiled... | [
"def",
"from_lines",
"(",
"cls",
",",
"pattern_factory",
",",
"lines",
")",
":",
"if",
"isinstance",
"(",
"pattern_factory",
",",
"string_types",
")",
":",
"pattern_factory",
"=",
"util",
".",
"lookup_pattern",
"(",
"pattern_factory",
")",
"if",
"not",
"callab... | Compiles the pattern lines.
*pattern_factory* can be either the name of a registered pattern
factory (:class:`str`), or a :class:`~collections.abc.Callable` used
to compile patterns. It must accept an uncompiled pattern (:class:`str`)
and return the compiled pattern (:class:`.Pattern`).
*lines* (:class:`~co... | [
"Compiles",
"the",
"pattern",
"lines",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pathspec.py#L50-L75 |
cpburnz/python-path-specification | pathspec/pathspec.py | PathSpec.match_file | def match_file(self, file, separators=None):
"""
Matches the file to this path-spec.
*file* (:class:`str`) is the file path to be matched against
:attr:`self.patterns <PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`)
optionally contains the path separators to normal... | python | def match_file(self, file, separators=None):
"""
Matches the file to this path-spec.
*file* (:class:`str`) is the file path to be matched against
:attr:`self.patterns <PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`)
optionally contains the path separators to normal... | [
"def",
"match_file",
"(",
"self",
",",
"file",
",",
"separators",
"=",
"None",
")",
":",
"norm_file",
"=",
"util",
".",
"normalize_file",
"(",
"file",
",",
"separators",
"=",
"separators",
")",
"return",
"util",
".",
"match_file",
"(",
"self",
".",
"patt... | Matches the file to this path-spec.
*file* (:class:`str`) is the file path to be matched against
:attr:`self.patterns <PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`)
optionally contains the path separators to normalize. See
:func:`~pathspec.util.normalize_file` for ... | [
"Matches",
"the",
"file",
"to",
"this",
"path",
"-",
"spec",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pathspec.py#L77-L91 |
cpburnz/python-path-specification | pathspec/pathspec.py | PathSpec.match_files | def match_files(self, files, separators=None):
"""
Matches the files to this path-spec.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be matched against :attr:`self.patterns
<PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`;
... | python | def match_files(self, files, separators=None):
"""
Matches the files to this path-spec.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be matched against :attr:`self.patterns
<PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`;
... | [
"def",
"match_files",
"(",
"self",
",",
"files",
",",
"separators",
"=",
"None",
")",
":",
"if",
"isinstance",
"(",
"files",
",",
"(",
"bytes",
",",
"unicode",
")",
")",
":",
"raise",
"TypeError",
"(",
"\"files:{!r} is not an iterable.\"",
".",
"format",
"... | Matches the files to this path-spec.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be matched against :attr:`self.patterns
<PathSpec.patterns>`.
*separators* (:class:`~collections.abc.Collection` of :class:`str`;
or :data:`None`) optionally contains the path separat... | [
"Matches",
"the",
"files",
"to",
"this",
"path",
"-",
"spec",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pathspec.py#L93-L115 |
cpburnz/python-path-specification | pathspec/pathspec.py | PathSpec.match_tree | def match_tree(self, root, on_error=None, follow_links=None):
"""
Walks the specified root path for all files and matches them to this
path-spec.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler... | python | def match_tree(self, root, on_error=None, follow_links=None):
"""
Walks the specified root path for all files and matches them to this
path-spec.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler... | [
"def",
"match_tree",
"(",
"self",
",",
"root",
",",
"on_error",
"=",
"None",
",",
"follow_links",
"=",
"None",
")",
":",
"files",
"=",
"util",
".",
"iter_tree",
"(",
"root",
",",
"on_error",
"=",
"on_error",
",",
"follow_links",
"=",
"follow_links",
")",... | Walks the specified root path for all files and matches them to this
path-spec.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler for file-system exceptions. See
:func:`~pathspec.util.iter_tree` fo... | [
"Walks",
"the",
"specified",
"root",
"path",
"for",
"all",
"files",
"and",
"matches",
"them",
"to",
"this",
"path",
"-",
"spec",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pathspec.py#L117-L137 |
cpburnz/python-path-specification | pathspec/patterns/gitwildmatch.py | GitWildMatchPattern.pattern_to_regex | def pattern_to_regex(cls, pattern):
"""
Convert the pattern into a regular expression.
*pattern* (:class:`unicode` or :class:`bytes`) is the pattern to
convert into a regular expression.
Returns the uncompiled regular expression (:class:`unicode`, :class:`bytes`,
or :data:`None`), and whether matched file... | python | def pattern_to_regex(cls, pattern):
"""
Convert the pattern into a regular expression.
*pattern* (:class:`unicode` or :class:`bytes`) is the pattern to
convert into a regular expression.
Returns the uncompiled regular expression (:class:`unicode`, :class:`bytes`,
or :data:`None`), and whether matched file... | [
"def",
"pattern_to_regex",
"(",
"cls",
",",
"pattern",
")",
":",
"if",
"isinstance",
"(",
"pattern",
",",
"unicode",
")",
":",
"return_type",
"=",
"unicode",
"elif",
"isinstance",
"(",
"pattern",
",",
"bytes",
")",
":",
"return_type",
"=",
"bytes",
"patter... | Convert the pattern into a regular expression.
*pattern* (:class:`unicode` or :class:`bytes`) is the pattern to
convert into a regular expression.
Returns the uncompiled regular expression (:class:`unicode`, :class:`bytes`,
or :data:`None`), and whether matched files should be included
(:data:`True`), exclu... | [
"Convert",
"the",
"pattern",
"into",
"a",
"regular",
"expression",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/patterns/gitwildmatch.py#L30-L174 |
cpburnz/python-path-specification | pathspec/patterns/gitwildmatch.py | GitWildMatchPattern._translate_segment_glob | def _translate_segment_glob(pattern):
"""
Translates the glob pattern to a regular expression. This is used in
the constructor to translate a path segment glob pattern to its
corresponding regular expression.
*pattern* (:class:`str`) is the glob pattern.
Returns the regular expression (:class:`str`).
""... | python | def _translate_segment_glob(pattern):
"""
Translates the glob pattern to a regular expression. This is used in
the constructor to translate a path segment glob pattern to its
corresponding regular expression.
*pattern* (:class:`str`) is the glob pattern.
Returns the regular expression (:class:`str`).
""... | [
"def",
"_translate_segment_glob",
"(",
"pattern",
")",
":",
"# NOTE: This is derived from `fnmatch.translate()` and is similar to",
"# the POSIX function `fnmatch()` with the `FNM_PATHNAME` flag set.",
"escape",
"=",
"False",
"regex",
"=",
"''",
"i",
",",
"end",
"=",
"0",
",",
... | Translates the glob pattern to a regular expression. This is used in
the constructor to translate a path segment glob pattern to its
corresponding regular expression.
*pattern* (:class:`str`) is the glob pattern.
Returns the regular expression (:class:`str`). | [
"Translates",
"the",
"glob",
"pattern",
"to",
"a",
"regular",
"expression",
".",
"This",
"is",
"used",
"in",
"the",
"constructor",
"to",
"translate",
"a",
"path",
"segment",
"glob",
"pattern",
"to",
"its",
"corresponding",
"regular",
"expression",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/patterns/gitwildmatch.py#L177-L280 |
cpburnz/python-path-specification | pathspec/patterns/gitwildmatch.py | GitIgnorePattern.pattern_to_regex | def pattern_to_regex(cls, *args, **kw):
"""
Warn about deprecation.
"""
cls._deprecated()
return super(GitIgnorePattern, cls).pattern_to_regex(*args, **kw) | python | def pattern_to_regex(cls, *args, **kw):
"""
Warn about deprecation.
"""
cls._deprecated()
return super(GitIgnorePattern, cls).pattern_to_regex(*args, **kw) | [
"def",
"pattern_to_regex",
"(",
"cls",
",",
"*",
"args",
",",
"*",
"*",
"kw",
")",
":",
"cls",
".",
"_deprecated",
"(",
")",
"return",
"super",
"(",
"GitIgnorePattern",
",",
"cls",
")",
".",
"pattern_to_regex",
"(",
"*",
"args",
",",
"*",
"*",
"kw",
... | Warn about deprecation. | [
"Warn",
"about",
"deprecation",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/patterns/gitwildmatch.py#L306-L311 |
cpburnz/python-path-specification | pathspec/util.py | iter_tree | def iter_tree(root, on_error=None, follow_links=None):
"""
Walks the specified directory for all files.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler for file-system exceptions. It will be
called w... | python | def iter_tree(root, on_error=None, follow_links=None):
"""
Walks the specified directory for all files.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler for file-system exceptions. It will be
called w... | [
"def",
"iter_tree",
"(",
"root",
",",
"on_error",
"=",
"None",
",",
"follow_links",
"=",
"None",
")",
":",
"if",
"on_error",
"is",
"not",
"None",
"and",
"not",
"callable",
"(",
"on_error",
")",
":",
"raise",
"TypeError",
"(",
"\"on_error:{!r} is not callable... | Walks the specified directory for all files.
*root* (:class:`str`) is the root directory to search for files.
*on_error* (:class:`~collections.abc.Callable` or :data:`None`)
optionally is the error handler for file-system exceptions. It will be
called with the exception (:exc:`OSError`). Reraise the exception to
... | [
"Walks",
"the",
"specified",
"directory",
"for",
"all",
"files",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L27-L55 |
cpburnz/python-path-specification | pathspec/util.py | _iter_tree_next | def _iter_tree_next(root_full, dir_rel, memo, on_error, follow_links):
"""
Scan the directory for all descendant files.
*root_full* (:class:`str`) the absolute path to the root directory.
*dir_rel* (:class:`str`) the path to the directory to scan relative to
*root_full*.
*memo* (:class:`dict`) keeps track of a... | python | def _iter_tree_next(root_full, dir_rel, memo, on_error, follow_links):
"""
Scan the directory for all descendant files.
*root_full* (:class:`str`) the absolute path to the root directory.
*dir_rel* (:class:`str`) the path to the directory to scan relative to
*root_full*.
*memo* (:class:`dict`) keeps track of a... | [
"def",
"_iter_tree_next",
"(",
"root_full",
",",
"dir_rel",
",",
"memo",
",",
"on_error",
",",
"follow_links",
")",
":",
"dir_full",
"=",
"os",
".",
"path",
".",
"join",
"(",
"root_full",
",",
"dir_rel",
")",
"dir_real",
"=",
"os",
".",
"path",
".",
"r... | Scan the directory for all descendant files.
*root_full* (:class:`str`) the absolute path to the root directory.
*dir_rel* (:class:`str`) the path to the directory to scan relative to
*root_full*.
*memo* (:class:`dict`) keeps track of ancestor directories
encountered. Maps each ancestor real path (:class:`str``... | [
"Scan",
"the",
"directory",
"for",
"all",
"descendant",
"files",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L57-L126 |
cpburnz/python-path-specification | pathspec/util.py | match_file | def match_file(patterns, file):
"""
Matches the file to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*file* (:class:`str`) is the normalized file path to be matched
against *patterns*.
Returns :data:`True` if *file* matched; ... | python | def match_file(patterns, file):
"""
Matches the file to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*file* (:class:`str`) is the normalized file path to be matched
against *patterns*.
Returns :data:`True` if *file* matched; ... | [
"def",
"match_file",
"(",
"patterns",
",",
"file",
")",
":",
"matched",
"=",
"False",
"for",
"pattern",
"in",
"patterns",
":",
"if",
"pattern",
".",
"include",
"is",
"not",
"None",
":",
"if",
"file",
"in",
"pattern",
".",
"match",
"(",
"(",
"file",
"... | Matches the file to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*file* (:class:`str`) is the normalized file path to be matched
against *patterns*.
Returns :data:`True` if *file* matched; otherwise, :data:`False`. | [
"Matches",
"the",
"file",
"to",
"the",
"patterns",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L139-L156 |
cpburnz/python-path-specification | pathspec/util.py | match_files | def match_files(patterns, files):
"""
Matches the files to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the normalized file paths to be matched against *patt... | python | def match_files(patterns, files):
"""
Matches the files to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the normalized file paths to be matched against *patt... | [
"def",
"match_files",
"(",
"patterns",
",",
"files",
")",
":",
"all_files",
"=",
"files",
"if",
"isinstance",
"(",
"files",
",",
"collection_type",
")",
"else",
"list",
"(",
"files",
")",
"return_files",
"=",
"set",
"(",
")",
"for",
"pattern",
"in",
"pat... | Matches the files to the patterns.
*patterns* (:class:`~collections.abc.Iterable` of :class:`~pathspec.pattern.Pattern`)
contains the patterns to use.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the normalized file paths to be matched against *patterns*.
Returns the matched files (:cla... | [
"Matches",
"the",
"files",
"to",
"the",
"patterns",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L158-L179 |
cpburnz/python-path-specification | pathspec/util.py | normalize_file | def normalize_file(file, separators=None):
"""
Normalizes the file path to use the POSIX path separator (i.e., ``'/'``).
*file* (:class:`str`) is the file path.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally contains the path separators to normalize.
This does no... | python | def normalize_file(file, separators=None):
"""
Normalizes the file path to use the POSIX path separator (i.e., ``'/'``).
*file* (:class:`str`) is the file path.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally contains the path separators to normalize.
This does no... | [
"def",
"normalize_file",
"(",
"file",
",",
"separators",
"=",
"None",
")",
":",
"# Normalize path separators.",
"if",
"separators",
"is",
"None",
":",
"separators",
"=",
"NORMALIZE_PATH_SEPS",
"norm_file",
"=",
"file",
"for",
"sep",
"in",
"separators",
":",
"nor... | Normalizes the file path to use the POSIX path separator (i.e., ``'/'``).
*file* (:class:`str`) is the file path.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally contains the path separators to normalize.
This does not need to include the POSIX path separator (``'/'... | [
"Normalizes",
"the",
"file",
"path",
"to",
"use",
"the",
"POSIX",
"path",
"separator",
"(",
"i",
".",
"e",
".",
"/",
")",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L181-L207 |
cpburnz/python-path-specification | pathspec/util.py | normalize_files | def normalize_files(files, separators=None):
"""
Normalizes the file paths to use the POSIX path separator.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be normalized.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally conta... | python | def normalize_files(files, separators=None):
"""
Normalizes the file paths to use the POSIX path separator.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be normalized.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally conta... | [
"def",
"normalize_files",
"(",
"files",
",",
"separators",
"=",
"None",
")",
":",
"norm_files",
"=",
"{",
"}",
"for",
"path",
"in",
"files",
":",
"norm_files",
"[",
"normalize_file",
"(",
"path",
",",
"separators",
"=",
"separators",
")",
"]",
"=",
"path... | Normalizes the file paths to use the POSIX path separator.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
the file paths to be normalized.
*separators* (:class:`~collections.abc.Collection` of :class:`str`; or
:data:`None`) optionally contains the path separators to normalize.
See :func:`n... | [
"Normalizes",
"the",
"file",
"paths",
"to",
"use",
"the",
"POSIX",
"path",
"separator",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L209-L226 |
cpburnz/python-path-specification | pathspec/util.py | register_pattern | def register_pattern(name, pattern_factory, override=None):
"""
Registers the specified pattern factory.
*name* (:class:`str`) is the name to register the pattern factory
under.
*pattern_factory* (:class:`~collections.abc.Callable`) is used to
compile patterns. It must accept an uncompiled pattern (:class:`str`... | python | def register_pattern(name, pattern_factory, override=None):
"""
Registers the specified pattern factory.
*name* (:class:`str`) is the name to register the pattern factory
under.
*pattern_factory* (:class:`~collections.abc.Callable`) is used to
compile patterns. It must accept an uncompiled pattern (:class:`str`... | [
"def",
"register_pattern",
"(",
"name",
",",
"pattern_factory",
",",
"override",
"=",
"None",
")",
":",
"if",
"not",
"isinstance",
"(",
"name",
",",
"string_types",
")",
":",
"raise",
"TypeError",
"(",
"\"name:{!r} is not a string.\"",
".",
"format",
"(",
"nam... | Registers the specified pattern factory.
*name* (:class:`str`) is the name to register the pattern factory
under.
*pattern_factory* (:class:`~collections.abc.Callable`) is used to
compile patterns. It must accept an uncompiled pattern (:class:`str`)
and return the compiled pattern (:class:`.Pattern`).
*overrid... | [
"Registers",
"the",
"specified",
"pattern",
"factory",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L228-L250 |
cpburnz/python-path-specification | pathspec/util.py | RecursionError.message | def message(self):
"""
*message* (:class:`str`) is the error message.
"""
return "Real path {real!r} was encountered at {first!r} and then {second!r}.".format(
real=self.real_path,
first=self.first_path,
second=self.second_path,
) | python | def message(self):
"""
*message* (:class:`str`) is the error message.
"""
return "Real path {real!r} was encountered at {first!r} and then {second!r}.".format(
real=self.real_path,
first=self.first_path,
second=self.second_path,
) | [
"def",
"message",
"(",
"self",
")",
":",
"return",
"\"Real path {real!r} was encountered at {first!r} and then {second!r}.\"",
".",
"format",
"(",
"real",
"=",
"self",
".",
"real_path",
",",
"first",
"=",
"self",
".",
"first_path",
",",
"second",
"=",
"self",
".",... | *message* (:class:`str`) is the error message. | [
"*",
"message",
"*",
"(",
":",
"class",
":",
"str",
")",
"is",
"the",
"error",
"message",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/util.py#L326-L334 |
cpburnz/python-path-specification | pathspec/pattern.py | Pattern.match | def match(self, files):
"""
Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
each file relative to the root directory (e.g., ``"relative/path/to/file"``).
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:... | python | def match(self, files):
"""
Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
each file relative to the root directory (e.g., ``"relative/path/to/file"``).
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:... | [
"def",
"match",
"(",
"self",
",",
"files",
")",
":",
"raise",
"NotImplementedError",
"(",
"\"{}.{} must override match().\"",
".",
"format",
"(",
"self",
".",
"__class__",
".",
"__module__",
",",
"self",
".",
"__class__",
".",
"__name__",
")",
")"
] | Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`) contains
each file relative to the root directory (e.g., ``"relative/path/to/file"``).
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:class:`str`). | [
"Matches",
"this",
"pattern",
"against",
"the",
"specified",
"files",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pattern.py#L35-L45 |
cpburnz/python-path-specification | pathspec/pattern.py | RegexPattern.match | def match(self, files):
"""
Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`)
contains each file relative to the root directory (e.g., "relative/path/to/file").
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:clas... | python | def match(self, files):
"""
Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`)
contains each file relative to the root directory (e.g., "relative/path/to/file").
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:clas... | [
"def",
"match",
"(",
"self",
",",
"files",
")",
":",
"if",
"self",
".",
"include",
"is",
"not",
"None",
":",
"for",
"path",
"in",
"files",
":",
"if",
"self",
".",
"regex",
".",
"match",
"(",
"path",
")",
"is",
"not",
"None",
":",
"yield",
"path"
... | Matches this pattern against the specified files.
*files* (:class:`~collections.abc.Iterable` of :class:`str`)
contains each file relative to the root directory (e.g., "relative/path/to/file").
Returns an :class:`~collections.abc.Iterable` yielding each matched
file path (:class:`str`). | [
"Matches",
"this",
"pattern",
"against",
"the",
"specified",
"files",
"."
] | train | https://github.com/cpburnz/python-path-specification/blob/6fc7567a58cb68ec7d72cc287e7fb97dbe22c017/pathspec/pattern.py#L116-L129 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.user_default_serializer | def user_default_serializer(self, obj):
"""Convert a User to a cached instance representation."""
if not obj:
return None
self.user_default_add_related_pks(obj)
return dict((
('id', obj.id),
('username', obj.username),
self.field_to_json('D... | python | def user_default_serializer(self, obj):
"""Convert a User to a cached instance representation."""
if not obj:
return None
self.user_default_add_related_pks(obj)
return dict((
('id', obj.id),
('username', obj.username),
self.field_to_json('D... | [
"def",
"user_default_serializer",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"obj",
":",
"return",
"None",
"self",
".",
"user_default_add_related_pks",
"(",
"obj",
")",
"return",
"dict",
"(",
"(",
"(",
"'id'",
",",
"obj",
".",
"id",
")",
",",
"(",... | Convert a User to a cached instance representation. | [
"Convert",
"a",
"User",
"to",
"a",
"cached",
"instance",
"representation",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L13-L24 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.user_default_loader | def user_default_loader(self, pk):
"""Load a User from the database."""
try:
obj = User.objects.get(pk=pk)
except User.DoesNotExist:
return None
else:
self.user_default_add_related_pks(obj)
return obj | python | def user_default_loader(self, pk):
"""Load a User from the database."""
try:
obj = User.objects.get(pk=pk)
except User.DoesNotExist:
return None
else:
self.user_default_add_related_pks(obj)
return obj | [
"def",
"user_default_loader",
"(",
"self",
",",
"pk",
")",
":",
"try",
":",
"obj",
"=",
"User",
".",
"objects",
".",
"get",
"(",
"pk",
"=",
"pk",
")",
"except",
"User",
".",
"DoesNotExist",
":",
"return",
"None",
"else",
":",
"self",
".",
"user_defau... | Load a User from the database. | [
"Load",
"a",
"User",
"from",
"the",
"database",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L26-L34 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.user_default_add_related_pks | def user_default_add_related_pks(self, obj):
"""Add related primary keys to a User instance."""
if not hasattr(obj, '_votes_pks'):
obj._votes_pks = list(obj.votes.values_list('pk', flat=True)) | python | def user_default_add_related_pks(self, obj):
"""Add related primary keys to a User instance."""
if not hasattr(obj, '_votes_pks'):
obj._votes_pks = list(obj.votes.values_list('pk', flat=True)) | [
"def",
"user_default_add_related_pks",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"hasattr",
"(",
"obj",
",",
"'_votes_pks'",
")",
":",
"obj",
".",
"_votes_pks",
"=",
"list",
"(",
"obj",
".",
"votes",
".",
"values_list",
"(",
"'pk'",
",",
"flat",
... | Add related primary keys to a User instance. | [
"Add",
"related",
"primary",
"keys",
"to",
"a",
"User",
"instance",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L36-L39 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.group_default_invalidator | def group_default_invalidator(self, obj):
"""Invalidated cached items when the Group changes."""
user_pks = User.objects.values_list('pk', flat=True)
return [('User', pk, False) for pk in user_pks] | python | def group_default_invalidator(self, obj):
"""Invalidated cached items when the Group changes."""
user_pks = User.objects.values_list('pk', flat=True)
return [('User', pk, False) for pk in user_pks] | [
"def",
"group_default_invalidator",
"(",
"self",
",",
"obj",
")",
":",
"user_pks",
"=",
"User",
".",
"objects",
".",
"values_list",
"(",
"'pk'",
",",
"flat",
"=",
"True",
")",
"return",
"[",
"(",
"'User'",
",",
"pk",
",",
"False",
")",
"for",
"pk",
"... | Invalidated cached items when the Group changes. | [
"Invalidated",
"cached",
"items",
"when",
"the",
"Group",
"changes",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L51-L54 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.question_default_serializer | def question_default_serializer(self, obj):
"""Convert a Question to a cached instance representation."""
if not obj:
return None
self.question_default_add_related_pks(obj)
return dict((
('id', obj.id),
('question_text', obj.question_text),
... | python | def question_default_serializer(self, obj):
"""Convert a Question to a cached instance representation."""
if not obj:
return None
self.question_default_add_related_pks(obj)
return dict((
('id', obj.id),
('question_text', obj.question_text),
... | [
"def",
"question_default_serializer",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"obj",
":",
"return",
"None",
"self",
".",
"question_default_add_related_pks",
"(",
"obj",
")",
"return",
"dict",
"(",
"(",
"(",
"'id'",
",",
"obj",
".",
"id",
")",
","... | Convert a Question to a cached instance representation. | [
"Convert",
"a",
"Question",
"to",
"a",
"cached",
"instance",
"representation",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L60-L71 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.question_default_loader | def question_default_loader(self, pk):
"""Load a Question from the database."""
try:
obj = Question.objects.get(pk=pk)
except Question.DoesNotExist:
return None
else:
self.question_default_add_related_pks(obj)
return obj | python | def question_default_loader(self, pk):
"""Load a Question from the database."""
try:
obj = Question.objects.get(pk=pk)
except Question.DoesNotExist:
return None
else:
self.question_default_add_related_pks(obj)
return obj | [
"def",
"question_default_loader",
"(",
"self",
",",
"pk",
")",
":",
"try",
":",
"obj",
"=",
"Question",
".",
"objects",
".",
"get",
"(",
"pk",
"=",
"pk",
")",
"except",
"Question",
".",
"DoesNotExist",
":",
"return",
"None",
"else",
":",
"self",
".",
... | Load a Question from the database. | [
"Load",
"a",
"Question",
"from",
"the",
"database",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L73-L81 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.question_default_add_related_pks | def question_default_add_related_pks(self, obj):
"""Add related primary keys to a Question instance."""
if not hasattr(obj, '_choice_pks'):
obj._choice_pks = list(obj.choices.values_list('pk', flat=True)) | python | def question_default_add_related_pks(self, obj):
"""Add related primary keys to a Question instance."""
if not hasattr(obj, '_choice_pks'):
obj._choice_pks = list(obj.choices.values_list('pk', flat=True)) | [
"def",
"question_default_add_related_pks",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"hasattr",
"(",
"obj",
",",
"'_choice_pks'",
")",
":",
"obj",
".",
"_choice_pks",
"=",
"list",
"(",
"obj",
".",
"choices",
".",
"values_list",
"(",
"'pk'",
",",
"f... | Add related primary keys to a Question instance. | [
"Add",
"related",
"primary",
"keys",
"to",
"a",
"Question",
"instance",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L83-L86 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.choice_default_serializer | def choice_default_serializer(self, obj):
"""Convert a Choice to a cached instance representation."""
if not obj:
return None
self.choice_default_add_related_pks(obj)
return dict((
('id', obj.id),
('choice_text', obj.choice_text),
self.fiel... | python | def choice_default_serializer(self, obj):
"""Convert a Choice to a cached instance representation."""
if not obj:
return None
self.choice_default_add_related_pks(obj)
return dict((
('id', obj.id),
('choice_text', obj.choice_text),
self.fiel... | [
"def",
"choice_default_serializer",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"obj",
":",
"return",
"None",
"self",
".",
"choice_default_add_related_pks",
"(",
"obj",
")",
"return",
"dict",
"(",
"(",
"(",
"'id'",
",",
"obj",
".",
"id",
")",
",",
... | Convert a Choice to a cached instance representation. | [
"Convert",
"a",
"Choice",
"to",
"a",
"cached",
"instance",
"representation",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L92-L104 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.choice_default_loader | def choice_default_loader(self, pk):
"""Load a Choice from the database."""
try:
obj = Choice.objects.get(pk=pk)
except Choice.DoesNotExist:
return None
else:
self.choice_default_add_related_pks(obj)
return obj | python | def choice_default_loader(self, pk):
"""Load a Choice from the database."""
try:
obj = Choice.objects.get(pk=pk)
except Choice.DoesNotExist:
return None
else:
self.choice_default_add_related_pks(obj)
return obj | [
"def",
"choice_default_loader",
"(",
"self",
",",
"pk",
")",
":",
"try",
":",
"obj",
"=",
"Choice",
".",
"objects",
".",
"get",
"(",
"pk",
"=",
"pk",
")",
"except",
"Choice",
".",
"DoesNotExist",
":",
"return",
"None",
"else",
":",
"self",
".",
"choi... | Load a Choice from the database. | [
"Load",
"a",
"Choice",
"from",
"the",
"database",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L106-L114 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.choice_default_add_related_pks | def choice_default_add_related_pks(self, obj):
"""Add related primary keys to a Choice instance."""
if not hasattr(obj, '_voter_pks'):
obj._voter_pks = obj.voters.values_list('pk', flat=True) | python | def choice_default_add_related_pks(self, obj):
"""Add related primary keys to a Choice instance."""
if not hasattr(obj, '_voter_pks'):
obj._voter_pks = obj.voters.values_list('pk', flat=True) | [
"def",
"choice_default_add_related_pks",
"(",
"self",
",",
"obj",
")",
":",
"if",
"not",
"hasattr",
"(",
"obj",
",",
"'_voter_pks'",
")",
":",
"obj",
".",
"_voter_pks",
"=",
"obj",
".",
"voters",
".",
"values_list",
"(",
"'pk'",
",",
"flat",
"=",
"True",... | Add related primary keys to a Choice instance. | [
"Add",
"related",
"primary",
"keys",
"to",
"a",
"Choice",
"instance",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L116-L119 |
jwhitlock/drf-cached-instances | sample_poll_app/cache.py | SampleCache.choice_default_invalidator | def choice_default_invalidator(self, obj):
"""Invalidated cached items when the Choice changes."""
invalid = [('Question', obj.question_id, True)]
for pk in obj.voters.values_list('pk', flat=True):
invalid.append(('User', pk, False))
return invalid | python | def choice_default_invalidator(self, obj):
"""Invalidated cached items when the Choice changes."""
invalid = [('Question', obj.question_id, True)]
for pk in obj.voters.values_list('pk', flat=True):
invalid.append(('User', pk, False))
return invalid | [
"def",
"choice_default_invalidator",
"(",
"self",
",",
"obj",
")",
":",
"invalid",
"=",
"[",
"(",
"'Question'",
",",
"obj",
".",
"question_id",
",",
"True",
")",
"]",
"for",
"pk",
"in",
"obj",
".",
"voters",
".",
"values_list",
"(",
"'pk'",
",",
"flat"... | Invalidated cached items when the Choice changes. | [
"Invalidated",
"cached",
"items",
"when",
"the",
"Choice",
"changes",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/cache.py#L121-L126 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.cache | def cache(self):
"""Get the Django cache interface.
This allows disabling the cache with
settings.USE_DRF_INSTANCE_CACHE=False. It also delays import so that
Django Debug Toolbar will record cache requests.
"""
if not self._cache:
use_cache = getattr(setting... | python | def cache(self):
"""Get the Django cache interface.
This allows disabling the cache with
settings.USE_DRF_INSTANCE_CACHE=False. It also delays import so that
Django Debug Toolbar will record cache requests.
"""
if not self._cache:
use_cache = getattr(setting... | [
"def",
"cache",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"_cache",
":",
"use_cache",
"=",
"getattr",
"(",
"settings",
",",
"'USE_DRF_INSTANCE_CACHE'",
",",
"True",
")",
"if",
"use_cache",
":",
"from",
"django",
".",
"core",
".",
"cache",
"import"... | Get the Django cache interface.
This allows disabling the cache with
settings.USE_DRF_INSTANCE_CACHE=False. It also delays import so that
Django Debug Toolbar will record cache requests. | [
"Get",
"the",
"Django",
"cache",
"interface",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L33-L45 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.delete_all_versions | def delete_all_versions(self, model_name, obj_pk):
"""Delete all versions of a cached instance."""
if self.cache:
for version in self.versions:
key = self.key_for(version, model_name, obj_pk)
self.cache.delete(key) | python | def delete_all_versions(self, model_name, obj_pk):
"""Delete all versions of a cached instance."""
if self.cache:
for version in self.versions:
key = self.key_for(version, model_name, obj_pk)
self.cache.delete(key) | [
"def",
"delete_all_versions",
"(",
"self",
",",
"model_name",
",",
"obj_pk",
")",
":",
"if",
"self",
".",
"cache",
":",
"for",
"version",
"in",
"self",
".",
"versions",
":",
"key",
"=",
"self",
".",
"key_for",
"(",
"version",
",",
"model_name",
",",
"o... | Delete all versions of a cached instance. | [
"Delete",
"all",
"versions",
"of",
"a",
"cached",
"instance",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L51-L56 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.model_function | def model_function(self, model_name, version, func_name):
"""Return the model-specific caching function."""
assert func_name in ('serializer', 'loader', 'invalidator')
name = "%s_%s_%s" % (model_name.lower(), version, func_name)
return getattr(self, name) | python | def model_function(self, model_name, version, func_name):
"""Return the model-specific caching function."""
assert func_name in ('serializer', 'loader', 'invalidator')
name = "%s_%s_%s" % (model_name.lower(), version, func_name)
return getattr(self, name) | [
"def",
"model_function",
"(",
"self",
",",
"model_name",
",",
"version",
",",
"func_name",
")",
":",
"assert",
"func_name",
"in",
"(",
"'serializer'",
",",
"'loader'",
",",
"'invalidator'",
")",
"name",
"=",
"\"%s_%s_%s\"",
"%",
"(",
"model_name",
".",
"lowe... | Return the model-specific caching function. | [
"Return",
"the",
"model",
"-",
"specific",
"caching",
"function",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L58-L62 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_function | def field_function(self, type_code, func_name):
"""Return the field function."""
assert func_name in ('to_json', 'from_json')
name = "field_%s_%s" % (type_code.lower(), func_name)
return getattr(self, name) | python | def field_function(self, type_code, func_name):
"""Return the field function."""
assert func_name in ('to_json', 'from_json')
name = "field_%s_%s" % (type_code.lower(), func_name)
return getattr(self, name) | [
"def",
"field_function",
"(",
"self",
",",
"type_code",
",",
"func_name",
")",
":",
"assert",
"func_name",
"in",
"(",
"'to_json'",
",",
"'from_json'",
")",
"name",
"=",
"\"field_%s_%s\"",
"%",
"(",
"type_code",
".",
"lower",
"(",
")",
",",
"func_name",
")"... | Return the field function. | [
"Return",
"the",
"field",
"function",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L64-L68 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_to_json | def field_to_json(self, type_code, key, *args, **kwargs):
"""Convert a field to a JSON-serializable representation."""
assert ':' not in key
to_json = self.field_function(type_code, 'to_json')
key_and_type = "%s:%s" % (key, type_code)
json_value = to_json(*args, **kwargs)
... | python | def field_to_json(self, type_code, key, *args, **kwargs):
"""Convert a field to a JSON-serializable representation."""
assert ':' not in key
to_json = self.field_function(type_code, 'to_json')
key_and_type = "%s:%s" % (key, type_code)
json_value = to_json(*args, **kwargs)
... | [
"def",
"field_to_json",
"(",
"self",
",",
"type_code",
",",
"key",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"assert",
"':'",
"not",
"in",
"key",
"to_json",
"=",
"self",
".",
"field_function",
"(",
"type_code",
",",
"'to_json'",
")",
"key_an... | Convert a field to a JSON-serializable representation. | [
"Convert",
"a",
"field",
"to",
"a",
"JSON",
"-",
"serializable",
"representation",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L70-L76 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_from_json | def field_from_json(self, key_and_type, json_value):
"""Convert a JSON-serializable representation back to a field."""
assert ':' in key_and_type
key, type_code = key_and_type.split(':', 1)
from_json = self.field_function(type_code, 'from_json')
value = from_json(json_value)
... | python | def field_from_json(self, key_and_type, json_value):
"""Convert a JSON-serializable representation back to a field."""
assert ':' in key_and_type
key, type_code = key_and_type.split(':', 1)
from_json = self.field_function(type_code, 'from_json')
value = from_json(json_value)
... | [
"def",
"field_from_json",
"(",
"self",
",",
"key_and_type",
",",
"json_value",
")",
":",
"assert",
"':'",
"in",
"key_and_type",
"key",
",",
"type_code",
"=",
"key_and_type",
".",
"split",
"(",
"':'",
",",
"1",
")",
"from_json",
"=",
"self",
".",
"field_fun... | Convert a JSON-serializable representation back to a field. | [
"Convert",
"a",
"JSON",
"-",
"serializable",
"representation",
"back",
"to",
"a",
"field",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L78-L84 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.get_instances | def get_instances(self, object_specs, version=None):
"""Get the cached native representation for one or more objects.
Keyword arguments:
object_specs - A sequence of triples (model name, pk, obj):
- model name - the name of the model
- pk - the primary key of the instance
... | python | def get_instances(self, object_specs, version=None):
"""Get the cached native representation for one or more objects.
Keyword arguments:
object_specs - A sequence of triples (model name, pk, obj):
- model name - the name of the model
- pk - the primary key of the instance
... | [
"def",
"get_instances",
"(",
"self",
",",
"object_specs",
",",
"version",
"=",
"None",
")",
":",
"ret",
"=",
"dict",
"(",
")",
"spec_keys",
"=",
"set",
"(",
")",
"cache_keys",
"=",
"[",
"]",
"version",
"=",
"version",
"or",
"self",
".",
"default_versio... | Get the cached native representation for one or more objects.
Keyword arguments:
object_specs - A sequence of triples (model name, pk, obj):
- model name - the name of the model
- pk - the primary key of the instance
- obj - the instance, or None to load it
version - The... | [
"Get",
"the",
"cached",
"native",
"representation",
"for",
"one",
"or",
"more",
"objects",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L86-L157 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.update_instance | def update_instance(
self, model_name, pk, instance=None, version=None,
update_only=False):
"""Create or update a cached instance.
Keyword arguments are:
model_name - The name of the model
pk - The primary key of the instance
instance - The Django model i... | python | def update_instance(
self, model_name, pk, instance=None, version=None,
update_only=False):
"""Create or update a cached instance.
Keyword arguments are:
model_name - The name of the model
pk - The primary key of the instance
instance - The Django model i... | [
"def",
"update_instance",
"(",
"self",
",",
"model_name",
",",
"pk",
",",
"instance",
"=",
"None",
",",
"version",
"=",
"None",
",",
"update_only",
"=",
"False",
")",
":",
"versions",
"=",
"[",
"version",
"]",
"if",
"version",
"else",
"self",
".",
"ver... | Create or update a cached instance.
Keyword arguments are:
model_name - The name of the model
pk - The primary key of the instance
instance - The Django model instance, or None to load it
versions - Version to update, or None for all
update_only - If False (default), the... | [
"Create",
"or",
"update",
"a",
"cached",
"instance",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L159-L228 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_date_to_json | def field_date_to_json(self, day):
"""Convert a date to a date triple."""
if isinstance(day, six.string_types):
day = parse_date(day)
return [day.year, day.month, day.day] if day else None | python | def field_date_to_json(self, day):
"""Convert a date to a date triple."""
if isinstance(day, six.string_types):
day = parse_date(day)
return [day.year, day.month, day.day] if day else None | [
"def",
"field_date_to_json",
"(",
"self",
",",
"day",
")",
":",
"if",
"isinstance",
"(",
"day",
",",
"six",
".",
"string_types",
")",
":",
"day",
"=",
"parse_date",
"(",
"day",
")",
"return",
"[",
"day",
".",
"year",
",",
"day",
".",
"month",
",",
... | Convert a date to a date triple. | [
"Convert",
"a",
"date",
"to",
"a",
"date",
"triple",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L238-L242 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_datetime_from_json | def field_datetime_from_json(self, json_val):
"""Convert a UTC timestamp to a UTC datetime."""
if type(json_val) == int:
seconds = int(json_val)
dt = datetime.fromtimestamp(seconds, utc)
elif json_val is None:
dt = None
else:
seconds, micro... | python | def field_datetime_from_json(self, json_val):
"""Convert a UTC timestamp to a UTC datetime."""
if type(json_val) == int:
seconds = int(json_val)
dt = datetime.fromtimestamp(seconds, utc)
elif json_val is None:
dt = None
else:
seconds, micro... | [
"def",
"field_datetime_from_json",
"(",
"self",
",",
"json_val",
")",
":",
"if",
"type",
"(",
"json_val",
")",
"==",
"int",
":",
"seconds",
"=",
"int",
"(",
"json_val",
")",
"dt",
"=",
"datetime",
".",
"fromtimestamp",
"(",
"seconds",
",",
"utc",
")",
... | Convert a UTC timestamp to a UTC datetime. | [
"Convert",
"a",
"UTC",
"timestamp",
"to",
"a",
"UTC",
"datetime",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L244-L255 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_datetime_to_json | def field_datetime_to_json(self, dt):
"""Convert a datetime to a UTC timestamp w/ microsecond resolution.
datetimes w/o timezone will be assumed to be in UTC
"""
if isinstance(dt, six.string_types):
dt = parse_datetime(dt)
if not dt:
return None
t... | python | def field_datetime_to_json(self, dt):
"""Convert a datetime to a UTC timestamp w/ microsecond resolution.
datetimes w/o timezone will be assumed to be in UTC
"""
if isinstance(dt, six.string_types):
dt = parse_datetime(dt)
if not dt:
return None
t... | [
"def",
"field_datetime_to_json",
"(",
"self",
",",
"dt",
")",
":",
"if",
"isinstance",
"(",
"dt",
",",
"six",
".",
"string_types",
")",
":",
"dt",
"=",
"parse_datetime",
"(",
"dt",
")",
"if",
"not",
"dt",
":",
"return",
"None",
"ts",
"=",
"timegm",
"... | Convert a datetime to a UTC timestamp w/ microsecond resolution.
datetimes w/o timezone will be assumed to be in UTC | [
"Convert",
"a",
"datetime",
"to",
"a",
"UTC",
"timestamp",
"w",
"/",
"microsecond",
"resolution",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L257-L270 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_timedelta_from_json | def field_timedelta_from_json(self, json_val):
"""Convert json_val to a timedelta object.
json_val contains total number of seconds in the timedelta.
If json_val is a string it will be converted to a float.
"""
if isinstance(json_val, str):
return timedelta(seconds=f... | python | def field_timedelta_from_json(self, json_val):
"""Convert json_val to a timedelta object.
json_val contains total number of seconds in the timedelta.
If json_val is a string it will be converted to a float.
"""
if isinstance(json_val, str):
return timedelta(seconds=f... | [
"def",
"field_timedelta_from_json",
"(",
"self",
",",
"json_val",
")",
":",
"if",
"isinstance",
"(",
"json_val",
",",
"str",
")",
":",
"return",
"timedelta",
"(",
"seconds",
"=",
"float",
"(",
"json_val",
")",
")",
"elif",
"json_val",
"is",
"None",
":",
... | Convert json_val to a timedelta object.
json_val contains total number of seconds in the timedelta.
If json_val is a string it will be converted to a float. | [
"Convert",
"json_val",
"to",
"a",
"timedelta",
"object",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L272-L283 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_timedelta_to_json | def field_timedelta_to_json(self, td):
"""Convert timedelta to value containing total number of seconds.
If there are fractions of a second the return value will be a
string, otherwise it will be an int.
"""
if isinstance(td, six.string_types):
td = parse_duration(td... | python | def field_timedelta_to_json(self, td):
"""Convert timedelta to value containing total number of seconds.
If there are fractions of a second the return value will be a
string, otherwise it will be an int.
"""
if isinstance(td, six.string_types):
td = parse_duration(td... | [
"def",
"field_timedelta_to_json",
"(",
"self",
",",
"td",
")",
":",
"if",
"isinstance",
"(",
"td",
",",
"six",
".",
"string_types",
")",
":",
"td",
"=",
"parse_duration",
"(",
"td",
")",
"if",
"not",
"td",
":",
"return",
"None",
"if",
"td",
".",
"mic... | Convert timedelta to value containing total number of seconds.
If there are fractions of a second the return value will be a
string, otherwise it will be an int. | [
"Convert",
"timedelta",
"to",
"value",
"containing",
"total",
"number",
"of",
"seconds",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L285-L298 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_pklist_from_json | def field_pklist_from_json(self, data):
"""Load a PkOnlyQueryset from a JSON dict.
This uses the same format as cached_queryset_from_json
"""
model = get_model(data['app'], data['model'])
return PkOnlyQueryset(self, model, data['pks']) | python | def field_pklist_from_json(self, data):
"""Load a PkOnlyQueryset from a JSON dict.
This uses the same format as cached_queryset_from_json
"""
model = get_model(data['app'], data['model'])
return PkOnlyQueryset(self, model, data['pks']) | [
"def",
"field_pklist_from_json",
"(",
"self",
",",
"data",
")",
":",
"model",
"=",
"get_model",
"(",
"data",
"[",
"'app'",
"]",
",",
"data",
"[",
"'model'",
"]",
")",
"return",
"PkOnlyQueryset",
"(",
"self",
",",
"model",
",",
"data",
"[",
"'pks'",
"]"... | Load a PkOnlyQueryset from a JSON dict.
This uses the same format as cached_queryset_from_json | [
"Load",
"a",
"PkOnlyQueryset",
"from",
"a",
"JSON",
"dict",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L300-L306 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_pklist_to_json | def field_pklist_to_json(self, model, pks):
"""Convert a list of primary keys to a JSON dict.
This uses the same format as cached_queryset_to_json
"""
app_label = model._meta.app_label
model_name = model._meta.model_name
return {
'app': app_label,
... | python | def field_pklist_to_json(self, model, pks):
"""Convert a list of primary keys to a JSON dict.
This uses the same format as cached_queryset_to_json
"""
app_label = model._meta.app_label
model_name = model._meta.model_name
return {
'app': app_label,
... | [
"def",
"field_pklist_to_json",
"(",
"self",
",",
"model",
",",
"pks",
")",
":",
"app_label",
"=",
"model",
".",
"_meta",
".",
"app_label",
"model_name",
"=",
"model",
".",
"_meta",
".",
"model_name",
"return",
"{",
"'app'",
":",
"app_label",
",",
"'model'"... | Convert a list of primary keys to a JSON dict.
This uses the same format as cached_queryset_to_json | [
"Convert",
"a",
"list",
"of",
"primary",
"keys",
"to",
"a",
"JSON",
"dict",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L308-L319 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_pk_from_json | def field_pk_from_json(self, data):
"""Load a PkOnlyModel from a JSON dict."""
model = get_model(data['app'], data['model'])
return PkOnlyModel(self, model, data['pk']) | python | def field_pk_from_json(self, data):
"""Load a PkOnlyModel from a JSON dict."""
model = get_model(data['app'], data['model'])
return PkOnlyModel(self, model, data['pk']) | [
"def",
"field_pk_from_json",
"(",
"self",
",",
"data",
")",
":",
"model",
"=",
"get_model",
"(",
"data",
"[",
"'app'",
"]",
",",
"data",
"[",
"'model'",
"]",
")",
"return",
"PkOnlyModel",
"(",
"self",
",",
"model",
",",
"data",
"[",
"'pk'",
"]",
")"
... | Load a PkOnlyModel from a JSON dict. | [
"Load",
"a",
"PkOnlyModel",
"from",
"a",
"JSON",
"dict",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L321-L324 |
jwhitlock/drf-cached-instances | drf_cached_instances/cache.py | BaseCache.field_pk_to_json | def field_pk_to_json(self, model, pk):
"""Convert a primary key to a JSON dict."""
app_label = model._meta.app_label
model_name = model._meta.model_name
return {
'app': app_label,
'model': model_name,
'pk': pk,
} | python | def field_pk_to_json(self, model, pk):
"""Convert a primary key to a JSON dict."""
app_label = model._meta.app_label
model_name = model._meta.model_name
return {
'app': app_label,
'model': model_name,
'pk': pk,
} | [
"def",
"field_pk_to_json",
"(",
"self",
",",
"model",
",",
"pk",
")",
":",
"app_label",
"=",
"model",
".",
"_meta",
".",
"app_label",
"model_name",
"=",
"model",
".",
"_meta",
".",
"model_name",
"return",
"{",
"'app'",
":",
"app_label",
",",
"'model'",
"... | Convert a primary key to a JSON dict. | [
"Convert",
"a",
"primary",
"key",
"to",
"a",
"JSON",
"dict",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/cache.py#L326-L334 |
jwhitlock/drf-cached-instances | sample_poll_app/models.py | choice_voters_changed_update_cache | def choice_voters_changed_update_cache(
sender, instance, action, reverse, model, pk_set, **kwargs):
"""Update cache when choice.voters changes."""
if action not in ('post_add', 'post_remove', 'post_clear'):
# post_clear is not handled, because clear is called in
# django.db.models.field... | python | def choice_voters_changed_update_cache(
sender, instance, action, reverse, model, pk_set, **kwargs):
"""Update cache when choice.voters changes."""
if action not in ('post_add', 'post_remove', 'post_clear'):
# post_clear is not handled, because clear is called in
# django.db.models.field... | [
"def",
"choice_voters_changed_update_cache",
"(",
"sender",
",",
"instance",
",",
"action",
",",
"reverse",
",",
"model",
",",
"pk_set",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"action",
"not",
"in",
"(",
"'post_add'",
",",
"'post_remove'",
",",
"'post_clea... | Update cache when choice.voters changes. | [
"Update",
"cache",
"when",
"choice",
".",
"voters",
"changes",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/models.py#L33-L60 |
jwhitlock/drf-cached-instances | sample_poll_app/models.py | post_delete_update_cache | def post_delete_update_cache(sender, instance, **kwargs):
"""Update the cache when an instance is deleted."""
name = sender.__name__
if name in cached_model_names:
from .tasks import update_cache_for_instance
update_cache_for_instance(name, instance.pk, instance) | python | def post_delete_update_cache(sender, instance, **kwargs):
"""Update the cache when an instance is deleted."""
name = sender.__name__
if name in cached_model_names:
from .tasks import update_cache_for_instance
update_cache_for_instance(name, instance.pk, instance) | [
"def",
"post_delete_update_cache",
"(",
"sender",
",",
"instance",
",",
"*",
"*",
"kwargs",
")",
":",
"name",
"=",
"sender",
".",
"__name__",
"if",
"name",
"in",
"cached_model_names",
":",
"from",
".",
"tasks",
"import",
"update_cache_for_instance",
"update_cach... | Update the cache when an instance is deleted. | [
"Update",
"the",
"cache",
"when",
"an",
"instance",
"is",
"deleted",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/models.py#L64-L69 |
jwhitlock/drf-cached-instances | sample_poll_app/models.py | post_save_update_cache | def post_save_update_cache(sender, instance, created, raw, **kwargs):
"""Update the cache when an instance is created or modified."""
if raw:
return
name = sender.__name__
if name in cached_model_names:
delay_cache = getattr(instance, '_delay_cache', False)
if not delay_cache:
... | python | def post_save_update_cache(sender, instance, created, raw, **kwargs):
"""Update the cache when an instance is created or modified."""
if raw:
return
name = sender.__name__
if name in cached_model_names:
delay_cache = getattr(instance, '_delay_cache', False)
if not delay_cache:
... | [
"def",
"post_save_update_cache",
"(",
"sender",
",",
"instance",
",",
"created",
",",
"raw",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"raw",
":",
"return",
"name",
"=",
"sender",
".",
"__name__",
"if",
"name",
"in",
"cached_model_names",
":",
"delay_cache"... | Update the cache when an instance is created or modified. | [
"Update",
"the",
"cache",
"when",
"an",
"instance",
"is",
"created",
"or",
"modified",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/models.py#L73-L82 |
jwhitlock/drf-cached-instances | drf_cached_instances/mixins.py | CachedViewMixin.get_queryset | def get_queryset(self):
"""Get the queryset for the action.
If action is read action, return a CachedQueryset
Otherwise, return a Django queryset
"""
queryset = super(CachedViewMixin, self).get_queryset()
if self.action in ('list', 'retrieve'):
return CachedQ... | python | def get_queryset(self):
"""Get the queryset for the action.
If action is read action, return a CachedQueryset
Otherwise, return a Django queryset
"""
queryset = super(CachedViewMixin, self).get_queryset()
if self.action in ('list', 'retrieve'):
return CachedQ... | [
"def",
"get_queryset",
"(",
"self",
")",
":",
"queryset",
"=",
"super",
"(",
"CachedViewMixin",
",",
"self",
")",
".",
"get_queryset",
"(",
")",
"if",
"self",
".",
"action",
"in",
"(",
"'list'",
",",
"'retrieve'",
")",
":",
"return",
"CachedQueryset",
"(... | Get the queryset for the action.
If action is read action, return a CachedQueryset
Otherwise, return a Django queryset | [
"Get",
"the",
"queryset",
"for",
"the",
"action",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/mixins.py#L17-L27 |
jwhitlock/drf-cached-instances | drf_cached_instances/mixins.py | CachedViewMixin.get_object | def get_object(self, queryset=None):
"""
Return the object the view is displaying.
Same as rest_framework.generics.GenericAPIView, but:
- Failed assertions instead of deprecations
"""
# Determine the base queryset to use.
assert queryset is None, "Passing a query... | python | def get_object(self, queryset=None):
"""
Return the object the view is displaying.
Same as rest_framework.generics.GenericAPIView, but:
- Failed assertions instead of deprecations
"""
# Determine the base queryset to use.
assert queryset is None, "Passing a query... | [
"def",
"get_object",
"(",
"self",
",",
"queryset",
"=",
"None",
")",
":",
"# Determine the base queryset to use.",
"assert",
"queryset",
"is",
"None",
",",
"\"Passing a queryset is disabled\"",
"queryset",
"=",
"self",
".",
"filter_queryset",
"(",
"self",
".",
"get_... | Return the object the view is displaying.
Same as rest_framework.generics.GenericAPIView, but:
- Failed assertions instead of deprecations | [
"Return",
"the",
"object",
"the",
"view",
"is",
"displaying",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/mixins.py#L33-L54 |
jwhitlock/drf-cached-instances | drf_cached_instances/mixins.py | CachedViewMixin.get_object_or_404 | def get_object_or_404(self, queryset, *filter_args, **filter_kwargs):
"""Return an object or raise a 404.
Same as Django's standard shortcut, but make sure to raise 404
if the filter_kwargs don't match the required types.
"""
if isinstance(queryset, CachedQueryset):
... | python | def get_object_or_404(self, queryset, *filter_args, **filter_kwargs):
"""Return an object or raise a 404.
Same as Django's standard shortcut, but make sure to raise 404
if the filter_kwargs don't match the required types.
"""
if isinstance(queryset, CachedQueryset):
... | [
"def",
"get_object_or_404",
"(",
"self",
",",
"queryset",
",",
"*",
"filter_args",
",",
"*",
"*",
"filter_kwargs",
")",
":",
"if",
"isinstance",
"(",
"queryset",
",",
"CachedQueryset",
")",
":",
"try",
":",
"return",
"queryset",
".",
"get",
"(",
"*",
"fi... | Return an object or raise a 404.
Same as Django's standard shortcut, but make sure to raise 404
if the filter_kwargs don't match the required types. | [
"Return",
"an",
"object",
"or",
"raise",
"a",
"404",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/mixins.py#L56-L69 |
skorokithakis/jsane | jsane/traversable.py | Empty.r | def r(self, **kwargs):
"""
Resolve the object.
This returns default (if present) or fails on an Empty.
"""
# by using kwargs we ensure that usage of positional arguments, as if
# this object were another kind of function, will fail-fast and raise
# a TypeError
... | python | def r(self, **kwargs):
"""
Resolve the object.
This returns default (if present) or fails on an Empty.
"""
# by using kwargs we ensure that usage of positional arguments, as if
# this object were another kind of function, will fail-fast and raise
# a TypeError
... | [
"def",
"r",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"# by using kwargs we ensure that usage of positional arguments, as if",
"# this object were another kind of function, will fail-fast and raise",
"# a TypeError",
"if",
"'default'",
"in",
"kwargs",
":",
"default",
"=",... | Resolve the object.
This returns default (if present) or fails on an Empty. | [
"Resolve",
"the",
"object",
"."
] | train | https://github.com/skorokithakis/jsane/blob/f63c01ab50899e207dae8427e4faef0e9d35e41e/jsane/traversable.py#L35-L54 |
skorokithakis/jsane | jsane/traversable.py | Traversable.r | def r(self, **kwargs):
"""
Resolve the object.
This will always succeed, since, if a lookup fails, an Empty
instance will be returned farther upstream.
"""
# by using kwargs we ensure that usage of positional arguments, as if
# this object were another kind of fu... | python | def r(self, **kwargs):
"""
Resolve the object.
This will always succeed, since, if a lookup fails, an Empty
instance will be returned farther upstream.
"""
# by using kwargs we ensure that usage of positional arguments, as if
# this object were another kind of fu... | [
"def",
"r",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"# by using kwargs we ensure that usage of positional arguments, as if",
"# this object were another kind of function, will fail-fast and raise",
"# a TypeError",
"kwargs",
".",
"pop",
"(",
"'default'",
",",
"None",
... | Resolve the object.
This will always succeed, since, if a lookup fails, an Empty
instance will be returned farther upstream. | [
"Resolve",
"the",
"object",
"."
] | train | https://github.com/skorokithakis/jsane/blob/f63c01ab50899e207dae8427e4faef0e9d35e41e/jsane/traversable.py#L125-L140 |
jwhitlock/drf-cached-instances | sample_poll_app/tasks.py | update_cache_for_instance | def update_cache_for_instance(
model_name, instance_pk, instance=None, version=None):
"""Update the cache for an instance, with cascading updates."""
cache = SampleCache()
invalid = cache.update_instance(model_name, instance_pk, instance, version)
for invalid_name, invalid_pk, invalid_version in... | python | def update_cache_for_instance(
model_name, instance_pk, instance=None, version=None):
"""Update the cache for an instance, with cascading updates."""
cache = SampleCache()
invalid = cache.update_instance(model_name, instance_pk, instance, version)
for invalid_name, invalid_pk, invalid_version in... | [
"def",
"update_cache_for_instance",
"(",
"model_name",
",",
"instance_pk",
",",
"instance",
"=",
"None",
",",
"version",
"=",
"None",
")",
":",
"cache",
"=",
"SampleCache",
"(",
")",
"invalid",
"=",
"cache",
".",
"update_instance",
"(",
"model_name",
",",
"i... | Update the cache for an instance, with cascading updates. | [
"Update",
"the",
"cache",
"for",
"an",
"instance",
"with",
"cascading",
"updates",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/sample_poll_app/tasks.py#L8-L15 |
jwhitlock/drf-cached-instances | drf_cached_instances/models.py | PkOnlyQueryset.values_list | def values_list(self, *args, **kwargs):
"""Return the primary keys as a list.
The only valid call is values_list('pk', flat=True)
"""
flat = kwargs.pop('flat', False)
assert flat is True
assert len(args) == 1
assert args[0] == self.model._meta.pk.name
ret... | python | def values_list(self, *args, **kwargs):
"""Return the primary keys as a list.
The only valid call is values_list('pk', flat=True)
"""
flat = kwargs.pop('flat', False)
assert flat is True
assert len(args) == 1
assert args[0] == self.model._meta.pk.name
ret... | [
"def",
"values_list",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"flat",
"=",
"kwargs",
".",
"pop",
"(",
"'flat'",
",",
"False",
")",
"assert",
"flat",
"is",
"True",
"assert",
"len",
"(",
"args",
")",
"==",
"1",
"assert",
"... | Return the primary keys as a list.
The only valid call is values_list('pk', flat=True) | [
"Return",
"the",
"primary",
"keys",
"as",
"a",
"list",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/models.py#L46-L55 |
jwhitlock/drf-cached-instances | drf_cached_instances/models.py | CachedQueryset.pks | def pks(self):
"""Lazy-load the primary keys."""
if self._primary_keys is None:
self._primary_keys = list(
self.queryset.values_list('pk', flat=True))
return self._primary_keys | python | def pks(self):
"""Lazy-load the primary keys."""
if self._primary_keys is None:
self._primary_keys = list(
self.queryset.values_list('pk', flat=True))
return self._primary_keys | [
"def",
"pks",
"(",
"self",
")",
":",
"if",
"self",
".",
"_primary_keys",
"is",
"None",
":",
"self",
".",
"_primary_keys",
"=",
"list",
"(",
"self",
".",
"queryset",
".",
"values_list",
"(",
"'pk'",
",",
"flat",
"=",
"True",
")",
")",
"return",
"self"... | Lazy-load the primary keys. | [
"Lazy",
"-",
"load",
"the",
"primary",
"keys",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/models.py#L95-L100 |
jwhitlock/drf-cached-instances | drf_cached_instances/models.py | CachedQueryset.count | def count(self):
"""Return a count of instances."""
if self._primary_keys is None:
return self.queryset.count()
else:
return len(self.pks) | python | def count(self):
"""Return a count of instances."""
if self._primary_keys is None:
return self.queryset.count()
else:
return len(self.pks) | [
"def",
"count",
"(",
"self",
")",
":",
"if",
"self",
".",
"_primary_keys",
"is",
"None",
":",
"return",
"self",
".",
"queryset",
".",
"count",
"(",
")",
"else",
":",
"return",
"len",
"(",
"self",
".",
"pks",
")"
] | Return a count of instances. | [
"Return",
"a",
"count",
"of",
"instances",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/models.py#L119-L124 |
jwhitlock/drf-cached-instances | drf_cached_instances/models.py | CachedQueryset.filter | def filter(self, **kwargs):
"""Filter the base queryset."""
assert not self._primary_keys
self.queryset = self.queryset.filter(**kwargs)
return self | python | def filter(self, **kwargs):
"""Filter the base queryset."""
assert not self._primary_keys
self.queryset = self.queryset.filter(**kwargs)
return self | [
"def",
"filter",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"assert",
"not",
"self",
".",
"_primary_keys",
"self",
".",
"queryset",
"=",
"self",
".",
"queryset",
".",
"filter",
"(",
"*",
"*",
"kwargs",
")",
"return",
"self"
] | Filter the base queryset. | [
"Filter",
"the",
"base",
"queryset",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/models.py#L126-L130 |
jwhitlock/drf-cached-instances | drf_cached_instances/models.py | CachedQueryset.get | def get(self, *args, **kwargs):
"""Return the single item from the filtered queryset."""
assert not args
assert list(kwargs.keys()) == ['pk']
pk = kwargs['pk']
model_name = self.model.__name__
object_spec = (model_name, pk, None)
instances = self.cache.get_instanc... | python | def get(self, *args, **kwargs):
"""Return the single item from the filtered queryset."""
assert not args
assert list(kwargs.keys()) == ['pk']
pk = kwargs['pk']
model_name = self.model.__name__
object_spec = (model_name, pk, None)
instances = self.cache.get_instanc... | [
"def",
"get",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"assert",
"not",
"args",
"assert",
"list",
"(",
"kwargs",
".",
"keys",
"(",
")",
")",
"==",
"[",
"'pk'",
"]",
"pk",
"=",
"kwargs",
"[",
"'pk'",
"]",
"model_name",
"... | Return the single item from the filtered queryset. | [
"Return",
"the",
"single",
"item",
"from",
"the",
"filtered",
"queryset",
"."
] | train | https://github.com/jwhitlock/drf-cached-instances/blob/ec4e8a6e1e83eeea6ec0b924b2eaa40a38d5963a/drf_cached_instances/models.py#L132-L147 |
thorgate/tg-react | tg_react/webpack.py | WebpackConstants.collect | def collect(cls):
""" Load all constant generators from settings.WEBPACK_CONSTANT_PROCESSORS
and concat their values.
"""
constants = {}
for method_path in WebpackConstants.get_constant_processors():
method = import_string(method_path)
if not callabl... | python | def collect(cls):
""" Load all constant generators from settings.WEBPACK_CONSTANT_PROCESSORS
and concat their values.
"""
constants = {}
for method_path in WebpackConstants.get_constant_processors():
method = import_string(method_path)
if not callabl... | [
"def",
"collect",
"(",
"cls",
")",
":",
"constants",
"=",
"{",
"}",
"for",
"method_path",
"in",
"WebpackConstants",
".",
"get_constant_processors",
"(",
")",
":",
"method",
"=",
"import_string",
"(",
"method_path",
")",
"if",
"not",
"callable",
"(",
"method"... | Load all constant generators from settings.WEBPACK_CONSTANT_PROCESSORS
and concat their values. | [
"Load",
"all",
"constant",
"generators",
"from",
"settings",
".",
"WEBPACK_CONSTANT_PROCESSORS",
"and",
"concat",
"their",
"values",
"."
] | train | https://github.com/thorgate/tg-react/blob/5a6e83d5a5c883f1a5ee4fda2226e81a468bdee3/tg_react/webpack.py#L23-L40 |
thorgate/tg-react | tg_react/api/accounts/serializers.py | phonenumber_validation | def phonenumber_validation(data):
""" Validates phonenumber
Similar to phonenumber_field.validators.validate_international_phonenumber() but uses a different message if the
country prefix is absent.
"""
from phonenumber_field.phonenumber import to_python
phone_number = to_python(data)
if no... | python | def phonenumber_validation(data):
""" Validates phonenumber
Similar to phonenumber_field.validators.validate_international_phonenumber() but uses a different message if the
country prefix is absent.
"""
from phonenumber_field.phonenumber import to_python
phone_number = to_python(data)
if no... | [
"def",
"phonenumber_validation",
"(",
"data",
")",
":",
"from",
"phonenumber_field",
".",
"phonenumber",
"import",
"to_python",
"phone_number",
"=",
"to_python",
"(",
"data",
")",
"if",
"not",
"phone_number",
":",
"return",
"data",
"elif",
"not",
"phone_number",
... | Validates phonenumber
Similar to phonenumber_field.validators.validate_international_phonenumber() but uses a different message if the
country prefix is absent. | [
"Validates",
"phonenumber"
] | train | https://github.com/thorgate/tg-react/blob/5a6e83d5a5c883f1a5ee4fda2226e81a468bdee3/tg_react/api/accounts/serializers.py#L98-L113 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.