repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1
value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 | partition stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
msoulier/tftpy | tftpy/TftpStates.py | TftpState.sendOACK | def sendOACK(self):
"""This method sends an OACK packet with the options from the current
context."""
log.debug("In sendOACK with options %s", self.context.options)
pkt = TftpPacketOACK()
pkt.options = self.context.options
self.context.sock.sendto(pkt.encode().buffer,
... | python | def sendOACK(self):
"""This method sends an OACK packet with the options from the current
context."""
log.debug("In sendOACK with options %s", self.context.options)
pkt = TftpPacketOACK()
pkt.options = self.context.options
self.context.sock.sendto(pkt.encode().buffer,
... | [
"def",
"sendOACK",
"(",
"self",
")",
":",
"log",
".",
"debug",
"(",
"\"In sendOACK with options %s\"",
",",
"self",
".",
"context",
".",
"options",
")",
"pkt",
"=",
"TftpPacketOACK",
"(",
")",
"pkt",
".",
"options",
"=",
"self",
".",
"context",
".",
"opt... | This method sends an OACK packet with the options from the current
context. | [
"This",
"method",
"sends",
"an",
"OACK",
"packet",
"with",
"the",
"options",
"from",
"the",
"current",
"context",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L142-L151 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpState.resendLast | def resendLast(self):
"Resend the last sent packet due to a timeout."
log.warning("Resending packet %s on sessions %s"
% (self.context.last_pkt, self))
self.context.metrics.resent_bytes += len(self.context.last_pkt.buffer)
self.context.metrics.add_dup(self.context.last_pkt)
... | python | def resendLast(self):
"Resend the last sent packet due to a timeout."
log.warning("Resending packet %s on sessions %s"
% (self.context.last_pkt, self))
self.context.metrics.resent_bytes += len(self.context.last_pkt.buffer)
self.context.metrics.add_dup(self.context.last_pkt)
... | [
"def",
"resendLast",
"(",
"self",
")",
":",
"log",
".",
"warning",
"(",
"\"Resending packet %s on sessions %s\"",
"%",
"(",
"self",
".",
"context",
".",
"last_pkt",
",",
"self",
")",
")",
"self",
".",
"context",
".",
"metrics",
".",
"resent_bytes",
"+=",
"... | Resend the last sent packet due to a timeout. | [
"Resend",
"the",
"last",
"sent",
"packet",
"due",
"to",
"a",
"timeout",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L153-L168 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpState.handleDat | def handleDat(self, pkt):
"""This method handles a DAT packet during a client download, or a
server upload."""
log.info("Handling DAT packet - block %d" % pkt.blocknumber)
log.debug("Expecting block %s", self.context.next_block)
if pkt.blocknumber == self.context.next_block:
... | python | def handleDat(self, pkt):
"""This method handles a DAT packet during a client download, or a
server upload."""
log.info("Handling DAT packet - block %d" % pkt.blocknumber)
log.debug("Expecting block %s", self.context.next_block)
if pkt.blocknumber == self.context.next_block:
... | [
"def",
"handleDat",
"(",
"self",
",",
"pkt",
")",
":",
"log",
".",
"info",
"(",
"\"Handling DAT packet - block %d\"",
"%",
"pkt",
".",
"blocknumber",
")",
"log",
".",
"debug",
"(",
"\"Expecting block %s\"",
",",
"self",
".",
"context",
".",
"next_block",
")"... | This method handles a DAT packet during a client download, or a
server upload. | [
"This",
"method",
"handles",
"a",
"DAT",
"packet",
"during",
"a",
"client",
"download",
"or",
"a",
"server",
"upload",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L170-L207 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpServerState.serverInitial | def serverInitial(self, pkt, raddress, rport):
"""This method performs initial setup for a server context transfer,
put here to refactor code out of the TftpStateServerRecvRRQ and
TftpStateServerRecvWRQ classes, since their initial setup is
identical. The method returns a boolean, sendoa... | python | def serverInitial(self, pkt, raddress, rport):
"""This method performs initial setup for a server context transfer,
put here to refactor code out of the TftpStateServerRecvRRQ and
TftpStateServerRecvWRQ classes, since their initial setup is
identical. The method returns a boolean, sendoa... | [
"def",
"serverInitial",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"options",
"=",
"pkt",
".",
"options",
"sendoack",
"=",
"False",
"if",
"not",
"self",
".",
"context",
".",
"tidport",
":",
"self",
".",
"context",
".",
"tidport",... | This method performs initial setup for a server context transfer,
put here to refactor code out of the TftpStateServerRecvRRQ and
TftpStateServerRecvWRQ classes, since their initial setup is
identical. The method returns a boolean, sendoack, to indicate whether
it is required to send an ... | [
"This",
"method",
"performs",
"initial",
"setup",
"for",
"a",
"server",
"context",
"transfer",
"put",
"here",
"to",
"refactor",
"code",
"out",
"of",
"the",
"TftpStateServerRecvRRQ",
"and",
"TftpStateServerRecvWRQ",
"classes",
"since",
"their",
"initial",
"setup",
... | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L219-L291 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateServerRecvRRQ.handle | def handle(self, pkt, raddress, rport):
"Handle an initial RRQ packet as a server."
log.debug("In TftpStateServerRecvRRQ.handle")
sendoack = self.serverInitial(pkt, raddress, rport)
path = self.full_path
log.info("Opening file %s for reading" % path)
if os.path.exists(pat... | python | def handle(self, pkt, raddress, rport):
"Handle an initial RRQ packet as a server."
log.debug("In TftpStateServerRecvRRQ.handle")
sendoack = self.serverInitial(pkt, raddress, rport)
path = self.full_path
log.info("Opening file %s for reading" % path)
if os.path.exists(pat... | [
"def",
"handle",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"log",
".",
"debug",
"(",
"\"In TftpStateServerRecvRRQ.handle\"",
")",
"sendoack",
"=",
"self",
".",
"serverInitial",
"(",
"pkt",
",",
"raddress",
",",
"rport",
")",
"path",... | Handle an initial RRQ packet as a server. | [
"Handle",
"an",
"initial",
"RRQ",
"packet",
"as",
"a",
"server",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L297-L344 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateServerRecvWRQ.make_subdirs | def make_subdirs(self):
"""The purpose of this method is to, if necessary, create all of the
subdirectories leading up to the file to the written."""
# Pull off everything below the root.
subpath = self.full_path[len(self.context.root):]
log.debug("make_subdirs: subpath is %s", s... | python | def make_subdirs(self):
"""The purpose of this method is to, if necessary, create all of the
subdirectories leading up to the file to the written."""
# Pull off everything below the root.
subpath = self.full_path[len(self.context.root):]
log.debug("make_subdirs: subpath is %s", s... | [
"def",
"make_subdirs",
"(",
"self",
")",
":",
"# Pull off everything below the root.",
"subpath",
"=",
"self",
".",
"full_path",
"[",
"len",
"(",
"self",
".",
"context",
".",
"root",
")",
":",
"]",
"log",
".",
"debug",
"(",
"\"make_subdirs: subpath is %s\"",
"... | The purpose of this method is to, if necessary, create all of the
subdirectories leading up to the file to the written. | [
"The",
"purpose",
"of",
"this",
"method",
"is",
"to",
"if",
"necessary",
"create",
"all",
"of",
"the",
"subdirectories",
"leading",
"up",
"to",
"the",
"file",
"to",
"the",
"written",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L352-L369 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateServerRecvWRQ.handle | def handle(self, pkt, raddress, rport):
"Handle an initial WRQ packet as a server."
log.debug("In TftpStateServerRecvWRQ.handle")
sendoack = self.serverInitial(pkt, raddress, rport)
path = self.full_path
if self.context.upload_open:
f = self.context.upload_open(path, ... | python | def handle(self, pkt, raddress, rport):
"Handle an initial WRQ packet as a server."
log.debug("In TftpStateServerRecvWRQ.handle")
sendoack = self.serverInitial(pkt, raddress, rport)
path = self.full_path
if self.context.upload_open:
f = self.context.upload_open(path, ... | [
"def",
"handle",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"log",
".",
"debug",
"(",
"\"In TftpStateServerRecvWRQ.handle\"",
")",
"sendoack",
"=",
"self",
".",
"serverInitial",
"(",
"pkt",
",",
"raddress",
",",
"rport",
")",
"path",... | Handle an initial WRQ packet as a server. | [
"Handle",
"an",
"initial",
"WRQ",
"packet",
"as",
"a",
"server",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L371-L407 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateExpectACK.handle | def handle(self, pkt, raddress, rport):
"Handle a packet, hopefully an ACK since we just sent a DAT."
if isinstance(pkt, TftpPacketACK):
log.debug("Received ACK for packet %d" % pkt.blocknumber)
# Is this an ack to the one we just sent?
if self.context.next_block == p... | python | def handle(self, pkt, raddress, rport):
"Handle a packet, hopefully an ACK since we just sent a DAT."
if isinstance(pkt, TftpPacketACK):
log.debug("Received ACK for packet %d" % pkt.blocknumber)
# Is this an ack to the one we just sent?
if self.context.next_block == p... | [
"def",
"handle",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"if",
"isinstance",
"(",
"pkt",
",",
"TftpPacketACK",
")",
":",
"log",
".",
"debug",
"(",
"\"Received ACK for packet %d\"",
"%",
"pkt",
".",
"blocknumber",
")",
"# Is this a... | Handle a packet, hopefully an ACK since we just sent a DAT. | [
"Handle",
"a",
"packet",
"hopefully",
"an",
"ACK",
"since",
"we",
"just",
"sent",
"a",
"DAT",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L438-L469 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateExpectDAT.handle | def handle(self, pkt, raddress, rport):
"""Handle the packet in response to an ACK, which should be a DAT."""
if isinstance(pkt, TftpPacketDAT):
return self.handleDat(pkt)
# Every other packet type is a problem.
elif isinstance(pkt, TftpPacketACK):
# Umm, we ACK,... | python | def handle(self, pkt, raddress, rport):
"""Handle the packet in response to an ACK, which should be a DAT."""
if isinstance(pkt, TftpPacketDAT):
return self.handleDat(pkt)
# Every other packet type is a problem.
elif isinstance(pkt, TftpPacketACK):
# Umm, we ACK,... | [
"def",
"handle",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"if",
"isinstance",
"(",
"pkt",
",",
"TftpPacketDAT",
")",
":",
"return",
"self",
".",
"handleDat",
"(",
"pkt",
")",
"# Every other packet type is a problem.",
"elif",
"isinst... | Handle the packet in response to an ACK, which should be a DAT. | [
"Handle",
"the",
"packet",
"in",
"response",
"to",
"an",
"ACK",
"which",
"should",
"be",
"a",
"DAT",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L473-L494 | train |
msoulier/tftpy | tftpy/TftpStates.py | TftpStateSentRRQ.handle | def handle(self, pkt, raddress, rport):
"""Handle the packet in response to an RRQ to the server."""
if not self.context.tidport:
self.context.tidport = rport
log.info("Set remote port for session to %s" % rport)
# Now check the packet type and dispatch it properly.
... | python | def handle(self, pkt, raddress, rport):
"""Handle the packet in response to an RRQ to the server."""
if not self.context.tidport:
self.context.tidport = rport
log.info("Set remote port for session to %s" % rport)
# Now check the packet type and dispatch it properly.
... | [
"def",
"handle",
"(",
"self",
",",
"pkt",
",",
"raddress",
",",
"rport",
")",
":",
"if",
"not",
"self",
".",
"context",
".",
"tidport",
":",
"self",
".",
"context",
".",
"tidport",
"=",
"rport",
"log",
".",
"info",
"(",
"\"Set remote port for session to ... | Handle the packet in response to an RRQ to the server. | [
"Handle",
"the",
"packet",
"in",
"response",
"to",
"an",
"RRQ",
"to",
"the",
"server",
"."
] | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpStates.py#L556-L611 | train |
msoulier/tftpy | tftpy/TftpServer.py | TftpServer.stop | def stop(self, now=False):
"""Stop the server gracefully. Do not take any new transfers,
but complete the existing ones. If force is True, drop everything
and stop. Note, immediately will not interrupt the select loop, it
will happen when the server returns on ready data, or a timeout.
... | python | def stop(self, now=False):
"""Stop the server gracefully. Do not take any new transfers,
but complete the existing ones. If force is True, drop everything
and stop. Note, immediately will not interrupt the select loop, it
will happen when the server returns on ready data, or a timeout.
... | [
"def",
"stop",
"(",
"self",
",",
"now",
"=",
"False",
")",
":",
"if",
"now",
":",
"self",
".",
"shutdown_immediately",
"=",
"True",
"else",
":",
"self",
".",
"shutdown_gracefully",
"=",
"True"
] | Stop the server gracefully. Do not take any new transfers,
but complete the existing ones. If force is True, drop everything
and stop. Note, immediately will not interrupt the select loop, it
will happen when the server returns on ready data, or a timeout.
ie. SOCK_TIMEOUT | [
"Stop",
"the",
"server",
"gracefully",
".",
"Do",
"not",
"take",
"any",
"new",
"transfers",
"but",
"complete",
"the",
"existing",
"ones",
".",
"If",
"force",
"is",
"True",
"drop",
"everything",
"and",
"stop",
".",
"Note",
"immediately",
"will",
"not",
"int... | af2f2fe89a3bf45748b78703820efb0986a8207a | https://github.com/msoulier/tftpy/blob/af2f2fe89a3bf45748b78703820efb0986a8207a/tftpy/TftpServer.py#L249-L258 | train |
avelkoski/FRB | fred/helpers/__init__.py | _fetch | def _fetch(url, ssl_verify = True):
"""
Helper funcation to fetch content from a given url.
"""
req = Request(url)
if ssl_verify:
page = urlopen(req)
else:
ctx = ssl.create_default_context()
ctx.check_hostname = False
ctx.verify_mode = ssl.CERT_NONE
page ... | python | def _fetch(url, ssl_verify = True):
"""
Helper funcation to fetch content from a given url.
"""
req = Request(url)
if ssl_verify:
page = urlopen(req)
else:
ctx = ssl.create_default_context()
ctx.check_hostname = False
ctx.verify_mode = ssl.CERT_NONE
page ... | [
"def",
"_fetch",
"(",
"url",
",",
"ssl_verify",
"=",
"True",
")",
":",
"req",
"=",
"Request",
"(",
"url",
")",
"if",
"ssl_verify",
":",
"page",
"=",
"urlopen",
"(",
"req",
")",
"else",
":",
"ctx",
"=",
"ssl",
".",
"create_default_context",
"(",
")",
... | Helper funcation to fetch content from a given url. | [
"Helper",
"funcation",
"to",
"fetch",
"content",
"from",
"a",
"given",
"url",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L25-L40 | train |
avelkoski/FRB | fred/helpers/__init__.py | _dict | def _dict(content):
"""
Helper funcation that converts text-based get response
to a python dictionary for additional manipulation.
"""
if _has_pandas:
data = _data_frame(content).to_dict(orient='records')
else:
response = loads(content)
key = [x for x in response.keys() i... | python | def _dict(content):
"""
Helper funcation that converts text-based get response
to a python dictionary for additional manipulation.
"""
if _has_pandas:
data = _data_frame(content).to_dict(orient='records')
else:
response = loads(content)
key = [x for x in response.keys() i... | [
"def",
"_dict",
"(",
"content",
")",
":",
"if",
"_has_pandas",
":",
"data",
"=",
"_data_frame",
"(",
"content",
")",
".",
"to_dict",
"(",
"orient",
"=",
"'records'",
")",
"else",
":",
"response",
"=",
"loads",
"(",
"content",
")",
"key",
"=",
"[",
"x... | Helper funcation that converts text-based get response
to a python dictionary for additional manipulation. | [
"Helper",
"funcation",
"that",
"converts",
"text",
"-",
"based",
"get",
"response",
"to",
"a",
"python",
"dictionary",
"for",
"additional",
"manipulation",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L61-L72 | train |
avelkoski/FRB | fred/helpers/__init__.py | _data_frame | def _data_frame(content):
"""
Helper funcation that converts text-based get response
to a pandas dataframe for additional manipulation.
"""
response = loads(content)
key = [x for x in response.keys() if x in c.response_data][0]
frame = DataFrame(response[key])
final_frame = _convert(fram... | python | def _data_frame(content):
"""
Helper funcation that converts text-based get response
to a pandas dataframe for additional manipulation.
"""
response = loads(content)
key = [x for x in response.keys() if x in c.response_data][0]
frame = DataFrame(response[key])
final_frame = _convert(fram... | [
"def",
"_data_frame",
"(",
"content",
")",
":",
"response",
"=",
"loads",
"(",
"content",
")",
"key",
"=",
"[",
"x",
"for",
"x",
"in",
"response",
".",
"keys",
"(",
")",
"if",
"x",
"in",
"c",
".",
"response_data",
"]",
"[",
"0",
"]",
"frame",
"="... | Helper funcation that converts text-based get response
to a pandas dataframe for additional manipulation. | [
"Helper",
"funcation",
"that",
"converts",
"text",
"-",
"based",
"get",
"response",
"to",
"a",
"pandas",
"dataframe",
"for",
"additional",
"manipulation",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L74-L83 | train |
avelkoski/FRB | fred/helpers/__init__.py | _tab | def _tab(content):
"""
Helper funcation that converts text-based get response
to tab separated values for additional manipulation.
"""
response = _data_frame(content).to_csv(index=False,sep='\t')
return response | python | def _tab(content):
"""
Helper funcation that converts text-based get response
to tab separated values for additional manipulation.
"""
response = _data_frame(content).to_csv(index=False,sep='\t')
return response | [
"def",
"_tab",
"(",
"content",
")",
":",
"response",
"=",
"_data_frame",
"(",
"content",
")",
".",
"to_csv",
"(",
"index",
"=",
"False",
",",
"sep",
"=",
"'\\t'",
")",
"return",
"response"
] | Helper funcation that converts text-based get response
to tab separated values for additional manipulation. | [
"Helper",
"funcation",
"that",
"converts",
"text",
"-",
"based",
"get",
"response",
"to",
"tab",
"separated",
"values",
"for",
"additional",
"manipulation",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L93-L99 | train |
avelkoski/FRB | fred/helpers/__init__.py | _pipe | def _pipe(content):
"""
Helper funcation that converts text-based get response
to pipe separated values for additional manipulation.
"""
response = _data_frame(content).to_csv(index=False,sep='|')
return response | python | def _pipe(content):
"""
Helper funcation that converts text-based get response
to pipe separated values for additional manipulation.
"""
response = _data_frame(content).to_csv(index=False,sep='|')
return response | [
"def",
"_pipe",
"(",
"content",
")",
":",
"response",
"=",
"_data_frame",
"(",
"content",
")",
".",
"to_csv",
"(",
"index",
"=",
"False",
",",
"sep",
"=",
"'|'",
")",
"return",
"response"
] | Helper funcation that converts text-based get response
to pipe separated values for additional manipulation. | [
"Helper",
"funcation",
"that",
"converts",
"text",
"-",
"based",
"get",
"response",
"to",
"pipe",
"separated",
"values",
"for",
"additional",
"manipulation",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L101-L107 | train |
avelkoski/FRB | fred/helpers/__init__.py | _get_request | def _get_request(url_root,api_key,path,response_type,params, ssl_verify):
"""
Helper funcation that requests a get response from FRED.
"""
url = _url_builder(url_root,api_key,path,params)
content = _fetch(url, ssl_verify)
response = _dispatch(response_type)(content)
return response | python | def _get_request(url_root,api_key,path,response_type,params, ssl_verify):
"""
Helper funcation that requests a get response from FRED.
"""
url = _url_builder(url_root,api_key,path,params)
content = _fetch(url, ssl_verify)
response = _dispatch(response_type)(content)
return response | [
"def",
"_get_request",
"(",
"url_root",
",",
"api_key",
",",
"path",
",",
"response_type",
",",
"params",
",",
"ssl_verify",
")",
":",
"url",
"=",
"_url_builder",
"(",
"url_root",
",",
"api_key",
",",
"path",
",",
"params",
")",
"content",
"=",
"_fetch",
... | Helper funcation that requests a get response from FRED. | [
"Helper",
"funcation",
"that",
"requests",
"a",
"get",
"response",
"from",
"FRED",
"."
] | 692bcf576e17bd1a81db2b7644f4f61aeb39e5c7 | https://github.com/avelkoski/FRB/blob/692bcf576e17bd1a81db2b7644f4f61aeb39e5c7/fred/helpers/__init__.py#L141-L148 | train |
NicolasLM/atoma | atoma/atom.py | parse_atom_file | def parse_atom_file(filename: str) -> AtomFeed:
"""Parse an Atom feed from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_atom(root) | python | def parse_atom_file(filename: str) -> AtomFeed:
"""Parse an Atom feed from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_atom(root) | [
"def",
"parse_atom_file",
"(",
"filename",
":",
"str",
")",
"->",
"AtomFeed",
":",
"root",
"=",
"parse_xml",
"(",
"filename",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_atom",
"(",
"root",
")"
] | Parse an Atom feed from a local XML file. | [
"Parse",
"an",
"Atom",
"feed",
"from",
"a",
"local",
"XML",
"file",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/atom.py#L275-L278 | train |
NicolasLM/atoma | atoma/atom.py | parse_atom_bytes | def parse_atom_bytes(data: bytes) -> AtomFeed:
"""Parse an Atom feed from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_atom(root) | python | def parse_atom_bytes(data: bytes) -> AtomFeed:
"""Parse an Atom feed from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_atom(root) | [
"def",
"parse_atom_bytes",
"(",
"data",
":",
"bytes",
")",
"->",
"AtomFeed",
":",
"root",
"=",
"parse_xml",
"(",
"BytesIO",
"(",
"data",
")",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_atom",
"(",
"root",
")"
] | Parse an Atom feed from a byte-string containing XML data. | [
"Parse",
"an",
"Atom",
"feed",
"from",
"a",
"byte",
"-",
"string",
"containing",
"XML",
"data",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/atom.py#L281-L284 | train |
NicolasLM/atoma | atoma/rss.py | _get_link | def _get_link(element: Element) -> Optional[str]:
"""Attempt to retrieve item link.
Use the GUID as a fallback if it is a permalink.
"""
link = get_text(element, 'link')
if link is not None:
return link
guid = get_child(element, 'guid')
if guid is not None and guid.attrib.get('isPe... | python | def _get_link(element: Element) -> Optional[str]:
"""Attempt to retrieve item link.
Use the GUID as a fallback if it is a permalink.
"""
link = get_text(element, 'link')
if link is not None:
return link
guid = get_child(element, 'guid')
if guid is not None and guid.attrib.get('isPe... | [
"def",
"_get_link",
"(",
"element",
":",
"Element",
")",
"->",
"Optional",
"[",
"str",
"]",
":",
"link",
"=",
"get_text",
"(",
"element",
",",
"'link'",
")",
"if",
"link",
"is",
"not",
"None",
":",
"return",
"link",
"guid",
"=",
"get_child",
"(",
"el... | Attempt to retrieve item link.
Use the GUID as a fallback if it is a permalink. | [
"Attempt",
"to",
"retrieve",
"item",
"link",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/rss.py#L118-L131 | train |
NicolasLM/atoma | atoma/rss.py | parse_rss_file | def parse_rss_file(filename: str) -> RSSChannel:
"""Parse an RSS feed from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_rss(root) | python | def parse_rss_file(filename: str) -> RSSChannel:
"""Parse an RSS feed from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_rss(root) | [
"def",
"parse_rss_file",
"(",
"filename",
":",
"str",
")",
"->",
"RSSChannel",
":",
"root",
"=",
"parse_xml",
"(",
"filename",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_rss",
"(",
"root",
")"
] | Parse an RSS feed from a local XML file. | [
"Parse",
"an",
"RSS",
"feed",
"from",
"a",
"local",
"XML",
"file",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/rss.py#L212-L215 | train |
NicolasLM/atoma | atoma/rss.py | parse_rss_bytes | def parse_rss_bytes(data: bytes) -> RSSChannel:
"""Parse an RSS feed from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_rss(root) | python | def parse_rss_bytes(data: bytes) -> RSSChannel:
"""Parse an RSS feed from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_rss(root) | [
"def",
"parse_rss_bytes",
"(",
"data",
":",
"bytes",
")",
"->",
"RSSChannel",
":",
"root",
"=",
"parse_xml",
"(",
"BytesIO",
"(",
"data",
")",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_rss",
"(",
"root",
")"
] | Parse an RSS feed from a byte-string containing XML data. | [
"Parse",
"an",
"RSS",
"feed",
"from",
"a",
"byte",
"-",
"string",
"containing",
"XML",
"data",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/rss.py#L218-L221 | train |
NicolasLM/atoma | atoma/json_feed.py | parse_json_feed_file | def parse_json_feed_file(filename: str) -> JSONFeed:
"""Parse a JSON feed from a local json file."""
with open(filename) as f:
try:
root = json.load(f)
except json.decoder.JSONDecodeError:
raise FeedJSONError('Not a valid JSON document')
return parse_json_feed(root) | python | def parse_json_feed_file(filename: str) -> JSONFeed:
"""Parse a JSON feed from a local json file."""
with open(filename) as f:
try:
root = json.load(f)
except json.decoder.JSONDecodeError:
raise FeedJSONError('Not a valid JSON document')
return parse_json_feed(root) | [
"def",
"parse_json_feed_file",
"(",
"filename",
":",
"str",
")",
"->",
"JSONFeed",
":",
"with",
"open",
"(",
"filename",
")",
"as",
"f",
":",
"try",
":",
"root",
"=",
"json",
".",
"load",
"(",
"f",
")",
"except",
"json",
".",
"decoder",
".",
"JSONDec... | Parse a JSON feed from a local json file. | [
"Parse",
"a",
"JSON",
"feed",
"from",
"a",
"local",
"json",
"file",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/json_feed.py#L205-L213 | train |
NicolasLM/atoma | atoma/json_feed.py | parse_json_feed_bytes | def parse_json_feed_bytes(data: bytes) -> JSONFeed:
"""Parse a JSON feed from a byte-string containing JSON data."""
try:
root = json.loads(data)
except json.decoder.JSONDecodeError:
raise FeedJSONError('Not a valid JSON document')
return parse_json_feed(root) | python | def parse_json_feed_bytes(data: bytes) -> JSONFeed:
"""Parse a JSON feed from a byte-string containing JSON data."""
try:
root = json.loads(data)
except json.decoder.JSONDecodeError:
raise FeedJSONError('Not a valid JSON document')
return parse_json_feed(root) | [
"def",
"parse_json_feed_bytes",
"(",
"data",
":",
"bytes",
")",
"->",
"JSONFeed",
":",
"try",
":",
"root",
"=",
"json",
".",
"loads",
"(",
"data",
")",
"except",
"json",
".",
"decoder",
".",
"JSONDecodeError",
":",
"raise",
"FeedJSONError",
"(",
"'Not a va... | Parse a JSON feed from a byte-string containing JSON data. | [
"Parse",
"a",
"JSON",
"feed",
"from",
"a",
"byte",
"-",
"string",
"containing",
"JSON",
"data",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/json_feed.py#L216-L223 | train |
NicolasLM/atoma | atoma/opml.py | parse_opml_file | def parse_opml_file(filename: str) -> OPML:
"""Parse an OPML document from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_opml(root) | python | def parse_opml_file(filename: str) -> OPML:
"""Parse an OPML document from a local XML file."""
root = parse_xml(filename).getroot()
return _parse_opml(root) | [
"def",
"parse_opml_file",
"(",
"filename",
":",
"str",
")",
"->",
"OPML",
":",
"root",
"=",
"parse_xml",
"(",
"filename",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_opml",
"(",
"root",
")"
] | Parse an OPML document from a local XML file. | [
"Parse",
"an",
"OPML",
"document",
"from",
"a",
"local",
"XML",
"file",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/opml.py#L82-L85 | train |
NicolasLM/atoma | atoma/opml.py | parse_opml_bytes | def parse_opml_bytes(data: bytes) -> OPML:
"""Parse an OPML document from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_opml(root) | python | def parse_opml_bytes(data: bytes) -> OPML:
"""Parse an OPML document from a byte-string containing XML data."""
root = parse_xml(BytesIO(data)).getroot()
return _parse_opml(root) | [
"def",
"parse_opml_bytes",
"(",
"data",
":",
"bytes",
")",
"->",
"OPML",
":",
"root",
"=",
"parse_xml",
"(",
"BytesIO",
"(",
"data",
")",
")",
".",
"getroot",
"(",
")",
"return",
"_parse_opml",
"(",
"root",
")"
] | Parse an OPML document from a byte-string containing XML data. | [
"Parse",
"an",
"OPML",
"document",
"from",
"a",
"byte",
"-",
"string",
"containing",
"XML",
"data",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/opml.py#L88-L91 | train |
NicolasLM/atoma | atoma/opml.py | get_feed_list | def get_feed_list(opml_obj: OPML) -> List[str]:
"""Walk an OPML document to extract the list of feed it contains."""
rv = list()
def collect(obj):
for outline in obj.outlines:
if outline.type == 'rss' and outline.xml_url:
rv.append(outline.xml_url)
if outlin... | python | def get_feed_list(opml_obj: OPML) -> List[str]:
"""Walk an OPML document to extract the list of feed it contains."""
rv = list()
def collect(obj):
for outline in obj.outlines:
if outline.type == 'rss' and outline.xml_url:
rv.append(outline.xml_url)
if outlin... | [
"def",
"get_feed_list",
"(",
"opml_obj",
":",
"OPML",
")",
"->",
"List",
"[",
"str",
"]",
":",
"rv",
"=",
"list",
"(",
")",
"def",
"collect",
"(",
"obj",
")",
":",
"for",
"outline",
"in",
"obj",
".",
"outlines",
":",
"if",
"outline",
".",
"type",
... | Walk an OPML document to extract the list of feed it contains. | [
"Walk",
"an",
"OPML",
"document",
"to",
"extract",
"the",
"list",
"of",
"feed",
"it",
"contains",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/opml.py#L94-L107 | train |
NicolasLM/atoma | atoma/simple.py | simple_parse_file | def simple_parse_file(filename: str) -> Feed:
"""Parse an Atom, RSS or JSON feed from a local file."""
pairs = (
(rss.parse_rss_file, _adapt_rss_channel),
(atom.parse_atom_file, _adapt_atom_feed),
(json_feed.parse_json_feed_file, _adapt_json_feed)
)
return _simple_parse(pairs, fi... | python | def simple_parse_file(filename: str) -> Feed:
"""Parse an Atom, RSS or JSON feed from a local file."""
pairs = (
(rss.parse_rss_file, _adapt_rss_channel),
(atom.parse_atom_file, _adapt_atom_feed),
(json_feed.parse_json_feed_file, _adapt_json_feed)
)
return _simple_parse(pairs, fi... | [
"def",
"simple_parse_file",
"(",
"filename",
":",
"str",
")",
"->",
"Feed",
":",
"pairs",
"=",
"(",
"(",
"rss",
".",
"parse_rss_file",
",",
"_adapt_rss_channel",
")",
",",
"(",
"atom",
".",
"parse_atom_file",
",",
"_adapt_atom_feed",
")",
",",
"(",
"json_f... | Parse an Atom, RSS or JSON feed from a local file. | [
"Parse",
"an",
"Atom",
"RSS",
"or",
"JSON",
"feed",
"from",
"a",
"local",
"file",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/simple.py#L207-L214 | train |
NicolasLM/atoma | atoma/simple.py | simple_parse_bytes | def simple_parse_bytes(data: bytes) -> Feed:
"""Parse an Atom, RSS or JSON feed from a byte-string containing data."""
pairs = (
(rss.parse_rss_bytes, _adapt_rss_channel),
(atom.parse_atom_bytes, _adapt_atom_feed),
(json_feed.parse_json_feed_bytes, _adapt_json_feed)
)
return _sim... | python | def simple_parse_bytes(data: bytes) -> Feed:
"""Parse an Atom, RSS or JSON feed from a byte-string containing data."""
pairs = (
(rss.parse_rss_bytes, _adapt_rss_channel),
(atom.parse_atom_bytes, _adapt_atom_feed),
(json_feed.parse_json_feed_bytes, _adapt_json_feed)
)
return _sim... | [
"def",
"simple_parse_bytes",
"(",
"data",
":",
"bytes",
")",
"->",
"Feed",
":",
"pairs",
"=",
"(",
"(",
"rss",
".",
"parse_rss_bytes",
",",
"_adapt_rss_channel",
")",
",",
"(",
"atom",
".",
"parse_atom_bytes",
",",
"_adapt_atom_feed",
")",
",",
"(",
"json_... | Parse an Atom, RSS or JSON feed from a byte-string containing data. | [
"Parse",
"an",
"Atom",
"RSS",
"or",
"JSON",
"feed",
"from",
"a",
"byte",
"-",
"string",
"containing",
"data",
"."
] | 16c6956112f975eb2ce774b2d5f8e9ddffde569f | https://github.com/NicolasLM/atoma/blob/16c6956112f975eb2ce774b2d5f8e9ddffde569f/atoma/simple.py#L217-L224 | train |
Atomistica/atomistica | src/python/atomistica/deformation.py | get_shear_distance | def get_shear_distance(a):
"""
Returns the distance a volume has moved during simple shear. Considers
either Lees-Edwards boundary conditions or sheared cells.
"""
cx, cy, cz = a.cell
if 'shear_dx' in a.info:
assert abs(cx[1]) < 1e-12, 'cx[1] = {0}'.format(cx[1])
assert abs(cx[2]... | python | def get_shear_distance(a):
"""
Returns the distance a volume has moved during simple shear. Considers
either Lees-Edwards boundary conditions or sheared cells.
"""
cx, cy, cz = a.cell
if 'shear_dx' in a.info:
assert abs(cx[1]) < 1e-12, 'cx[1] = {0}'.format(cx[1])
assert abs(cx[2]... | [
"def",
"get_shear_distance",
"(",
"a",
")",
":",
"cx",
",",
"cy",
",",
"cz",
"=",
"a",
".",
"cell",
"if",
"'shear_dx'",
"in",
"a",
".",
"info",
":",
"assert",
"abs",
"(",
"cx",
"[",
"1",
"]",
")",
"<",
"1e-12",
",",
"'cx[1] = {0}'",
".",
"format"... | Returns the distance a volume has moved during simple shear. Considers
either Lees-Edwards boundary conditions or sheared cells. | [
"Returns",
"the",
"distance",
"a",
"volume",
"has",
"moved",
"during",
"simple",
"shear",
".",
"Considers",
"either",
"Lees",
"-",
"Edwards",
"boundary",
"conditions",
"or",
"sheared",
"cells",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/deformation.py#L30-L50 | train |
Atomistica/atomistica | src/python/atomistica/atomic_strain.py | array_inverse | def array_inverse(A):
"""
Compute inverse for each matrix in a list of matrices.
This is faster than calling numpy.linalg.inv for each matrix.
"""
A = np.ascontiguousarray(A, dtype=float)
b = np.identity(A.shape[2], dtype=A.dtype)
n_eq = A.shape[1]
n_rhs = A.shape[2]
pivots = np.zer... | python | def array_inverse(A):
"""
Compute inverse for each matrix in a list of matrices.
This is faster than calling numpy.linalg.inv for each matrix.
"""
A = np.ascontiguousarray(A, dtype=float)
b = np.identity(A.shape[2], dtype=A.dtype)
n_eq = A.shape[1]
n_rhs = A.shape[2]
pivots = np.zer... | [
"def",
"array_inverse",
"(",
"A",
")",
":",
"A",
"=",
"np",
".",
"ascontiguousarray",
"(",
"A",
",",
"dtype",
"=",
"float",
")",
"b",
"=",
"np",
".",
"identity",
"(",
"A",
".",
"shape",
"[",
"2",
"]",
",",
"dtype",
"=",
"A",
".",
"dtype",
")",
... | Compute inverse for each matrix in a list of matrices.
This is faster than calling numpy.linalg.inv for each matrix. | [
"Compute",
"inverse",
"for",
"each",
"matrix",
"in",
"a",
"list",
"of",
"matrices",
".",
"This",
"is",
"faster",
"than",
"calling",
"numpy",
".",
"linalg",
".",
"inv",
"for",
"each",
"matrix",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/atomic_strain.py#L66-L86 | train |
Atomistica/atomistica | src/python/atomistica/atomic_strain.py | get_delta_plus_epsilon | def get_delta_plus_epsilon(nat, i_now, dr_now, dr_old):
"""
Calculate delta_ij+epsilon_ij, i.e. the deformation gradient matrix
"""
XIJ = get_XIJ(nat, i_now, dr_now, dr_old)
YIJ = get_YIJ(nat, i_now, dr_old)
YIJ_invert = array_inverse(YIJ)
# Perform sum_k X_ik Y_jk^-1
epsilon = np.sum(... | python | def get_delta_plus_epsilon(nat, i_now, dr_now, dr_old):
"""
Calculate delta_ij+epsilon_ij, i.e. the deformation gradient matrix
"""
XIJ = get_XIJ(nat, i_now, dr_now, dr_old)
YIJ = get_YIJ(nat, i_now, dr_old)
YIJ_invert = array_inverse(YIJ)
# Perform sum_k X_ik Y_jk^-1
epsilon = np.sum(... | [
"def",
"get_delta_plus_epsilon",
"(",
"nat",
",",
"i_now",
",",
"dr_now",
",",
"dr_old",
")",
":",
"XIJ",
"=",
"get_XIJ",
"(",
"nat",
",",
"i_now",
",",
"dr_now",
",",
"dr_old",
")",
"YIJ",
"=",
"get_YIJ",
"(",
"nat",
",",
"i_now",
",",
"dr_old",
")"... | Calculate delta_ij+epsilon_ij, i.e. the deformation gradient matrix | [
"Calculate",
"delta_ij",
"+",
"epsilon_ij",
"i",
".",
"e",
".",
"the",
"deformation",
"gradient",
"matrix"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/atomic_strain.py#L89-L101 | train |
Atomistica/atomistica | src/python/atomistica/atomic_strain.py | get_D_square_min | def get_D_square_min(atoms_now, atoms_old, i_now, j_now, delta_plus_epsilon=None):
"""
Calculate the D^2_min norm of Falk and Langer
"""
nat = len(atoms_now)
assert len(atoms_now) == len(atoms_old)
pos_now = atoms_now.positions
pos_old = atoms_old.positions
# Compute current and old di... | python | def get_D_square_min(atoms_now, atoms_old, i_now, j_now, delta_plus_epsilon=None):
"""
Calculate the D^2_min norm of Falk and Langer
"""
nat = len(atoms_now)
assert len(atoms_now) == len(atoms_old)
pos_now = atoms_now.positions
pos_old = atoms_old.positions
# Compute current and old di... | [
"def",
"get_D_square_min",
"(",
"atoms_now",
",",
"atoms_old",
",",
"i_now",
",",
"j_now",
",",
"delta_plus_epsilon",
"=",
"None",
")",
":",
"nat",
"=",
"len",
"(",
"atoms_now",
")",
"assert",
"len",
"(",
"atoms_now",
")",
"==",
"len",
"(",
"atoms_old",
... | Calculate the D^2_min norm of Falk and Langer | [
"Calculate",
"the",
"D^2_min",
"norm",
"of",
"Falk",
"and",
"Langer"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/atomic_strain.py#L104-L144 | train |
Atomistica/atomistica | src/python/atomistica/hardware.py | dhms | def dhms(secs):
"""return days,hours,minutes and seconds"""
dhms = [0, 0, 0, 0]
dhms[0] = int(secs // 86400)
s = secs % 86400
dhms[1] = int(s // 3600)
s = secs % 3600
dhms[2] = int(s // 60)
s = secs % 60
dhms[3] = int(s+.5)
return dhms | python | def dhms(secs):
"""return days,hours,minutes and seconds"""
dhms = [0, 0, 0, 0]
dhms[0] = int(secs // 86400)
s = secs % 86400
dhms[1] = int(s // 3600)
s = secs % 3600
dhms[2] = int(s // 60)
s = secs % 60
dhms[3] = int(s+.5)
return dhms | [
"def",
"dhms",
"(",
"secs",
")",
":",
"dhms",
"=",
"[",
"0",
",",
"0",
",",
"0",
",",
"0",
"]",
"dhms",
"[",
"0",
"]",
"=",
"int",
"(",
"secs",
"//",
"86400",
")",
"s",
"=",
"secs",
"%",
"86400",
"dhms",
"[",
"1",
"]",
"=",
"int",
"(",
... | return days,hours,minutes and seconds | [
"return",
"days",
"hours",
"minutes",
"and",
"seconds"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/hardware.py#L52-L62 | train |
Atomistica/atomistica | src/python/atomistica/hardware.py | hms | def hms(secs):
"""return hours,minutes and seconds"""
hms = [0, 0, 0]
hms[0] = int(secs // 3600)
s = secs % 3600
hms[1] = int(s // 60)
s = secs % 60
hms[2] = int(s+.5)
return hms | python | def hms(secs):
"""return hours,minutes and seconds"""
hms = [0, 0, 0]
hms[0] = int(secs // 3600)
s = secs % 3600
hms[1] = int(s // 60)
s = secs % 60
hms[2] = int(s+.5)
return hms | [
"def",
"hms",
"(",
"secs",
")",
":",
"hms",
"=",
"[",
"0",
",",
"0",
",",
"0",
"]",
"hms",
"[",
"0",
"]",
"=",
"int",
"(",
"secs",
"//",
"3600",
")",
"s",
"=",
"secs",
"%",
"3600",
"hms",
"[",
"1",
"]",
"=",
"int",
"(",
"s",
"//",
"60",... | return hours,minutes and seconds | [
"return",
"hours",
"minutes",
"and",
"seconds"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/hardware.py#L65-L73 | train |
Atomistica/atomistica | src/python/atomistica/analysis.py | get_enclosing_orthorhombic_box | def get_enclosing_orthorhombic_box(cell):
"""
Return lower and upper bounds of the orthorhombic box that encloses
the parallelepiped spanned by the three cell vectors of cell.
"""
# Cell vectors
cx, cy, cz = cell
# The cell has eight corners, one is at the origin, three at cx, cy, ... | python | def get_enclosing_orthorhombic_box(cell):
"""
Return lower and upper bounds of the orthorhombic box that encloses
the parallelepiped spanned by the three cell vectors of cell.
"""
# Cell vectors
cx, cy, cz = cell
# The cell has eight corners, one is at the origin, three at cx, cy, ... | [
"def",
"get_enclosing_orthorhombic_box",
"(",
"cell",
")",
":",
"# Cell vectors",
"cx",
",",
"cy",
",",
"cz",
"=",
"cell",
"# The cell has eight corners, one is at the origin, three at cx, cy, cz",
"# and the last ones are...",
"c1",
"=",
"cx",
"+",
"cy",
"c2",
"=",
"cx... | Return lower and upper bounds of the orthorhombic box that encloses
the parallelepiped spanned by the three cell vectors of cell. | [
"Return",
"lower",
"and",
"upper",
"bounds",
"of",
"the",
"orthorhombic",
"box",
"that",
"encloses",
"the",
"parallelepiped",
"spanned",
"by",
"the",
"three",
"cell",
"vectors",
"of",
"cell",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/analysis.py#L38-L58 | train |
Atomistica/atomistica | src/python/atomistica/analysis.py | stress_invariants | def stress_invariants(s):
"""Receives a list of stress tensors and returns the three invariants.
Return hydrostatic pressure, octahedral shear stress and J3
"""
s = np.asarray(s)
if s.shape == (6,):
s = s.reshape(1,-1)
elif s.shape == (3,3):
s = s.reshape(1,-1,-1)
if len(s... | python | def stress_invariants(s):
"""Receives a list of stress tensors and returns the three invariants.
Return hydrostatic pressure, octahedral shear stress and J3
"""
s = np.asarray(s)
if s.shape == (6,):
s = s.reshape(1,-1)
elif s.shape == (3,3):
s = s.reshape(1,-1,-1)
if len(s... | [
"def",
"stress_invariants",
"(",
"s",
")",
":",
"s",
"=",
"np",
".",
"asarray",
"(",
"s",
")",
"if",
"s",
".",
"shape",
"==",
"(",
"6",
",",
")",
":",
"s",
"=",
"s",
".",
"reshape",
"(",
"1",
",",
"-",
"1",
")",
"elif",
"s",
".",
"shape",
... | Receives a list of stress tensors and returns the three invariants.
Return hydrostatic pressure, octahedral shear stress and J3 | [
"Receives",
"a",
"list",
"of",
"stress",
"tensors",
"and",
"returns",
"the",
"three",
"invariants",
".",
"Return",
"hydrostatic",
"pressure",
"octahedral",
"shear",
"stress",
"and",
"J3"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/analysis.py#L181-L203 | train |
Atomistica/atomistica | tools/meta.py | scanmeta | def scanmeta(f):
"""Scan file headers for @meta ... @endmeta information and store that into
a dictionary.
"""
print(f)
if isinstance(f, str):
f = io.open(f, mode='r', encoding='latin-1')
done = False
l = f.readline()
s = None
while l and s is None:
i = l.find('!... | python | def scanmeta(f):
"""Scan file headers for @meta ... @endmeta information and store that into
a dictionary.
"""
print(f)
if isinstance(f, str):
f = io.open(f, mode='r', encoding='latin-1')
done = False
l = f.readline()
s = None
while l and s is None:
i = l.find('!... | [
"def",
"scanmeta",
"(",
"f",
")",
":",
"print",
"(",
"f",
")",
"if",
"isinstance",
"(",
"f",
",",
"str",
")",
":",
"f",
"=",
"io",
".",
"open",
"(",
"f",
",",
"mode",
"=",
"'r'",
",",
"encoding",
"=",
"'latin-1'",
")",
"done",
"=",
"False",
"... | Scan file headers for @meta ... @endmeta information and store that into
a dictionary. | [
"Scan",
"file",
"headers",
"for"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/tools/meta.py#L14-L67 | train |
Atomistica/atomistica | src/python/atomistica/snippets.py | mic | def mic(dr, cell, pbc=None):
"""
Apply minimum image convention to an array of distance vectors.
"""
# Check where distance larger than 1/2 cell. Particles have crossed
# periodic boundaries then and need to be unwrapped.
rec = np.linalg.inv(cell)
if pbc is not None:
rec *= np.array(... | python | def mic(dr, cell, pbc=None):
"""
Apply minimum image convention to an array of distance vectors.
"""
# Check where distance larger than 1/2 cell. Particles have crossed
# periodic boundaries then and need to be unwrapped.
rec = np.linalg.inv(cell)
if pbc is not None:
rec *= np.array(... | [
"def",
"mic",
"(",
"dr",
",",
"cell",
",",
"pbc",
"=",
"None",
")",
":",
"# Check where distance larger than 1/2 cell. Particles have crossed",
"# periodic boundaries then and need to be unwrapped.",
"rec",
"=",
"np",
".",
"linalg",
".",
"inv",
"(",
"cell",
")",
"if",... | Apply minimum image convention to an array of distance vectors. | [
"Apply",
"minimum",
"image",
"convention",
"to",
"an",
"array",
"of",
"distance",
"vectors",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/snippets.py#L31-L43 | train |
Atomistica/atomistica | src/python/tools/a_run.py | s_from_dhms | def s_from_dhms(time):
"""return seconds from dhms"""
dhms_s = { 's' : 1, 'm' : 60, 'h' : 3600, 'd' : 86400 }
time = time.lower()
word_list = re.findall('\d*[^\d]*',time)
seconds=0
for word in word_list:
if word != '':
sec = 1
for t in list(dhms_s.keys()):
... | python | def s_from_dhms(time):
"""return seconds from dhms"""
dhms_s = { 's' : 1, 'm' : 60, 'h' : 3600, 'd' : 86400 }
time = time.lower()
word_list = re.findall('\d*[^\d]*',time)
seconds=0
for word in word_list:
if word != '':
sec = 1
for t in list(dhms_s.keys()):
... | [
"def",
"s_from_dhms",
"(",
"time",
")",
":",
"dhms_s",
"=",
"{",
"'s'",
":",
"1",
",",
"'m'",
":",
"60",
",",
"'h'",
":",
"3600",
",",
"'d'",
":",
"86400",
"}",
"time",
"=",
"time",
".",
"lower",
"(",
")",
"word_list",
"=",
"re",
".",
"findall"... | return seconds from dhms | [
"return",
"seconds",
"from",
"dhms"
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/tools/a_run.py#L33-L52 | train |
Atomistica/atomistica | src/python/atomistica/join_calculators.py | JoinCalculators.get_stress | def get_stress(self, a):
"""Calculate stress tensor."""
s = np.zeros( 6, dtype=float )
for c in self.calcs:
s += c.get_stress(a)
return s | python | def get_stress(self, a):
"""Calculate stress tensor."""
s = np.zeros( 6, dtype=float )
for c in self.calcs:
s += c.get_stress(a)
return s | [
"def",
"get_stress",
"(",
"self",
",",
"a",
")",
":",
"s",
"=",
"np",
".",
"zeros",
"(",
"6",
",",
"dtype",
"=",
"float",
")",
"for",
"c",
"in",
"self",
".",
"calcs",
":",
"s",
"+=",
"c",
".",
"get_stress",
"(",
"a",
")",
"return",
"s"
] | Calculate stress tensor. | [
"Calculate",
"stress",
"tensor",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/join_calculators.py#L66-L71 | train |
Atomistica/atomistica | src/python/atomistica/join_calculators.py | JoinCalculators.set_atoms | def set_atoms(self, a):
"""Assign an atoms object."""
for c in self.calcs:
if hasattr(c, "set_atoms"):
c.set_atoms(a) | python | def set_atoms(self, a):
"""Assign an atoms object."""
for c in self.calcs:
if hasattr(c, "set_atoms"):
c.set_atoms(a) | [
"def",
"set_atoms",
"(",
"self",
",",
"a",
")",
":",
"for",
"c",
"in",
"self",
".",
"calcs",
":",
"if",
"hasattr",
"(",
"c",
",",
"\"set_atoms\"",
")",
":",
"c",
".",
"set_atoms",
"(",
"a",
")"
] | Assign an atoms object. | [
"Assign",
"an",
"atoms",
"object",
"."
] | 5ed79d776c92b91a566be22615bfb304ecc75db7 | https://github.com/Atomistica/atomistica/blob/5ed79d776c92b91a566be22615bfb304ecc75db7/src/python/atomistica/join_calculators.py#L79-L83 | train |
thieman/py-dag | dag/__init__.py | DAG.rename_edges | def rename_edges(self, old_task_name, new_task_name, graph=None):
""" Change references to a task in existing edges. """
if not graph:
graph = self.graph
for node, edges in graph.items():
if node == old_task_name:
graph[new_task_name] = copy(edges)
... | python | def rename_edges(self, old_task_name, new_task_name, graph=None):
""" Change references to a task in existing edges. """
if not graph:
graph = self.graph
for node, edges in graph.items():
if node == old_task_name:
graph[new_task_name] = copy(edges)
... | [
"def",
"rename_edges",
"(",
"self",
",",
"old_task_name",
",",
"new_task_name",
",",
"graph",
"=",
"None",
")",
":",
"if",
"not",
"graph",
":",
"graph",
"=",
"self",
".",
"graph",
"for",
"node",
",",
"edges",
"in",
"graph",
".",
"items",
"(",
")",
":... | Change references to a task in existing edges. | [
"Change",
"references",
"to",
"a",
"task",
"in",
"existing",
"edges",
"."
] | 5b5eed396c930751576bdf0d45907a665aac000b | https://github.com/thieman/py-dag/blob/5b5eed396c930751576bdf0d45907a665aac000b/dag/__init__.py#L77-L90 | train |
thieman/py-dag | dag/__init__.py | DAG.predecessors | def predecessors(self, node, graph=None):
""" Returns a list of all predecessors of the given node """
if graph is None:
graph = self.graph
return [key for key in graph if node in graph[key]] | python | def predecessors(self, node, graph=None):
""" Returns a list of all predecessors of the given node """
if graph is None:
graph = self.graph
return [key for key in graph if node in graph[key]] | [
"def",
"predecessors",
"(",
"self",
",",
"node",
",",
"graph",
"=",
"None",
")",
":",
"if",
"graph",
"is",
"None",
":",
"graph",
"=",
"self",
".",
"graph",
"return",
"[",
"key",
"for",
"key",
"in",
"graph",
"if",
"node",
"in",
"graph",
"[",
"key",
... | Returns a list of all predecessors of the given node | [
"Returns",
"a",
"list",
"of",
"all",
"predecessors",
"of",
"the",
"given",
"node"
] | 5b5eed396c930751576bdf0d45907a665aac000b | https://github.com/thieman/py-dag/blob/5b5eed396c930751576bdf0d45907a665aac000b/dag/__init__.py#L92-L96 | train |
buriburisuri/sugartensor | sugartensor/sg_initializer.py | constant | def constant(name, shape, value=0, dtype=tf.sg_floatx, summary=True, regularizer=None, trainable=True):
r"""Creates a tensor variable of which initial values are `value` and shape is `shape`.
Args:
name: The name of new variable.
shape: A tuple/list of integers or an integer.
If shape is a... | python | def constant(name, shape, value=0, dtype=tf.sg_floatx, summary=True, regularizer=None, trainable=True):
r"""Creates a tensor variable of which initial values are `value` and shape is `shape`.
Args:
name: The name of new variable.
shape: A tuple/list of integers or an integer.
If shape is a... | [
"def",
"constant",
"(",
"name",
",",
"shape",
",",
"value",
"=",
"0",
",",
"dtype",
"=",
"tf",
".",
"sg_floatx",
",",
"summary",
"=",
"True",
",",
"regularizer",
"=",
"None",
",",
"trainable",
"=",
"True",
")",
":",
"shape",
"=",
"shape",
"if",
"is... | r"""Creates a tensor variable of which initial values are `value` and shape is `shape`.
Args:
name: The name of new variable.
shape: A tuple/list of integers or an integer.
If shape is an integer, it is converted to a list.
value: A Python scalar. All elements of the initialized variable... | [
"r",
"Creates",
"a",
"tensor",
"variable",
"of",
"which",
"initial",
"values",
"are",
"value",
"and",
"shape",
"is",
"shape",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_initializer.py#L10-L36 | train |
buriburisuri/sugartensor | sugartensor/sg_queue.py | sg_producer_func | def sg_producer_func(func):
r"""Decorates a function `func` as sg_producer_func.
Args:
func: A function to decorate.
"""
@wraps(func)
def wrapper(**kwargs):
r"""Manages arguments of `tf.sg_opt`.
Args:
**kwargs:
source: A source queue list to enqueue
... | python | def sg_producer_func(func):
r"""Decorates a function `func` as sg_producer_func.
Args:
func: A function to decorate.
"""
@wraps(func)
def wrapper(**kwargs):
r"""Manages arguments of `tf.sg_opt`.
Args:
**kwargs:
source: A source queue list to enqueue
... | [
"def",
"sg_producer_func",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"wrapper",
"(",
"*",
"*",
"kwargs",
")",
":",
"r\"\"\"Manages arguments of `tf.sg_opt`.\n\n Args:\n **kwargs:\n source: A source queue list to enqueue\n ... | r"""Decorates a function `func` as sg_producer_func.
Args:
func: A function to decorate. | [
"r",
"Decorates",
"a",
"function",
"func",
"as",
"sg_producer_func",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_queue.py#L11-L77 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_transpose | def sg_transpose(tensor, opt):
r"""Permutes the dimensions according to `opt.perm`.
See `tf.transpose()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
perm: A permutation of the dimensions of `tensor`. The target shape.
name: If provided, replace ... | python | def sg_transpose(tensor, opt):
r"""Permutes the dimensions according to `opt.perm`.
See `tf.transpose()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
perm: A permutation of the dimensions of `tensor`. The target shape.
name: If provided, replace ... | [
"def",
"sg_transpose",
"(",
"tensor",
",",
"opt",
")",
":",
"assert",
"opt",
".",
"perm",
"is",
"not",
"None",
",",
"'perm is mandatory'",
"return",
"tf",
".",
"transpose",
"(",
"tensor",
",",
"opt",
".",
"perm",
",",
"name",
"=",
"opt",
".",
"name",
... | r"""Permutes the dimensions according to `opt.perm`.
See `tf.transpose()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
perm: A permutation of the dimensions of `tensor`. The target shape.
name: If provided, replace current tensor's name.
Returns... | [
"r",
"Permutes",
"the",
"dimensions",
"according",
"to",
"opt",
".",
"perm",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L161-L176 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_argmin | def sg_argmin(tensor, opt):
r"""Returns the indices of the minimum values along the specified axis.
See `tf.argin()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis: Target axis. Default is the last one.
name: If provided, replace current tenso... | python | def sg_argmin(tensor, opt):
r"""Returns the indices of the minimum values along the specified axis.
See `tf.argin()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis: Target axis. Default is the last one.
name: If provided, replace current tenso... | [
"def",
"sg_argmin",
"(",
"tensor",
",",
"opt",
")",
":",
"opt",
"+=",
"tf",
".",
"sg_opt",
"(",
"axis",
"=",
"tensor",
".",
"get_shape",
"(",
")",
".",
"ndims",
"-",
"1",
")",
"return",
"tf",
".",
"argmin",
"(",
"tensor",
",",
"opt",
".",
"axis",... | r"""Returns the indices of the minimum values along the specified axis.
See `tf.argin()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis: Target axis. Default is the last one.
name: If provided, replace current tensor's name.
Returns:
A ... | [
"r",
"Returns",
"the",
"indices",
"of",
"the",
"minimum",
"values",
"along",
"the",
"specified",
"axis",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L199-L214 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_concat | def sg_concat(tensor, opt):
r"""Concatenates tensors along a axis.
See `tf.concat()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
target: A `Tensor`. Must have the same rank as `tensor`, and
all dimensions except `opt.dim` must be equal.
... | python | def sg_concat(tensor, opt):
r"""Concatenates tensors along a axis.
See `tf.concat()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
target: A `Tensor`. Must have the same rank as `tensor`, and
all dimensions except `opt.dim` must be equal.
... | [
"def",
"sg_concat",
"(",
"tensor",
",",
"opt",
")",
":",
"assert",
"opt",
".",
"target",
"is",
"not",
"None",
",",
"'target is mandatory.'",
"opt",
"+=",
"tf",
".",
"sg_opt",
"(",
"axis",
"=",
"tensor",
".",
"get_shape",
"(",
")",
".",
"ndims",
"-",
... | r"""Concatenates tensors along a axis.
See `tf.concat()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
target: A `Tensor`. Must have the same rank as `tensor`, and
all dimensions except `opt.dim` must be equal.
axis : Target axis. Default is... | [
"r",
"Concatenates",
"tensors",
"along",
"a",
"axis",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L218-L237 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_log | def sg_log(tensor, opt):
r"""Log transform a dense tensor
See `tf.log()` in tensorflow.
Args:
tensor: A `Tensor` ( automatically given by chain )
opt:
name: If provided, replace current tensor's name.
Returns:
A `Tensor`.
"""
return tf.log(tensor + tf.sg_eps, name=... | python | def sg_log(tensor, opt):
r"""Log transform a dense tensor
See `tf.log()` in tensorflow.
Args:
tensor: A `Tensor` ( automatically given by chain )
opt:
name: If provided, replace current tensor's name.
Returns:
A `Tensor`.
"""
return tf.log(tensor + tf.sg_eps, name=... | [
"def",
"sg_log",
"(",
"tensor",
",",
"opt",
")",
":",
"return",
"tf",
".",
"log",
"(",
"tensor",
"+",
"tf",
".",
"sg_eps",
",",
"name",
"=",
"opt",
".",
"name",
")"
] | r"""Log transform a dense tensor
See `tf.log()` in tensorflow.
Args:
tensor: A `Tensor` ( automatically given by chain )
opt:
name: If provided, replace current tensor's name.
Returns:
A `Tensor`. | [
"r",
"Log",
"transform",
"a",
"dense",
"tensor"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L281-L294 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_prod | def sg_prod(tensor, opt):
r"""Computes the product of elements across axis of a tensor.
See `tf.reduce_prod()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, ret... | python | def sg_prod(tensor, opt):
r"""Computes the product of elements across axis of a tensor.
See `tf.reduce_prod()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, ret... | [
"def",
"sg_prod",
"(",
"tensor",
",",
"opt",
")",
":",
"return",
"tf",
".",
"reduce_prod",
"(",
"tensor",
",",
"axis",
"=",
"opt",
".",
"axis",
",",
"keep_dims",
"=",
"opt",
".",
"keep_dims",
",",
"name",
"=",
"opt",
".",
"name",
")"
] | r"""Computes the product of elements across axis of a tensor.
See `tf.reduce_prod()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retains reduced dimensions with l... | [
"r",
"Computes",
"the",
"product",
"of",
"elements",
"across",
"axis",
"of",
"a",
"tensor",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L357-L372 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_min | def sg_min(tensor, opt):
r"""Computes the minimum of elements across axis of a tensor.
See `tf.reduce_min()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retai... | python | def sg_min(tensor, opt):
r"""Computes the minimum of elements across axis of a tensor.
See `tf.reduce_min()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retai... | [
"def",
"sg_min",
"(",
"tensor",
",",
"opt",
")",
":",
"return",
"tf",
".",
"reduce_min",
"(",
"tensor",
",",
"axis",
"=",
"opt",
".",
"axis",
",",
"keep_dims",
"=",
"opt",
".",
"keep_dims",
",",
"name",
"=",
"opt",
".",
"name",
")"
] | r"""Computes the minimum of elements across axis of a tensor.
See `tf.reduce_min()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retains reduced dimensions with le... | [
"r",
"Computes",
"the",
"minimum",
"of",
"elements",
"across",
"axis",
"of",
"a",
"tensor",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L376-L391 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_max | def sg_max(tensor, opt):
r"""Computes the maximum of elements across axis of a tensor.
See `tf.reduce_max()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retai... | python | def sg_max(tensor, opt):
r"""Computes the maximum of elements across axis of a tensor.
See `tf.reduce_max()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retai... | [
"def",
"sg_max",
"(",
"tensor",
",",
"opt",
")",
":",
"return",
"tf",
".",
"reduce_max",
"(",
"tensor",
",",
"axis",
"=",
"opt",
".",
"axis",
",",
"keep_dims",
"=",
"opt",
".",
"keep_dims",
",",
"name",
"=",
"opt",
".",
"name",
")"
] | r"""Computes the maximum of elements across axis of a tensor.
See `tf.reduce_max()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retains reduced dimensions with le... | [
"r",
"Computes",
"the",
"maximum",
"of",
"elements",
"across",
"axis",
"of",
"a",
"tensor",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L395-L410 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_any | def sg_any(tensor, opt):
r"""Computes the "logical or" of elements across axis of a tensor.
See `tf.reduce_any()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, ... | python | def sg_any(tensor, opt):
r"""Computes the "logical or" of elements across axis of a tensor.
See `tf.reduce_any()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, ... | [
"def",
"sg_any",
"(",
"tensor",
",",
"opt",
")",
":",
"return",
"tf",
".",
"reduce_any",
"(",
"tensor",
",",
"axis",
"=",
"opt",
".",
"axis",
",",
"keep_dims",
"=",
"opt",
".",
"keep_dims",
",",
"name",
"=",
"opt",
".",
"name",
")"
] | r"""Computes the "logical or" of elements across axis of a tensor.
See `tf.reduce_any()` in tensorflow.
Args:
tensor: A `Tensor` (automatically given by chain).
opt:
axis : A tuple/list of integers or an integer. The axis to reduce.
keep_dims: If true, retains reduced dimensions wi... | [
"r",
"Computes",
"the",
"logical",
"or",
"of",
"elements",
"across",
"axis",
"of",
"a",
"tensor",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L433-L448 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_lookup | def sg_lookup(tensor, opt):
r"""Looks up the `tensor`, which is the embedding matrix.
Args:
tensor: A tensor ( automatically given by chain )
opt:
emb: A 2-D `Tensor`. An embedding matrix.
name: If provided, replace current tensor's name.
Returns:
A `Tensor`.
... | python | def sg_lookup(tensor, opt):
r"""Looks up the `tensor`, which is the embedding matrix.
Args:
tensor: A tensor ( automatically given by chain )
opt:
emb: A 2-D `Tensor`. An embedding matrix.
name: If provided, replace current tensor's name.
Returns:
A `Tensor`.
... | [
"def",
"sg_lookup",
"(",
"tensor",
",",
"opt",
")",
":",
"assert",
"opt",
".",
"emb",
"is",
"not",
"None",
",",
"'emb is mandatory.'",
"return",
"tf",
".",
"nn",
".",
"embedding_lookup",
"(",
"opt",
".",
"emb",
",",
"tensor",
",",
"name",
"=",
"opt",
... | r"""Looks up the `tensor`, which is the embedding matrix.
Args:
tensor: A tensor ( automatically given by chain )
opt:
emb: A 2-D `Tensor`. An embedding matrix.
name: If provided, replace current tensor's name.
Returns:
A `Tensor`. | [
"r",
"Looks",
"up",
"the",
"tensor",
"which",
"is",
"the",
"embedding",
"matrix",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L533-L547 | train |
buriburisuri/sugartensor | sugartensor/sg_transform.py | sg_reverse_seq | def sg_reverse_seq(tensor, opt):
r"""Reverses variable length slices.
Before applying the pure tensorflow function tf.reverse_sequence,
this function calculates sequence lengths by counting non-zeros.
For example,
```
tensor = [[1, 2, 3, 0, 0], [4, 5, 0, 0, 0]]
tensor.sg_reverse_seq... | python | def sg_reverse_seq(tensor, opt):
r"""Reverses variable length slices.
Before applying the pure tensorflow function tf.reverse_sequence,
this function calculates sequence lengths by counting non-zeros.
For example,
```
tensor = [[1, 2, 3, 0, 0], [4, 5, 0, 0, 0]]
tensor.sg_reverse_seq... | [
"def",
"sg_reverse_seq",
"(",
"tensor",
",",
"opt",
")",
":",
"# default sequence dimension",
"opt",
"+=",
"tf",
".",
"sg_opt",
"(",
"axis",
"=",
"1",
")",
"seq_len",
"=",
"tf",
".",
"not_equal",
"(",
"tensor",
",",
"tf",
".",
"zeros_like",
"(",
"tensor"... | r"""Reverses variable length slices.
Before applying the pure tensorflow function tf.reverse_sequence,
this function calculates sequence lengths by counting non-zeros.
For example,
```
tensor = [[1, 2, 3, 0, 0], [4, 5, 0, 0, 0]]
tensor.sg_reverse_seq()
=> [[3 2 1 0 0]
[5 4 0... | [
"r",
"Reverses",
"variable",
"length",
"slices",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_transform.py#L551-L578 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_gpus | def sg_gpus():
r""" Gets current available GPU nums
Returns:
A integer : total # of GPUs available
"""
global _gpus
if _gpus is None:
local_device_protos = device_lib.list_local_devices()
_gpus = len([x.name for x in local_device_protos if x.device_type == 'GPU'])
return... | python | def sg_gpus():
r""" Gets current available GPU nums
Returns:
A integer : total # of GPUs available
"""
global _gpus
if _gpus is None:
local_device_protos = device_lib.list_local_devices()
_gpus = len([x.name for x in local_device_protos if x.device_type == 'GPU'])
return... | [
"def",
"sg_gpus",
"(",
")",
":",
"global",
"_gpus",
"if",
"_gpus",
"is",
"None",
":",
"local_device_protos",
"=",
"device_lib",
".",
"list_local_devices",
"(",
")",
"_gpus",
"=",
"len",
"(",
"[",
"x",
".",
"name",
"for",
"x",
"in",
"local_device_protos",
... | r""" Gets current available GPU nums
Returns:
A integer : total # of GPUs available | [
"r",
"Gets",
"current",
"available",
"GPU",
"nums"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L64-L76 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_context | def sg_context(**kwargs):
r"""Context helper for computational graph building.
Makes all elements within the with Block share the parameters.
For example, in the following example, the default value of parameter `bn` will be set to True
in the all layers within the with block.
```
with tf.sg_c... | python | def sg_context(**kwargs):
r"""Context helper for computational graph building.
Makes all elements within the with Block share the parameters.
For example, in the following example, the default value of parameter `bn` will be set to True
in the all layers within the with block.
```
with tf.sg_c... | [
"def",
"sg_context",
"(",
"*",
"*",
"kwargs",
")",
":",
"global",
"_context",
"# set options when enter",
"context_now",
"=",
"tf",
".",
"sg_opt",
"(",
"kwargs",
")",
"_context",
"+=",
"[",
"context_now",
"]",
"# if named context",
"if",
"context_now",
".",
"n... | r"""Context helper for computational graph building.
Makes all elements within the with Block share the parameters.
For example, in the following example, the default value of parameter `bn` will be set to True
in the all layers within the with block.
```
with tf.sg_context(bn=True):
...
... | [
"r",
"Context",
"helper",
"for",
"computational",
"graph",
"building",
".",
"Makes",
"all",
"elements",
"within",
"the",
"with",
"Block",
"share",
"the",
"parameters",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L87-L132 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_get_context | def sg_get_context():
r"""Get current context information
Returns:
tf.sg_opt class object which contains all context information
"""
global _context
# merge current context
res = tf.sg_opt()
for c in _context:
res += c
return res | python | def sg_get_context():
r"""Get current context information
Returns:
tf.sg_opt class object which contains all context information
"""
global _context
# merge current context
res = tf.sg_opt()
for c in _context:
res += c
return res | [
"def",
"sg_get_context",
"(",
")",
":",
"global",
"_context",
"# merge current context",
"res",
"=",
"tf",
".",
"sg_opt",
"(",
")",
"for",
"c",
"in",
"_context",
":",
"res",
"+=",
"c",
"return",
"res"
] | r"""Get current context information
Returns:
tf.sg_opt class object which contains all context information | [
"r",
"Get",
"current",
"context",
"information"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L135-L149 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_sugar_func | def sg_sugar_func(func):
r""" Decorates a function `func` so that it can be a sugar function.
Sugar function can be used in a chainable manner.
Args:
func: function to decorate
Returns:
A sugar function.
"""
@wraps(func)
def wrapper(tensor, **kwargs):
# call sugar fu... | python | def sg_sugar_func(func):
r""" Decorates a function `func` so that it can be a sugar function.
Sugar function can be used in a chainable manner.
Args:
func: function to decorate
Returns:
A sugar function.
"""
@wraps(func)
def wrapper(tensor, **kwargs):
# call sugar fu... | [
"def",
"sg_sugar_func",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"wrapper",
"(",
"tensor",
",",
"*",
"*",
"kwargs",
")",
":",
"# call sugar function",
"out",
"=",
"func",
"(",
"tensor",
",",
"tf",
".",
"sg_opt",
"(",
"kwargs",
")... | r""" Decorates a function `func` so that it can be a sugar function.
Sugar function can be used in a chainable manner.
Args:
func: function to decorate
Returns:
A sugar function. | [
"r",
"Decorates",
"a",
"function",
"func",
"so",
"that",
"it",
"can",
"be",
"a",
"sugar",
"function",
".",
"Sugar",
"function",
"can",
"be",
"used",
"in",
"a",
"chainable",
"manner",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L156-L177 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_reuse | def sg_reuse(tensor, **opt):
r""" Reconstruct computational graph of `tensor` so all the parameters
can be reused and replace its input tensor with `opt.input`.
Args:
tensor: A `Tensor` (automatically given by chaining).
**opt:
input: A `Tensor` that will replace the original input tens... | python | def sg_reuse(tensor, **opt):
r""" Reconstruct computational graph of `tensor` so all the parameters
can be reused and replace its input tensor with `opt.input`.
Args:
tensor: A `Tensor` (automatically given by chaining).
**opt:
input: A `Tensor` that will replace the original input tens... | [
"def",
"sg_reuse",
"(",
"tensor",
",",
"*",
"*",
"opt",
")",
":",
"opt",
"=",
"tf",
".",
"sg_opt",
"(",
"opt",
")",
"assert",
"hasattr",
"(",
"tensor",
",",
"'_sugar'",
")",
",",
"'cannot reuse this node.'",
"assert",
"opt",
".",
"input",
"is",
"not",
... | r""" Reconstruct computational graph of `tensor` so all the parameters
can be reused and replace its input tensor with `opt.input`.
Args:
tensor: A `Tensor` (automatically given by chaining).
**opt:
input: A `Tensor` that will replace the original input tensor.
Returns:
Reconstru... | [
"r",
"Reconstruct",
"computational",
"graph",
"of",
"tensor",
"so",
"all",
"the",
"parameters",
"can",
"be",
"reused",
"and",
"replace",
"its",
"input",
"tensor",
"with",
"opt",
".",
"input",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L452-L487 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_input | def sg_input(shape=None, dtype=sg_floatx, name=None):
r"""Creates a placeholder.
Args:
shape: A tuple/list of integers. If an integers is given, it will turn to a list.
dtype: A data type. Default is float32.
name: A name for the placeholder.
Returns:
A wrapped placeholder `Tensor`... | python | def sg_input(shape=None, dtype=sg_floatx, name=None):
r"""Creates a placeholder.
Args:
shape: A tuple/list of integers. If an integers is given, it will turn to a list.
dtype: A data type. Default is float32.
name: A name for the placeholder.
Returns:
A wrapped placeholder `Tensor`... | [
"def",
"sg_input",
"(",
"shape",
"=",
"None",
",",
"dtype",
"=",
"sg_floatx",
",",
"name",
"=",
"None",
")",
":",
"if",
"shape",
"is",
"None",
":",
"return",
"tf",
".",
"placeholder",
"(",
"dtype",
",",
"shape",
"=",
"None",
",",
"name",
"=",
"name... | r"""Creates a placeholder.
Args:
shape: A tuple/list of integers. If an integers is given, it will turn to a list.
dtype: A data type. Default is float32.
name: A name for the placeholder.
Returns:
A wrapped placeholder `Tensor`. | [
"r",
"Creates",
"a",
"placeholder",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L494-L510 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_inject | def sg_inject(path, mod_name):
r"""Converts all functions in the given Python module to sugar functions
so that they can be used in a chainable manner.
Args:
path: A string. Path to the Python module
mod_name: A string. The name of the Python module to inject.
Returns:
None
"""
... | python | def sg_inject(path, mod_name):
r"""Converts all functions in the given Python module to sugar functions
so that they can be used in a chainable manner.
Args:
path: A string. Path to the Python module
mod_name: A string. The name of the Python module to inject.
Returns:
None
"""
... | [
"def",
"sg_inject",
"(",
"path",
",",
"mod_name",
")",
":",
"# import module",
"import",
"sys",
"if",
"path",
"not",
"in",
"list",
"(",
"sys",
".",
"path",
")",
":",
"sys",
".",
"path",
".",
"append",
"(",
"path",
")",
"globals",
"(",
")",
"[",
"mo... | r"""Converts all functions in the given Python module to sugar functions
so that they can be used in a chainable manner.
Args:
path: A string. Path to the Python module
mod_name: A string. The name of the Python module to inject.
Returns:
None | [
"r",
"Converts",
"all",
"functions",
"in",
"the",
"given",
"Python",
"module",
"to",
"sugar",
"functions",
"so",
"that",
"they",
"can",
"be",
"used",
"in",
"a",
"chainable",
"manner",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L517-L540 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_queue_context | def sg_queue_context(sess=None):
r"""Context helper for queue routines.
Args:
sess: A session to open queues. If not specified, a new session is created.
Returns:
None
"""
# default session
sess = tf.get_default_session() if sess is None else sess
# thread coordinator
coo... | python | def sg_queue_context(sess=None):
r"""Context helper for queue routines.
Args:
sess: A session to open queues. If not specified, a new session is created.
Returns:
None
"""
# default session
sess = tf.get_default_session() if sess is None else sess
# thread coordinator
coo... | [
"def",
"sg_queue_context",
"(",
"sess",
"=",
"None",
")",
":",
"# default session",
"sess",
"=",
"tf",
".",
"get_default_session",
"(",
")",
"if",
"sess",
"is",
"None",
"else",
"sess",
"# thread coordinator",
"coord",
"=",
"tf",
".",
"train",
".",
"Coordinat... | r"""Context helper for queue routines.
Args:
sess: A session to open queues. If not specified, a new session is created.
Returns:
None | [
"r",
"Context",
"helper",
"for",
"queue",
"routines",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L565-L588 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_arg | def sg_arg():
r"""Gets current command line options
Returns:
tf.sg_opt instance that is updated with current commandd line options.
"""
if not tf.app.flags.FLAGS.__dict__['__parsed']:
tf.app.flags.FLAGS._parse_flags()
return tf.sg_opt(tf.app.flags.FLAGS.__dict__['__flags']) | python | def sg_arg():
r"""Gets current command line options
Returns:
tf.sg_opt instance that is updated with current commandd line options.
"""
if not tf.app.flags.FLAGS.__dict__['__parsed']:
tf.app.flags.FLAGS._parse_flags()
return tf.sg_opt(tf.app.flags.FLAGS.__dict__['__flags']) | [
"def",
"sg_arg",
"(",
")",
":",
"if",
"not",
"tf",
".",
"app",
".",
"flags",
".",
"FLAGS",
".",
"__dict__",
"[",
"'__parsed'",
"]",
":",
"tf",
".",
"app",
".",
"flags",
".",
"FLAGS",
".",
"_parse_flags",
"(",
")",
"return",
"tf",
".",
"sg_opt",
"... | r"""Gets current command line options
Returns:
tf.sg_opt instance that is updated with current commandd line options. | [
"r",
"Gets",
"current",
"command",
"line",
"options"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L632-L640 | train |
buriburisuri/sugartensor | sugartensor/sg_main.py | sg_arg_def | def sg_arg_def(**kwargs):
r"""Defines command line options
Args:
**kwargs:
key: A name for the option.
value : Default value or a tuple of (default value, description).
Returns:
None
For example,
```
# Either of the following two lines will define `--n_epoch` comm... | python | def sg_arg_def(**kwargs):
r"""Defines command line options
Args:
**kwargs:
key: A name for the option.
value : Default value or a tuple of (default value, description).
Returns:
None
For example,
```
# Either of the following two lines will define `--n_epoch` comm... | [
"def",
"sg_arg_def",
"(",
"*",
"*",
"kwargs",
")",
":",
"for",
"k",
",",
"v",
"in",
"kwargs",
".",
"items",
"(",
")",
":",
"if",
"type",
"(",
"v",
")",
"is",
"tuple",
"or",
"type",
"(",
"v",
")",
"is",
"list",
":",
"v",
",",
"c",
"=",
"v",
... | r"""Defines command line options
Args:
**kwargs:
key: A name for the option.
value : Default value or a tuple of (default value, description).
Returns:
None
For example,
```
# Either of the following two lines will define `--n_epoch` command line argument and set its ... | [
"r",
"Defines",
"command",
"line",
"options"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_main.py#L643-L675 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_loss | def sg_summary_loss(tensor, prefix='losses', name=None):
r"""Register `tensor` to summary report as `loss`
Args:
tensor: A `Tensor` to log as loss
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
... | python | def sg_summary_loss(tensor, prefix='losses', name=None):
r"""Register `tensor` to summary report as `loss`
Args:
tensor: A `Tensor` to log as loss
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
... | [
"def",
"sg_summary_loss",
"(",
"tensor",
",",
"prefix",
"=",
"'losses'",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"name",
"=",
"prefix",
"+",
"_pr... | r"""Register `tensor` to summary report as `loss`
Args:
tensor: A `Tensor` to log as loss
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
None | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"loss"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L19-L36 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_gradient | def sg_summary_gradient(tensor, gradient, prefix=None, name=None):
r"""Register `tensor` to summary report as `gradient`
Args:
tensor: A `Tensor` to log as gradient
gradient: A 0-D `Tensor`. A gradient to log
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `s... | python | def sg_summary_gradient(tensor, gradient, prefix=None, name=None):
r"""Register `tensor` to summary report as `gradient`
Args:
tensor: A `Tensor` to log as gradient
gradient: A 0-D `Tensor`. A gradient to log
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `s... | [
"def",
"sg_summary_gradient",
"(",
"tensor",
",",
"gradient",
",",
"prefix",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"name",
"=",
"p... | r"""Register `tensor` to summary report as `gradient`
Args:
tensor: A `Tensor` to log as gradient
gradient: A 0-D `Tensor`. A gradient to log
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
... | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"gradient"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L59-L78 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_activation | def sg_summary_activation(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `activation`
Args:
tensor: A `Tensor` to log as activation
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
... | python | def sg_summary_activation(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `activation`
Args:
tensor: A `Tensor` to log as activation
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
... | [
"def",
"sg_summary_activation",
"(",
"tensor",
",",
"prefix",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"name",
"=",
"prefix",
"+",
"_... | r"""Register `tensor` to summary report as `activation`
Args:
tensor: A `Tensor` to log as activation
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
None | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"activation"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L81-L99 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_param | def sg_summary_param(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `parameters`
Args:
tensor: A `Tensor` to log as parameters
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
... | python | def sg_summary_param(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `parameters`
Args:
tensor: A `Tensor` to log as parameters
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
... | [
"def",
"sg_summary_param",
"(",
"tensor",
",",
"prefix",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"name",
"=",
"prefix",
"+",
"_prett... | r"""Register `tensor` to summary report as `parameters`
Args:
tensor: A `Tensor` to log as parameters
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
None | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"parameters"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L102-L119 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_image | def sg_summary_image(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `image`
Args:
tensor: A tensor to log as image
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
... | python | def sg_summary_image(tensor, prefix=None, name=None):
r"""Register `tensor` to summary report as `image`
Args:
tensor: A tensor to log as image
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
... | [
"def",
"sg_summary_image",
"(",
"tensor",
",",
"prefix",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"name",
"=",
"prefix",
"+",
"_prett... | r"""Register `tensor` to summary report as `image`
Args:
tensor: A tensor to log as image
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
Returns:
None | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"image"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L122-L139 | train |
buriburisuri/sugartensor | sugartensor/sg_logging.py | sg_summary_audio | def sg_summary_audio(tensor, sample_rate=16000, prefix=None, name=None):
r"""Register `tensor` to summary report as audio
Args:
tensor: A `Tensor` to log as audio
sample_rate : An int. Sample rate to report. Default is 16000.
prefix: A `string`. A prefix to display in the tensor board web UI.... | python | def sg_summary_audio(tensor, sample_rate=16000, prefix=None, name=None):
r"""Register `tensor` to summary report as audio
Args:
tensor: A `Tensor` to log as audio
sample_rate : An int. Sample rate to report. Default is 16000.
prefix: A `string`. A prefix to display in the tensor board web UI.... | [
"def",
"sg_summary_audio",
"(",
"tensor",
",",
"sample_rate",
"=",
"16000",
",",
"prefix",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"# defaults",
"prefix",
"=",
"''",
"if",
"prefix",
"is",
"None",
"else",
"prefix",
"+",
"'/'",
"# summary name",
"... | r"""Register `tensor` to summary report as audio
Args:
tensor: A `Tensor` to log as audio
sample_rate : An int. Sample rate to report. Default is 16000.
prefix: A `string`. A prefix to display in the tensor board web UI.
name: A `string`. A name to display in the tensor board web UI.
R... | [
"r",
"Register",
"tensor",
"to",
"summary",
"report",
"as",
"audio"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_logging.py#L142-L160 | train |
buriburisuri/sugartensor | sugartensor/sg_train.py | sg_train | def sg_train(**kwargs):
r"""Trains the model.
Args:
**kwargs:
optim: A name for optimizer. 'MaxProp' (default), 'AdaMax', 'Adam', 'RMSProp' or 'sgd'.
loss: A 0-D `Tensor` containing the value to minimize.
lr: A Python Scalar (optional). Learning rate. Default is .001.
beta... | python | def sg_train(**kwargs):
r"""Trains the model.
Args:
**kwargs:
optim: A name for optimizer. 'MaxProp' (default), 'AdaMax', 'Adam', 'RMSProp' or 'sgd'.
loss: A 0-D `Tensor` containing the value to minimize.
lr: A Python Scalar (optional). Learning rate. Default is .001.
beta... | [
"def",
"sg_train",
"(",
"*",
"*",
"kwargs",
")",
":",
"opt",
"=",
"tf",
".",
"sg_opt",
"(",
"kwargs",
")",
"assert",
"opt",
".",
"loss",
"is",
"not",
"None",
",",
"'loss is mandatory.'",
"# default training options",
"opt",
"+=",
"tf",
".",
"sg_opt",
"("... | r"""Trains the model.
Args:
**kwargs:
optim: A name for optimizer. 'MaxProp' (default), 'AdaMax', 'Adam', 'RMSProp' or 'sgd'.
loss: A 0-D `Tensor` containing the value to minimize.
lr: A Python Scalar (optional). Learning rate. Default is .001.
beta1: A Python Scalar (optional... | [
"r",
"Trains",
"the",
"model",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_train.py#L13-L69 | train |
buriburisuri/sugartensor | sugartensor/sg_train.py | sg_restore | def sg_restore(sess, save_path, category=''):
r""" Restores previously saved variables.
Args:
sess: A `Session` to use to restore the parameters.
save_path: Path where parameters were previously saved.
category: A `String` to filter variables starts with given category.
Returns:
"""... | python | def sg_restore(sess, save_path, category=''):
r""" Restores previously saved variables.
Args:
sess: A `Session` to use to restore the parameters.
save_path: Path where parameters were previously saved.
category: A `String` to filter variables starts with given category.
Returns:
"""... | [
"def",
"sg_restore",
"(",
"sess",
",",
"save_path",
",",
"category",
"=",
"''",
")",
":",
"# to list",
"if",
"not",
"isinstance",
"(",
"category",
",",
"(",
"tuple",
",",
"list",
")",
")",
":",
"category",
"=",
"[",
"category",
"]",
"# make variable list... | r""" Restores previously saved variables.
Args:
sess: A `Session` to use to restore the parameters.
save_path: Path where parameters were previously saved.
category: A `String` to filter variables starts with given category.
Returns: | [
"r",
"Restores",
"previously",
"saved",
"variables",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_train.py#L124-L148 | train |
buriburisuri/sugartensor | sugartensor/sg_train.py | sg_regularizer_loss | def sg_regularizer_loss(scale=1.0):
r""" Get regularizer losss
Args:
scale: A scalar. A weight applied to regularizer loss
"""
return scale * tf.reduce_mean(tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)) | python | def sg_regularizer_loss(scale=1.0):
r""" Get regularizer losss
Args:
scale: A scalar. A weight applied to regularizer loss
"""
return scale * tf.reduce_mean(tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)) | [
"def",
"sg_regularizer_loss",
"(",
"scale",
"=",
"1.0",
")",
":",
"return",
"scale",
"*",
"tf",
".",
"reduce_mean",
"(",
"tf",
".",
"get_collection",
"(",
"tf",
".",
"GraphKeys",
".",
"REGULARIZATION_LOSSES",
")",
")"
] | r""" Get regularizer losss
Args:
scale: A scalar. A weight applied to regularizer loss | [
"r",
"Get",
"regularizer",
"losss"
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_train.py#L376-L382 | train |
buriburisuri/sugartensor | sugartensor/sg_net.py | sg_densenet_layer | def sg_densenet_layer(x, opt):
r"""Applies basic architecture of densenet layer.
Note that the fc layers in the original architecture
will be replaced with fully convolutional layers.
For convenience, We still call them fc layers, though.
Args:
x: A `Tensor`.
opt:
dim: An... | python | def sg_densenet_layer(x, opt):
r"""Applies basic architecture of densenet layer.
Note that the fc layers in the original architecture
will be replaced with fully convolutional layers.
For convenience, We still call them fc layers, though.
Args:
x: A `Tensor`.
opt:
dim: An... | [
"def",
"sg_densenet_layer",
"(",
"x",
",",
"opt",
")",
":",
"assert",
"opt",
".",
"dim",
"is",
"not",
"None",
",",
"'dim is mandatory.'",
"assert",
"opt",
".",
"num",
"is",
"not",
"None",
",",
"'num is mandatory.'",
"# default stride",
"opt",
"+=",
"tf",
"... | r"""Applies basic architecture of densenet layer.
Note that the fc layers in the original architecture
will be replaced with fully convolutional layers.
For convenience, We still call them fc layers, though.
Args:
x: A `Tensor`.
opt:
dim: An integer. Dimension for this resnet... | [
"r",
"Applies",
"basic",
"architecture",
"of",
"densenet",
"layer",
"."
] | d2c039954777c7fbe3eb0c2ae40c45c9854deb40 | https://github.com/buriburisuri/sugartensor/blob/d2c039954777c7fbe3eb0c2ae40c45c9854deb40/sugartensor/sg_net.py#L432-L480 | train |
thanethomson/statik | statik/utils.py | deep_merge_dict | def deep_merge_dict(a, b):
"""Deep merges dictionary b into dictionary a."""
if not isinstance(a, dict):
raise TypeError("a must be a dict, but found %s" % a.__class__.__name__)
if not isinstance(b, dict):
raise TypeError("b must be a dict, but found %s" % b.__class__.__name__)
_a = cop... | python | def deep_merge_dict(a, b):
"""Deep merges dictionary b into dictionary a."""
if not isinstance(a, dict):
raise TypeError("a must be a dict, but found %s" % a.__class__.__name__)
if not isinstance(b, dict):
raise TypeError("b must be a dict, but found %s" % b.__class__.__name__)
_a = cop... | [
"def",
"deep_merge_dict",
"(",
"a",
",",
"b",
")",
":",
"if",
"not",
"isinstance",
"(",
"a",
",",
"dict",
")",
":",
"raise",
"TypeError",
"(",
"\"a must be a dict, but found %s\"",
"%",
"a",
".",
"__class__",
".",
"__name__",
")",
"if",
"not",
"isinstance"... | Deep merges dictionary b into dictionary a. | [
"Deep",
"merges",
"dictionary",
"b",
"into",
"dictionary",
"a",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L102-L124 | train |
thanethomson/statik | statik/utils.py | copy_file_if_modified | def copy_file_if_modified(src_path, dest_path):
"""Only copies the file from the source path to the destination path if it doesn't exist yet or it has
been modified. Intended to provide something of an optimisation when a project has large trees of assets."""
# if the destination path is a directory, delet... | python | def copy_file_if_modified(src_path, dest_path):
"""Only copies the file from the source path to the destination path if it doesn't exist yet or it has
been modified. Intended to provide something of an optimisation when a project has large trees of assets."""
# if the destination path is a directory, delet... | [
"def",
"copy_file_if_modified",
"(",
"src_path",
",",
"dest_path",
")",
":",
"# if the destination path is a directory, delete it completely - we assume here we are",
"# writing a file to the filesystem",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"dest_path",
")",
":",
"shu... | Only copies the file from the source path to the destination path if it doesn't exist yet or it has
been modified. Intended to provide something of an optimisation when a project has large trees of assets. | [
"Only",
"copies",
"the",
"file",
"from",
"the",
"source",
"path",
"to",
"the",
"destination",
"path",
"if",
"it",
"doesn",
"t",
"exist",
"yet",
"or",
"it",
"has",
"been",
"modified",
".",
"Intended",
"to",
"provide",
"something",
"of",
"an",
"optimisation"... | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L141-L163 | train |
thanethomson/statik | statik/utils.py | get_url_file_ext | def get_url_file_ext(url):
"""Attempts to extract the file extension from the given URL."""
# get the last part of the path component
filename = url.split('/')[-1]
name, ext = os.path.splitext(filename)
# handle case of files with leading dot
if not ext and name and name[0] == '.':
ext ... | python | def get_url_file_ext(url):
"""Attempts to extract the file extension from the given URL."""
# get the last part of the path component
filename = url.split('/')[-1]
name, ext = os.path.splitext(filename)
# handle case of files with leading dot
if not ext and name and name[0] == '.':
ext ... | [
"def",
"get_url_file_ext",
"(",
"url",
")",
":",
"# get the last part of the path component",
"filename",
"=",
"url",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"1",
"]",
"name",
",",
"ext",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"#... | Attempts to extract the file extension from the given URL. | [
"Attempts",
"to",
"extract",
"the",
"file",
"extension",
"from",
"the",
"given",
"URL",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L195-L205 | train |
thanethomson/statik | statik/utils.py | generate_quickstart | def generate_quickstart(project_path):
"""Generates all of the basic paths for a Statik project within the given project path. If the project path
doesn't exist, it will be created."""
ensure_path_exists(project_path)
ensure_file_exists(os.path.join(project_path, "config.yml"), DEFAULT_CONFIG_CONTENT)
... | python | def generate_quickstart(project_path):
"""Generates all of the basic paths for a Statik project within the given project path. If the project path
doesn't exist, it will be created."""
ensure_path_exists(project_path)
ensure_file_exists(os.path.join(project_path, "config.yml"), DEFAULT_CONFIG_CONTENT)
... | [
"def",
"generate_quickstart",
"(",
"project_path",
")",
":",
"ensure_path_exists",
"(",
"project_path",
")",
"ensure_file_exists",
"(",
"os",
".",
"path",
".",
"join",
"(",
"project_path",
",",
"\"config.yml\"",
")",
",",
"DEFAULT_CONFIG_CONTENT",
")",
"ensure_path_... | Generates all of the basic paths for a Statik project within the given project path. If the project path
doesn't exist, it will be created. | [
"Generates",
"all",
"of",
"the",
"basic",
"paths",
"for",
"a",
"Statik",
"project",
"within",
"the",
"given",
"project",
"path",
".",
"If",
"the",
"project",
"path",
"doesn",
"t",
"exist",
"it",
"will",
"be",
"created",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L208-L219 | train |
thanethomson/statik | statik/utils.py | get_project_config_file | def get_project_config_file(path, default_config_file_name):
"""Attempts to extract the project config file's absolute path from the given path. If the path is a
directory, it automatically assumes a "config.yml" file will be in that directory. If the path is to
a .yml file, it assumes that that is the root... | python | def get_project_config_file(path, default_config_file_name):
"""Attempts to extract the project config file's absolute path from the given path. If the path is a
directory, it automatically assumes a "config.yml" file will be in that directory. If the path is to
a .yml file, it assumes that that is the root... | [
"def",
"get_project_config_file",
"(",
"path",
",",
"default_config_file_name",
")",
":",
"_path",
",",
"_config_file_path",
"=",
"None",
",",
"None",
"path",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"path",
")",
"if",
"os",
".",
"path",
".",
"isdir",
... | Attempts to extract the project config file's absolute path from the given path. If the path is a
directory, it automatically assumes a "config.yml" file will be in that directory. If the path is to
a .yml file, it assumes that that is the root configuration file for the project. | [
"Attempts",
"to",
"extract",
"the",
"project",
"config",
"file",
"s",
"absolute",
"path",
"from",
"the",
"given",
"path",
".",
"If",
"the",
"path",
"is",
"a",
"directory",
"it",
"automatically",
"assumes",
"a",
"config",
".",
"yml",
"file",
"will",
"be",
... | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L259-L276 | train |
thanethomson/statik | statik/utils.py | strip_el_text | def strip_el_text(el, max_depth=0, cur_depth=0):
"""Recursively strips the plain text out of the given XML etree element up to the desired depth.
Args:
el: The etree element to scan.
max_depth: The depth to which to recursively strip text (default: 0).
cur_depth: The current recursive d... | python | def strip_el_text(el, max_depth=0, cur_depth=0):
"""Recursively strips the plain text out of the given XML etree element up to the desired depth.
Args:
el: The etree element to scan.
max_depth: The depth to which to recursively strip text (default: 0).
cur_depth: The current recursive d... | [
"def",
"strip_el_text",
"(",
"el",
",",
"max_depth",
"=",
"0",
",",
"cur_depth",
"=",
"0",
")",
":",
"# text in front of any child elements",
"el_text",
"=",
"strip_str",
"(",
"el",
".",
"text",
"if",
"el",
".",
"text",
"is",
"not",
"None",
"else",
"\"\"",... | Recursively strips the plain text out of the given XML etree element up to the desired depth.
Args:
el: The etree element to scan.
max_depth: The depth to which to recursively strip text (default: 0).
cur_depth: The current recursive depth to which we've scanned so far.
Returns:
... | [
"Recursively",
"strips",
"the",
"plain",
"text",
"out",
"of",
"the",
"given",
"XML",
"etree",
"element",
"up",
"to",
"the",
"desired",
"depth",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L303-L333 | train |
thanethomson/statik | statik/utils.py | find_first_file_with_ext | def find_first_file_with_ext(base_paths, prefix, exts):
"""Runs through the given list of file extensions and returns the first file with the given base
path and extension combination that actually exists.
Args:
base_paths: The base paths in which to search for files.
prefix: The filename p... | python | def find_first_file_with_ext(base_paths, prefix, exts):
"""Runs through the given list of file extensions and returns the first file with the given base
path and extension combination that actually exists.
Args:
base_paths: The base paths in which to search for files.
prefix: The filename p... | [
"def",
"find_first_file_with_ext",
"(",
"base_paths",
",",
"prefix",
",",
"exts",
")",
":",
"for",
"base_path",
"in",
"base_paths",
":",
"for",
"ext",
"in",
"exts",
":",
"filename",
"=",
"os",
".",
"path",
".",
"join",
"(",
"base_path",
",",
"\"%s%s\"",
... | Runs through the given list of file extensions and returns the first file with the given base
path and extension combination that actually exists.
Args:
base_paths: The base paths in which to search for files.
prefix: The filename prefix of the file for which to search.
exts: An ordered... | [
"Runs",
"through",
"the",
"given",
"list",
"of",
"file",
"extensions",
"and",
"returns",
"the",
"first",
"file",
"with",
"the",
"given",
"base",
"path",
"and",
"extension",
"combination",
"that",
"actually",
"exists",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L336-L357 | train |
thanethomson/statik | statik/utils.py | find_duplicates_in_array | def find_duplicates_in_array(array):
"""Runs through the array and returns the elements that contain
more than one duplicate
Args:
array: The array to check for duplicates.
Returns:
Array of the elements that are duplicates. Returns empty list if
there are no duplicates.
""... | python | def find_duplicates_in_array(array):
"""Runs through the array and returns the elements that contain
more than one duplicate
Args:
array: The array to check for duplicates.
Returns:
Array of the elements that are duplicates. Returns empty list if
there are no duplicates.
""... | [
"def",
"find_duplicates_in_array",
"(",
"array",
")",
":",
"duplicates",
"=",
"[",
"]",
"non_duplicates",
"=",
"[",
"]",
"if",
"len",
"(",
"array",
")",
"!=",
"len",
"(",
"set",
"(",
"array",
")",
")",
":",
"for",
"item",
"in",
"array",
":",
"if",
... | Runs through the array and returns the elements that contain
more than one duplicate
Args:
array: The array to check for duplicates.
Returns:
Array of the elements that are duplicates. Returns empty list if
there are no duplicates. | [
"Runs",
"through",
"the",
"array",
"and",
"returns",
"the",
"elements",
"that",
"contain",
"more",
"than",
"one",
"duplicate"
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/utils.py#L365-L386 | train |
thanethomson/statik | setup.py | read_requirements | def read_requirements(filename):
"""
Parse a requirements file.
Accepts vcs+ links, and places the URL into
`DEPENDENCY_LINKS`.
:return: list of str for each package
"""
data = []
for line in read_file(filename):
line = line.strip()
if not line or line.startswith('#'):
... | python | def read_requirements(filename):
"""
Parse a requirements file.
Accepts vcs+ links, and places the URL into
`DEPENDENCY_LINKS`.
:return: list of str for each package
"""
data = []
for line in read_file(filename):
line = line.strip()
if not line or line.startswith('#'):
... | [
"def",
"read_requirements",
"(",
"filename",
")",
":",
"data",
"=",
"[",
"]",
"for",
"line",
"in",
"read_file",
"(",
"filename",
")",
":",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"if",
"not",
"line",
"or",
"line",
".",
"startswith",
"(",
"'#'",
... | Parse a requirements file.
Accepts vcs+ links, and places the URL into
`DEPENDENCY_LINKS`.
:return: list of str for each package | [
"Parse",
"a",
"requirements",
"file",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/setup.py#L22-L49 | train |
thanethomson/statik | statik/models.py | StatikModel.find_additional_rels | def find_additional_rels(self, all_models):
"""Attempts to scan for additional relationship fields for this model based on all of the other models'
structures and relationships.
"""
for model_name, model in iteritems(all_models):
if model_name != self.name:
fo... | python | def find_additional_rels(self, all_models):
"""Attempts to scan for additional relationship fields for this model based on all of the other models'
structures and relationships.
"""
for model_name, model in iteritems(all_models):
if model_name != self.name:
fo... | [
"def",
"find_additional_rels",
"(",
"self",
",",
"all_models",
")",
":",
"for",
"model_name",
",",
"model",
"in",
"iteritems",
"(",
"all_models",
")",
":",
"if",
"model_name",
"!=",
"self",
".",
"name",
":",
"for",
"field_name",
"in",
"model",
".",
"field_... | Attempts to scan for additional relationship fields for this model based on all of the other models'
structures and relationships. | [
"Attempts",
"to",
"scan",
"for",
"additional",
"relationship",
"fields",
"for",
"this",
"model",
"based",
"on",
"all",
"of",
"the",
"other",
"models",
"structures",
"and",
"relationships",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/models.py#L73-L96 | train |
thanethomson/statik | statik/database.py | StatikDatabase.create_db | def create_db(self, models):
"""Creates the in-memory SQLite database from the model
configuration."""
# first create the table definitions
self.tables = dict(
[
(model_name, self.create_model_table(model))
for model_name, model in iteritems(mo... | python | def create_db(self, models):
"""Creates the in-memory SQLite database from the model
configuration."""
# first create the table definitions
self.tables = dict(
[
(model_name, self.create_model_table(model))
for model_name, model in iteritems(mo... | [
"def",
"create_db",
"(",
"self",
",",
"models",
")",
":",
"# first create the table definitions",
"self",
".",
"tables",
"=",
"dict",
"(",
"[",
"(",
"model_name",
",",
"self",
".",
"create_model_table",
"(",
"model",
")",
")",
"for",
"model_name",
",",
"mode... | Creates the in-memory SQLite database from the model
configuration. | [
"Creates",
"the",
"in",
"-",
"memory",
"SQLite",
"database",
"from",
"the",
"model",
"configuration",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/database.py#L106-L125 | train |
thanethomson/statik | statik/database.py | StatikDatabase.sort_models | def sort_models(self):
"""Sorts the database models appropriately based on their relationships so that we load our data
in the appropriate order.
Returns:
A sorted list containing the names of the models.
"""
model_names = [
table.name for table in self.B... | python | def sort_models(self):
"""Sorts the database models appropriately based on their relationships so that we load our data
in the appropriate order.
Returns:
A sorted list containing the names of the models.
"""
model_names = [
table.name for table in self.B... | [
"def",
"sort_models",
"(",
"self",
")",
":",
"model_names",
"=",
"[",
"table",
".",
"name",
"for",
"table",
"in",
"self",
".",
"Base",
".",
"metadata",
".",
"sorted_tables",
"if",
"table",
".",
"name",
"in",
"self",
".",
"models",
"]",
"logger",
".",
... | Sorts the database models appropriately based on their relationships so that we load our data
in the appropriate order.
Returns:
A sorted list containing the names of the models. | [
"Sorts",
"the",
"database",
"models",
"appropriately",
"based",
"on",
"their",
"relationships",
"so",
"that",
"we",
"load",
"our",
"data",
"in",
"the",
"appropriate",
"order",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/database.py#L141-L178 | train |
thanethomson/statik | statik/database.py | StatikDatabase.create_model_table | def create_model_table(self, model):
"""Creates the table for the given model.
Args:
model: A StatikModel instance.
Returns:
A SQLAlchemy model instance for the table corresponding to this
particular model.
"""
try:
return db_mode... | python | def create_model_table(self, model):
"""Creates the table for the given model.
Args:
model: A StatikModel instance.
Returns:
A SQLAlchemy model instance for the table corresponding to this
particular model.
"""
try:
return db_mode... | [
"def",
"create_model_table",
"(",
"self",
",",
"model",
")",
":",
"try",
":",
"return",
"db_model_factory",
"(",
"self",
".",
"Base",
",",
"model",
",",
"self",
".",
"models",
")",
"except",
"Exception",
"as",
"exc",
":",
"raise",
"ModelError",
"(",
"mod... | Creates the table for the given model.
Args:
model: A StatikModel instance.
Returns:
A SQLAlchemy model instance for the table corresponding to this
particular model. | [
"Creates",
"the",
"table",
"for",
"the",
"given",
"model",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/database.py#L180-L198 | train |
thanethomson/statik | statik/database.py | StatikDatabase.load_model_data | def load_model_data(self, path, model):
"""Loads the data for the specified model from the given path.
"""
if os.path.isdir(path):
# try find a model data collection
if os.path.isfile(os.path.join(path, '_all.yml')):
self.load_model_data_collection(path, m... | python | def load_model_data(self, path, model):
"""Loads the data for the specified model from the given path.
"""
if os.path.isdir(path):
# try find a model data collection
if os.path.isfile(os.path.join(path, '_all.yml')):
self.load_model_data_collection(path, m... | [
"def",
"load_model_data",
"(",
"self",
",",
"path",
",",
"model",
")",
":",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"path",
")",
":",
"# try find a model data collection",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"os",
".",
"path",
".",
"join",... | Loads the data for the specified model from the given path. | [
"Loads",
"the",
"data",
"for",
"the",
"specified",
"model",
"from",
"the",
"given",
"path",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/database.py#L200-L208 | train |
thanethomson/statik | statik/database.py | StatikDatabase.query | def query(self, query, additional_locals=None, safe_mode=False):
"""Executes the given SQLAlchemy query string.
Args:
query: The SQLAlchemy ORM query (or Python code) to be executed.
additional_locals: Any additional local variables to inject into the execution context
... | python | def query(self, query, additional_locals=None, safe_mode=False):
"""Executes the given SQLAlchemy query string.
Args:
query: The SQLAlchemy ORM query (or Python code) to be executed.
additional_locals: Any additional local variables to inject into the execution context
... | [
"def",
"query",
"(",
"self",
",",
"query",
",",
"additional_locals",
"=",
"None",
",",
"safe_mode",
"=",
"False",
")",
":",
"logger",
".",
"debug",
"(",
"\"Attempting to execute database query: %s\"",
",",
"query",
")",
"if",
"safe_mode",
"and",
"not",
"isinst... | Executes the given SQLAlchemy query string.
Args:
query: The SQLAlchemy ORM query (or Python code) to be executed.
additional_locals: Any additional local variables to inject into the execution context
when executing the query.
safe_mode: Boolean value indica... | [
"Executes",
"the",
"given",
"SQLAlchemy",
"query",
"string",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/database.py#L328-L367 | train |
thanethomson/statik | statik/generator.py | generate | def generate(input_path, output_path=None, in_memory=False, safe_mode=False, error_context=None):
"""Executes the Statik site generator using the given parameters.
"""
project = StatikProject(input_path, safe_mode=safe_mode, error_context=error_context)
return project.generate(output_path=output_path, i... | python | def generate(input_path, output_path=None, in_memory=False, safe_mode=False, error_context=None):
"""Executes the Statik site generator using the given parameters.
"""
project = StatikProject(input_path, safe_mode=safe_mode, error_context=error_context)
return project.generate(output_path=output_path, i... | [
"def",
"generate",
"(",
"input_path",
",",
"output_path",
"=",
"None",
",",
"in_memory",
"=",
"False",
",",
"safe_mode",
"=",
"False",
",",
"error_context",
"=",
"None",
")",
":",
"project",
"=",
"StatikProject",
"(",
"input_path",
",",
"safe_mode",
"=",
"... | Executes the Statik site generator using the given parameters. | [
"Executes",
"the",
"Statik",
"site",
"generator",
"using",
"the",
"given",
"parameters",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/generator.py#L11-L15 | train |
thanethomson/statik | statik/project.py | StatikProject.generate | def generate(self, output_path=None, in_memory=False):
"""Executes the Statik project generator.
Args:
output_path: The path to which to write output files.
in_memory: Whether or not to generate the results in memory. If True, this will
generate the output result... | python | def generate(self, output_path=None, in_memory=False):
"""Executes the Statik project generator.
Args:
output_path: The path to which to write output files.
in_memory: Whether or not to generate the results in memory. If True, this will
generate the output result... | [
"def",
"generate",
"(",
"self",
",",
"output_path",
"=",
"None",
",",
"in_memory",
"=",
"False",
")",
":",
"result",
"=",
"dict",
"(",
")",
"if",
"in_memory",
"else",
"0",
"logger",
".",
"info",
"(",
"\"Generating Statik build...\"",
")",
"try",
":",
"if... | Executes the Statik project generator.
Args:
output_path: The path to which to write output files.
in_memory: Whether or not to generate the results in memory. If True, this will
generate the output result as a dictionary. If False, this will write the output
... | [
"Executes",
"the",
"Statik",
"project",
"generator",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/project.py#L71-L156 | train |
thanethomson/statik | statik/project.py | StatikProject.load_views | def load_views(self):
"""Loads the views for this project from the project directory
structure."""
view_path = os.path.join(self.path, StatikProject.VIEWS_DIR)
logger.debug("Loading views from: %s", view_path)
if not os.path.isdir(view_path):
raise MissingProjectFolde... | python | def load_views(self):
"""Loads the views for this project from the project directory
structure."""
view_path = os.path.join(self.path, StatikProject.VIEWS_DIR)
logger.debug("Loading views from: %s", view_path)
if not os.path.isdir(view_path):
raise MissingProjectFolde... | [
"def",
"load_views",
"(",
"self",
")",
":",
"view_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"self",
".",
"path",
",",
"StatikProject",
".",
"VIEWS_DIR",
")",
"logger",
".",
"debug",
"(",
"\"Loading views from: %s\"",
",",
"view_path",
")",
"if",
"... | Loads the views for this project from the project directory
structure. | [
"Loads",
"the",
"views",
"for",
"this",
"project",
"from",
"the",
"project",
"directory",
"structure",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/project.py#L181-L203 | train |
thanethomson/statik | statik/project.py | StatikProject.process_views | def process_views(self):
"""Processes the loaded views to generate the required output data."""
output = {}
logger.debug("Processing %d view(s)...", len(self.views))
for view_name, view in iteritems(self.views):
try:
output = deep_merge_dict(
... | python | def process_views(self):
"""Processes the loaded views to generate the required output data."""
output = {}
logger.debug("Processing %d view(s)...", len(self.views))
for view_name, view in iteritems(self.views):
try:
output = deep_merge_dict(
... | [
"def",
"process_views",
"(",
"self",
")",
":",
"output",
"=",
"{",
"}",
"logger",
".",
"debug",
"(",
"\"Processing %d view(s)...\"",
",",
"len",
"(",
"self",
".",
"views",
")",
")",
"for",
"view_name",
",",
"view",
"in",
"iteritems",
"(",
"self",
".",
... | Processes the loaded views to generate the required output data. | [
"Processes",
"the",
"loaded",
"views",
"to",
"generate",
"the",
"required",
"output",
"data",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/project.py#L243-L268 | train |
thanethomson/statik | statik/project.py | StatikProject.dump_in_memory_result | def dump_in_memory_result(self, result, output_path):
"""Recursively dumps the result of our processing into files within the
given output path.
Args:
result: The in-memory result of our processing.
output_path: Full path to the folder into which to dump the files.
... | python | def dump_in_memory_result(self, result, output_path):
"""Recursively dumps the result of our processing into files within the
given output path.
Args:
result: The in-memory result of our processing.
output_path: Full path to the folder into which to dump the files.
... | [
"def",
"dump_in_memory_result",
"(",
"self",
",",
"result",
",",
"output_path",
")",
":",
"file_count",
"=",
"0",
"logger",
".",
"debug",
"(",
"\"Dumping in-memory processing results to output folder: %s\"",
",",
"output_path",
")",
"for",
"k",
",",
"v",
"in",
"it... | Recursively dumps the result of our processing into files within the
given output path.
Args:
result: The in-memory result of our processing.
output_path: Full path to the folder into which to dump the files.
Returns:
The number of files generated (integer). | [
"Recursively",
"dumps",
"the",
"result",
"of",
"our",
"processing",
"into",
"files",
"within",
"the",
"given",
"output",
"path",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/project.py#L270-L300 | train |
thanethomson/statik | statik/project.py | StatikProject.copy_assets | def copy_assets(self, output_path):
"""Copies all asset files from the source path to the destination
path. If no such source path exists, no asset copying will be performed.
"""
src_paths = []
# if we have a theme
if self.config.theme is not None:
# assume i... | python | def copy_assets(self, output_path):
"""Copies all asset files from the source path to the destination
path. If no such source path exists, no asset copying will be performed.
"""
src_paths = []
# if we have a theme
if self.config.theme is not None:
# assume i... | [
"def",
"copy_assets",
"(",
"self",
",",
"output_path",
")",
":",
"src_paths",
"=",
"[",
"]",
"# if we have a theme",
"if",
"self",
".",
"config",
".",
"theme",
"is",
"not",
"None",
":",
"# assume it's in the folder: \"themes/theme_name/assets\"",
"src_paths",
".",
... | Copies all asset files from the source path to the destination
path. If no such source path exists, no asset copying will be performed. | [
"Copies",
"all",
"asset",
"files",
"from",
"the",
"source",
"path",
"to",
"the",
"destination",
"path",
".",
"If",
"no",
"such",
"source",
"path",
"exists",
"no",
"asset",
"copying",
"will",
"be",
"performed",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/project.py#L302-L340 | train |
thanethomson/statik | statik/autogen.py | autogen | def autogen(project_path):
"""Autogenerates views and templates for all the models in the project."""
generate_quickstart(project_path)
project = StatikProject(project_path)
project.config = StatikConfig(project.config_file_path)
models = list(project.load_models().values())
logger.info('Crea... | python | def autogen(project_path):
"""Autogenerates views and templates for all the models in the project."""
generate_quickstart(project_path)
project = StatikProject(project_path)
project.config = StatikConfig(project.config_file_path)
models = list(project.load_models().values())
logger.info('Crea... | [
"def",
"autogen",
"(",
"project_path",
")",
":",
"generate_quickstart",
"(",
"project_path",
")",
"project",
"=",
"StatikProject",
"(",
"project_path",
")",
"project",
".",
"config",
"=",
"StatikConfig",
"(",
"project",
".",
"config_file_path",
")",
"models",
"=... | Autogenerates views and templates for all the models in the project. | [
"Autogenerates",
"views",
"and",
"templates",
"for",
"all",
"the",
"models",
"in",
"the",
"project",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/autogen.py#L17-L51 | train |
thanethomson/statik | statik/autogen.py | generate_yaml_file | def generate_yaml_file(filename, contents):
"""Creates a yaml file with the given content."""
with open(filename, 'w') as file:
file.write(yaml.dump(contents, default_flow_style=False)) | python | def generate_yaml_file(filename, contents):
"""Creates a yaml file with the given content."""
with open(filename, 'w') as file:
file.write(yaml.dump(contents, default_flow_style=False)) | [
"def",
"generate_yaml_file",
"(",
"filename",
",",
"contents",
")",
":",
"with",
"open",
"(",
"filename",
",",
"'w'",
")",
"as",
"file",
":",
"file",
".",
"write",
"(",
"yaml",
".",
"dump",
"(",
"contents",
",",
"default_flow_style",
"=",
"False",
")",
... | Creates a yaml file with the given content. | [
"Creates",
"a",
"yaml",
"file",
"with",
"the",
"given",
"content",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/autogen.py#L54-L57 | train |
thanethomson/statik | statik/autogen.py | generate_index_file | def generate_index_file(filename):
"""Constructs a default home page for the project."""
with open(filename, 'w') as file:
content = open(os.path.join(os.path.dirname(__file__), 'templates/index_page.html'), 'r').read()
file.write(content) | python | def generate_index_file(filename):
"""Constructs a default home page for the project."""
with open(filename, 'w') as file:
content = open(os.path.join(os.path.dirname(__file__), 'templates/index_page.html'), 'r').read()
file.write(content) | [
"def",
"generate_index_file",
"(",
"filename",
")",
":",
"with",
"open",
"(",
"filename",
",",
"'w'",
")",
"as",
"file",
":",
"content",
"=",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
... | Constructs a default home page for the project. | [
"Constructs",
"a",
"default",
"home",
"page",
"for",
"the",
"project",
"."
] | 56b1b5a2cb05a97afa81f428bfcefc833e935b8d | https://github.com/thanethomson/statik/blob/56b1b5a2cb05a97afa81f428bfcefc833e935b8d/statik/autogen.py#L60-L64 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.