code stringlengths 4 4.48k | docstring stringlengths 1 6.45k | _id stringlengths 24 24 |
|---|---|---|
def compute_IOU(self,bb1,bb2): <NEW_LINE> <INDENT> up_left_corner=np.array((max(bb1[0],bb2[0]),max(bb1[1],bb2[1]))) <NEW_LINE> low_right_corner=np.array((min(bb1[2],bb2[2]),min(bb1[3],bb2[3]))) <NEW_LINE> dist=low_right_corner-up_left_corner <NEW_LINE> if np.any(dist<=0): <NEW_LINE> <INDENT> inter_area=0 <NEW_LINE> <DE... | returns the IOU of bb1 with bb2 | 625941b424f1403a92600948 |
def animate_bisection(_f, xmin, xmax): <NEW_LINE> <INDENT> yrange = _f(xmin), _f(xmax) <NEW_LINE> ymin, ymax = min(yrange), max(yrange) <NEW_LINE> _x = np.linspace(xmin, xmax) <NEW_LINE> _y = _f(_x) <NEW_LINE> fig = plt.figure() <NEW_LINE> _ax = plt.axes(xlim=(xmin-0.1, xmax+0.1), ylim=(ymin, ymax)) <NEW_LINE> curve, =... | animates the bisection method | 625941b49f2886367277a66f |
def _as_deces(sal, _P): <NEW_LINE> <INDENT> return 0 | Assurance sociale - décès # TODO: à compléter | 625941b43317a56b86939a42 |
def _generate_item_for_reply_or_forward(self, folder, subject, body): <NEW_LINE> <INDENT> item = folder.create_item( subject=subject, body=body, from_=self.store.user, to=self.from_.email, message_class=self.message_class ) <NEW_LINE> source_message_info = self._create_source_message_info('reply').encode('ascii') <NEW_... | Generate item object for 'reply' or 'forward' action.
Args:
folder (Folder): folder object.
subject (str): subject text.
body (str): body text.
Returns:
Item: created item for 'reply' or 'forward' action. | 625941b430c21e258bdfa274 |
def pylab_setup(): <NEW_LINE> <INDENT> if backend.startswith('module://'): <NEW_LINE> <INDENT> backend_name = backend[9:] <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> backend_name = 'backend_'+backend <NEW_LINE> backend_name = backend_name.lower() <NEW_LINE> backend_name = 'matplotlib.backends.%s'%backend_name.lower()... | return new_figure_manager, draw_if_interactive and show for pylab | 625941b460cbc95b062c6320 |
def getBalanceInQuote(crypto_enum): <NEW_LINE> <INDENT> return LEDGERS[crypto_enum].balance * TRADES[crypto_enum].current_row.price | Convert to quote the balance of the LEDGERS[crypto_enum] | 625941b4e5267d203edcda79 |
def generate_xml_file(self): <NEW_LINE> <INDENT> unique_id = unicode(uuid.uuid4()) <NEW_LINE> xml_content = self.tmpl.render({ 'uniq_string_value': unique_id, 'rand_number': unicode(self.__get_rand_number(self.number_range)), 'random_strings': self.__get_random_strings(), }) <NEW_LINE> file_name = unique_id + '.xml' <N... | Generate XML file
@return:
XML filename and XML file data. | 625941b4a79ad161976cbf1d |
def get_pixel_range(self, row, col): <NEW_LINE> <INDENT> range = c_double() <NEW_LINE> error_code = dll.BVTColorImage_GetPixelRange(self._handle, row, col, byref(range)) <NEW_LINE> if (0 != error_code): <NEW_LINE> <INDENT> raise sdkerror.SDKError(error_code) <NEW_LINE> <DEDENT> return range.value | Retrieve the range (from the sonar head) of the specified pixel (in meters) | 625941b43539df3088e2e123 |
def match(self, name, s): <NEW_LINE> <INDENT> return re.match(self.get_pattern(name, r".*"), s) | :param name: Name of the regex pattern
:param s: Target string to try to match with | 625941b421a7993f00bc7ac0 |
def add_transaction(self, sender, receiver, amount): <NEW_LINE> <INDENT> self.transactions.append({'sender': sender, 'receiver': receiver, 'amount': amount}) <NEW_LINE> previous_block = self.get_previous_block() <NEW_LINE> return previous_block['index'] + 1 | Add transaction to the block | 625941b494891a1f4081b87f |
def populate_uuids(apps, schema_editor): <NEW_LINE> <INDENT> Document = apps.get_model('wiki', 'Document') <NEW_LINE> docs = Document.objects.filter(uuid__isnull=True) <NEW_LINE> for document_id in docs.values_list('id', flat=True).iterator(): <NEW_LINE> <INDENT> Document.objects.filter(id=document_id).update(uuid=uuid... | Populate Document.uuid, without bumping last modified. | 625941b41d351010ab8558fd |
def on_reset_delivery(self, delivery: QuicDeliveryState) -> None: <NEW_LINE> <INDENT> if delivery == QuicDeliveryState.ACKED: <NEW_LINE> <INDENT> self.is_finished = True <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> self.reset_pending = True | Callback when a reset is ACK'd. | 625941b4d164cc6175782b25 |
def index(self, value: float) -> int: <NEW_LINE> <INDENT> cursor = self.head <NEW_LINE> index = 0 <NEW_LINE> while cursor: <NEW_LINE> <INDENT> if cursor.val == value: <NEW_LINE> <INDENT> return index <NEW_LINE> <DEDENT> cursor = cursor.next <NEW_LINE> index += 1 <NEW_LINE> <DEDENT> return -1 | Find value index position in linked list
:param value: float
:return: index position | 625941b48e7ae83300e4ada3 |
def ratios(self): <NEW_LINE> <INDENT> emp = EmploymentData(self.data, self.dpath) <NEW_LINE> emp.id_extend() <NEW_LINE> emp.ratio_gen() <NEW_LINE> self.data = emp.data_merge() <NEW_LINE> self.data = self.data[pd.notnull(self.data['id'])] <NEW_LINE> self.data = self.data[pd.notnull(self.data['school'])] <NEW_LINE> for n... | Generate transparency ratios | 625941b497e22403b379cd70 |
def test_timeseries(): <NEW_LINE> <INDENT> pass | >>> import wafo.data
>>> import wafo.objects as wo
>>> x = wafo.data.sea()
>>> ts = wo.mat2timeseries(x)
>>> ts.sampling_period()
0.25
Estimate spectrum
>>> S = ts.tospecdata()
>>> S.data[:10]
array([ 0.00913087, 0.00881073, 0.00791944, 0.00664244, 0.00522429,
0.00389816, 0.00282753, 0.00207843, 0.00162... | 625941b4aad79263cf390812 |
def no_file_warning(file): <NEW_LINE> <INDENT> messagebox.showwarning("Not Found", file + " could not be opened") | Warn if the given file path does not exist | 625941b49b70327d1c4e0bac |
def _disallow_batch_hooks_in_ps_strategy(self): <NEW_LINE> <INDENT> strategy = tf.distribute.get_strategy() <NEW_LINE> if strategy._should_use_with_coordinator: <NEW_LINE> <INDENT> unsupported_callbacks = [] <NEW_LINE> for cb in self.callbacks: <NEW_LINE> <INDENT> if getattr(cb, '_supports_tf_logs', False): <NEW_LINE> ... | Error out if batch-level callbacks are passed with PSStrategy. | 625941b40c0af96317bb7fc1 |
def test_read_default(self): <NEW_LINE> <INDENT> mock = MagicMock() <NEW_LINE> with patch.dict(macdefaults.__salt__, {'cmd.run': mock}): <NEW_LINE> <INDENT> macdefaults.read('com.apple.CrashReporter', 'Crash') <NEW_LINE> mock.assert_called_once_with('defaults read "com.apple.CrashReporter" "Crash"', runas=None) | Test reading a default setting | 625941b4187af65679ca4efc |
def setup(self, static, moving): <NEW_LINE> <INDENT> self.static = static <NEW_LINE> self.moving = moving | Setup static and moving sets of streamlines.
Parameters
----------
static : streamlines
Fixed or reference set of streamlines.
moving : streamlines
Moving streamlines.
Notes
-----
Call this after the object is initiated and before distance.
Num_threads is not used in this class. Use ``BundleMinDistanceMetric... | 625941b4d18da76e235322a8 |
def setUp(self): <NEW_LINE> <INDENT> self.table_names = [f for f in os.listdir('python/tests/tables/') if os.path.isfile(os.path.join('python/tests/tables/', f))] <NEW_LINE> self.doc_names = [f for f in os.listdir('python/tests/docs/') if os.path.isfile(os.path.join('python/tests/docs/', f))] <NEW_LINE> self.sbtab_docs... | setup SBtabTable class with files from test directory | 625941b431939e2706e4cc49 |
def grad2deg(grad): <NEW_LINE> <INDENT> return (grad * 9) / 10 | grad2deg(100.00) --> 90.0 | 625941b4d268445f265b4c4d |
def runSimulation(N,gridX,gridY,gridP): <NEW_LINE> <INDENT> results = [] <NEW_LINE> for take in range(N): <NEW_LINE> <INDENT> grid = Grid(gridX,gridY,gridP) <NEW_LINE> results.append(grid.letItRain()) <NEW_LINE> <DEDENT> return sum(results)/float(len(results)) | takes Grid object, lets water propagate through from top to bottom
returns list of results (number of ) | 625941b43d592f4c4ed1ce59 |
def amber_energies(prmtop, crd, input, amb_path): <NEW_LINE> <INDENT> logger.info('Evaluating energy of {0}'.format(crd)) <NEW_LINE> directory, _ = os.path.split(os.path.abspath(prmtop)) <NEW_LINE> if not prmtop: <NEW_LINE> <INDENT> prmtop = os.path.join(directory, 'parm.prmtop') <NEW_LINE> <DEDENT> if not crd: <NEW_LI... | Compute single-point energies using AMBER.
Args:
prmtop (str):
crd (str):
input (str)
amb_path (str):
Returns:
e_out:
ener_xvg: | 625941b43eb6a72ae02ec2b2 |
def inuse(self): <NEW_LINE> <INDENT> self.inuse_bit_at_offset(self.size & ~SIZE_BITS) | extract p's inuse bit | 625941b4498bea3a759b988a |
def __init__(self, hash: chainhash.Hash, height: int, bits: int, block_size: int, block_weight: int, num_txns: int, total_txns: int, median_time: int): <NEW_LINE> <INDENT> self.hash = hash <NEW_LINE> self.height = height <NEW_LINE> self.bits = bits <NEW_LINE> self.block_size = block_size <NEW_LINE> self.block_weight = ... | :param chainhash.Hash *chainhash.Hash hash:
:param int int32 height:
:param int uint32 bits:
:param int uint64 block_size:
:param int uint64 block_weight:
:param int uint64 num_txns:
:param int uint64 total_txns:
:param int time.Time median_time: | 625941b44527f215b584c234 |
def init_kernel(self): <NEW_LINE> <INDENT> shell_stream = ZMQStream(self.shell_socket) <NEW_LINE> control_stream = ZMQStream(self.control_socket) <NEW_LINE> kernel_factory = self.kernel_class.instance <NEW_LINE> kernel = kernel_factory(parent=self, session=self.session, shell_streams=[shell_stream, control_stream], iop... | Create the Kernel object itself | 625941b4b7558d58953c4cf4 |
def test_get_schema(self): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> Draft4Validator.check_schema(snapshot.SnapshotView.GET_SCHEMA) <NEW_LINE> schema_valid = True <NEW_LINE> <DEDENT> except RuntimeError: <NEW_LINE> <INDENT> schema_valid = False <NEW_LINE> <DEDENT> self.assertTrue(schema_valid) | The schema defined for GET on base end point is valid | 625941b48a43f66fc4b53e42 |
def test_main_discover(mocker, mock_config, mock_context): <NEW_LINE> <INDENT> mocker.patch( 'tap_intacct.singer.utils.parse_args', return_value=argparse.Namespace(config=mock_config, discover=True, state=None), ) <NEW_LINE> mock_do_discover = mocker.patch('tap_intacct.do_discover', autospec=True) <NEW_LINE> tap_intacc... | Ensure that the correct functions are called when tap is executed in discovery mode. | 625941b41b99ca400220a888 |
def edit_reservation(hotel_id, database): <NEW_LINE> <INDENT> _, options = get_reservations(hotel_id, database) <NEW_LINE> while True: <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> user_choice = int( input('Which reservation do you want to manage? >')) <NEW_LINE> try: <NEW_LINE> <INDENT> reservation_id = options[user_ch... | Edits a reservation.
Args:
hotel_id(int): The id of the hotel the reservation belongs to.
database(str): The database to connect to.
Raises:
ReservationDoesNotExist: Returns if the reservation can not be found.
Returns:
(obj): A reservation object. | 625941b45510c4643540f1d2 |
def sorted(self): <NEW_LINE> <INDENT> cache_sorted = self._cache.get('sorted') <NEW_LINE> if cache_sorted is not None: <NEW_LINE> <INDENT> return cache_sorted <NEW_LINE> <DEDENT> shallowElements = copy.copy(self._elements) <NEW_LINE> shallowEndElements = copy.copy(self._endElements) <NEW_LINE> s = copy.copy(self) <NEW_... | (TL;DR: you probably do not need to call this method unless you have turned `.autoSort` to
off.)
Returns a new Stream where all the elements are sorted according to offset time, then
priority, then classSortOrder (so that, for instance, a Clef at offset 0 appears before
a Note at offset 0).
If this Stream is not flat... | 625941b4004d5f362079a110 |
def move(self, direction): <NEW_LINE> <INDENT> self.direction = direction <NEW_LINE> self.write("S") | Move the stepper motor continuously in the given direction until a stop command is sent
or a limit switch is reached. This method corresponds to the 'slew' command in the DP
series instrument manuals.
:param direction: value to set on the direction property before moving the motor. | 625941b47b180e01f3dc45e0 |
def tearDown(self): <NEW_LINE> <INDENT> db.session.remove() <NEW_LINE> db.drop_all() | Get rid of the database again after each test. | 625941b43617ad0b5ed67cd7 |
def Process_Loading_Render_Painting_Network(self): <NEW_LINE> <INDENT> with open(self.trace, 'r') as content_file: <NEW_LINE> <INDENT> content = content_file.read() <NEW_LINE> <DEDENT> trace_events = json.loads(content) <NEW_LINE> for trace_event in trace_events: <NEW_LINE> <INDENT> trace_event = json.loads(trace_event... | f = None
try:
file_name, ext = os.path.splitext(trace)
if ext.lower() == '.gz':
f = gzip.open(trace, 'rb')
else:
f = open(trace, 'r')
for line in f:
try:
trace_event = json.loads(line.strip("[]
,"))
cat = trace_event['cat']
name = trace_event... | 625941b4293b9510aa2c3072 |
@pytest.mark.django_db <NEW_LINE> def test_empty_object_returned_with_authentication(api_client, activities_url): <NEW_LINE> <INDENT> sender = _auth_sender(activities_url) <NEW_LINE> response = api_client.get( activities_url, content_type='', HTTP_AUTHORIZATION=sender.request_header, HTTP_X_FORWARDED_FOR='1.2.3.4, 123.... | If the Authorization and X-Forwarded-For headers are correct, then
the correct, and authentic, data is returned | 625941b4f7d966606f6a9de1 |
def on_select(self, event): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> self.idx = int(self.listbox.curselection()[0]) <NEW_LINE> <DEDENT> except IndexError: <NEW_LINE> <INDENT> self.idx = None <NEW_LINE> <DEDENT> if self.idx is not None: <NEW_LINE> <INDENT> self.included_lines_box.delete(0, tk.END) <NEW_LINE> if self... | Handle selection of a new line.
| 625941b45e10d32532c5ed08 |
def _parse_allotment_line(tr): <NEW_LINE> <INDENT> if not tr: <NEW_LINE> <INDENT> return None <NEW_LINE> <DEDENT> tds = tr.find_all('td') <NEW_LINE> if tds and len(tds) >= 9: <NEW_LINE> <INDENT> return { '公告日期': tds[0].text.strip(), '配股方案(每10股配股股数)': tds[1].text.strip(), '配股价格(元)': tds[2].text.strip(), '基准股本(万股)': tds[... | 解析配股数据tr行 | 625941b4d10714528d5ffab7 |
def __init__(self, problem, deadline): <NEW_LINE> <INDENT> SimulatedAnnealing.__init__(self, problem, deadline) <NEW_LINE> self.heuristic = "all" <NEW_LINE> self.prob_l = self._get_problem_length() | Create an Algorithm according to the heuristic. | 625941b492d797404e303f63 |
def source_element_initializer(self): <NEW_LINE> <INDENT> if self.dynamic_initializer: <NEW_LINE> <INDENT> return Statement(self.data_object_manager.get_namespace() + "::" + self.name + ".init()") <NEW_LINE> <DEDENT> return [] | Builds a dynamic initialization statement.
:returns: A Statement invoking the init() method of the object | 625941b4a8370b771705267a |
def tycat(*things): <NEW_LINE> <INDENT> print("[", Displayer.file_count, "]") <NEW_LINE> user = getpass.getuser() <NEW_LINE> directory = "/tmp/{}".format(user) <NEW_LINE> if not os.path.exists(directory): <NEW_LINE> <INDENT> os.makedirs(directory) <NEW_LINE> <DEDENT> filename = "{}/{}.svg".format(directory, str(Display... | graphically displays all objects given.
each argument will be displayed in a different color.
requires :
- the terminology terminal emulator
- each object either implements
* bounding_quadrant
* svg_content
or is an iterable on things implementing it. | 625941b499cbb53fe67929c0 |
def get_nn_domain_from_constraints(nn_type, constraint_dict): <NEW_LINE> <INDENT> constraint_dict = copy(constraint_dict) <NEW_LINE> mandatory_keys = ['max_num_layers', 'max_mass'] <NEW_LINE> optional_key_vals = [('min_num_layers', 5), ('min_mass', 0), ('max_out_degree', np.inf), ('max_in_degree', np.inf), ('max_num_ed... | nn_type is the type of the network.
See CNNConstraintChecker, MLPConstraintChecker, NNConstraintChecker constructors
for args and kwargs. | 625941b4b7558d58953c4cf5 |
def testAmenityIds(self): <NEW_LINE> <INDENT> self.assertIs(type(Place.amenity_ids), list) <NEW_LINE> self.assertIs(type(Place.amenity_ids[0]), str) | Tests the amenity_ids attribute | 625941b4d58c6744b4257a39 |
def MakeEncoder(InputFunction,EncoderArchitecture): <NEW_LINE> <INDENT> inputEncoder=InputFunction <NEW_LINE> en=Dense(EncoderArchitecture[0])(inputEncoder) <NEW_LINE> en=Activation('relu')(en) <NEW_LINE> for j in range(len(EncoderArchitecture)-1): <NEW_LINE> <INDENT> en=Dense(EncoderArchitecture[j+1])(en) <NEW_LINE> e... | Generates the encoder network using the functional API from keras
Its Intended as a wrapper function for TrainAutoencoder
InputFunction Input function from the keras functional API
EncoderArchitecture A list with the number of dense units in the layer,
the lenght of the list is the number of la... | 625941b44f6381625f11481f |
def report_version(self, data): <NEW_LINE> <INDENT> self.firmata_version.append(data[0]) <NEW_LINE> self.firmata_version.append(data[1]) | This method processes the report version message, sent asynchronously by Firmata when it starts up
or after refresh_report_version() is called
Use the api method api_get_version to retrieve this information
:param data: Message data from Firmata
:return: No return value. | 625941b4b830903b967e96f1 |
def DoDragDrop(self,data,allowedEffects): <NEW_LINE> <INDENT> pass | DoDragDrop(self: ToolStripItem,data: object,allowedEffects: DragDropEffects) -> DragDropEffects
Begins a drag-and-drop operation.
data: The object to be dragged.
allowedEffects: The drag operations that can occur.
Returns: One of the System.Windows.Forms.DragDropEffects values. | 625941b431939e2706e4cc4a |
def const_s(v): <NEW_LINE> <INDENT> result = Stream(v, lambda: result) <NEW_LINE> return result | The constant stream of Vs. | 625941b46fb2d068a760ee7a |
def test_sell_btc_jpy(): <NEW_LINE> <INDENT> time.sleep(1) <NEW_LINE> sell_value = 2 * market_info.get('bid') <NEW_LINE> o1 = order.Order(access_key=settings.access_key, secret_key=settings.secret_key) <NEW_LINE> ok_(o1.sell_btc_jpy(amount = 0.01, rate = sell_value).get('success')) | sell BTC | 625941b41f037a2d8b945fd8 |
def to_indexer(self, cutoff=None, from_cutoff=True): <NEW_LINE> <INDENT> if from_cutoff: <NEW_LINE> <INDENT> return self.to_relative(cutoff).to_pandas() - 1 <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> relative = self.to_relative(cutoff) <NEW_LINE> return relative - relative.to_pandas()[0] | Return zero-based indexer values for easy indexing into arrays.
Parameters
----------
cutoff : pd.Period, pd.Timestamp, int, optional (default=None)
Cutoff value required to convert a relative forecasting
horizon to an absolute one and vice versa.
from_cutoff : bool, optional (default=True)
- If True, zero... | 625941b4004d5f362079a111 |
def __init__(self, form, *children, use_csrf=True, standalone=True, **kwargs): <NEW_LINE> <INDENT> self.standalone = standalone <NEW_LINE> attributes = {"method": "POST", "autocomplete": "off"} <NEW_LINE> attributes.update(kwargs) <NEW_LINE> if ( attributes["method"].upper() == "POST" and use_csrf is not False and stan... | form: lazy evaluated value which should resolve to the form object
children: any child elements, can be formfields or other
use_csrf: add a CSRF input, but only for POST submission and standalone forms
standalone: if true, will add a CSRF token and will render enclosing FORM-element | 625941b48c0ade5d55d3e798 |
def call_finished(location_uri, headers, timeout): <NEW_LINE> <INDENT> logger = logging.getLogger(__name__) <NEW_LINE> logger.debug(f"GET API call: {location_uri}, timeout {timeout} seconds and headers: {headers}") <NEW_LINE> response = requests.get(location_uri, headers=headers, proxies=get_proxies(location_uri), time... | :param location_uri: URI to check the status of API call
:param headers: HTTP headers
:param timeout: connect timeout
:return indication and response tuple | 625941b48a349b6b435e7f4e |
def get_servers_input(self): <NEW_LINE> <INDENT> if not self.so: return <NEW_LINE> sockdata= str() <NEW_LINE> while True: <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> sockdata,addr= self.so.recvfrom(data_size) <NEW_LINE> sockdata = sockdata.decode('utf-8') <NEW_LINE> <DEDENT> except socket.error as emsg: <NEW_LINE> <IN... | Server's input is stored in a ServerState object | 625941b450812a4eaa59c0ff |
@click.command() <NEW_LINE> @decorators.catch_all <NEW_LINE> @decorators.json_output <NEW_LINE> @decorators.capture_usage <NEW_LINE> def inbox(ctx): <NEW_LINE> <INDENT> return _inbox(ctx.obj['config'], ctx.obj['client']) | View notifications from 21.co. | 625941b44a966d76dd550de4 |
def main(args=sys.argv[1:]): <NEW_LINE> <INDENT> opt = docopt(main.__doc__.strip(), args, options_first=True) <NEW_LINE> config_logging(opt['--verbose']) <NEW_LINE> if opt['check']: <NEW_LINE> <INDENT> check_backends(opt['--title']) <NEW_LINE> <DEDENT> elif opt['extract']: <NEW_LINE> <INDENT> handler = fulltext.get <NE... | Extract text from a file.
Commands:
extract - extract text from path
check - make sure all deps are installed
Usage:
fulltext extract [-v] [-f] <path>...
fulltext check [-t]
Options:
-f, --file Open file first.
-t, --title Check deps for title.
-v, --verbose Mo... | 625941b4baa26c4b54cb0efd |
def total_description_type_support(self, string): <NEW_LINE> <INDENT> support = self.local_description_type_support(string) <NEW_LINE> return round((support + self.activation) / 2.0) | Return the total description type support with string.
A function of the local description type support and the node's
activation. | 625941b4283ffb24f3c556e7 |
def append_df_to_excel(filename, df, sheet_name='Sheet1', startrow=None, truncate_sheet=False, **to_excel_kwargs): <NEW_LINE> <INDENT> from openpyxl import load_workbook <NEW_LINE> import pandas as pd <NEW_LINE> if 'engine' in to_excel_kwargs: <NEW_LINE> <INDENT> to_excel_kwargs.pop('engine') <NEW_LINE> <DEDENT> writer... | Append a DataFrame [df] to existing Excel file [filename]
into [sheet_name] Sheet.
If [filename] doesn't exist, then this function will create it.
Parameters:
filename : File path or existing ExcelWriter
(Example: '/path/to/file.xlsx')
df : dataframe to save to workbook
sheet_name : Name of sheet wh... | 625941b4d7e4931a7ee9dcf5 |
def bindAnim(self, animName, partName = None, lodName = None, allowAsyncBind = False): <NEW_LINE> <INDENT> self.getAnimControls(animName = animName, partName = partName, lodName = lodName, allowAsyncBind = allowAsyncBind) | Binds the named animation to the named part and/or lod. If
allowAsyncBind is False, this guarantees that the animation is
bound immediately--the animation is never bound in a
sub-thread; it will be loaded and bound in the main thread, so
it will be available by the time this method returns.
The parameters are the sam... | 625941b44e696a04525c922e |
def Qmix_mixer_checkers(agent_qs, state_env, state, goals_all, state_dim, goal_dim, n_agents, f1=4, k1=[3,5]): <NEW_LINE> <INDENT> conv = convnet_1(state_env, f1=f1, k1=k1, s1=[1,1], scope='conv') <NEW_LINE> embed_dim = 128 <NEW_LINE> state_goals_dim = state_dim + n_agents*goal_dim + conv.get_shape().as_list()[1] <NEW_... | Args:
agent_qs: shape [batch, n_agents]
state_env: shape [batch, rows, cols, channels]
state: shape [batch, state_dim]
goals_all: shape [batch, n_agents*goal_dim] | 625941b430bbd722463cbb9c |
@pypub.command() <NEW_LINE> def changelog(): <NEW_LINE> <INDENT> shell(f"{Publisher.CMD_CHANGELOG} -u | less") | Preview the changelog. | 625941b49f2886367277a671 |
def get_conn(self): <NEW_LINE> <INDENT> if self.conn_queue.qsize() == 0: <NEW_LINE> <INDENT> cn = self._build_conn() <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> while True: <NEW_LINE> <INDENT> cn = self.conn_queue.get_nowait() <NEW_LINE> if not cn.stat().startswith('Uptime:'): <NEW_LINE> <IND... | 获取mysql连接 | 625941b47b180e01f3dc45e1 |
def _set_status(self, status): <NEW_LINE> <INDENT> self._lock.acquire() <NEW_LINE> try: <NEW_LINE> <INDENT> self._status = status <NEW_LINE> if (status > 0) and self._context._verbose: <NEW_LINE> <INDENT> action = STATUSMAP[status] <NEW_LINE> print("Yoton: %s at %s:%s." % (action, self._hostname1, self._port1)) <NEW_LI... | Used to change the status. Subclasses can reimplement this
to get the desired behavior. | 625941b426238365f5f0ec42 |
def get_group(self, user, user_partition, assign=True): <NEW_LINE> <INDENT> return user_partition.scheme.get_group_for_user( self._course_id, user, user_partition, assign=assign, ) | Returns the group from the specified user partition to which the user is assigned.
If the user has not yet been assigned, a group will be chosen for them based upon
the partition's scheme. | 625941b42eb69b55b151c683 |
def limit_to_chromosomes(df, chromosomes): <NEW_LINE> <INDENT> if chromosomes: <NEW_LINE> <INDENT> df = df[df['Chromosome'].isin(chromosomes)] <NEW_LINE> <DEDENT> return df | Returns only records that are in chromosomes (if "chromosomes" is not empty) | 625941b445492302aab5e099 |
def deQueue(self) -> bool: <NEW_LINE> <INDENT> if self.isEmpty(): <NEW_LINE> <INDENT> return False <NEW_LINE> <DEDENT> self.queue[self.front] = None <NEW_LINE> if self.front != self.rear: <NEW_LINE> <INDENT> self.front = (self.front + 1) % self.max <NEW_LINE> <DEDENT> return True | Delete an element from the circular queue. Return true if the operation is successful. | 625941b4be383301e01b526f |
def in_place_qs(lis): <NEW_LINE> <INDENT> pivot = lis[0] <NEW_LINE> l, r = 1, len(lis) - 1 <NEW_LINE> while l <= r: <NEW_LINE> <INDENT> if lis[l] < pivot: <NEW_LINE> <INDENT> l += 1 <NEW_LINE> <DEDENT> elif lis[i] > pivot: <NEW_LINE> <INDENT> r -= 1 <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> lis[r], lis[l] = lis[l],... | dicks | 625941b485dfad0860c3ac32 |
def test_never_mutate(self): <NEW_LINE> <INDENT> mutator = ConversionMutation(mutation_rate = 0.0) <NEW_LINE> self._never_mutate(mutator) | Make sure we do not mutate at unexpected times.
| 625941b48e71fb1e9831d588 |
def update(self, message): <NEW_LINE> <INDENT> self.name = str(message.name) <NEW_LINE> if message.HasField('inherit'): <NEW_LINE> <INDENT> self.inherit = bool(message.inherit) <NEW_LINE> <DEDENT> if message.HasField('inherited'): <NEW_LINE> <INDENT> self.inherited = bool(message.inherited) <NEW_LINE> <DEDENT> if messa... | Update a ACL information, based on the incoming message | 625941b499cbb53fe67929c1 |
def epcr_parse(self): <NEW_LINE> <INDENT> logging.info('Parsing ePCR outputs') <NEW_LINE> for sample in self.metadata: <NEW_LINE> <INDENT> if sample.general.bestassemblyfile != 'NA': <NEW_LINE> <INDENT> toxin_set = set() <NEW_LINE> if os.path.isfile(sample[self.analysistype].resultsfile): <NEW_LINE> <INDENT> with open(... | Parse the ePCR outputs | 625941b430bbd722463cbb9d |
def __scan_ports(self, ip, delay, message): <NEW_LINE> <INDENT> output = {} <NEW_LINE> thread = threading.Thread(target=self.__scan_ports_helper, args=(ip, delay, output, message)) <NEW_LINE> thread.start() <NEW_LINE> while len(output) < len(self.target_ports): <NEW_LINE> <INDENT> continue <NEW_LINE> <DEDENT> for port ... | Controller of the __scan_ports_helper() function
:param ip: the ip address that is being scanned
:type ip: str
:param delay: the time in seconds that a TCP socket waits until timeout
:type delay: int
:param message: the message that is going to be included in the scanning packets,
in order to prevent ethical problem, ... | 625941b4596a8972360898a5 |
def align_labels(tiers_list, precision=None, regex=r'[^\s]+'): <NEW_LINE> <INDENT> if precision is not None: <NEW_LINE> <INDENT> precision_old = Time._precision <NEW_LINE> Time._precision = precision <NEW_LINE> <DEDENT> if len(tiers_list) < 2: <NEW_LINE> <INDENT> raise Exception('At least two tiers need to be provided'... | Create a list of lists for all time-aligned Interval
or Point object in tiers_list, whose text matches regex.
For example:
[[label_1-tier_1, label_1-tier_2, label_1-tier_3],
[label_2-tier_1, label_2-tier_2, label_2-tier_3],
...
[label_n-tier_n, label_n-tier_n, label_n-tier_n]]
The allowed mismatch between object's... | 625941b466656f66f7cbbf84 |
def repo_contents(self, owner: str, repo: str, path: str, ref: str = 'master') -> dict: <NEW_LINE> <INDENT> return self.request('repos/{}/{}/contents/{}'.format(owner, repo, path), ref=ref) | Get repository's contents
https://developer.github.com/v3/repos/contents/#get-contents | 625941b4287bf620b61d384b |
def highlight_text(raw_text_document, lexer): <NEW_LINE> <INDENT> highlighted_text = HighlightedText(raw_text_document, lexer) <NEW_LINE> document = TextDocumentModel() <NEW_LINE> text_block = TextBlock(raw_text_document.line_slice) <NEW_LINE> document.append(text_block) <NEW_LINE> for highlighted_fragment in highlight... | Highlight a text.
The parameter *raw_text_document* is a :class:`DiffViewer.RawTextDocument` instance and the
parameter *lexer* is a Pygments lexer instance.
Return an :class:`DiffViewer.TextDocumentModel` instance. The document has one text block that
contains all the fragments. Text fragments use light views. | 625941b497e22403b379cd72 |
def four_digits_num(n: int) -> int: <NEW_LINE> <INDENT> total = 0 <NEW_LINE> while n > 0: <NEW_LINE> <INDENT> total += n % 10 <NEW_LINE> n //= 10 <NEW_LINE> <DEDENT> return(total) | >>> four_digits_num(9000)
9
>>> four_digits_num(1234)
10
>>> four_digits_num(1925)
17 | 625941b4be8e80087fb20a29 |
def highlight(self, sleep_and_stop=2): <NEW_LINE> <INDENT> self.original_style = self.get_attribute('style') <NEW_LINE> self.set_attribute('style', '"border: 2px solid red;"') <NEW_LINE> if sleep_and_stop > 0: <NEW_LINE> <INDENT> sleep(sleep_and_stop) <NEW_LINE> self.set_attribute('style', self.original_style) | turn current element's border into solid red 2px.
return original style after 'sleep_and_stop' seconds of time.sleep
if 'sleep_and_stop' is 0 - border stays and no sleep.
original stlyle saved to self.original_style | 625941b48e7ae83300e4ada6 |
def _is_episode_terminated(self): <NEW_LINE> <INDENT> dist = self.distance_from_light() <NEW_LINE> done = self.current_energy <= self.min_energy or dist < self.min_light_distance or dist > self.max_light_distance <NEW_LINE> return bool(done) | Returns a True if the episode is finished.
:return: True if the termination criteria is met, false otherwise | 625941b40c0af96317bb7fc3 |
def load_workspace(self, file_paths, merge=False, force_overwrite=False): <NEW_LINE> <INDENT> with show_busy(self._view): <NEW_LINE> <INDENT> ws_names = [os.path.splitext(os.path.basename(base))[0] for base in file_paths] <NEW_LINE> if merge: <NEW_LINE> <INDENT> if not self.file_types_match(file_paths): <NEW_LINE> <IND... | Loads one or more workspaces.
:param file_paths: list of paths to files to load
:param merge: boolean - whether to combine files into a single workspace
:param force_overwrite: int - 0: asks for overwrite, 1 always overwrite, -1 never overwrite | 625941b494891a1f4081b882 |
def __init__(self,X,k=20): <NEW_LINE> <INDENT> self.X=np.array(X) <NEW_LINE> self.k=k <NEW_LINE> self.ave=np.mean(self.X[:,2]) <NEW_LINE> print("the input data size is ",self.X.shape ) <NEW_LINE> self.bi={} <NEW_LINE> self.bu={} <NEW_LINE> self.qi={} <NEW_LINE> self.pu={} <NEW_LINE> self.movie_user={} <NEW_LINE> self.... | ''
k is the length of vector | 625941b4aad79263cf390815 |
def __init__(self, k=2, n_estimators=50, max_features=5, min_features=2, scale_features=True, verbose=0, random_state=None, voting_rule="hard", **kwargs): <NEW_LINE> <INDENT> self.n_estimators = n_estimators <NEW_LINE> self.max_features = max_features <NEW_LINE> self.min_features = min_features <NEW_LINE> self.scale_fe... | Parameters
----------
k : integer, optional (default=2)
number of clusters in k-means model
n_estimators: integer, optional (default=50)
number of k-means model to run
max_features: integer, optional (default=5)
maximum number of features to use in each model; would be
overwritten if max_features is g... | 625941b4187af65679ca4efe |
def is_permitted(body_request): <NEW_LINE> <INDENT> perm = whoami(body_request.get('authname')) <NEW_LINE> try: <NEW_LINE> <INDENT> login, permission, authed, tstamp, ip = perm[0], perm[1], perm[2], perm[3], perm[4] <NEW_LINE> <DEDENT> except TypeError: <NEW_LINE> <INDENT> return False <NEW_LINE> <DEDENT> if login in p... | Tryes to checks ttl, ipaddress and authname that graps from body
then access will be granted or forbidden | 625941b44527f215b584c236 |
def health_check(cancelled_event): <NEW_LINE> <INDENT> health_url = '{0}_info/'.format(datalab_address) <NEW_LINE> healthy = False <NEW_LINE> print('Waiting for Datalab to be reachable at ' + datalab_address) <NEW_LINE> while not cancelled_event.is_set(): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> health_resp = urlli... | Check if the Datalab instance is reachable via the connection.
After the instance is reachable, the `on_ready` method is called.
This method is meant to be suitable for running in a separate thread,
and takes an event argument to indicate when that thread should exit.
Args:
cancelled_event: A threading.Event insta... | 625941b41f5feb6acb0c492f |
def logoff_and_openrecord(self, username, password, menu1, menu2, menu3, recordname): <NEW_LINE> <INDENT> self.goback() <NEW_LINE> self.switch_account(username, password) <NEW_LINE> self.close_message() <NEW_LINE> self.open_m(menu1, menu2, menu3) <NEW_LINE> self.openagain_record(recordname) | 退出当前登录,切换账号并打开菜单记录for视图 | 625941b423e79379d52ee343 |
def deleteAll(self, exchange_id, yes_I_mean_it=False): <NEW_LINE> <INDENT> if not yes_I_mean_it: <NEW_LINE> <INDENT> raise RuntimeError("Cowardly refusing to delete all an exch's history without the 'yes_I_mean_it' flag") <NEW_LINE> <DEDENT> sql = "DELETE FROM HistTrades WHERE exchange_id=?" <NEW_LINE> cur = self.db.cu... | Delete ALL trades held for exchange with id `exchange_id`. This is useful
when you're about to batch-load a new set. | 625941b482261d6c526ab27d |
def perform_pca(data, n_components=3, y=None): <NEW_LINE> <INDENT> features = data.copy(deep=True) <NEW_LINE> if (y is not None): <NEW_LINE> <INDENT> features = data.drop(y, axis=1) <NEW_LINE> <DEDENT> x = StandardScaler().fit_transform(features) <NEW_LINE> pca = PCA(n_components=n_components) <NEW_LINE> principalCompo... | data = Data on which to perform PCA
n_components = 3. This is the number of Principal Components to
consider for the analysis | 625941b4004d5f362079a112 |
@greet.command() <NEW_LINE> def list_available(): <NEW_LINE> <INDENT> indices.list_available_index_files() | List the indices available for download.
A list is printed that shows which index locations have been
implemented with their URLs for downloading. | 625941b4be7bc26dc91cd3e1 |
def get_structuralresources_content_constraints(limit=25, offset=0, query='', orderby=''): <NEW_LINE> <INDENT> path = 'contentConstraints' <NEW_LINE> url = services.build_entrypoint_url( API, path, limit=limit, offset=offset, query=query, orderBy=orderby ) <NEW_LINE> return services.get_content(url) | Get content constraints
This function returns the content from ``/v1.0/contentConstraints``
Args:
limit (int): Results limit. By default ``limit = 25``.
offset (int): Displacement. Result from which it is returned. By default
``offset = 0``.
query (string): Query to filter the results.
orderby... | 625941b45510c4643540f1d4 |
def lnprior(pars): <NEW_LINE> <INDENT> logprob = uniform_prior(pars[0], 0., np.inf) + uniform_prior(pars[1], -1, 5) <NEW_LINE> return logprob | Return probability of parameter values according to prior knowledge.
Parameter limits should be done here through uniform prior ditributions | 625941b4a934411ee3751476 |
def GeneralFastOCR(self, request): <NEW_LINE> <INDENT> try: <NEW_LINE> <INDENT> params = request._serialize() <NEW_LINE> body = self.call("GeneralFastOCR", params) <NEW_LINE> response = json.loads(body) <NEW_LINE> if "Error" not in response["Response"]: <NEW_LINE> <INDENT> model = models.GeneralFastOCRResponse() <NEW_L... | 本接口支持图片中整体文字的检测和识别,返回文字框位置与文字内容。相比通用印刷体识别接口,识别速度更快。
:param request: Request instance for GeneralFastOCR.
:type request: :class:`tencentcloud.ocr.v20181119.models.GeneralFastOCRRequest`
:rtype: :class:`tencentcloud.ocr.v20181119.models.GeneralFastOCRResponse` | 625941b4e5267d203edcda7c |
def _get_standard_pressure_levels2(self): <NEW_LINE> <INDENT> d = dict() <NEW_LINE> for sec, standard_pressure_level in self.data.standard_pressure_levels.items(): <NEW_LINE> <INDENT> d[standard_pressure_level.press] = standard_pressure_level <NEW_LINE> <DEDENT> return d | Return the same dict, but where keys are pressure levels instead of secs. This way we reduce
the number of loops per pressure reading in the wind_data | 625941b41b99ca400220a88b |
def create(self): <NEW_LINE> <INDENT> self.elem = etree.makeEtreeElement(str(self.qname), SOAPHeaderBase.DEFAULT_ELEMENT_NS_PREFIX, SOAPHeaderBase.DEFAULT_ELEMENT_NS) | Create header ElementTree element | 625941b45e10d32532c5ed0a |
def _prepare_invoice(self): <NEW_LINE> <INDENT> vals = super()._prepare_invoice() <NEW_LINE> self._get_payment_mode_vals(vals) <NEW_LINE> return vals | Copy bank partner from sale order to invoice | 625941b4ad47b63b2c509d66 |
def takepicture(self): <NEW_LINE> <INDENT> camera = cv2.VideoCapture(0) <NEW_LINE> while True: <NEW_LINE> <INDENT> check, image = camera.read() <NEW_LINE> cv2.imshow('image', image) <NEW_LINE> if cv2.waitKey(1) & 0xFF == ord('s'): <NEW_LINE> <INDENT> cv2.imwrite(str(self.textname.get(1.0, END))+str(self.textsurname.get... | takes a picture of a guest | 625941b4ec188e330fd5a582 |
def get_olx_hash_for_definition_key(def_key): <NEW_LINE> <INDENT> if def_key.bundle_version: <NEW_LINE> <INDENT> files_list = get_bundle_version_files_cached(def_key.bundle_uuid, def_key.bundle_version) <NEW_LINE> <DEDENT> else: <NEW_LINE> <INDENT> files_list = get_bundle_draft_files_cached(def_key.bundle_uuid, def_key... | Given a BundleDefinitionLocator, which identifies a specific version of an
OLX file, return the hash of the OLX file as given by the Blockstore API. | 625941b44428ac0f6e5ba5d4 |
@pytest.fixture() <NEW_LINE> def cleandir() -> str: <NEW_LINE> <INDENT> with tempfile.TemporaryDirectory() as tmpdirname: <NEW_LINE> <INDENT> old_dir = os.getcwd() <NEW_LINE> os.chdir(tmpdirname) <NEW_LINE> yield tmpdirname <NEW_LINE> os.chdir(old_dir) | This fixture will use the stdlib `tempfile` module to
move the current working directory to a tmp-dir for the
duration of the test.
Afterwards, the session returns to its previous working
directory, and the temporary directory and its contents
are removed.
Yields
------
str
The name of the temporary directory. | 625941b4d58c6744b4257a3b |
def test_coordinates(self): <NEW_LINE> <INDENT> pass | Ensure the POINTS are respecting lat/long from returned data | 625941b4ac7a0e7691ed3eb5 |
def __init__(self, feat_appear_limit=20): <NEW_LINE> <INDENT> self._dic = None <NEW_LINE> self._count = None <NEW_LINE> self._feat_appear_limit = feat_appear_limit | @feat_appear_limit: int | 625941b4c4546d3d9de72813 |
def read_topology_from_file(file_name): <NEW_LINE> <INDENT> import json <NEW_LINE> with open(file_name, 'r') as f: <NEW_LINE> <INDENT> return json.loads(f.read()) | Read saved topology.
:param string file_name: Path to saved topology file.
:return: dictionary in which agents names are the keys and lists of neighbours agents names are values.
:rtype: dictionary | 625941b430dc7b7665901746 |
def compute_vanderijt_edge_balance( dgraph: nx.DiGraph) -> Dict[tuple, Dict[str, int]]: <NEW_LINE> <INDENT> edge_sign = {} <NEW_LINE> for edge in dgraph.edges(): <NEW_LINE> <INDENT> nodes3 = 0 <NEW_LINE> balanced_node3 = 0 <NEW_LINE> nodes = list(set(dgraph.nodes()) - set(edge)) <NEW_LINE> xij = dgraph.get_edge_data(ed... | Computes edge balance based on Van De Rijt (2011).
Args:
dgraph: Directed weighted graph to apply edge balance.
Returns:
Dictionary of edges mapped to the number of balanced
and total number of triads the edge is involved in.
Raises:
None | 625941b46fb2d068a760ee7c |
def fit(self, dataset, n_episodes_per_epoch=1000, nb_epochs=1, n_pos=1, n_neg=9, log_every_n_samples=10, **kwargs): <NEW_LINE> <INDENT> time_start = time.time() <NEW_LINE> n_tasks = len(dataset.get_task_names()) <NEW_LINE> n_test = self.test_batch_size <NEW_LINE> feed_total, run_total = 0, 0 <NEW_LINE> for epoch in ran... | Fits model on dataset using cached supports.
For each epcoh, sample n_episodes_per_epoch (support, test) pairs and does
gradient descent.
Parameters
----------
dataset: dc.data.Dataset
Dataset to fit model on.
nb_epochs: int, optional
number of epochs of training.
n_episodes_per_epoch: int, optional
Number of (... | 625941b4b830903b967e96f3 |
def updateTopologyViews(self, **kwargs): <NEW_LINE> <INDENT> allParams = ['applicationUuid', 'pageUuid', 'topologyViewDtoList'] <NEW_LINE> params = locals() <NEW_LINE> for (key, val) in list(params['kwargs'].items()): <NEW_LINE> <INDENT> if key not in allParams: <NEW_LINE> <INDENT> raise TypeError("Got an unexpected ke... | updateTopologyViews
Args:
applicationUuid, str: Topology Application Uuid (required)
pageUuid, str: Topology Application Page Uuid (required)
topologyViewDtoList, list[TopologyViewDto]: page (required)
Returns: TaskIdResult | 625941b42ae34c7f2600cf0c |
def projectRerunRequest(self, project, item, clearError): <NEW_LINE> <INDENT> cmdstring="project-rerun" <NEW_LINE> fields = [] <NEW_LINE> fields.append(Input('cmd', cmdstring)) <NEW_LINE> fields.append(Input('version', "1")) <NEW_LINE> fields.append(Input('item', item)) <NEW_LINE> if clearError: <NEW_LINE> <INDENT> fie... | Force a rerun and optionally clear an error. | 625941b41f037a2d8b945fda |
def delta_0f(beta): <NEW_LINE> <INDENT> beta = ZZ(round(beta)) <NEW_LINE> return BKZ._delta_0f(beta) | Compute root-Hermite factor `δ_0` from block size `β`. | 625941b4462c4b4f79d1d4ab |
def parse_parameter(self, testblock_name, metric_name, params): <NEW_LINE> <INDENT> metric_type = "time" <NEW_LINE> unit = "s" <NEW_LINE> split_name = metric_name.split("::") <NEW_LINE> if len(split_name) != 2: <NEW_LINE> <INDENT> raise ATFConfigurationError("no valid metric name for metric '%s' in testblock '%s'" %(me... | Method that returns the metric method with the given parameter.
:param params: Parameter | 625941b47c178a314d6ef232 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.