source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(s, tokenize=True, encoding='utf-8', **kwargs): tags = [] for sentence in parse(s, tokenize, True, False, False, False, encoding, **kwargs).split(): for token in sentence: tags.append((token[0], token[1])) return tags
tag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, tokenize=True, encoding='utf-8', **kwargs):\n tags = []\n for sentence in parse(s, tokenize, True, False, False, False, encoding,\n **kwargs).s...
def x(s, top=10, **kwargs): return parser.find_keywords(s, top=top, frequency=parser.frequency)
keywords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, top=10, **kwargs):\n return parser.find_keywords(s, top=top, frequency=parser.frequency)\n", "role": "user" }, { "content": "keywords", "...
def x(text, tokenizer=None): _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer() return SentenceTokenizer(tokenizer=_tokenizer).itokenize(text)
sent_tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, tokenizer=None):\n _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()\n return SentenceTokenizer(tokenizer=_tokenizer).itoke...
def x(text, tokenizer=None, include_punc=True, *args, **kwargs): _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer() words = chain.from_iterable(WordTokenizer(tokenizer=_tokenizer). itokenize(sentence, include_punc, *args, **kwargs) for sentence in sent_tokenize(text, tokenizer=_tokenizer)) return words
word_tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, tokenizer=None, include_punc=True, *args, **kwargs):\n _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()\n words = chain.fr...
def x(self, text, include_punc=True, nested=False): self.tokens = [w for w in (self.word_tokenize(s, include_punc) for s in self.sent_tokenize(text))] if nested: return self.tokens else: return list(chain.from_iterable(self.tokens))
tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, include_punc=True, nested=False):\n self.tokens = [w for w in (self.word_tokenize(s, include_punc) for s in\n self.sent_tokenize(text))]\...
def x(self, text, **kwargs): sentences = self.sent_tok.tokenize(text, realign_boundaries=kwargs.get( 'realign_boundaries', True)) return sentences
sent_tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, **kwargs):\n sentences = self.sent_tok.tokenize(text, realign_boundaries=kwargs.get(\n 'realign_boundaries', True))\n return sentences...
def x(self, text, include_punc=True): if text.strip() == '': return [] _tokens = self.word_tok.tokenize(text) if len(_tokens) == 1: if _tokens[0] in PUNCTUATION: if include_punc: return _tokens else: return [] if include_punc: return _tokens else: words = [(word if word.startswith("'") else strip_punc(word, all= False)) for word in _tokens if strip_punc(word, all=False)] return list(words)
word_tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, include_punc=True):\n if text.strip() == '':\n return []\n _tokens = self.word_tok.tokenize(text)\n if len(_tokens) == 1:\n ...
def x(self, text, **kwargs): sentences = find_sentences(text, punctuation=kwargs.get('punctuation', PUNCTUATION), abbreviations=kwargs.get('abbreviations', ABBREVIATIONS_DE), replace=kwargs.get('replace', replacements), linebreak='\\n{2,}') return sentences
sent_tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, **kwargs):\n sentences = find_sentences(text, punctuation=kwargs.get('punctuation',\n PUNCTUATION), abbreviations=kwargs.get('abbreviatio...
def x(self, text, include_punc=True, **kwargs): return self.tokenizer.word_tokenize(text, include_punc, **kwargs)
tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, include_punc=True, **kwargs):\n return self.tokenizer.word_tokenize(text, include_punc, **kwargs)\n", "role": "user" }, { "content":...
def x(self, text): if text.strip() == '': return '' elif text.strip() in PUNCTUATION: _sym = text.strip() if _sym in tuple('.?!'): _tag = '.' else: _tag = _sym if self.lemmata: return '{0}/{1}/O/O/{0}'.format(_sym, _tag) else: return '{0}/{1}/O/O'.format(_sym, _tag) if self.tokenize: _tokenized = ' '.join(self.tokenizer.tokenize(text)) else: _tokenized = text _parsed = pattern_parse(_tokenized, tokenize=False, tags=self.tags, chunks=self.chunks, relations=self.relations, lemmata=self.lemmata, encoding=self.encoding, tagset=self.tagset) if self.pprint: _parsed = pattern_pprint(_parsed) return _parsed
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text):\n if text.strip() == '':\n return ''\n elif text.strip() in PUNCTUATION:\n _sym = text.strip()\n if _sym in tuple('.?!'...
def x(self, text): _extracted = [] if text.strip() == '': return _extracted parsed_sentences = self._parse_text(text) for s in parsed_sentences: tokens = s.split() new_np = [] for t in tokens: w, tag, phrase, role = t.split('/') if 'NP' in phrase and not self._is_verb(w, tag): if len(new_np) > 0 and w.lower() in START_NEW_NP: _extracted.append(' '.join(new_np)) new_np = [w] else: new_np.append(w.lower() if tokens[0].startswith(w) and not tag.startswith('N') else w) else: if len(new_np) > 0: _extracted.append(' '.join(new_np)) new_np = [] return self._filter_extracted(_extracted)
extract
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text):\n _extracted = []\n if text.strip() == '':\n return _extracted\n parsed_sentences = self._parse_text(text)\n for s in parsed_se...
def x(self, extracted_list): _filtered = [] for np in extracted_list: _np = np.split() if _np[0] in INSIGNIFICANT: _np.pop(0) try: if _np[-1] in INSIGNIFICANT: _np.pop(-1) if _np[0] in INSIGNIFICANT: _np.pop(0) except IndexError: _np = [] if len(_np) > 0: _filtered.append(' '.join(_np)) return _filtered
filter_extracted
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, extracted_list):\n _filtered = []\n for np in extracted_list:\n _np = np.split()\n if _np[0] in INSIGNIFICANT:\n _np.pop(0...
def x(self, text): if isinstance(self.tokenizer, PatternTokenizer): parsed_text = pattern_parse(text, tokenize=True, lemmata=False) else: _tokenized = [] _sentences = sent_tokenize(text, tokenizer=self.tokenizer) for s in _sentences: _tokenized.append(' '.join(self.tokenizer.tokenize(s))) parsed_text = pattern_parse(_tokenized, tokenize=False, lemmata=False) return parsed_text.split('\n')
parse_text
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text):\n if isinstance(self.tokenizer, PatternTokenizer):\n parsed_text = pattern_parse(text, tokenize=True, lemmata=False)\n else:\n ...
def x(self, sentence, tokenize=True): if sentence.strip() == '': return [] elif sentence.strip() in PUNCTUATION: if self.include_punc: _sym = sentence.strip() if _sym in tuple('.?!'): _tag = '.' else: _tag = _sym return [(_sym, _tag)] else: return [] if tokenize: _tokenized = ' '.join(self.tokenizer.tokenize(sentence)) sentence = _tokenized _tagged = pattern_tag(sentence, tokenize=False, encoding=self.encoding, tagset=self.tagset) if self.include_punc: return _tagged else: _tagged = [(word, t) for word, t in _tagged if not PUNCTUATION_REGEX.match(unicode(t))] return _tagged
tag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sentence, tokenize=True):\n if sentence.strip() == '':\n return []\n elif sentence.strip() in PUNCTUATION:\n if self.include_punc:\n ...
def x(v, encoding='utf-8'): if isinstance(encoding, basestring): encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore')) if isinstance(v, binary_type): for e in encoding: try: return v.decode(*e) except: pass return v return unicode(v)
decode_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v, encoding='utf-8'):\n if isinstance(encoding, basestring):\n encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))\n if isinstanc...
def x(v, encoding='utf-8'): if isinstance(encoding, basestring): encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore')) if isinstance(v, unicode): for e in encoding: try: return v.encode(*e) except: pass return v return str(v)
encode_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v, encoding='utf-8'):\n if isinstance(encoding, basestring):\n encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))\n if isinstanc...
def x(cmd, mode=os.F_OK | os.X_OK, path=None): def x(fn, mode): return os.path.exists(fn) and os.access(fn, mode ) and not os.path.isdir(fn) if os.path.dirname(cmd): if _access_check(cmd, mode): return cmd return None if path is None: path = os.environ.get('PATH', os.defpath) if not path: return None path = path.split(os.pathsep) if sys.platform == 'win32': if not os.curdir in path: path.insert(0, os.curdir) pathext = os.environ.get('PATHEXT', '').split(os.pathsep) if any([cmd.lower().endswith(ext.lower()) for ext in pathext]): files = [cmd] else: files = [(cmd + ext) for ext in pathext] else: files = [cmd] seen = set() for dir in path: normdir = os.path.normcase(dir) if normdir not in seen: seen.add(normdir) for thefile in files: name = os.path.join(dir, thefile) if _access_check(name, mode): return name return None
shutil_which
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, mode=os.F_OK | os.X_OK, path=None):\n\n def x(fn, mode):\n return os.path.exists(fn) and os.access(fn, mode\n ) and not os.path.isdir...
def x(self, from_lang=None, to='de'): if from_lang is None: from_lang = self.translator.detect(self.string) return self.translator.translate(self.string, from_lang=from_lang, to_lang=to)
translate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, from_lang=None, to='de'):\n if from_lang is None:\n from_lang = self.translator.detect(self.string)\n return self.translator.translate(self....
def x(self): _lemmatizer = PatternParserLemmatizer(tokenizer=NLTKPunktTokenizer()) _raw = ' '.join(self) + '.' _lemmas = _lemmatizer.lemmatize(_raw) return self.__class__([Word(l, t) for l, t in _lemmas])
lemmatize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n _lemmatizer = PatternParserLemmatizer(tokenizer=NLTKPunktTokenizer())\n _raw = ' '.join(self) + '.'\n _lemmas = _lemmatizer.lemmatize(_raw)\n...
def x(self, tokenizer=None): t = tokenizer if tokenizer is not None else self.tokenizer return WordList(t.tokenize(self.raw))
tokenize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokenizer=None):\n t = tokenizer if tokenizer is not None else self.tokenizer\n return WordList(t.tokenize(self.raw))\n", "role": "user" }, ...
def x(self): return WordList([phrase.strip() for phrase in self.np_extractor.extract (self.raw) if len(phrase.split()) > 1])
noun_phrases
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return WordList([phrase.strip() for phrase in self.np_extractor.extract\n (self.raw) if len(phrase.split()) > 1])\n", "role": "user" }...
def x(self): return [(Word(word, pos_tag=t), unicode(t)) for word, t in self. pos_tagger.tag(self.raw)]
pos_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [(Word(word, pos_tag=t), unicode(t)) for word, t in self.\n pos_tagger.tag(self.raw)]\n", "role": "user" }, { "content": ...
def x(self): counts = defaultdict(int) stripped_words = [lowerstrip(word) for word in self.words] for word in stripped_words: counts[word] += 1 return counts
word_counts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n counts = defaultdict(int)\n stripped_words = [lowerstrip(word) for word in self.words]\n for word in stripped_words:\n counts[word] +=...
def x(self): return {'raw': self.raw, 'start_index': self.start_index, 'end_index': self.end_index, 'stripped': self.stripped, 'noun_phrases': self. noun_phrases, 'polarity': self.polarity, 'subjectivity': self. subjectivity}
dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {'raw': self.raw, 'start_index': self.start_index, 'end_index':\n self.end_index, 'stripped': self.stripped, 'noun_phrases': self.\n ...
def x(self): return WordList(word_tokenize(self.raw, self.tokenizer, include_punc=False) )
words
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return WordList(word_tokenize(self.raw, self.tokenizer, include_punc=False)\n )\n", "role": "user" }, { "content": "words", ...
def x(self): _RETURN_TYPE = namedtuple('Sentiment', ['polarity', 'subjectivity']) _polarity = 0 _subjectivity = 0 for s in self.sentences: _polarity += s.polarity _subjectivity += s.subjectivity try: polarity = _polarity / len(self.sentences) except ZeroDivisionError: polarity = 0.0 try: subjectivity = _subjectivity / len(self.sentences) except ZeroDivisionError: subjectivity = 0.0 return _RETURN_TYPE(polarity, subjectivity)
sentiment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n _RETURN_TYPE = namedtuple('Sentiment', ['polarity', 'subjectivity'])\n _polarity = 0\n _subjectivity = 0\n for s in self.sentences:\n ...
def x(self, *args, **kwargs): return json.dumps(self.serialized, *args, **kwargs)
to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n return json.dumps(self.serialized, *args, **kwargs)\n", "role": "user" }, { "content": "to_json", "role": "assista...
def x(self): sentence_objects = [] sentences = sent_tokenize(self.raw, tokenizer=self.tokenizer) char_index = 0 for sent in sentences: try: start_index = self.raw.index(sent, char_index) char_index += len(sent) end_index = start_index + len(sent) except ValueError: start_index = None end_index = None s = Sentence(sent, start_index=start_index, end_index=end_index, tokenizer=self.tokenizer, np_extractor=self.np_extractor, pos_tagger=self.pos_tagger, analyzer=self.analyzer, parser=self .parser, classifier=self.classifier) sentence_objects.append(s) return sentence_objects
create_sentence_objects
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n sentence_objects = []\n sentences = sent_tokenize(self.raw, tokenizer=self.tokenizer)\n char_index = 0\n for sent in sentences:\n t...
def x(string, n=3, punctuation=PUNCTUATION, continuous=False): def x(s, punctuation=set(punctuation)): return [w for w in s if (isinstance(w, Word) and w.string or w) not in punctuation] if n <= 0: return [] if isinstance(string, basestring): s = [strip_punctuation(s.split(' ')) for s in tokenize(string)] if isinstance(string, Sentence): s = [strip_punctuation(string)] if isinstance(string, Text): s = [strip_punctuation(s) for s in string] if continuous: s = [sum(s, [])] g = [] for s in s: g.extend([tuple(s[i:i + n]) for i in range(len(s) - n + 1)]) return g
ngrams
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, n=3, punctuation=PUNCTUATION, continuous=False):\n\n def x(s, punctuation=set(punctuation)):\n return [w for w in s if (isinstance(w, Word) a...
def x(s, n=3): if n == 0: return s[0:0] return re.sub('((.)\\2{%s,})' % (n - 1), lambda m: m.group(1)[0] * n, s)
deflood
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, n=3):\n if n == 0:\n return s[0:0]\n return re.sub('((.)\\\\2{%s,})' % (n - 1), lambda m: m.group(1)[0] * n, s)\n", "role": "user" }, {...
def x(string, token=[WORD, POS, CHUNK, PNP], column=4): if isinstance(string, basestring): print('\n\n'.join([table(sentence, fill=column) for sentence in Text(string, token)])) if isinstance(string, Text): print('\n\n'.join([table(sentence, fill=column) for sentence in string])) if isinstance(string, Sentence): print(table(string, fill=column))
pprint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, token=[WORD, POS, CHUNK, PNP], column=4):\n if isinstance(string, basestring):\n print('\\n\\n'.join([table(sentence, fill=column) for senten...
def x(path, encoding='utf-8', comment=';;;'): if path: if isinstance(path, basestring) and os.path.exists(path): if PY2: f = codecs.open(path, 'r', encoding='utf-8') else: f = open(path, 'r', encoding='utf-8') elif isinstance(path, basestring): f = path.splitlines() else: f = path for i, line in enumerate(f): line = line.strip(codecs.BOM_UTF8) if i == 0 and isinstance(line, binary_type) else line line = line.strip() line = decode_utf8(line, encoding) if not line or comment and line.startswith(comment): continue yield line return
read
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, encoding='utf-8', comment=';;;'):\n if path:\n if isinstance(path, basestring) and os.path.exists(path):\n if PY2:\n ...
def x(token, tag): if tag.startswith(('NNP-', 'NNPS-')): return token, '%s-%s' % (NOUN, tag.split('-')[-1]) if tag in ('NN', 'NNS', 'NNP', 'NNPS', 'NP'): return token, NOUN if tag in ('MD', 'VB', 'VBD', 'VBG', 'VBN', 'VBP', 'VBZ'): return token, VERB if tag in ('JJ', 'JJR', 'JJS'): return token, ADJ if tag in ('RB', 'RBR', 'RBS', 'WRB'): return token, ADV if tag in ('PRP', 'PRP$', 'WP', 'WP$'): return token, PRON if tag in ('DT', 'PDT', 'WDT', 'EX'): return token, DET if tag in ('IN',): return token, PREP if tag in ('CD',): return token, NUM if tag in ('CC',): return token, CONJ if tag in ('UH',): return token, INTJ if tag in ('POS', 'RP', 'TO'): return token, PRT if tag in ('SYM', 'LS', '.', '!', '?', ',', ':', '(', ')', '"', '#', '$'): return token, PUNC return token, X
penntreebank2universal
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(token, tag):\n if tag.startswith(('NNP-', 'NNPS-')):\n return token, '%s-%s' % (NOUN, tag.split('-')[-1])\n if tag in ('NN', 'NNS', 'NNP', 'NNPS',...
def x(string, punctuation=PUNCTUATION, abbreviations=ABBREVIATIONS, replace =replacements, linebreak='\\n{2,}'): punctuation = tuple(punctuation.replace('.', '')) for a, b in replace.items(): string = re.sub(a, b, string) if isinstance(string, unicode): string = string.replace(u'“', u' “ ') string = string.replace(u'”', u' ” ') string = string.replace(u'‘', u' ‘ ') string = string.replace(u'’', u' ’ ') string = re.sub('\r\n', '\n', string) string = re.sub(linebreak, ' %s ' % EOS, string) string = re.sub('\\s+', ' ', string) tokens = [] for t in TOKEN.findall(string + ' '): if len(t) > 0: tail = [] while t.startswith(punctuation) and not t in replace: if t.startswith(punctuation): tokens.append(t[0]) t = t[1:] while t.endswith(punctuation + ('.',)) and not t in replace: if t.endswith(punctuation): tail.append(t[-1]) t = t[:-1] if t.endswith('...'): tail.append('...') t = t[:-3].rstrip('.') if t.endswith('.'): if t in abbreviations or RE_ABBR1.match(t ) is not None or RE_ABBR2.match(t ) is not None or RE_ABBR3.match(t) is not None: break else: tail.append(t[-1]) t = t[:-1] if t != '': tokens.append(t) tokens.extend(reversed(tail)) sentences, i, j = [[]], 0, 0 while j < len(tokens): if tokens[j] in ('...', '.', '!', '?', EOS): while j < len(tokens) and tokens[j] in ("'", '"', u'”', u'’', '...', '.', '!', '?', ')', EOS): if tokens[j] in ("'", '"') and sentences[-1].count(tokens[j] ) % 2 == 0: break j += 1 sentences[-1].extend(t for t in tokens[i:j] if t != EOS) sentences.append([]) i = j j += 1 sentences[-1].extend(tokens[i:j]) sentences = (' '.join(s) for s in sentences if len(s) > 0) sentences = (RE_SARCASM.sub('(!)', s) for s in sentences) sentences = [RE_EMOTICONS.sub(lambda m: m.group(1).replace(' ', '') + m .group(2), s) for s in sentences] return sentences
find_tokens
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, punctuation=PUNCTUATION, abbreviations=ABBREVIATIONS, replace\n =replacements, linebreak='\\\\n{2,}'):\n punctuation = tuple(punctuation.replace(...
def x(token, tag='NN'): if isinstance(token, (list, tuple)): token, tag = token if token.endswith('ing'): tag = 'VBG' if token.endswith('ly'): tag = 'RB' if token.endswith('s') and not token.endswith(('is', 'ous', 'ss')): tag = 'NNS' if token.endswith(('able', 'al', 'ful', 'ible', 'ient', 'ish', 'ive', 'less', 'tic', 'ous')) or '-' in token: tag = 'JJ' if token.endswith('ed'): tag = 'VBN' if token.endswith(('ate', 'ify', 'ise', 'ize')): tag = 'VBP' return [token, tag]
suffix_rules
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(token, tag='NN'):\n if isinstance(token, (list, tuple)):\n token, tag = token\n if token.endswith('ing'):\n tag = 'VBG'\n if token.endsw...
def x(tokens, lexicon={}, model=None, morphology=None, context=None, entities=None, default=('NN', 'NNP', 'CD'), language='en', map=None, ** kwargs): tagged = [] for i, token in enumerate(tokens): tagged.append([token, lexicon.get(token, i == 0 and lexicon.get( token.lower()) or None)]) for i, (token, tag) in enumerate(tagged): prev, next = (None, None), (None, None) if i > 0: prev = tagged[i - 1] if i < len(tagged) - 1: next = tagged[i + 1] if tag is None or token in (model is not None and model.unknown or ()): if model is not None: tagged[i] = model.apply([token, None], prev, next) elif token.istitle() and language != 'de': tagged[i] = [token, default[1]] elif CD.match(token) is not None: tagged[i] = [token, default[2]] elif morphology is not None: tagged[i] = morphology.apply([token, default[0]], prev, next) elif language == 'en': tagged[i] = _suffix_rules([token, default[0]]) else: tagged[i] = [token, default[0]] if context is not None and model is None: tagged = context.apply(tagged) if entities is not None: tagged = entities.apply(tagged) if map is not None: tagged = [(list(map(token, tag)) or [token, default[0]]) for token, tag in tagged] return tagged
find_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tokens, lexicon={}, model=None, morphology=None, context=None,\n entities=None, default=('NN', 'NNP', 'CD'), language='en', map=None, **\n kwargs):\n ...
def x(tagged, language='en'): chunked = [x for x in tagged] tags = ''.join('%s%s' % (tag, SEPARATOR) for token, tag in tagged) for tag, rule in CHUNKS[int(language in ('ca', 'es', 'pt', 'fr', 'it', 'pt', 'ro'))]: for m in rule.finditer(tags): i = m.start() j = tags[:i].count(SEPARATOR) n = m.group(0).count(SEPARATOR) for k in range(j, j + n): if len(chunked[k]) == 3: continue if len(chunked[k]) < 3: if k == j and chunked[k][1] in ('CC', 'CJ', ','): j += 1 elif k == j: chunked[k].append('B-' + tag) else: chunked[k].append('I-' + tag) for chink in filter(lambda x: len(x) < 3, chunked): chink.append('O') for i, (word, tag, chunk) in enumerate(chunked): if tag.startswith('RB') and chunk == 'B-NP': if i < len(chunked) - 1 and not chunked[i + 1][1].startswith('JJ'): chunked[i + 0][2] = 'B-ADVP' chunked[i + 1][2] = 'B-NP' if i < len(chunked) - 1 and chunked[i + 1][1] in ('CC', 'CJ', ','): chunked[i + 1][2] = 'O' if i < len(chunked) - 2 and chunked[i + 1][2] == 'O': chunked[i + 2][2] = 'B-NP' return chunked
find_chunks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tagged, language='en'):\n chunked = [x for x in tagged]\n tags = ''.join('%s%s' % (tag, SEPARATOR) for token, tag in tagged)\n for tag, rule in CHUNKS...
def x(chunked): for ch in chunked: ch.append('O') for i, chunk in enumerate(chunked): if chunk[2].endswith('PP') and chunk[-1] == 'O': if i < len(chunked) - 1 and (chunked[i + 1][2].endswith(('NP', 'PP')) or chunked[i + 1][1] in ('VBG', 'VBN')): chunk[-1] = 'B-PNP' pp = True for ch in chunked[i + 1:]: if not (ch[2].endswith(('NP', 'PP')) or ch[1] in ('VBG', 'VBN')): break if ch[2].endswith('PP') and pp: ch[-1] = 'I-PNP' if not ch[2].endswith('PP'): ch[-1] = 'I-PNP' pp = False return chunked
find_prepositions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chunked):\n for ch in chunked:\n ch.append('O')\n for i, chunk in enumerate(chunked):\n if chunk[2].endswith('PP') and chunk[-1] == 'O':\n ...
def x(chunked): tag = lambda token: token[2].split('-')[-1] chunks = [] for token in chunked: if len(chunks) == 0 or token[2].startswith('B-') or tag(token) != tag( chunks[-1][-1]): chunks.append([]) chunks[-1].append(token + ['O']) id = 0 for i, chunk in enumerate(chunks): if tag(chunk[-1]) == 'VP' and i > 0 and tag(chunks[i - 1][-1]) == 'NP': if chunk[-1][-1] == 'O': id += 1 for token in chunk: token[-1] = 'VP-' + str(id) for token in chunks[i - 1]: token[-1] += '*NP-SBJ-' + str(id) token[-1] = token[-1].lstrip('O-*') if tag(chunk[-1]) == 'VP' and i < len(chunks) - 1 and tag(chunks[i + 1][-1]) == 'NP': if chunk[-1][-1] == 'O': id += 1 for token in chunk: token[-1] = 'VP-' + str(id) for token in chunks[i + 1]: token[-1] = '*NP-OBJ-' + str(id) token[-1] = token[-1].lstrip('O-*') for i, chunk in enumerate(chunks): if 0 < i < len(chunks) - 1 and len(chunk) == 1 and chunk[-1][-1 ] == 'O': t0, t1, t2 = chunks[i - 1][-1], chunks[i][0], chunks[i + 1][0] if tag(t1) == 'PP' and t2[1] in ('DT', 'PR', 'PRP$'): if t0[0] in BE and t1[0] in ('in', 'at'): t1[-1] = 'PP-LOC' if t0[0] in GO and t1[0] in ('to', 'towards'): t1[-1] = 'PP-DIR' related = [] [related.extend(chunk) for chunk in chunks] return related
find_relations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chunked):\n tag = lambda token: token[2].split('-')[-1]\n chunks = []\n for token in chunked:\n if len(chunks) == 0 or token[2].startswith('B-'...
def x(string, parser, top=10, frequency={}, **kwargs): lemmata = kwargs.pop('lemmata', kwargs.pop('stem', True)) chunks = [] wordcount = 0 for sentence in parser.parse(string, chunks=True, lemmata=lemmata).split(): for w in sentence: if w[2] == 'B-NP': chunks.append([w]) wordcount += 1 elif w[2] == 'I-NP' and w[1][:3] == chunks[-1][-1][1][:3] == 'NNP': chunks[-1][-1][+0] += ' ' + w[+0] chunks[-1][-1][-1] += ' ' + w[-1] elif w[2] == 'I-NP': chunks[-1].append(w) wordcount += 1 m = {} for i, chunk in enumerate(chunks): head = True if parser.language not in ('ca', 'es', 'pt', 'fr', 'it', 'pt', 'ro'): chunk = list(reversed(chunk)) for w in chunk: if w[1].startswith('NN'): if lemmata: k = w[-1] else: k = w[0].lower() if not k in m: m[k] = [0.0, set(), 1.0, 1.0, 1.0] m[k][0] += 1 / float(wordcount) m[k][1].add(' '.join(map(lambda x: x[0], chunk)).lower()) m[k][2] += 1 if float(i) / len(chunks) <= 0.25 else 0 m[k][3] += 1 if w[3] == 'O' else 0 m[k][4] += 1 if head else 0 head = False for k in m: if frequency: df = frequency.get(k, 0.0) df = max(df, 1e-10) df = log(1.0 / df, 2.71828) else: df = 1.0 m[k][0] = max(1e-10, m[k][0] * df) m[k][1] = 1 + float(len(m[k][1])) hmean = lambda a: len(a) / sum(1.0 / x for x in a) m = [(hmean(m[k]), k) for k in m] m = sorted(m, key=lambda x: x[1]) m = sorted(m, key=lambda x: x[0], reverse=True) m = [k for score, k in m] return m[:top]
find_keywords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, parser, top=10, frequency={}, **kwargs):\n lemmata = kwargs.pop('lemmata', kwargs.pop('stem', True))\n chunks = []\n wordcount = 0\n for se...
def x(*args, **kwargs): if len(args) == 1 and isinstance(args[0], (list, tuple)): if args[0] not in ((PRESENT, PARTICIPLE), (PAST, PARTICIPLE)): args = args[0] if len(args) == 0 and len(kwargs) == 0: t = INFINITIVE else: t = PRESENT tense = kwargs.get('tense', args[0] if len(args) > 0 else t) person = kwargs.get('person', args[1] if len(args) > 1 else 3) or None number = kwargs.get('number', args[2] if len(args) > 2 else SINGULAR) mood = kwargs.get('mood', args[3] if len(args) > 3 else INDICATIVE) aspect = kwargs.get('aspect', args[4] if len(args) > 4 else IMPERFECTIVE) negated = kwargs.get('negated', args[5] if len(args) > 5 else False) if mood in (PERFECTIVE, IMPERFECTIVE): mood, aspect = INDICATIVE, mood if tense == INFINITIVE: person = number = mood = aspect = None negated = False if tense in ((PRESENT, PARTICIPLE), PRESENT + PARTICIPLE, PARTICIPLE, GERUND): tense, aspect = PRESENT, PROGRESSIVE if tense in ((PAST, PARTICIPLE), PAST + PARTICIPLE): tense, aspect = PAST, PROGRESSIVE if tense == IMPERFECT: tense, aspect = PAST, IMPERFECTIVE if tense == PRETERITE: tense, aspect = PAST, PERFECTIVE if aspect in (CONTINUOUS, PARTICIPLE, GERUND): aspect = PROGRESSIVE if aspect == PROGRESSIVE: person = number = None if tense == CONDITIONAL and mood == INDICATIVE: tense, mood = PRESENT, CONDITIONAL return TENSES_ID.get(tense.lower(), TENSES_ID.get((tense, person, number, mood, aspect, negated)))
tense_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n if len(args) == 1 and isinstance(args[0], (list, tuple)):\n if args[0] not in ((PRESENT, PARTICIPLE), (PAST, PARTICIPLE)):\n ...
def x(function, *args, **kwargs): return getattr(_module(kwargs.pop('language', 'en')), function)(*args, **kwargs)
multilingual
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(function, *args, **kwargs):\n return getattr(_module(kwargs.pop('language', 'en')), function)(*args,\n **kwargs)\n", "role": "user" }, { ...
def x(s): s = decode_utf8(s) s = set(w.strip(PUNCTUATION) for w in s.replace("'", "' ").split()) n = float(len(s) or 1) p = {} for xx in LANGUAGES: lexicon = _module(xx).__dict__['lexicon'] p[xx] = sum(1 for w in s if w in lexicon) / n return max(p.items(), key=lambda kv: (kv[1], int(kv[0] == 'en')))
language
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n s = decode_utf8(s)\n s = set(w.strip(PUNCTUATION) for w in s.replace(\"'\", \"' \").split())\n n = float(len(s) or 1)\n p = {}\n for xx in...
def x(self, method, *args): if list.__len__(self) == 0: self.load() setattr(self, method, types.MethodType(getattr(list, method), self)) return getattr(list, method)(self, *args)
lazy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, method, *args):\n if list.__len__(self) == 0:\n self.load()\n setattr(self, method, types.MethodType(getattr(list, method), self))\n ...
def x(self, token, tag, previous=None, next=None): self._classifier.train(self._v(token, previous, next), type=tag)
train
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, tag, previous=None, next=None):\n self._classifier.train(self._v(token, previous, next), type=tag)\n", "role": "user" }, { "content...
def x(self, token, previous=None, next=None, **kwargs): return self._classifier.classify(self._v(token, previous, next), **kwargs)
classify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, previous=None, next=None, **kwargs):\n return self._classifier.classify(self._v(token, previous, next), **kwargs)\n", "role": "user" }, ...
def x(self, token, previous=(None, None), next=(None, None)): return [token[0], self._classifier.classify(self._v(token[0], previous, next))]
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, previous=(None, None), next=(None, None)):\n return [token[0], self._classifier.classify(self._v(token[0], previous,\n next))]\n", "...
def x(self, token, previous=None, next=None): def x(v, s1, s2): if s2: v[s1 + ' ' + s2] = 1 p, n = previous, next p = ('', '') if not p else (p[0] or '', p[1] or '') n = ('', '') if not n else (n[0] or '', n[1] or '') v = {} f(v, 'b', 'b') f(v, 'h', token[0]) f(v, 'w', token[-6:] if token not in self.known or token in self. unknown else '') f(v, 'x', token[-3:]) f(v, '-x', p[0][-3:]) f(v, '+x', n[0][-3:]) f(v, '-t', p[1]) f(v, '-+', p[1] + n[1]) f(v, '+t', n[1]) return v
v
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, previous=None, next=None):\n\n def x(v, s1, s2):\n if s2:\n v[s1 + ' ' + s2] = 1\n p, n = previous, next\n p = ('', '')...
def x(self, token, previous=(None, None), next=(None, None)): w = token[0] for r in self: if r[1] in self._cmd: f, x, pos, cmd = bool(0), r[0], r[-2], r[1].lower() if r[2] in self._cmd: f, x, pos, cmd = bool(1), r[1], r[-2], r[2].lower().lstrip('f') if f and token[1] != r[0]: continue if (cmd == 'word' and x == w or cmd == 'char' and x in w or cmd == 'haspref' and w.startswith(x) or cmd == 'hassuf' and w.endswith (x) or cmd == 'addpref' and x + w in self.known or cmd == 'addsuf' and w + x in self.known or cmd == 'deletepref' and w. startswith(x) and w[len(x):] in self.known or cmd == 'deletesuf' and w.endswith(x) and w[:-len(x)] in self.known or cmd == 'goodleft' and x == next[0] or cmd == 'goodright' and x == previous[0]): token[1] = pos return token
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, previous=(None, None), next=(None, None)):\n w = token[0]\n for r in self:\n if r[1] in self._cmd:\n f, x, pos, cmd = bool...
def x(self, i, tag, affix, cmd='hassuf', tagged=None): if affix.startswith('-') and affix.endswith('-'): affix, cmd = affix[+1:-1], 'char' if affix.startswith('-'): affix, cmd = affix[+1:-0], 'hassuf' if affix.endswith('-'): affix, cmd = affix[+0:-1], 'haspref' if tagged: r = [tagged, affix, 'f' + cmd.lstrip('f'), tag, 'x'] else: r = [affix, cmd.lstrip('f'), tag, 'x'] lazylist.insert(self, i, r)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, i, tag, affix, cmd='hassuf', tagged=None):\n if affix.startswith('-') and affix.endswith('-'):\n affix, cmd = affix[+1:-1], 'char'\n if affi...
def x(self, tokens): o = [('STAART', 'STAART')] * 3 t = o + tokens + o for i, token in enumerate(t): for r in self: if token[1] == 'STAART': continue if token[1] != r[0] and r[0] != '*': continue cmd, x, y = r[2], r[3], r[4] if len(r) > 4 else '' cmd = cmd.lower() if cmd == 'prevtag' and x == t[i - 1][1 ] or cmd == 'nexttag' and x == t[i + 1][1 ] or cmd == 'prev2tag' and x == t[i - 2][1 ] or cmd == 'next2tag' and x == t[i + 2][1 ] or cmd == 'prev1or2tag' and x in (t[i - 1][1], t[i - 2][1] ) or cmd == 'next1or2tag' and x in (t[i + 1][1], t[i + 2][1] ) or cmd == 'prev1or2or3tag' and x in (t[i - 1][1], t[i - 2 ][1], t[i - 3][1]) or cmd == 'next1or2or3tag' and x in (t[i + 1][1], t[i + 2][1], t[i + 3][1] ) or cmd == 'surroundtag' and x == t[i - 1][1] and y == t[i + 1 ][1] or cmd == 'curwd' and x == t[i + 0][0 ] or cmd == 'prevwd' and x == t[i - 1][0 ] or cmd == 'nextwd' and x == t[i + 1][0 ] or cmd == 'prev1or2wd' and x in (t[i - 1][0], t[i - 2][0] ) or cmd == 'next1or2wd' and x in (t[i + 1][0], t[i + 2][0] ) or cmd == 'prevwdtag' and x == t[i - 1][0] and y == t[i - 1][ 1] or cmd == 'nextwdtag' and x == t[i + 1][0] and y == t[i + 1 ][1] or cmd == 'wdprevtag' and x == t[i - 1][1] and y == t[ i + 0][0] or cmd == 'wdnexttag' and x == t[i + 0][0 ] and y == t[i + 1][1] or cmd == 'wdand2aft' and x == t[i + 0][ 0] and y == t[i + 2][0] or cmd == 'wdand2tagbfr' and x == t[ i - 2][1] and y == t[i + 0][0 ] or cmd == 'wdand2tagaft' and x == t[i + 0][0] and y == t[ i + 2][1] or cmd == 'lbigram' and x == t[i - 1][0] and y == t[ i + 0][0] or cmd == 'rbigram' and x == t[i + 0][0] and y == t[ i + 1][0] or cmd == 'prevbigram' and x == t[i - 2][1 ] and y == t[i - 1][1] or cmd == 'nextbigram' and x == t[i + 1 ][1] and y == t[i + 2][1]: t[i] = [t[i][0], r[1]] return t[len(o):-len(o)]
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokens):\n o = [('STAART', 'STAART')] * 3\n t = o + tokens + o\n for i, token in enumerate(t):\n for r in self:\n if token[1] ...
def x(self, i, tag1, tag2, cmd='prevtag', x=None, y=None): if ' < ' in tag1 and not x and not y: tag1, x = tag1.split(' < ') cmd = 'prevtag' if ' > ' in tag1 and not x and not y: x, tag1 = tag1.split(' > ') cmd = 'nexttag' lazylist.insert(self, i, [tag1, tag2, cmd, x or '', y or ''])
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, i, tag1, tag2, cmd='prevtag', x=None, y=None):\n if ' < ' in tag1 and not x and not y:\n tag1, x = tag1.split(' < ')\n cmd = 'prevtag'\n...
def x(self, tokens): i = 0 while i < len(tokens): w = tokens[i][0].lower() if RE_ENTITY1.match(w) or RE_ENTITY2.match(w) or RE_ENTITY3.match(w): tokens[i][1] = self.tag if w in self: for e in self[w]: e, tag = (e[:-1], '-' + e[-1].upper()) if e[-1 ] in self._cmd else (e, '') b = True for j, e in enumerate(e): if i + j >= len(tokens) or tokens[i + j][0].lower() != e: b = False break if b: for token in tokens[i:i + j + 1]: token[1] = token[1] if token[1].startswith(self.tag ) else self.tag token[1] += tag i += j break i += 1 return tokens
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokens):\n i = 0\n while i < len(tokens):\n w = tokens[i][0].lower()\n if RE_ENTITY1.match(w) or RE_ENTITY2.match(w) or RE_ENTITY3.ma...
def x(self, entity, name='pers'): e = map(lambda s: s.lower(), entity.split(' ') + [name]) self.setdefault(e[0], []).append(e)
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity, name='pers'):\n e = map(lambda s: s.lower(), entity.split(' ') + [name])\n self.setdefault(e[0], []).append(e)\n", "role": "user" }, ...
def x(self, string, **kwargs): return find_keywords(string, parser=self, top=kwargs.pop('top', 10), frequency=kwargs.pop('frequency', {}), **kwargs)
find_keywords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, string, **kwargs):\n return find_keywords(string, parser=self, top=kwargs.pop('top', 10),\n frequency=kwargs.pop('frequency', {}), **kwargs)\n"...
def x(self, string, **kwargs): return find_tokens(string, punctuation=kwargs.get('punctuation', PUNCTUATION), abbreviations=kwargs.get('abbreviations', ABBREVIATIONS), replace=kwargs.get('replace', replacements), linebreak='\\n{2,}')
find_tokens
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, string, **kwargs):\n return find_tokens(string, punctuation=kwargs.get('punctuation',\n PUNCTUATION), abbreviations=kwargs.get('abbreviations',...
def x(self, tokens, **kwargs): return find_tags(tokens, lexicon=kwargs.get('lexicon', self.lexicon or {}), model=kwargs.get('model', self.model), morphology=kwargs.get( 'morphology', self.morphology), context=kwargs.get('context', self. context), entities=kwargs.get('entities', self.entities), language= kwargs.get('language', self.language), default=kwargs.get('default', self.default), map=kwargs.get('map', None))
find_tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokens, **kwargs):\n return find_tags(tokens, lexicon=kwargs.get('lexicon', self.lexicon or\n {}), model=kwargs.get('model', self.model), morph...
def x(self, tokens, **kwargs): return find_prepositions(find_chunks(tokens, language=kwargs.get( 'language', self.language)))
find_chunks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokens, **kwargs):\n return find_prepositions(find_chunks(tokens, language=kwargs.get(\n 'language', self.language)))\n", "role": "user" ...
def x(self, s, tokenize=True, tags=True, chunks=True, relations=False, lemmata=False, encoding='utf-8', **kwargs): if tokenize is True: s = self.find_tokens(s, **kwargs) if isinstance(s, (list, tuple)): s = [(isinstance(s, basestring) and s.split(' ') or s) for s in s] if isinstance(s, basestring): s = [s.split(' ') for s in s.split('\n')] for i in range(len(s)): for j in range(len(s[i])): if isinstance(s[i][j], str): s[i][j] = decode_string(s[i][j], encoding) if tags or chunks or relations or lemmata: s[i] = self.find_tags(s[i], **kwargs) else: s[i] = [[w] for w in s[i]] if chunks or relations: s[i] = self.find_chunks(s[i], **kwargs) if relations: s[i] = self.find_labels(s[i], **kwargs) if lemmata: s[i] = self.find_lemmata(s[i], **kwargs) if not kwargs.get('collapse', True) or kwargs.get('split', False): return s format = ['word'] if tags: format.append('part-of-speech') if chunks: format.extend(('chunk', 'preposition')) if relations: format.append('relation') if lemmata: format.append('lemma') for i in range(len(s)): for j in range(len(s[i])): s[i][j][0] = s[i][j][0].replace('/', '&slash;') s[i][j] = '/'.join(s[i][j]) s[i] = ' '.join(s[i]) s = '\n'.join(s) s = TaggedString(s, format, language=kwargs.get('language', self.language)) return s
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, s, tokenize=True, tags=True, chunks=True, relations=False,\n lemmata=False, encoding='utf-8', **kwargs):\n if tokenize is True:\n s = self.f...
def x(self, sep=TOKENS): if sep != TOKENS: return unicode.split(self, sep) if len(self) == 0: return [] return [[[x.replace('&slash;', '/') for x in token.split('/')] for token in sentence.split(' ')] for sentence in unicode.split(self, '\n') ]
split
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sep=TOKENS):\n if sep != TOKENS:\n return unicode.split(self, sep)\n if len(self) == 0:\n return []\n return [[[x.replace('&slash;...
def x(self): a = set(TENSES[id] for id in self._format) a = a.union(set(TENSES[id] for id in self._default.keys())) a = a.union(set(TENSES[id] for id in self._default.values())) a = sorted(x[:-2] for x in a if x[-2] is False) return a
TENSES
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n a = set(TENSES[id] for id in self._format)\n a = a.union(set(TENSES[id] for id in self._default.keys()))\n a = a.union(set(TENSES[id] for id ...
def x(self, verb, parse=True): if dict.__len__(self) == 0: self.load() if verb.lower() in self._inverse: return self._inverse[verb.lower()] if verb in self._inverse: return self._inverse[verb] if parse is True: return self.find_lemma(verb)
lemma
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verb, parse=True):\n if dict.__len__(self) == 0:\n self.load()\n if verb.lower() in self._inverse:\n return self._inverse[verb.lower(...
def x(self, verb, parse=True): a = [] b = self.lemma(verb, parse=parse) if b in self: a = [x for x in self[b] if x != ''] elif parse is True: a = self.find_lexeme(b) u = [] [u.append(x) for x in a if x not in u] return u
lexeme
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verb, parse=True):\n a = []\n b = self.lemma(verb, parse=parse)\n if b in self:\n a = [x for x in self[b] if x != '']\n elif parse is ...
def x(self, verb, *args, **kwargs): id = tense_id(*args, **kwargs) i1 = self._format.get(id) i2 = self._format.get(self._default.get(id)) i3 = self._format.get(self._default.get(self._default.get(id))) b = self.lemma(verb, parse=kwargs.get('parse', True)) v = [] if b in self: v = self[b] for i in (i1, i2, i3): if i is not None and 0 <= i < len(v) and v[i]: return v[i] if kwargs.get('parse', True) is True: v = self.find_lexeme(b) for i in (i1, i2, i3): if i is not None and 0 <= i < len(v) and v[i]: return v[i]
conjugate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verb, *args, **kwargs):\n id = tense_id(*args, **kwargs)\n i1 = self._format.get(id)\n i2 = self._format.get(self._default.get(id))\n i3 = se...
def x(self, verb, parse=True): verb = verb.lower() a = set() b = self.lemma(verb, parse=parse) v = [] if b in self: v = self[b] elif parse is True: v = self.find_lexeme(b) for i, tense in enumerate(v): if tense == verb: for id, index in self._format.items(): if i == index: a.add(id) for id1, id2 in self._default.items(): if id2 in a: a.add(id1) for id1, id2 in self._default.items(): if id2 in a: a.add(id1) a = (TENSES[id][:-2] for id in a) a = Tenses(sorted(a)) return a
tenses
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verb, parse=True):\n verb = verb.lower()\n a = set()\n b = self.lemma(verb, parse=parse)\n v = []\n if b in self:\n v = self[b]\n ...
def x(self, path=None): if not path: path = self._path if not os.path.exists(path): return words, synsets, labels = {}, {}, {} xml = cElementTree.parse(path) xml = xml.getroot() for w in xml.findall('word'): if self._confidence is None or self._confidence <= float(w.attrib. get('confidence', 0.0)): w, pos, p, s, i, label, synset = w.attrib.get('form' ), w.attrib.get('pos'), w.attrib.get('polarity', 0.0 ), w.attrib.get('subjectivity', 0.0), w.attrib.get('intensity', 1.0), w.attrib.get('label'), w.attrib.get(self._synset) psi = float(p), float(s), float(i) if w: words.setdefault(w, {}).setdefault(pos, []).append(psi) if w and label: labels[w] = label if synset: synsets.setdefault(synset, []).append(psi) self._language = xml.attrib.get('language', self._language) for w in words: words[w] = dict((pos, [avg(each) for each in zip(*psi)]) for pos, psi in words[w].items()) for w, pos in list(words.items()): words[w][None] = [avg(each) for each in zip(*pos.values())] for id, psi in synsets.items(): synsets[id] = [avg(each) for each in zip(*psi)] dict.update(self, words) dict.update(self.labeler, labels) dict.update(self._synsets, synsets)
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path=None):\n if not path:\n path = self._path\n if not os.path.exists(path):\n return\n words, synsets, labels = {}, {}, {}\n ...
def x(self, id, pos=ADJECTIVE): id = str(id).zfill(8) if not id.startswith(('n-', 'v-', 'a-', 'r-')): if pos == NOUN: id = 'n-' + id if pos == VERB: id = 'v-' + id if pos == ADJECTIVE: id = 'a-' + id if pos == ADVERB: id = 'r-' + id if dict.__len__(self) == 0: self.load() try: return tuple(self._synsets[id])[:2] except KeyError: return tuple(self._synsets.get(re.sub('-0+', '-', id), (0.0, 0.0))[:2])
synset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, pos=ADJECTIVE):\n id = str(id).zfill(8)\n if not id.startswith(('n-', 'v-', 'a-', 'r-')):\n if pos == NOUN:\n id = 'n-' + id\...
def x(self, words=[], negation=True): a = [] m = None n = None for w, pos in words: if w is None: continue if w in self and pos in self[w]: p, s, i = self[w][pos] if m is None: a.append(dict(w=[w], p=p, s=s, i=i, n=1, x=self.labeler.get(w)) ) if m is not None: a[-1]['w'].append(w) a[-1]['p'] = max(-1.0, min(p * a[-1]['i'], +1.0)) a[-1]['s'] = max(-1.0, min(s * a[-1]['i'], +1.0)) a[-1]['i'] = i a[-1]['x'] = self.labeler.get(w) if n is not None: a[-1]['w'].insert(0, n) a[-1]['i'] = 1.0 / a[-1]['i'] a[-1]['n'] = -1 m = None n = None if pos and pos in self.modifiers or any(map(self[w]. __contains__, self.modifiers)): m = w, pos if negation and w in self.negations: n = w else: if negation and w in self.negations: n = w elif n and len(w.strip("'")) > 1: n = None if n is not None and m is not None and (pos in self.modifiers or self.modifier(m[0])): a[-1]['w'].append(n) a[-1]['n'] = -1 n = None elif m and len(w) > 2: m = None if w == '!' and len(a) > 0: a[-1]['w'].append('!') a[-1]['p'] = max(-1.0, min(a[-1]['p'] * 1.25, +1.0)) if w == '(!)': a.append(dict(w=[w], p=0.0, s=1.0, i=1.0, n=1, x=IRONY)) if w.isalpha() is False and len(w) <= 5 and w not in PUNCTUATION: for (type, p), e in EMOTICONS.items(): if w in map(lambda e: e.lower(), e): a.append(dict(w=[w], p=p, s=1.0, i=1.0, n=1, x=MOOD)) break for i in range(len(a)): w = a[i]['w'] p = a[i]['p'] s = a[i]['s'] n = a[i]['n'] x = a[i]['x'] a[i] = w, p * -0.5 if n < 0 else p, s, x return a
assessments
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, words=[], negation=True):\n a = []\n m = None\n n = None\n for w, pos in words:\n if w is None:\n continue\n if w in...
def x(self, word, pos=None, polarity=0.0, subjectivity=0.0, intensity=1.0, label=None): w = self.setdefault(word, {}) w[pos] = w[None] = polarity, subjectivity, intensity if label: self.labeler[word] = label
annotate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, word, pos=None, polarity=0.0, subjectivity=0.0, intensity=1.0,\n label=None):\n w = self.setdefault(word, {})\n w[pos] = w[None] = polarity, sub...
def x(self, s, path='spelling.txt'): model = {} for w in re.findall('[a-z]+', s.lower()): model[w] = w in model and model[w] + 1 or 1 model = ('%s %s' % (k, v) for k, v in sorted(model.items())) model = '\n'.join(model) f = open(path, 'w') f.write(model) f.close()
train
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, s, path='spelling.txt'):\n model = {}\n for w in re.findall('[a-z]+', s.lower()):\n model[w] = w in model and model[w] + 1 or 1\n model =...
def x(self, w): split = [(w[:i], w[i:]) for i in range(len(w) + 1)] delete, transpose, replace, insert = [(a + b[1:]) for a, b in split if b ], [(a + b[1] + b[0] + b[2:]) for a, b in split if len(b) > 1], [(a + c + b[1:]) for a, b in split for c in Spelling.ALPHA if b], [(a + c + b[0:]) for a, b in split for c in Spelling.ALPHA] return set(delete + transpose + replace + insert)
edit1
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, w):\n split = [(w[:i], w[i:]) for i in range(len(w) + 1)]\n delete, transpose, replace, insert = [(a + b[1:]) for a, b in split if b\n ], [(...
def x(self, w): return set(e2 for e1 in self._edit1(w) for e2 in self._edit1(e1) if e2 in self)
edit2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, w):\n return set(e2 for e1 in self._edit1(w) for e2 in self._edit1(e1) if e2 in\n self)\n", "role": "user" }, { "content": "edit2",...
def x(self, w): if len(self) == 0: self.load() if len(w) == 1: return [(w, 1.0)] if w in PUNCTUATION: return [(w, 1.0)] if w.replace('.', '').isdigit(): return [(w, 1.0)] candidates = self._known([w]) or self._known(self._edit1(w) ) or self._known(self._edit2(w)) or [w] candidates = [(self.get(c, 0.0), c) for c in candidates] s = float(sum(p for p, w in candidates) or 1) candidates = sorted(((p / s, w) for p, w in candidates), reverse=True) candidates = [(w.istitle() and x.title() or x, p) for p, x in candidates] return candidates
suggest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, w):\n if len(self) == 0:\n self.load()\n if len(w) == 1:\n return [(w, 1.0)]\n if w in PUNCTUATION:\n return [(w, 1.0)]\n ...
def x(*args, **kwargs): args = [list(iterable) for iterable in args] n = max(map(len, args)) v = kwargs.get('default', None) return _zip(*[(i + [v] * (n - len(i))) for i in args])
zip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n args = [list(iterable) for iterable in args]\n n = max(map(len, args))\n v = kwargs.get('default', None)\n return _zip(*[(i + [...
def x(sentence): chunks = [] for word in sentence: if word.chunk is not None: if len(chunks) == 0 or chunks[-1] != word.chunk: chunks.append(word.chunk) else: ch = Chink(sentence) ch.append(word.copy(ch)) chunks.append(ch) return chunks
chunked
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sentence):\n chunks = []\n for word in sentence:\n if word.chunk is not None:\n if len(chunks) == 0 or chunks[-1] != word.chunk:\n ...
def x(string, token=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]): return Text(string, token)
tree
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string, token=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n return Text(string, token)\n", "role": "user" }, { "content": "tree", "role": ...
def x(string): string = string.replace('&', '&amp;') string = string.replace('<', '&lt;') string = string.replace('>', '&gt;') string = string.replace('"', '&quot;') string = string.replace(SLASH, '/') return string
xml_encode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n string = string.replace('&', '&amp;')\n string = string.replace('<', '&lt;')\n string = string.replace('>', '&gt;')\n string = string.re...
def x(string): string = string.replace('&amp;', '&') string = string.replace('&lt;', '<') string = string.replace('&gt;', '>') string = string.replace('&quot;', '"') string = string.replace('/', SLASH) return string
xml_decode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n string = string.replace('&amp;', '&')\n string = string.replace('&lt;', '<')\n string = string.replace('&gt;', '>')\n string = string.re...
def x(sentence, tab='\t', id=''): uid = lambda *parts: ''.join([str(id), _UID_SEPARATOR] + [str(x) for x in parts]).lstrip(_UID_SEPARATOR) push = lambda indent: indent + tab pop = lambda indent: indent[:-len(tab)] indent = tab xml = [] xml.append('<%s%s %s="%s" %s="%s">' % (XML_SENTENCE, XML_ID and ' %s="%s"' % (XML_ID, str(id)) or '', XML_TOKEN, ', '.join(sentence .token), XML_LANGUAGE, sentence.language)) anchors = {} for chunk in sentence.chunks: if chunk.attachments: anchors[chunk.start] = len(anchors) + 1 for word in sentence.words: chunk = word.chunk pnp = word.chunk and word.chunk.pnp or None if pnp and pnp.start == chunk.start: a = pnp.anchor and ' %s="%s"' % (XML_OF, uid('A', anchors.get( pnp.anchor.start, ''))) or '' xml.append(indent + '<%s %s="PNP"%s>' % (XML_CHUNK, XML_TYPE, a)) indent = push(indent) if chunk and chunk.start == word.index: if chunk.relations: r1 = unzip(0, chunk.relations) r2 = unzip(1, chunk.relations) r1 = [(x is None and '-' or uid(x)) for x in r1] r2 = [(x is None and '-' or x) for x in r2] r1 = not len(unique(r1)) == 1 and '|'.join(r1) or (r1 + [None] )[0] r2 = not len(unique(r2)) == 1 and '|'.join(r2) or (r2 + [None] )[0] xml.append(indent + '<%s%s%s%s%s%s>' % (XML_CHUNK, chunk.type and ' %s="%s"' % (XML_TYPE, chunk.type) or '', chunk.relations and chunk.role != None and ' %s="%s"' % (XML_RELATION, r2) or '', chunk.relation and chunk.type == 'VP' and ' %s="%s"' % (XML_ID, uid(chunk.relation)) or '', chunk.relation and chunk.type != 'VP' and ' %s="%s"' % (XML_OF, r1) or '', chunk.attachments and ' %s="%s"' % (XML_ANCHOR, uid('A', anchors[chunk.start])) or '')) indent = push(indent) if not chunk: xml.append(indent + '<%s>' % XML_CHINK) indent = push(indent) xml.append(indent + '<%s%s%s%s>%s</%s>' % (XML_WORD, word.type and ' %s="%s"' % (XML_TYPE, xml_encode(word.type)) or '', word. lemma and ' %s="%s"' % (XML_LEMMA, xml_encode(word.lemma)) or '', (' ' + ' '.join([('%s="%s"' % (k, v)) for k, v in word. custom_tags.items() if v != None])).rstrip(), xml_encode( unicode(word)), XML_WORD)) if not chunk: indent = pop(indent) xml.append(indent + '</%s>' % XML_CHINK) if chunk and chunk.stop - 1 == word.index: indent = pop(indent) xml.append(indent + '</%s>' % XML_CHUNK) if pnp and pnp.stop - 1 == word.index: indent = pop(indent) xml.append(indent + '</%s>' % XML_CHUNK) xml.append('</%s>' % XML_SENTENCE) return '\n'.join(xml).encode('utf-8')
parse_xml
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sentence, tab='\\t', id=''):\n uid = lambda *parts: ''.join([str(id), _UID_SEPARATOR] + [str(x) for x in\n parts]).lstrip(_UID_SEPARATOR)\n push =...
def x(xml): string = '' dom = XML(xml) for sentence in dom(XML_SENTENCE): _anchors.clear() _attachments.clear() language = sentence.get(XML_LANGUAGE, 'en') format = sentence.get(XML_TOKEN, [WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]) format = not isinstance(format, basestring ) and format or format.replace(' ', '').split(',') tokens = [] for chunk in sentence: tokens.extend(_parse_tokens(chunk, format)) if ANCHOR in format: A, P, a, i = _anchors, _attachments, 1, format.index(ANCHOR) for id in sorted(A.keys()): for token in A[id]: token[i] += '-' + '-'.join([('A' + str(a + p)) for p in range(len(P[id]))]) token[i] = token[i].strip('O-') for p, pnp in enumerate(P[id]): for token in pnp: token[i] += '-' + 'P' + str(a + p) token[i] = token[i].strip('O-') a += len(P[id]) tokens = ['/'.join([tag for tag in token]) for token in tokens] tokens = ' '.join(tokens) string += tokens + '\n' try: if MBSP: from mbsp import TokenString return TokenString(string.strip(), tags=format, language=language) except: return TaggedString(string.strip(), tags=format, language=language)
parse_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(xml):\n string = ''\n dom = XML(xml)\n for sentence in dom(XML_SENTENCE):\n _anchors.clear()\n _attachments.clear()\n language = ...
def x(chunk, format=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]): tokens = [] if not (chunk.tag == XML_CHUNK or chunk.tag == XML_CHINK): return [] type = chunk.get(XML_TYPE, 'O') if type == 'PNP': for ch in chunk: tokens.extend(_parse_tokens(ch, format)) if PNP in format: i = format.index(PNP) for j, token in enumerate(tokens): token[i] = (j == 0 and 'B-' or 'I-') + 'PNP' a = chunk.get(XML_OF).split(_UID_SEPARATOR)[-1] if a: _attachments.setdefault(a, []) _attachments[a].append(tokens) return tokens relation = _parse_relation(chunk, type) for word in filter(lambda n: n.tag == XML_WORD, chunk): tokens.append(_parse_token(word, chunk=type, relation=relation, format=format)) if CHUNK in format: i = format.index(CHUNK) for j, token in enumerate(tokens): token[i] = token[i] != 'O' and (j == 0 and 'B-' or 'I-') + token[i ] or 'O' a = chunk.get(XML_ANCHOR, '').split(_UID_SEPARATOR)[-1] if a: _anchors[a] = tokens return tokens
parse_tokens
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chunk, format=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n tokens = []\n if not (chunk.tag == XML_CHUNK or chunk.tag == XML_CHINK):\n return []...
def x(chunk, type='O'): r1 = chunk.get(XML_RELATION) r2 = chunk.get(XML_ID, chunk.get(XML_OF)) r1 = [(x != '-' and x or None) for x in r1.split('|')] or [None] r2 = [(x != '-' and x or None) for x in r2.split('|')] or [None] r2 = [(x is not None and x.split(_UID_SEPARATOR)[-1] or x) for x in r2] if len(r1) < len(r2): r1 = r1 + r1 * (len(r2) - len(r1)) if len(r2) < len(r1): r2 = r2 + r2 * (len(r1) - len(r2)) return ';'.join(['-'.join([x for x in (type, r1, r2) if x]) for r1, r2 in zip(r1, r2)])
parse_relation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chunk, type='O'):\n r1 = chunk.get(XML_RELATION)\n r2 = chunk.get(XML_ID, chunk.get(XML_OF))\n r1 = [(x != '-' and x or None) for x in r1.split('|')] ...
def x(word, chunk='O', pnp='O', relation='O', anchor='O', format=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]): tags = [] for tag in format: if tag == WORD: tags.append(xml_decode(word.value)) elif tag == POS: tags.append(xml_decode(word.get(XML_TYPE, 'O'))) elif tag == CHUNK: tags.append(chunk) elif tag == PNP: tags.append(pnp) elif tag == REL: tags.append(relation) elif tag == ANCHOR: tags.append(anchor) elif tag == LEMMA: tags.append(xml_decode(word.get(XML_LEMMA, ''))) else: tags.append(xml_decode(word.get(tag, 'O'))) return tags
parse_token
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(word, chunk='O', pnp='O', relation='O', anchor='O', format=[WORD, POS,\n CHUNK, PNP, REL, ANCHOR, LEMMA]):\n tags = []\n for tag in format:\n i...
def x(sentence): from nltk import tree def x(pnp): s = ' '.join([do_chunk(ch) for ch in pnp.chunks]) return '(PNP %s)' % s def x(ch): s = ' '.join([('(%s %s)' % (w.pos, w.string)) for w in ch.words]) s += ' '.join([do_pnp(pnp) for pnp in ch.attachments]) return '(%s %s)' % (ch.type, s) T = ['(S'] v = [] for ch in sentence.chunked(): if not ch.pnp and isinstance(ch, Chink): T.append('(%s %s)' % (ch.words[0].pos, ch.words[0].string)) elif not ch.pnp: T.append(do_chunk(ch)) elif ch.pnp.anchor is None and ch.pnp not in v: T.append(do_pnp(ch.pnp)) v.append(ch.pnp) T.append(')') return tree.bracket_parse(' '.join(T))
nltk_tree
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sentence):\n from nltk import tree\n\n def x(pnp):\n s = ' '.join([do_chunk(ch) for ch in pnp.chunks])\n return '(PNP %s)' % s\n\n def x...
def x(sentence, font='Arial', colors=BLUE): s = 'digraph sentence {\n' s += '\tranksep=0.75;\n' s += '\tnodesep=0.15;\n' s += ( '\tnode [penwidth=1, fontname="%s", shape=record, margin=0.1, height=0.35];\n' % font) s += '\tedge [penwidth=1];\n' s += '\t{ rank=same;\n' for w in sentence.words: s += '\t\tword%s [label="<f0>%s|<f1>%s"%s];\n' % (w.index, w.string, w.type, _colorize(w, colors)) for w in sentence.words[:-1]: s += '\t\tword%s -> word%s [color=none];\n' % (w.index, w.index + 1) s += '\t}\n' s += '\t{ rank=same;\n' for i, ch in enumerate(sentence.chunks): s += '\t\tchunk%s [label="<f0>%s"%s];\n' % (i + 1, '-'.join([x for x in (ch.type, ch.role, str(ch.relation or '')) if x]) or '-', _colorize(ch, colors)) for i, ch in enumerate(sentence.chunks[:-1]): s += '\t\tchunk%s -> chunk%s [color=none];\n' % (i + 1, i + 2) s += '}\n' s += '\t{ rank=same;\n' for i, ch in enumerate(sentence.pnp): s += '\t\tpnp%s [label="<f0>PNP"%s];\n' % (i + 1, _colorize(ch, colors) ) s += '\t}\n' s += '\t{ rank=same;\n S [shape=circle, margin=0.25, penwidth=2]; }\n' for i, ch in enumerate(sentence.chunks): for w in ch: s += '\tword%s -> chunk%s;\n' % (w.index, i + 1) if ch.pnp: s += '\tchunk%s -> pnp%s;\n' % (i + 1, sentence.pnp.index(ch. pnp) + 1) else: s += '\tchunk%s -> S;\n' % (i + 1) if ch.type == 'VP': for r in ch.related: s += ( '\tchunk%s -> chunk%s [style=dotted, arrowhead=none];\n' % (i + 1, sentence.chunks.index(r) + 1)) for i, ch in enumerate(sentence.pnp): if ch.anchor: s += '\tpnp%s -> chunk%s;\n' % (i + 1, sentence.chunks.index(ch .anchor) + 1) s += '\tpnp%s -> S [color=none];\n' % (i + 1) else: s += '\tpnp%s -> S;\n' % (i + 1) s += '}' return s
graphviz_dot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sentence, font='Arial', colors=BLUE):\n s = 'digraph sentence {\\n'\n s += '\\tranksep=0.75;\\n'\n s += '\\tnodesep=0.15;\\n'\n s += (\n '\\...
def x(sentence, fill=1, placeholder='-'): tags = [WORD, POS, IOB, CHUNK, ROLE, REL, PNP, ANCHOR, LEMMA] tags += [tag for tag in sentence.token if tag not in tags] def x(token, tag): if tag == WORD: s = token.string elif tag == POS: s = token.type elif tag == IOB: s = token.chunk and (token.index == token.chunk.start and 'B' or 'I') elif tag == CHUNK: s = token.chunk and token.chunk.type elif tag == ROLE: s = token.chunk and token.chunk.role elif tag == REL: s = token.chunk and token.chunk.relation and str(token.chunk. relation) elif tag == PNP: s = token.chunk and token.chunk.pnp and token.chunk.pnp.type elif tag == ANCHOR: s = token.chunk and token.chunk.anchor_id elif tag == LEMMA: s = token.lemma else: s = token.custom_tags.get(tag) return s or placeholder def x(column, fill=1, padding=3, align='left'): n = max([len(x) for x in column] + [fill]) if align == 'left': return [(x + ' ' * (n - len(x)) + ' ' * padding) for x in column] if align == 'right': return [(' ' * (n - len(x)) + x + ' ' * padding) for x in column] columns = [[format(token, tag) for token in sentence] for tag in tags] columns[3] = [(columns[3][i] + (iob == 'I' and ' ^' or '')) for i, iob in enumerate(columns[2])] del columns[2] for i, header in enumerate(['word', 'tag', 'chunk', 'role', 'id', 'pnp', 'anchor', 'lemma'] + tags[9:]): columns[i].insert(0, '') columns[i].insert(0, header.upper()) for i, column in enumerate(columns): columns[i] = outline(column, fill + 10 * (i == 0), align=('left', 'right')[i == 0]) if not MBSP: del columns[6] return '\n'.join([''.join([x[i] for x in columns]) for i in range(len( columns[0]))])
table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sentence, fill=1, placeholder='-'):\n tags = [WORD, POS, IOB, CHUNK, ROLE, REL, PNP, ANCHOR, LEMMA]\n tags += [tag for tag in sentence.token if tag not i...
def x(self): ch, I, O, B = self.chunk, INSIDE + '-', OUTSIDE, BEGIN + '-' tags = [OUTSIDE for i in range(len(self.sentence.token))] for i, tag in enumerate(self.sentence.token): if tag == WORD: tags[i] = encode_entities(self.string) elif tag == POS and self.type: tags[i] = self.type elif tag == CHUNK and ch and ch.type: tags[i] = (self == ch[0] and B or I) + ch.type elif tag == PNP and self.pnp: tags[i] = (self == self.pnp[0] and B or I) + 'PNP' elif tag == REL and ch and len(ch.relations) > 0: tags[i] = ['-'.join([str(x) for x in [ch.type] + list(reversed( r)) if x]) for r in ch.relations] tags[i] = '*'.join(tags[i]) elif tag == ANCHOR and ch: tags[i] = ch.anchor_id or OUTSIDE elif tag == LEMMA: tags[i] = encode_entities(self.lemma or '') elif tag in self.custom_tags: tags[i] = self.custom_tags.get(tag) or OUTSIDE return tags
tags
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ch, I, O, B = self.chunk, INSIDE + '-', OUTSIDE, BEGIN + '-'\n tags = [OUTSIDE for i in range(len(self.sentence.token))]\n for i, tag in enum...
def x(self, type=None): i = self.index + 1 s = self.sentence while i < len(s): if type in (s[i].type, None): return s[i] i += 1
next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type=None):\n i = self.index + 1\n s = self.sentence\n while i < len(s):\n if type in (s[i].type, None):\n return s[i]\n ...
def x(self, type=None): i = self.index - 1 s = self.sentence while i > 0: if type in (s[i].type, None): return s[i] i -= 1
previous
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type=None):\n i = self.index - 1\n s = self.sentence\n while i > 0:\n if type in (s[i].type, None):\n return s[i]\n i -...
def x(self): if self.type == 'NP' and any(w.type.startswith('NNP') for w in self): w = find(lambda w: w.type.startswith('NNP'), reversed(self)) elif self.type == 'NP': w = find(lambda w: w.type.startswith('NN'), reversed(self)) elif self.type == 'VP': w = find(lambda w: w.type.startswith('VB'), reversed(self)) elif self.type == 'PP': w = find(lambda w: w.type.startswith(('IN', 'PP')), self) elif self.type == 'PNP': w = find(lambda w: w.type.startswith('NN'), reversed(self)) else: w = None if w is None: w = self[-1] return w
head
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.type == 'NP' and any(w.type.startswith('NNP') for w in self):\n w = find(lambda w: w.type.startswith('NNP'), reversed(self))\n el...
def x(self): return [ch for ch in self.sentence.chunks if ch != self and intersects( unzip(0, ch.relations), unzip(0, self.relations))]
related
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [ch for ch in self.sentence.chunks if ch != self and intersects(\n unzip(0, ch.relations), unzip(0, self.relations))]\n", "role":...
def x(self): id = '' f = lambda ch: filter(lambda k: self.sentence._anchors[k] == ch, self. sentence._anchors) if self.pnp and self.pnp.anchor: id += '-' + '-'.join(f(self.pnp)) if self.anchor: id += '-' + '-'.join(f(self)) if self.attachments: id += '-' + '-'.join(f(self)) return id.strip('-') or None
anchor_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n id = ''\n f = lambda ch: filter(lambda k: self.sentence._anchors[k] == ch, self.\n sentence._anchors)\n if self.pnp and self.pnp.ancho...
def x(self): if self._modifiers is None: is_modifier = lambda ch: ch.type in ('ADJP', 'ADVP' ) and ch.relation is None for chunk in self.sentence.chunks: chunk._modifiers = [] for chunk in filter(is_modifier, self.sentence.chunks): anchor = chunk.nearest('VP') if anchor: anchor._modifiers.append(chunk) return self._modifiers
modifiers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._modifiers is None:\n is_modifier = lambda ch: ch.type in ('ADJP', 'ADVP'\n ) and ch.relation is None\n for chunk ...
def x(self, type='VP'): candidate, d = None, len(self.sentence.chunks) if isinstance(self, PNPChunk): i = self.sentence.chunks.index(self.chunks[0]) else: i = self.sentence.chunks.index(self) for j, chunk in enumerate(self.sentence.chunks): if chunk.type.startswith(type) and abs(i - j) < d: candidate, d = chunk, abs(i - j) return candidate
nearest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type='VP'):\n candidate, d = None, len(self.sentence.chunks)\n if isinstance(self, PNPChunk):\n i = self.sentence.chunks.index(self.chunks[0...
def x(self, type=None): i = self.stop s = self.sentence while i < len(s): if s[i].chunk is not None and type in (s[i].chunk.type, None): return s[i].chunk i += 1
next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type=None):\n i = self.stop\n s = self.sentence\n while i < len(s):\n if s[i].chunk is not None and type in (s[i].chunk.type, None):\n ...
def x(self, type=None): i = self.start - 1 s = self.sentence while i > 0: if s[i].chunk is not None and type in (s[i].chunk.type, None): return s[i].chunk i -= 1
previous
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type=None):\n i = self.start - 1\n s = self.sentence\n while i > 0:\n if s[i].chunk is not None and type in (s[i].chunk.type, None):\n ...
def x(self, word, lemma=None, type=None, chunk=None, role=None, relation= None, pnp=None, anchor=None, iob=None, custom={}): self._do_word(word, lemma, type) self._do_chunk(chunk, role, relation, iob) self._do_conjunction() self._do_relation() self._do_pnp(pnp, anchor) self._do_anchor(anchor) self._do_custom(custom)
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, word, lemma=None, type=None, chunk=None, role=None, relation=\n None, pnp=None, anchor=None, iob=None, custom={}):\n self._do_word(word, lemma, typ...
def x(self, token, tags=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]): p = {WORD: '', POS: None, IOB: None, CHUNK: None, PNP: None, REL: None, ROLE: None, ANCHOR: None, LEMMA: None} custom = {} for k, v in izip(tags, token.split('/')): if SLASH0 in v: v = v.replace(SLASH, '/') if k not in p: custom[k] = None if v != OUTSIDE or k == WORD or k == LEMMA: (p if k not in custom else custom)[k] = v ch = p[CHUNK] if ch is not None and ch.startswith(('B-', 'I-')): p[IOB], p[CHUNK] = ch[:1], ch[2:] if p[REL] is not None: ch, p[REL], p[ROLE] = self._parse_relation(p[REL]) if ch == 'PP' and self._previous and self._previous[REL] == p[REL ] and self._previous[ROLE] == p[ROLE]: ch = 'NP' if p[CHUNK] is None and ch != OUTSIDE: p[CHUNK] = ch self._previous = p return p[WORD], p[LEMMA], p[POS], p[CHUNK], p[ROLE], p[REL], p[PNP], p[ ANCHOR], p[IOB], custom
parse_token
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, token, tags=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n p = {WORD: '', POS: None, IOB: None, CHUNK: None, PNP: None, REL: None,\n ROLE: Non...
def x(self, tag): chunk, relation, role = None, [], [] if ';' in tag: id = tag.split('*')[0][-2:] id = id if id.startswith('-') else '' tag = tag.replace(';', id + '*') if '*' in tag: tag = tag.split('*') else: tag = [tag] for s in tag: s = s.split('-') n = len(s) if n == 1: chunk = s[0] if n == 2: chunk = s[0] relation.append(s[1]) role.append(None) if n >= 3: chunk = s[0] relation.append(s[2]) role.append(s[1]) if n > 1: id = relation[-1] if id.isdigit(): relation[-1] = int(id) else: relation[-1], role[-1] = None, id return chunk, relation, role
parse_relation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag):\n chunk, relation, role = None, [], []\n if ';' in tag:\n id = tag.split('*')[0][-2:]\n id = id if id.startswith('-') else ''\n...
def x(self, word, lemma=None, type=None): if lemma == "'s" and type in ('VB', 'VBZ'): lemma = 'be' self.words.append(Word(self, word, lemma, type, index=len(self.words)))
do_word
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, word, lemma=None, type=None):\n if lemma == \"'s\" and type in ('VB', 'VBZ'):\n lemma = 'be'\n self.words.append(Word(self, word, lemma, typ...