lines
sequencelengths 1
383
| raw_lines
sequencelengths 1
383
| label
sequencelengths 1
383
| type
sequencelengths 1
383
|
---|---|---|---|
[
"@override_settings(LOGIN_REDIRECT_URL='/custom/')...\n",
"self.assertLoginRedirectURLEqual('/custom/')\n"
] | [
"@override_settings(LOGIN_REDIRECT_URL='/custom/')...\n",
"self.assertLoginRedirectURLEqual('/custom/')\n"
] | [
0,
0
] | [
"Condition",
"Expr'"
] |
[
"def FUNC_44(VAR_15):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_108 = VAR_15.replace('\\\\', '/')\n",
"if os.path.isabs(VAR_108) and os.path.isfile(VAR_108):\n",
"VAR_168, VAR_5 = os.path.split(VAR_108)\n",
"return ''\n",
"VAR_11, VAR_169 = os.path.splitext(VAR_5)\n",
"VAR_3 = FUNC_5()\n",
"VAR_170 = {'controllers': '.py', 'models': '.py', 'views': '.html'}\n",
"for VAR_143 in VAR_170.keys():\n",
"VAR_194 = VAR_168.endswith('%s/%s' % (VAR_3, VAR_143))\n",
"if VAR_169.lower() == VAR_170[VAR_143] and VAR_194:\n",
"return to_native(A('\"' + VAR_108 + '\"', _href=URL(r=request, f=\n 'edit/%s/%s/%s' % (app, key, filename))).xml())\n"
] | [
"def make_link(path):...\n",
"\"\"\"docstring\"\"\"\n",
"tryFile = path.replace('\\\\', '/')\n",
"if os.path.isabs(tryFile) and os.path.isfile(tryFile):\n",
"folder, filename = os.path.split(tryFile)\n",
"return ''\n",
"base, ext = os.path.splitext(filename)\n",
"app = get_app()\n",
"editable = {'controllers': '.py', 'models': '.py', 'views': '.html'}\n",
"for key in editable.keys():\n",
"check_extension = folder.endswith('%s/%s' % (app, key))\n",
"if ext.lower() == editable[key] and check_extension:\n",
"return to_native(A('\"' + tryFile + '\"', _href=URL(r=request, f=\n 'edit/%s/%s/%s' % (app, key, filename))).xml())\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_0(VAR_2, VAR_3):...\n",
"\"\"\"docstring\"\"\"\n",
"for i in range(min(5, VAR_3)):\n",
"yield VAR_2 + i\n",
"for i in range(VAR_3 - 5):\n",
"yield max(1, VAR_2 + random.randint(-2 * VAR_3, 2 * VAR_3))\n"
] | [
"def random_ports(port, n):...\n",
"\"\"\"docstring\"\"\"\n",
"for i in range(min(5, n)):\n",
"yield port + i\n",
"for i in range(n - 5):\n",
"yield max(1, port + random.randint(-2 * n, 2 * n))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"For",
"Expr'",
"For",
"Expr'"
] |
[
"async def FUNC_45(self, VAR_50: List[str], VAR_11: str) ->Optional[dict]:...\n",
"\"\"\"docstring\"\"\"\n",
"for host in VAR_50:\n",
"VAR_168 = await self.federation_client.get_room_complexity(host, VAR_11)\n",
"return None\n",
"if VAR_168:\n",
"return VAR_168\n"
] | [
"async def get_room_complexity(self, remote_room_hosts: List[str], room_id: str...\n",
"\"\"\"docstring\"\"\"\n",
"for host in remote_room_hosts:\n",
"res = await self.federation_client.get_room_complexity(host, room_id)\n",
"return None\n",
"if res:\n",
"return res\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"AsyncFunctionDef'",
"Docstring",
"For",
"Assign'",
"Return'",
"Condition",
"Return'"
] |
[
"def FUNC_4(VAR_14, VAR_18=4):...\n",
"VAR_14 = VAR_3.sub('\\n', VAR_14)\n",
"VAR_14 = VAR_14.expandtabs(VAR_18)\n",
"VAR_14 = VAR_14.replace('', '\\n')\n",
"VAR_21 = re.compile('^ +$', re.M)\n",
"return VAR_21.sub('', VAR_14)\n"
] | [
"def preprocessing(text, tab=4):...\n",
"text = _newline_pattern.sub('\\n', text)\n",
"text = text.expandtabs(tab)\n",
"text = text.replace('', '\\n')\n",
"pattern = re.compile('^ +$', re.M)\n",
"return pattern.sub('', text)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_2(self, VAR_5, VAR_6, *VAR_21, **VAR_22):...\n",
"VAR_31 = self.rel.to\n",
"VAR_36 = VAR_31._meta.app_label, VAR_31._meta.model_name\n",
"self.widget.choices = self.choices\n",
"VAR_27 = [self.widget.render(VAR_5, VAR_6, *VAR_21, **kwargs)]\n",
"if self.can_add_related:\n",
"VAR_38 = reverse('admin:%s_%s_add' % VAR_36, current_app=self.admin_site.name)\n",
"return mark_safe(''.join(VAR_27))\n",
"VAR_27.append(\n '<a href=\"%s\" class=\"add-another\" id=\"add_id_%s\" onclick=\"return showAddAnotherPopup(this);\"> '\n % (VAR_38, VAR_5))\n",
"VAR_27.append('<img src=\"%s\" width=\"10\" height=\"10\" alt=\"%s\"/></a>' % (\n static('admin/img/icon_addlink.gif'), _('Add Another')))\n"
] | [
"def render(self, name, value, *args, **kwargs):...\n",
"rel_to = self.rel.to\n",
"info = rel_to._meta.app_label, rel_to._meta.model_name\n",
"self.widget.choices = self.choices\n",
"output = [self.widget.render(name, value, *args, **kwargs)]\n",
"if self.can_add_related:\n",
"related_url = reverse('admin:%s_%s_add' % info, current_app=self.admin_site\n .name)\n",
"return mark_safe(''.join(output))\n",
"output.append(\n '<a href=\"%s\" class=\"add-another\" id=\"add_id_%s\" onclick=\"return showAddAnotherPopup(this);\"> '\n % (related_url, name))\n",
"output.append('<img src=\"%s\" width=\"10\" height=\"10\" alt=\"%s\"/></a>' % (\n static('admin/img/icon_addlink.gif'), _('Add Another')))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Expr'",
"Expr'"
] |
[
"def FUNC_1(self, VAR_4, VAR_5, VAR_6):...\n",
"self.room_id = self.helper.create_room_as(self.user_id)\n",
"self.helper.join(self.room_id, VAR_2='@jim:red')\n"
] | [
"def prepare(self, reactor, clock, hs):...\n",
"self.room_id = self.helper.create_room_as(self.user_id)\n",
"self.helper.join(self.room_id, user='@jim:red')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_27(self):...\n",
"self.assert_expected(self.folder.t, 'InterpolationInContent.html')\n"
] | [
"def testInterpolationInContent(self):...\n",
"self.assert_expected(self.folder.t, 'InterpolationInContent.html')\n"
] | [
0,
0
] | [
"FunctionDef'",
"Expr'"
] |
[
"def FUNC_4(self):...\n",
"from openapi_python_client.parser.properties import EnumProperty\n",
"VAR_23 = ['abc', '123', 'a23', '1bc']\n",
"VAR_24 = VAR_27.values_from_list(VAR_23)\n",
"assert VAR_24 == {'ABC': 'abc', 'VALUE_1': '123', 'A23': 'a23', 'VALUE_3':\n '1bc'}\n"
] | [
"def test_values_from_list(self):...\n",
"from openapi_python_client.parser.properties import EnumProperty\n",
"data = ['abc', '123', 'a23', '1bc']\n",
"result = EnumProperty.values_from_list(data)\n",
"assert result == {'ABC': 'abc', 'VALUE_1': '123', 'A23': 'a23', 'VALUE_3':\n '1bc'}\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assert'"
] |
[
"@pytest.mark.posix...\n",
"assert filescheme.is_root(VAR_9) == VAR_10\n"
] | [
"@pytest.mark.posix...\n",
"assert filescheme.is_root(directory) == is_root\n"
] | [
0,
0
] | [
"Condition",
"Assert'"
] |
[
"def FUNC_73(self, VAR_2, VAR_8, VAR_175):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_203 = VAR_8.getEventContext()\n",
"VAR_24 = {}\n",
"for VAR_167 in ['sessionId', 'sessionUuid', 'userId', 'userName', 'groupId',\n",
"if hasattr(VAR_203, VAR_167):\n",
"return JsonResponse({'success': True, 'eventContext': VAR_24})\n",
"VAR_24[VAR_167] = getattr(VAR_203, VAR_167)\n"
] | [
"def handle_logged_in(self, request, conn, connector):...\n",
"\"\"\"docstring\"\"\"\n",
"c = conn.getEventContext()\n",
"ctx = {}\n",
"for a in ['sessionId', 'sessionUuid', 'userId', 'userName', 'groupId',\n",
"if hasattr(c, a):\n",
"return JsonResponse({'success': True, 'eventContext': ctx})\n",
"ctx[a] = getattr(c, a)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"For",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_57(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.loggedIn = True\n"
] | [
"def logIn(self):...\n",
"\"\"\"docstring\"\"\"\n",
"self.loggedIn = True\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'"
] |
[
"def FUNC_1(VAR_7, **VAR_8):...\n",
"if not VAR_7.user.is_authenticated:\n",
"return redirect(VAR_7.GET.get('next', reverse(settings.LOGIN_URL)))\n",
"if VAR_7.method == 'POST':\n",
"return VAR_2(VAR_7, **kwargs)\n",
"return render(VAR_7, 'spirit/user/auth/logout.html')\n"
] | [
"def custom_logout(request, **kwargs):...\n",
"if not request.user.is_authenticated:\n",
"return redirect(request.GET.get('next', reverse(settings.LOGIN_URL)))\n",
"if request.method == 'POST':\n",
"return _logout_view(request, **kwargs)\n",
"return render(request, 'spirit/user/auth/logout.html')\n"
] | [
0,
0,
4,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"def FUNC_11(self, VAR_1):...\n",
"VAR_29 = VAR_1.patch(f'{VAR_0}._string_based_property')\n",
"VAR_5 = VAR_1.MagicMock()\n",
"VAR_26 = VAR_1.MagicMock()\n",
"VAR_23 = oai.Schema.construct(type='string')\n",
"from openapi_python_client.parser.properties import property_from_data\n",
"VAR_4 = property_from_data(VAR_5=name, VAR_26=required, VAR_23=data)\n",
"assert VAR_4 == VAR_29.return_value\n",
"VAR_29.assert_called_once_with(VAR_5=name, VAR_26=required, VAR_23=data)\n"
] | [
"def test_property_from_data_string(self, mocker):...\n",
"_string_based_property = mocker.patch(f'{MODULE_NAME}._string_based_property')\n",
"name = mocker.MagicMock()\n",
"required = mocker.MagicMock()\n",
"data = oai.Schema.construct(type='string')\n",
"from openapi_python_client.parser.properties import property_from_data\n",
"p = property_from_data(name=name, required=required, data=data)\n",
"assert p == _string_based_property.return_value\n",
"_string_based_property.assert_called_once_with(name=name, required=required,\n data=data)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"ImportFrom'",
"Assign'",
"Assert'",
"Expr'"
] |
[
"def FUNC_11(self, VAR_55):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_81 = []\n",
"self._kill_elements(VAR_55, lambda VAR_15: VAR_6.search(VAR_15.text), etree\n .Comment)\n"
] | [
"def kill_conditional_comments(self, doc):...\n",
"\"\"\"docstring\"\"\"\n",
"bad = []\n",
"self._kill_elements(doc, lambda el: _conditional_comment_re.search(el.text),\n etree.Comment)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'"
] |
[
"def FUNC_5(VAR_11):...\n",
"return os.path.join(VAR_3, VAR_11)\n"
] | [
"def _get_obj_absolute_path(obj_path):...\n",
"return os.path.join(DATAROOT, obj_path)\n"
] | [
0,
1
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_2(VAR_0, VAR_2, VAR_3):...\n",
"VAR_7 = Topic.objects.get_public_or_404(VAR_2, VAR_0.user)\n",
"if VAR_7.slug != VAR_3:\n",
"return HttpResponsePermanentRedirect(VAR_7.get_absolute_url())\n",
"utils.topic_viewed(VAR_0=request, VAR_7=topic)\n",
"VAR_8 = Comment.objects.for_topic(VAR_7=topic).with_likes(VAR_4=request.user\n ).with_polls(VAR_4=request.user).order_by('date')\n",
"VAR_8 = paginate(VAR_8, per_page=config.comments_per_page, page_number=\n request.GET.get('page', 1))\n",
"return render(VAR_0=request, template_name='spirit/topic/detail.html',\n context={'topic': topic, 'comments': comments})\n"
] | [
"def detail(request, pk, slug):...\n",
"topic = Topic.objects.get_public_or_404(pk, request.user)\n",
"if topic.slug != slug:\n",
"return HttpResponsePermanentRedirect(topic.get_absolute_url())\n",
"utils.topic_viewed(request=request, topic=topic)\n",
"comments = Comment.objects.for_topic(topic=topic).with_likes(user=request.user\n ).with_polls(user=request.user).order_by('date')\n",
"comments = paginate(comments, per_page=config.comments_per_page,\n page_number=request.GET.get('page', 1))\n",
"return render(request=request, template_name='spirit/topic/detail.html',\n context={'topic': topic, 'comments': comments})\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_88(*VAR_79, **VAR_42):...\n",
"\"\"\"docstring\"\"\"\n",
"import frappe.utils.formatters\n",
"return frappe.utils.formatters.format_value(*VAR_79, **kwargs)\n"
] | [
"def format_value(*args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"import frappe.utils.formatters\n",
"return frappe.utils.formatters.format_value(*args, **kwargs)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Import'",
"Return'"
] |
[
"@VAR_1.route('/ajax/editbooks/<param>', methods=['POST'])...\n",
"VAR_79 = VAR_32.form.to_dict()\n",
"VAR_17 = calibre_db.get_book(VAR_79['pk'])\n",
"VAR_80 = ''\n",
"if VAR_43 == 'series_index':\n",
"FUNC_18(VAR_79['value'], VAR_17)\n",
"if VAR_43 == 'tags':\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.\n series_index}), mimetype='application/json')\n",
"FUNC_16(VAR_79['value'], VAR_17)\n",
"if VAR_43 == 'series':\n",
"VAR_17.last_modified = datetime.utcnow()\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([tag.\n name for tag in VAR_17.tags])}), mimetype='application/json')\n",
"FUNC_17(VAR_79['value'], VAR_17)\n",
"if VAR_43 == 'publishers':\n",
"calibre_db.session.commit()\n",
"calibre_db.session.rollback()\n",
"return VAR_80\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([serie\n .name for serie in VAR_17.series])}), mimetype='application/json')\n",
"FUNC_21(VAR_79['value'], VAR_17)\n",
"if VAR_43 == 'languages':\n",
"if VAR_43 == 'title' and VAR_79.get('checkT') == 'false':\n",
"VAR_2.error('Database error: %s', e)\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([\n VAR_86.name for VAR_86 in VAR_17.publishers])}), mimetype=\n 'application/json')\n",
"VAR_27 = list()\n",
"if VAR_43 == 'author_sort':\n",
"VAR_17.sort = VAR_124\n",
"FUNC_20(VAR_79['value'], VAR_17, VAR_27=invalid)\n",
"VAR_17.author_sort = VAR_79['value']\n",
"if VAR_43 == 'title':\n",
"calibre_db.session.commit()\n",
"if VAR_27:\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.\n author_sort}), mimetype='application/json')\n",
"VAR_124 = VAR_17.sort\n",
"if VAR_43 == 'sort':\n",
"VAR_80 = Response(json.dumps({'success': False, 'msg':\n 'Invalid languages in request: {}'.format(','.join(VAR_27))}), mimetype\n ='application/json')\n",
"VAR_121 = list()\n",
"FUNC_27(VAR_17, VAR_79.get('value', ''))\n",
"VAR_17.sort = VAR_79['value']\n",
"if VAR_43 == 'authors':\n",
"for lang in VAR_17.languages:\n",
"helper.update_dir_stucture(VAR_17.id, config.config_calibre_dir)\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.sort}),\n mimetype='application/json')\n",
"VAR_62, VAR_125 = FUNC_28(VAR_17, VAR_79['value'], VAR_79.get('checkA',\n None) == 'true')\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join(\n VAR_121)}), mimetype='application/json')\n",
"VAR_121.append(LC.parse(lang.lang_code).get_language_name(get_locale()))\n",
"VAR_121.append(_(isoLanguages.get(part3=lang.lang_code).name))\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.title}),\n mimetype='application/json')\n",
"helper.update_dir_stucture(VAR_17.id, config.config_calibre_dir, VAR_62[0])\n",
"VAR_80 = Response(json.dumps({'success': True, 'newValue': ' & '.join([\n author.replace('|', ',') for author in VAR_62])}), mimetype=\n 'application/json')\n"
] | [
"@editbook.route('/ajax/editbooks/<param>', methods=['POST'])...\n",
"vals = request.form.to_dict()\n",
"book = calibre_db.get_book(vals['pk'])\n",
"ret = ''\n",
"if param == 'series_index':\n",
"edit_book_series_index(vals['value'], book)\n",
"if param == 'tags':\n",
"ret = Response(json.dumps({'success': True, 'newValue': book.series_index}),\n mimetype='application/json')\n",
"edit_book_tags(vals['value'], book)\n",
"if param == 'series':\n",
"book.last_modified = datetime.utcnow()\n",
"ret = Response(json.dumps({'success': True, 'newValue': ', '.join([tag.name for\n tag in book.tags])}), mimetype='application/json')\n",
"edit_book_series(vals['value'], book)\n",
"if param == 'publishers':\n",
"calibre_db.session.commit()\n",
"calibre_db.session.rollback()\n",
"return ret\n",
"ret = Response(json.dumps({'success': True, 'newValue': ', '.join([serie.\n name for serie in book.series])}), mimetype='application/json')\n",
"edit_book_publisher(vals['value'], book)\n",
"if param == 'languages':\n",
"if param == 'title' and vals.get('checkT') == 'false':\n",
"log.error('Database error: %s', e)\n",
"ret = Response(json.dumps({'success': True, 'newValue': ', '.join([\n publisher.name for publisher in book.publishers])}), mimetype=\n 'application/json')\n",
"invalid = list()\n",
"if param == 'author_sort':\n",
"book.sort = sort\n",
"edit_book_languages(vals['value'], book, invalid=invalid)\n",
"book.author_sort = vals['value']\n",
"if param == 'title':\n",
"calibre_db.session.commit()\n",
"if invalid:\n",
"ret = Response(json.dumps({'success': True, 'newValue': book.author_sort}),\n mimetype='application/json')\n",
"sort = book.sort\n",
"if param == 'sort':\n",
"ret = Response(json.dumps({'success': False, 'msg':\n 'Invalid languages in request: {}'.format(','.join(invalid))}),\n mimetype='application/json')\n",
"lang_names = list()\n",
"handle_title_on_edit(book, vals.get('value', ''))\n",
"book.sort = vals['value']\n",
"if param == 'authors':\n",
"for lang in book.languages:\n",
"helper.update_dir_stucture(book.id, config.config_calibre_dir)\n",
"ret = Response(json.dumps({'success': True, 'newValue': book.sort}),\n mimetype='application/json')\n",
"input_authors, __ = handle_author_on_edit(book, vals['value'], vals.get(\n 'checkA', None) == 'true')\n",
"ret = Response(json.dumps({'success': True, 'newValue': ', '.join(\n lang_names)}), mimetype='application/json')\n",
"lang_names.append(LC.parse(lang.lang_code).get_language_name(get_locale()))\n",
"lang_names.append(_(isoLanguages.get(part3=lang.lang_code).name))\n",
"ret = Response(json.dumps({'success': True, 'newValue': book.title}),\n mimetype='application/json')\n",
"helper.update_dir_stucture(book.id, config.config_calibre_dir, input_authors[0]\n )\n",
"ret = Response(json.dumps({'success': True, 'newValue': ' & '.join([author.\n replace('|', ',') for author in input_authors])}), mimetype=\n 'application/json')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Return'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"For",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'"
] |
[
"def __post_init__(self) ->None:...\n",
"super().__post_init__()\n",
"if self.default is not None:\n",
"self.default = (\n f'field(default_factory=lambda: cast({self.get_type_string()}, {self.default}))'\n )\n"
] | [
"def __post_init__(self) ->None:...\n",
"super().__post_init__()\n",
"if self.default is not None:\n",
"self.default = (\n f'field(default_factory=lambda: cast({self.get_type_string()}, {self.default}))'\n )\n"
] | [
0,
5,
5,
5
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Assign'"
] |
[
"def FUNC_13(VAR_13):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_31 = LOG_LEVELS.get(str(VAR_13.get('log_level')).lower(), logging.NOTSET)\n",
"if VAR_31 < logging.INFO:\n",
"VAR_0.warning(\n 'Insecure logging configuration detected! Sensitive data may be logged.')\n"
] | [
"def verify_log(opts):...\n",
"\"\"\"docstring\"\"\"\n",
"level = LOG_LEVELS.get(str(opts.get('log_level')).lower(), logging.NOTSET)\n",
"if level < logging.INFO:\n",
"log.warning(\n 'Insecure logging configuration detected! Sensitive data may be logged.')\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Expr'"
] |
[
"@defer.inlineCallbacks...\n",
"VAR_7 = {'room': {'state': {'types': ['m.*']}}}\n",
"VAR_8 = yield defer.ensureDeferred(self.filtering.add_user_filter(VAR_0=\n user_localpart, VAR_10=user_filter_json))\n",
"self.assertEquals(VAR_8, 0)\n",
"self.assertEquals(VAR_7, (yield defer.ensureDeferred(self.datastore.\n get_user_filter(VAR_0=user_localpart, VAR_8=0))))\n"
] | [
"@defer.inlineCallbacks...\n",
"user_filter_json = {'room': {'state': {'types': ['m.*']}}}\n",
"filter_id = yield defer.ensureDeferred(self.filtering.add_user_filter(\n user_localpart=user_localpart, user_filter=user_filter_json))\n",
"self.assertEquals(filter_id, 0)\n",
"self.assertEquals(user_filter_json, (yield defer.ensureDeferred(self.\n datastore.get_user_filter(user_localpart=user_localpart, filter_id=0))))\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_11(self, VAR_11):...\n",
"VAR_60 = self.download_check_files(VAR_11)\n",
"if VAR_60 == 'not_permitted':\n",
"return 'You are not allowed to download files.'\n",
"if VAR_60 == 'invalid_file':\n",
"return 'Error: invalid filename found in {list}'.format(list=filelist)\n",
"if VAR_60 == 'too_big':\n",
"VAR_59 = cherry.config['media.maximum_download_size']\n",
"if VAR_60 == 'ok':\n",
"return 'string'.format(maxsize=size_limit / 1024 / 1024)\n",
"return VAR_60\n",
"VAR_125 = 'Error status check for download: {status!r}'.format(VAR_60=status)\n",
"log.e(VAR_125)\n",
"return VAR_125\n"
] | [
"def api_downloadcheck(self, filelist):...\n",
"status = self.download_check_files(filelist)\n",
"if status == 'not_permitted':\n",
"return 'You are not allowed to download files.'\n",
"if status == 'invalid_file':\n",
"return 'Error: invalid filename found in {list}'.format(list=filelist)\n",
"if status == 'too_big':\n",
"size_limit = cherry.config['media.maximum_download_size']\n",
"if status == 'ok':\n",
"return \"\"\"Can't download: Playlist is bigger than {maxsize} mB.\n The server administrator can change this configuration.\n \"\"\".format(\n maxsize=size_limit / 1024 / 1024)\n",
"return status\n",
"message = 'Error status check for download: {status!r}'.format(status=status)\n",
"log.e(message)\n",
"return message\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'",
"Condition",
"Return'",
"Condition",
"Assign'",
"Condition",
"Return'",
"Return'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_52(self):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_69 = self.get_doc_before_save()\n",
"if self._action == 'save':\n",
"self.run_method('on_update')\n",
"if self._action == 'submit':\n",
"self.clear_cache()\n",
"self.run_method('on_update')\n",
"if self._action == 'cancel':\n",
"self.notify_update()\n",
"self.run_method('on_submit')\n",
"self.run_method('on_cancel')\n",
"if self._action == 'update_after_submit':\n",
"update_global_search(self)\n",
"self.check_no_back_links_exist()\n",
"self.run_method('on_update_after_submit')\n",
"self.save_version()\n",
"self.run_method('on_change')\n",
"if (self.doctype, self.name) in frappe.flags.currently_saving:\n",
"frappe.flags.currently_saving.remove((self.doctype, self.name))\n",
"self.latest = None\n"
] | [
"def run_post_save_methods(self):...\n",
"\"\"\"docstring\"\"\"\n",
"doc_before_save = self.get_doc_before_save()\n",
"if self._action == 'save':\n",
"self.run_method('on_update')\n",
"if self._action == 'submit':\n",
"self.clear_cache()\n",
"self.run_method('on_update')\n",
"if self._action == 'cancel':\n",
"self.notify_update()\n",
"self.run_method('on_submit')\n",
"self.run_method('on_cancel')\n",
"if self._action == 'update_after_submit':\n",
"update_global_search(self)\n",
"self.check_no_back_links_exist()\n",
"self.run_method('on_update_after_submit')\n",
"self.save_version()\n",
"self.run_method('on_change')\n",
"if (self.doctype, self.name) in frappe.flags.currently_saving:\n",
"frappe.flags.currently_saving.remove((self.doctype, self.name))\n",
"self.latest = None\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"Assign'"
] |
[
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_87 = FUNC_0(VAR_2, 'page', 1)\n",
"return HttpResponseBadRequest('Invalid parameter value')\n",
"if not VAR_5.isValidGroup(VAR_20):\n",
"VAR_88 = FUNC_0(VAR_2, 'limit', settings.PAGE)\n",
"return HttpResponseForbidden('Not a member of Group: %s' % VAR_20)\n",
"VAR_25 = 'share_id' in VAR_6 and VAR_241(VAR_6['share_id']) or None\n",
"VAR_20 = FUNC_0(VAR_2, 'group', -1)\n",
"VAR_106 = tree.marshal_images(VAR_5=conn, VAR_247=orphaned, VAR_11=\n experimenter_id, VAR_246=dataset_id, VAR_25=share_id, VAR_248=\n load_pixels, VAR_20=group_id, VAR_87=page, VAR_250=date, VAR_249=\n thumb_version, VAR_88=limit)\n",
"return HttpResponseBadRequest(e.serverStackTrace)\n",
"return JsonResponse({'images': VAR_106})\n",
"VAR_246 = FUNC_0(VAR_2, 'id', None)\n",
"VAR_247 = FUNC_3(VAR_2, 'orphaned', False)\n",
"VAR_248 = FUNC_3(VAR_2, 'sizeXYZ', False)\n",
"VAR_249 = FUNC_3(VAR_2, 'thumbVersion', False)\n",
"VAR_250 = FUNC_3(VAR_2, 'date', False)\n",
"VAR_11 = FUNC_0(VAR_2, 'experimenter_id', -1)\n"
] | [
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"page = get_long_or_default(request, 'page', 1)\n",
"return HttpResponseBadRequest('Invalid parameter value')\n",
"if not conn.isValidGroup(group_id):\n",
"limit = get_long_or_default(request, 'limit', settings.PAGE)\n",
"return HttpResponseForbidden('Not a member of Group: %s' % group_id)\n",
"share_id = 'share_id' in kwargs and long(kwargs['share_id']) or None\n",
"group_id = get_long_or_default(request, 'group', -1)\n",
"images = tree.marshal_images(conn=conn, orphaned=orphaned, experimenter_id=\n experimenter_id, dataset_id=dataset_id, share_id=share_id, load_pixels=\n load_pixels, group_id=group_id, page=page, date=date, thumb_version=\n thumb_version, limit=limit)\n",
"return HttpResponseBadRequest(e.serverStackTrace)\n",
"return JsonResponse({'images': images})\n",
"dataset_id = get_long_or_default(request, 'id', None)\n",
"orphaned = get_bool_or_default(request, 'orphaned', False)\n",
"load_pixels = get_bool_or_default(request, 'sizeXYZ', False)\n",
"thumb_version = get_bool_or_default(request, 'thumbVersion', False)\n",
"date = get_bool_or_default(request, 'date', False)\n",
"experimenter_id = get_long_or_default(request, 'experimenter_id', -1)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'"
] |
[
"@staticmethod...\n",
"VAR_26 = VAR_13(VAR_24.text)\n",
"if not VAR_26:\n"
] | [
"@staticmethod...\n",
"is_valid = function(document.text)\n",
"if not is_valid:\n"
] | [
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition"
] |
[
"def FUNC_36(self):...\n",
"return self._queryset\n"
] | [
"def _get_queryset(self):...\n",
"return self._queryset\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_11(VAR_37):...\n",
"VAR_4, VAR_39 = VAR_37\n",
"VAR_25.write(VAR_4)\n",
"return VAR_39\n"
] | [
"def write_to(r):...\n",
"data, response = r\n",
"output_stream.write(data)\n",
"return response\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_4():...\n",
"return VAR_7\n"
] | [
"def getTrustedEngine():...\n",
"return _trusted_engine\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@FUNC_0...\n",
"return EventSources(self)\n"
] | [
"@cache_in_self...\n",
"return EventSources(self)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"from security_monkey import app, db\n",
"from flask_wtf.csrf import generate_csrf\n",
"from security_monkey.auth.models import RBACRole\n",
"from security_monkey.decorators import crossdomain\n",
"from flask_restful import fields, marshal, Resource, reqparse\n",
"from flask_login import current_user\n",
"VAR_0 = ['https://{}:{}'.format(app.config.get('FQDN'), app.config.get(\n 'WEB_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.config\n .get('API_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.\n config.get('NGINX_PORT')), 'https://{}:80'.format(app.config.get('FQDN'))]\n",
"VAR_1 = {'id': fields.Integer, 'date_created': fields.String,\n 'date_last_ephemeral_change': fields.String, 'active': fields.Boolean,\n 'item_id': fields.Integer}\n",
"VAR_2 = {'id': fields.Integer, 'region': fields.String, 'name': fields.String}\n",
"VAR_3 = {'id': fields.Integer, 'score': fields.Integer, 'issue': fields.\n String, 'notes': fields.String, 'justified': fields.Boolean,\n 'justification': fields.String, 'justified_date': fields.String,\n 'item_id': fields.Integer}\n",
"VAR_4 = {'id': fields.Integer, 'revision_id': fields.Integer,\n 'date_created': fields.String, 'text': fields.String}\n",
"VAR_5 = {'id': fields.Integer, 'date_created': fields.String, 'text':\n fields.String, 'item_id': fields.Integer}\n",
"VAR_6 = {'daily_audit_email': fields.Boolean, 'change_reports': fields.String}\n",
"VAR_7 = {'id': fields.Integer, 'name': fields.String, 'identifier': fields.\n String, 'notes': fields.String, 'active': fields.Boolean, 'third_party':\n fields.Boolean, 'account_type': fields.String}\n",
"VAR_8 = {'id': fields.Integer, 'active': fields.Boolean, 'email': fields.\n String, 'role': fields.String, 'confirmed_at': fields.String,\n 'daily_audit_email': fields.Boolean, 'change_reports': fields.String,\n 'last_login_at': fields.String, 'current_login_at': fields.String,\n 'login_count': fields.Integer, 'last_login_ip': fields.String,\n 'current_login_ip': fields.String}\n",
"VAR_9 = {'id': fields.Integer, 'name': fields.String, 'description': fields\n .String}\n",
"VAR_10 = {'id': fields.Integer, 'name': fields.String, 'notes': fields.\n String, 'cidr': fields.String}\n",
"VAR_11 = {'id': fields.Integer, 'prefix': fields.String, 'notes': fields.String\n }\n",
"VAR_12 = {'id': fields.Integer, 'disabled': fields.Boolean, 'issue_text':\n fields.String}\n",
"VAR_13 = {'id': fields.Integer, 'name': fields.String}\n",
"def __init__(self):...\n",
"self.reqparse = reqparse.RequestParser()\n",
"super(CLASS_0, self).__init__()\n",
"self.auth_dict = dict()\n",
"if current_user.is_authenticated():\n",
"VAR_15 = []\n",
"if app.config.get('FRONTED_BY_NGINX'):\n",
"for role in current_user.roles:\n",
"VAR_16 = 'https://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'NGINX_PORT'), '/login')\n",
"VAR_16 = 'http://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'API_PORT'), '/login')\n",
"VAR_15.append(marshal(role.__dict__, VAR_9))\n",
"VAR_15.append({'name': current_user.role})\n",
"self.auth_dict = {'authenticated': False, 'user': None, 'url': VAR_16}\n",
"for role in RBACRole.roles[current_user.role].get_parents():\n",
"@app.after_request...\n",
"VAR_15.append({'name': role.name})\n",
"self.auth_dict = {'authenticated': True, 'user': current_user.email,\n 'roles': VAR_15}\n",
"VAR_14.set_cookie('XSRF-COOKIE', generate_csrf())\n",
"return VAR_14\n"
] | [
"from security_monkey import app, db\n",
"from flask_wtf.csrf import generate_csrf\n",
"from security_monkey.auth.models import RBACRole\n",
"from security_monkey.decorators import crossdomain\n",
"from flask_restful import fields, marshal, Resource, reqparse\n",
"from flask_login import current_user\n",
"ORIGINS = ['https://{}:{}'.format(app.config.get('FQDN'), app.config.get(\n 'WEB_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.config\n .get('API_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.\n config.get('NGINX_PORT')), 'https://{}:80'.format(app.config.get('FQDN'))]\n",
"REVISION_FIELDS = {'id': fields.Integer, 'date_created': fields.String,\n 'date_last_ephemeral_change': fields.String, 'active': fields.Boolean,\n 'item_id': fields.Integer}\n",
"ITEM_FIELDS = {'id': fields.Integer, 'region': fields.String, 'name':\n fields.String}\n",
"AUDIT_FIELDS = {'id': fields.Integer, 'score': fields.Integer, 'issue':\n fields.String, 'notes': fields.String, 'justified': fields.Boolean,\n 'justification': fields.String, 'justified_date': fields.String,\n 'item_id': fields.Integer}\n",
"REVISION_COMMENT_FIELDS = {'id': fields.Integer, 'revision_id': fields.\n Integer, 'date_created': fields.String, 'text': fields.String}\n",
"ITEM_COMMENT_FIELDS = {'id': fields.Integer, 'date_created': fields.String,\n 'text': fields.String, 'item_id': fields.Integer}\n",
"USER_SETTINGS_FIELDS = {'daily_audit_email': fields.Boolean,\n 'change_reports': fields.String}\n",
"ACCOUNT_FIELDS = {'id': fields.Integer, 'name': fields.String, 'identifier':\n fields.String, 'notes': fields.String, 'active': fields.Boolean,\n 'third_party': fields.Boolean, 'account_type': fields.String}\n",
"USER_FIELDS = {'id': fields.Integer, 'active': fields.Boolean, 'email':\n fields.String, 'role': fields.String, 'confirmed_at': fields.String,\n 'daily_audit_email': fields.Boolean, 'change_reports': fields.String,\n 'last_login_at': fields.String, 'current_login_at': fields.String,\n 'login_count': fields.Integer, 'last_login_ip': fields.String,\n 'current_login_ip': fields.String}\n",
"ROLE_FIELDS = {'id': fields.Integer, 'name': fields.String, 'description':\n fields.String}\n",
"WHITELIST_FIELDS = {'id': fields.Integer, 'name': fields.String, 'notes':\n fields.String, 'cidr': fields.String}\n",
"IGNORELIST_FIELDS = {'id': fields.Integer, 'prefix': fields.String, 'notes':\n fields.String}\n",
"AUDITORSETTING_FIELDS = {'id': fields.Integer, 'disabled': fields.Boolean,\n 'issue_text': fields.String}\n",
"ITEM_LINK_FIELDS = {'id': fields.Integer, 'name': fields.String}\n",
"def __init__(self):...\n",
"self.reqparse = reqparse.RequestParser()\n",
"super(AuthenticatedService, self).__init__()\n",
"self.auth_dict = dict()\n",
"if current_user.is_authenticated():\n",
"roles_marshal = []\n",
"if app.config.get('FRONTED_BY_NGINX'):\n",
"for role in current_user.roles:\n",
"url = 'https://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'NGINX_PORT'), '/login')\n",
"url = 'http://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'API_PORT'), '/login')\n",
"roles_marshal.append(marshal(role.__dict__, ROLE_FIELDS))\n",
"roles_marshal.append({'name': current_user.role})\n",
"self.auth_dict = {'authenticated': False, 'user': None, 'url': url}\n",
"for role in RBACRole.roles[current_user.role].get_parents():\n",
"@app.after_request...\n",
"roles_marshal.append({'name': role.name})\n",
"self.auth_dict = {'authenticated': True, 'user': current_user.email,\n 'roles': roles_marshal}\n",
"response.set_cookie('XSRF-COOKIE', generate_csrf())\n",
"return response\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"For",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"For",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_23(VAR_48, VAR_49=VAR_29):...\n",
"return VAR_49 is not None and (len(VAR_48) > VAR_49 and '%s...' % VAR_48[:\n max(0, VAR_49 - 3)]) or VAR_48\n"
] | [
"def trim_url(x, limit=trim_url_limit):...\n",
"return limit is not None and (len(x) > limit and '%s...' % x[:max(0, limit -\n 3)]) or x\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_6(self):...\n",
"return self.kw_component\n"
] | [
"def get_kwargs(self):...\n",
"return self.kw_component\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_14(VAR_7, VAR_8):...\n",
"VAR_10 = [db.Books.timestamp.desc()]\n",
"if VAR_7 == 'stored':\n",
"VAR_7 = VAR_87.get_view_property(VAR_8, 'stored')\n",
"VAR_87.set_view_property(VAR_8, 'stored', VAR_7)\n",
"if VAR_7 == 'pubnew':\n",
"VAR_10 = [db.Books.pubdate.desc()]\n",
"if VAR_7 == 'pubold':\n",
"VAR_10 = [db.Books.pubdate]\n",
"if VAR_7 == 'abc':\n",
"VAR_10 = [db.Books.sort]\n",
"if VAR_7 == 'zyx':\n",
"VAR_10 = [db.Books.sort.desc()]\n",
"if VAR_7 == 'new':\n",
"VAR_10 = [db.Books.timestamp.desc()]\n",
"if VAR_7 == 'old':\n",
"VAR_10 = [db.Books.timestamp]\n",
"if VAR_7 == 'authaz':\n",
"VAR_10 = [db.Books.author_sort.asc(), db.Series.name, db.Books.series_index]\n",
"if VAR_7 == 'authza':\n",
"VAR_10 = [db.Books.author_sort.desc(), db.Series.name.desc(), db.Books.\n series_index.desc()]\n",
"if VAR_7 == 'seriesasc':\n",
"VAR_10 = [db.Books.series_index.asc()]\n",
"if VAR_7 == 'seriesdesc':\n",
"VAR_10 = [db.Books.series_index.desc()]\n",
"if VAR_7 == 'hotdesc':\n",
"VAR_10 = [func.count(ub.Downloads.book_id).desc()]\n",
"if VAR_7 == 'hotasc':\n",
"VAR_10 = [func.count(ub.Downloads.book_id).asc()]\n",
"if VAR_7 is None:\n",
"VAR_7 = 'new'\n",
"return VAR_10, VAR_7\n"
] | [
"def get_sort_function(sort, data):...\n",
"order = [db.Books.timestamp.desc()]\n",
"if sort == 'stored':\n",
"sort = current_user.get_view_property(data, 'stored')\n",
"current_user.set_view_property(data, 'stored', sort)\n",
"if sort == 'pubnew':\n",
"order = [db.Books.pubdate.desc()]\n",
"if sort == 'pubold':\n",
"order = [db.Books.pubdate]\n",
"if sort == 'abc':\n",
"order = [db.Books.sort]\n",
"if sort == 'zyx':\n",
"order = [db.Books.sort.desc()]\n",
"if sort == 'new':\n",
"order = [db.Books.timestamp.desc()]\n",
"if sort == 'old':\n",
"order = [db.Books.timestamp]\n",
"if sort == 'authaz':\n",
"order = [db.Books.author_sort.asc(), db.Series.name, db.Books.series_index]\n",
"if sort == 'authza':\n",
"order = [db.Books.author_sort.desc(), db.Series.name.desc(), db.Books.\n series_index.desc()]\n",
"if sort == 'seriesasc':\n",
"order = [db.Books.series_index.asc()]\n",
"if sort == 'seriesdesc':\n",
"order = [db.Books.series_index.desc()]\n",
"if sort == 'hotdesc':\n",
"order = [func.count(ub.Downloads.book_id).desc()]\n",
"if sort == 'hotasc':\n",
"order = [func.count(ub.Downloads.book_id).asc()]\n",
"if sort is None:\n",
"sort = 'new'\n",
"return order, sort\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Expr'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Return'"
] |
[
"def FUNC_0(self):...\n",
"return self.post()\n"
] | [
"def get(self):...\n",
"return self.post()\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@wraps(VAR_4)...\n",
"if VAR_87.role_viewer():\n",
"return VAR_4(*VAR_49, **kwargs)\n",
"abort(403)\n"
] | [
"@wraps(f)...\n",
"if current_user.role_viewer():\n",
"return f(*args, **kwargs)\n",
"abort(403)\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Return'",
"Expr'"
] |
[
"def FUNC_69(VAR_12, VAR_29):...\n",
"VAR_94 = re.search('([a-zA-Z0-9-]+)=%s', VAR_29, re.IGNORECASE | re.UNICODE)\n",
"if VAR_94:\n",
"return VAR_94.group(1)\n"
] | [
"def extract_dynamic_field_from_filter(user, filtr):...\n",
"match = re.search('([a-zA-Z0-9-]+)=%s', filtr, re.IGNORECASE | re.UNICODE)\n",
"if match:\n",
"return match.group(1)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Condition",
"Return'"
] |
[
"def FUNC_1(self, VAR_5, VAR_7, VAR_8=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if self.ignore_login_fail:\n",
"return HttpResponse('Connection Failed')\n",
"if self.login_redirect is not None:\n",
"return super(CLASS_0, self).on_not_logged_in(VAR_5, VAR_7, VAR_8)\n",
"VAR_7 = reverse(self.login_redirect)\n"
] | [
"def on_not_logged_in(self, request, url, error=None):...\n",
"\"\"\"docstring\"\"\"\n",
"if self.ignore_login_fail:\n",
"return HttpResponse('Connection Failed')\n",
"if self.login_redirect is not None:\n",
"return super(login_required, self).on_not_logged_in(request, url, error)\n",
"url = reverse(self.login_redirect)\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_0(self, VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"return yaml.load(self.vault.decrypt(VAR_1))\n"
] | [
"def load(self, stream):...\n",
"\"\"\"docstring\"\"\"\n",
"return yaml.load(self.vault.decrypt(stream))\n"
] | [
0,
0,
5
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_9(self, VAR_5):...\n",
"for VAR_20 in self._subexprs:\n",
"return 0\n",
"VAR_20(VAR_5)\n",
"return 1\n"
] | [
"def _exists(self, econtext):...\n",
"for expr in self._subexprs:\n",
"return 0\n",
"expr(econtext)\n",
"return 1\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"For",
"Return'",
"Expr'",
"Return'"
] |
[
"@ensure_csrf_cookie...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n",
"VAR_121 = RegistrationCodeRedemption.objects.select_related('registration_code'\n ).filter(registration_code__course_id=course_id)\n",
"VAR_122 = []\n",
"if VAR_121.exists():\n",
"VAR_184 = [VAR_82.registration_code.code for VAR_82 in VAR_121]\n",
"VAR_34 = 'spent'\n",
"VAR_122 = CourseRegistrationCode.objects.filter(VAR_10=course_id, code__in=\n redeemed_registration_codes).order_by('invoice_item__invoice__company_name'\n ).select_related('invoice_item__invoice')\n",
"return FUNC_34('Spent_Registration_Codes.csv', VAR_122, VAR_34)\n",
"VAR_88 = VAR_9.POST['spent_company_name']\n",
"if VAR_88:\n",
"VAR_122 = VAR_122.filter(invoice_item__invoice__company_name=company_name)\n"
] | [
"@ensure_csrf_cookie...\n",
"\"\"\"docstring\"\"\"\n",
"course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n",
"code_redemption_set = RegistrationCodeRedemption.objects.select_related(\n 'registration_code').filter(registration_code__course_id=course_id)\n",
"spent_codes_list = []\n",
"if code_redemption_set.exists():\n",
"redeemed_registration_codes = [code.registration_code.code for code in\n code_redemption_set]\n",
"csv_type = 'spent'\n",
"spent_codes_list = CourseRegistrationCode.objects.filter(course_id=\n course_id, code__in=redeemed_registration_codes).order_by(\n 'invoice_item__invoice__company_name').select_related(\n 'invoice_item__invoice')\n",
"return registration_codes_csv('Spent_Registration_Codes.csv',\n spent_codes_list, csv_type)\n",
"company_name = request.POST['spent_company_name']\n",
"if company_name:\n",
"spent_codes_list = spent_codes_list.filter(invoice_item__invoice__company_name\n =company_name)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Return'",
"Assign'",
"Condition",
"Assign'"
] |
[
"def FUNC_12(self, VAR_0, VAR_13):...\n",
"VAR_33 = 'C:\\\\' if utils.is_windows else '/'\n",
"VAR_25 = VAR_13(VAR_33)\n",
"assert not VAR_25.parent\n"
] | [
"def test_root_dir(self, tmpdir, parser):...\n",
"root_dir = 'C:\\\\' if utils.is_windows else '/'\n",
"parsed = parser(root_dir)\n",
"assert not parsed.parent\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assert'"
] |
[
"def FUNC_15(self, VAR_31, VAR_32):...\n",
"return CLASS_7(VAR_31, VAR_32)\n"
] | [
"def createErrorInfo(self, err, position):...\n",
"return ErrorInfo(err, position)\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"def FUNC_5(self, VAR_7):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_12, VAR_13 = self.make_request('GET', '/joined_groups'.encode('ascii'),\n VAR_7=access_token)\n",
"self.assertEqual(200, int(VAR_13.result['code']), msg=channel.result['body'])\n",
"return VAR_13.json_body['groups']\n"
] | [
"def _get_groups_user_is_in(self, access_token):...\n",
"\"\"\"docstring\"\"\"\n",
"request, channel = self.make_request('GET', '/joined_groups'.encode('ascii'\n ), access_token=access_token)\n",
"self.assertEqual(200, int(channel.result['code']), msg=channel.result['body'])\n",
"return channel.json_body['groups']\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Expr'",
"Return'"
] |
[
"async def FUNC_18(VAR_3, **VAR_4):...\n",
"VAR_17 = RedirectException(b'/no/over/there', 304)\n",
"VAR_17.cookies.append(b'session=yespls')\n"
] | [
"async def callback(request, **kwargs):...\n",
"e = RedirectException(b'/no/over/there', 304)\n",
"e.cookies.append(b'session=yespls')\n"
] | [
0,
0,
0
] | [
"AsyncFunctionDef'",
"Assign'",
"Expr'"
] |
[
"@login_required...\n",
"if not is_ajax(VAR_0):\n",
"return Http404()\n",
"VAR_6 = TopicNotification.objects.for_access(VAR_0.user).order_by('is_read',\n '-date').with_related_data()\n",
"VAR_6 = VAR_6[:settings.ST_NOTIFICATIONS_PER_PAGE]\n",
"VAR_6 = [{'user': escape(n.comment.user.st.nickname), 'action': n.action,\n 'title': escape(n.topic.title), 'url': n.get_absolute_url(), 'is_read':\n n.is_read} for n in VAR_6]\n",
"return HttpResponse(json.dumps({'n': VAR_6}), content_type='application/json')\n"
] | [
"@login_required...\n",
"if not is_ajax(request):\n",
"return Http404()\n",
"notifications = TopicNotification.objects.for_access(request.user).order_by(\n 'is_read', '-date').with_related_data()\n",
"notifications = notifications[:settings.ST_NOTIFICATIONS_PER_PAGE]\n",
"notifications = [{'user': escape(n.comment.user.st.nickname), 'action': n.\n action, 'title': escape(n.topic.title), 'url': n.get_absolute_url(),\n 'is_read': n.is_read} for n in notifications]\n",
"return HttpResponse(json.dumps({'n': notifications}), content_type=\n 'application/json')\n"
] | [
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Return'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_12(VAR_2, VAR_3, VAR_5, VAR_10, VAR_11, VAR_12, VAR_13=None,...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_4 = saved_model_utils.get_meta_graph_def(VAR_2, VAR_3)\n",
"VAR_24 = FUNC_2(VAR_4, VAR_5)\n",
"for input_key_name in VAR_10.keys():\n",
"if input_key_name not in VAR_24:\n",
"VAR_36 = {VAR_24[key].name: tensor for key, tensor in VAR_10.items()}\n",
"VAR_25 = FUNC_3(VAR_4, VAR_5)\n",
"VAR_37 = sorted(VAR_25.keys())\n",
"VAR_38 = [VAR_25[tensor_key].name for tensor_key in VAR_37]\n",
"if VAR_14:\n",
"print('Initializing TPU System ...')\n",
"loader.load(VAR_77, VAR_3.split(','), VAR_2)\n",
"VAR_77.run(tpu.initialize_system())\n",
"if VAR_15:\n",
"VAR_77 = local_cli_wrapper.LocalCLIDebugWrapperSession(VAR_77)\n",
"VAR_66 = VAR_77.run(VAR_38, feed_dict=inputs_feed_dict)\n",
"for i, output in enumerate(VAR_66):\n",
"VAR_78 = VAR_37[i]\n",
"print(\"\"\"Result for output key %s:\n%s\"\"\" % (VAR_78, output))\n",
"if VAR_11:\n",
"if not os.path.isdir(VAR_11):\n",
"os.makedirs(VAR_11)\n",
"VAR_79 = os.path.join(VAR_11, VAR_78 + '.npy')\n",
"if not VAR_12 and os.path.exists(VAR_79):\n",
"np.save(VAR_79, output)\n",
"print('Output %s is saved to %s' % (VAR_78, VAR_79))\n"
] | [
"def run_saved_model_with_feed_dict(saved_model_dir, tag_set,...\n",
"\"\"\"docstring\"\"\"\n",
"meta_graph_def = saved_model_utils.get_meta_graph_def(saved_model_dir, tag_set)\n",
"inputs_tensor_info = _get_inputs_tensor_info_from_meta_graph_def(meta_graph_def\n , signature_def_key)\n",
"for input_key_name in input_tensor_key_feed_dict.keys():\n",
"if input_key_name not in inputs_tensor_info:\n",
"inputs_feed_dict = {inputs_tensor_info[key].name: tensor for key, tensor in\n input_tensor_key_feed_dict.items()}\n",
"outputs_tensor_info = _get_outputs_tensor_info_from_meta_graph_def(\n meta_graph_def, signature_def_key)\n",
"output_tensor_keys_sorted = sorted(outputs_tensor_info.keys())\n",
"output_tensor_names_sorted = [outputs_tensor_info[tensor_key].name for\n tensor_key in output_tensor_keys_sorted]\n",
"if init_tpu:\n",
"print('Initializing TPU System ...')\n",
"loader.load(sess, tag_set.split(','), saved_model_dir)\n",
"sess.run(tpu.initialize_system())\n",
"if tf_debug:\n",
"sess = local_cli_wrapper.LocalCLIDebugWrapperSession(sess)\n",
"outputs = sess.run(output_tensor_names_sorted, feed_dict=inputs_feed_dict)\n",
"for i, output in enumerate(outputs):\n",
"output_tensor_key = output_tensor_keys_sorted[i]\n",
"print(\"\"\"Result for output key %s:\n%s\"\"\" % (output_tensor_key, output))\n",
"if outdir:\n",
"if not os.path.isdir(outdir):\n",
"os.makedirs(outdir)\n",
"output_full_path = os.path.join(outdir, output_tensor_key + '.npy')\n",
"if not overwrite_flag and os.path.exists(output_full_path):\n",
"np.save(output_full_path, output)\n",
"print('Output %s is saved to %s' % (output_tensor_key, output_full_path))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Assign'",
"For",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Expr'"
] |
[
"import pytest\n",
"from werkzeug.routing import BuildError\n",
"from flask_unchained.bundles.controller import Controller, Resource\n",
"from flask_unchained.bundles.controller.utils import controller_name, get_param_tuples, join, method_name_to_url, url_for, _validate_redirect_url\n",
"from py_meta_utils import deep_getattr\n",
"def FUNC_0():...\n",
"VAR_0 = {'a': 'clsdict'}\n",
"VAR_4 = 'first'\n",
"VAR_5 = 'first'\n",
"VAR_5 = 'second'\n",
"VAR_6 = 'second'\n",
"VAR_1 = CLASS_6, CLASS_7\n",
"assert deep_getattr(VAR_0, VAR_1, 'a') == 'clsdict'\n",
"assert deep_getattr(VAR_0, VAR_1, 'b') == 'first'\n",
"assert deep_getattr(VAR_0, VAR_1, 'c') == 'second'\n",
"deep_getattr(VAR_0, VAR_1, 'd')\n",
"assert deep_getattr(VAR_0, VAR_1, 'a', 'default') == 'clsdict'\n",
"assert deep_getattr(VAR_0, VAR_1, 'b', 'default') == 'first'\n",
"assert deep_getattr(VAR_0, VAR_1, 'c', 'default') == 'second'\n",
"assert deep_getattr(VAR_0, VAR_1, 'd', 'default') == 'default'\n",
"def FUNC_1(self):...\n",
"assert controller_name(CLASS_8) == 'user'\n",
"def FUNC_2(self):...\n",
"assert controller_name(CLASS_9) == 'api'\n",
"def FUNC_3(self):...\n",
"assert controller_name(CLASS_10) == 'some'\n",
"def FUNC_4(self):...\n",
"assert controller_name(CLASS_11) == 'more_than_one_word'\n",
"def FUNC_5(self):...\n",
"assert controller_name(CLASS_12) == 'user'\n",
"def FUNC_6(self):...\n",
"assert controller_name(CLASS_13) == 'role'\n",
"def FUNC_7(self):...\n",
"assert controller_name(CLASS_14) == 'role_view_controller'\n",
"def FUNC_8(self):...\n",
"assert controller_name(CLASS_15) == 'some_ctrl'\n",
"def FUNC_9(self):...\n",
"assert get_param_tuples('<int:id>') == [('int', 'id')]\n",
"def FUNC_10(self):...\n",
"assert get_param_tuples(None) == []\n",
"def FUNC_11(self):...\n",
"VAR_7 = '/users/<int:user_id>/roles/<string:slug>'\n",
"assert get_param_tuples(VAR_7) == [('int', 'user_id'), ('string', 'slug')]\n",
"def FUNC_12(self):...\n",
"assert url_for('/foobar') == '/foobar'\n",
"def FUNC_13(self):...\n",
"assert url_for(None) is None\n",
"def FUNC_14(self, VAR_2):...\n",
"VAR_2.config.from_mapping({'MY_KEY': '/my-key'})\n",
"assert url_for('MY_KEY') == '/my-key'\n",
"def FUNC_15(self, VAR_2):...\n",
"VAR_2.config.from_mapping({'MY_KEY': 'some.endpoint'})\n",
"assert url_for('MY_KEY')\n",
"VAR_2.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n",
"assert url_for('MY_KEY') == '/some-endpoint'\n",
"def FUNC_16(self, VAR_2):...\n",
"assert url_for('some.endpoint')\n",
"VAR_2.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n",
"assert url_for('some.endpoint') == '/some-endpoint'\n",
"def FUNC_17(self, VAR_2):...\n",
"def FUNC_29(self):...\n",
"VAR_2.add_url_rule('/about-us', endpoint='site_controller.about_us')\n",
"assert url_for('about_us', _cls=SiteController) == '/about-us'\n",
"def FUNC_18(self, VAR_2):...\n",
"def FUNC_30(self, VAR_8):...\n",
"VAR_2.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n",
"assert url_for('get', VAR_8=1, _cls=SiteResource) == '/sites/1'\n",
"VAR_2.add_url_rule('/foo/<string:slug>', endpoint='some.endpoint')\n",
"assert url_for('some.endpoint', slug='hi') == '/foo/hi'\n",
"def FUNC_19(self, VAR_2):...\n",
"def FUNC_30(self, VAR_8):...\n",
"VAR_2.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n",
"url_for('delete', VAR_8=1, _cls=SiteResource)\n",
"def FUNC_13(self):...\n",
"assert join(None) == '/'\n",
"assert join(None, None, '', 0) == '/'\n",
"def FUNC_20(self):...\n",
"assert join('/', 'foo', None, 'bar', '', 'baz') == '/foo/bar/baz'\n",
"def FUNC_21(self):...\n",
"assert join('/', '/foo', '/', '/bar') == '/foo/bar'\n",
"def FUNC_22(self):...\n",
"assert join('/', '/') == '/'\n",
"assert join(None, '/') == '/'\n",
"assert join('/', None) == '/'\n",
"def FUNC_23(self):...\n",
"assert join('/foo/bar/') == '/foo/bar'\n",
"assert join('/foo/bar/', None) == '/foo/bar'\n",
"assert join('/foo/bar/', '/') == '/foo/bar'\n",
"assert join('/foo', 'bar/') == '/foo/bar'\n",
"def FUNC_24(self):...\n",
"assert join('/', trailing_slash=True) == '/'\n",
"assert join('/foo', 'baz', None, trailing_slash=True) == '/foo/baz/'\n",
"assert join('/foo', 'baz/', trailing_slash=True) == '/foo/baz/'\n",
"def FUNC_9(self):...\n",
"assert method_name_to_url('fooBar') == '/foo-bar'\n",
"assert method_name_to_url('foo_bar') == '/foo-bar'\n",
"assert method_name_to_url('fooBar_baz') == '/foo-bar-baz'\n",
"assert method_name_to_url('_FooBar_baz-booFoo_') == '/foo-bar-baz-boo-foo'\n",
"def FUNC_25(self):...\n",
"assert _validate_redirect_url(None) is False\n",
"assert _validate_redirect_url(' ') is False\n",
"def FUNC_26(self, VAR_2, VAR_3):...\n",
"VAR_3.setattr('flask.request.host_url', 'http://example.com')\n",
"assert _validate_redirect_url('http://fail.com') is False\n",
"VAR_3.undo()\n",
"@pytest.mark.options(EXTERNAL_SERVER_NAME='works.com')...\n",
"VAR_3.setattr('flask.request.host_url', 'http://example.com')\n",
"assert _validate_redirect_url('http://works.com') is True\n",
"VAR_3.undo()\n",
"def FUNC_28(self, VAR_2, VAR_3):...\n",
"VAR_3.setattr('flask.request.host_url', 'http://example.com')\n",
"VAR_9 = _validate_redirect_url('http://works.com', _external_host='works.com')\n",
"assert VAR_9 is True\n",
"VAR_3.undo()\n"
] | [
"import pytest\n",
"from werkzeug.routing import BuildError\n",
"from flask_unchained.bundles.controller import Controller, Resource\n",
"from flask_unchained.bundles.controller.utils import controller_name, get_param_tuples, join, method_name_to_url, url_for, _validate_redirect_url\n",
"from py_meta_utils import deep_getattr\n",
"def test_deep_getattr():...\n",
"clsdict = {'a': 'clsdict'}\n",
"a = 'first'\n",
"b = 'first'\n",
"b = 'second'\n",
"c = 'second'\n",
"bases = First, Second\n",
"assert deep_getattr(clsdict, bases, 'a') == 'clsdict'\n",
"assert deep_getattr(clsdict, bases, 'b') == 'first'\n",
"assert deep_getattr(clsdict, bases, 'c') == 'second'\n",
"deep_getattr(clsdict, bases, 'd')\n",
"assert deep_getattr(clsdict, bases, 'a', 'default') == 'clsdict'\n",
"assert deep_getattr(clsdict, bases, 'b', 'default') == 'first'\n",
"assert deep_getattr(clsdict, bases, 'c', 'default') == 'second'\n",
"assert deep_getattr(clsdict, bases, 'd', 'default') == 'default'\n",
"def test_it_strips_controller(self):...\n",
"assert controller_name(UserController) == 'user'\n",
"def test_it_handles_acronyms(self):...\n",
"assert controller_name(APIController) == 'api'\n",
"def test_it_strips_view(self):...\n",
"assert controller_name(SomeView) == 'some'\n",
"def test_it_works_with_more_than_one_word(self):...\n",
"assert controller_name(MoreThanOneWordController) == 'more_than_one_word'\n",
"def test_it_strips_resource(self):...\n",
"assert controller_name(UserResource) == 'user'\n",
"def test_it_strips_method_view(self):...\n",
"assert controller_name(RoleMethodView) == 'role'\n",
"def test_it_only_strips_one_suffix(self):...\n",
"assert controller_name(RoleViewControllerResource) == 'role_view_controller'\n",
"def test_it_works_without_stripping_any_suffixes(self):...\n",
"assert controller_name(SomeCtrl) == 'some_ctrl'\n",
"def test_it_works(self):...\n",
"assert get_param_tuples('<int:id>') == [('int', 'id')]\n",
"def test_it_works_on_garbage(self):...\n",
"assert get_param_tuples(None) == []\n",
"def test_multiple(self):...\n",
"path = '/users/<int:user_id>/roles/<string:slug>'\n",
"assert get_param_tuples(path) == [('int', 'user_id'), ('string', 'slug')]\n",
"def test_it_works_with_already_formed_path(self):...\n",
"assert url_for('/foobar') == '/foobar'\n",
"def test_it_works_with_garbage(self):...\n",
"assert url_for(None) is None\n",
"def test_it_works_with_config_keys_returning_path(self, app):...\n",
"app.config.from_mapping({'MY_KEY': '/my-key'})\n",
"assert url_for('MY_KEY') == '/my-key'\n",
"def test_it_works_with_config_keys_returning_endpoints(self, app):...\n",
"app.config.from_mapping({'MY_KEY': 'some.endpoint'})\n",
"assert url_for('MY_KEY')\n",
"app.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n",
"assert url_for('MY_KEY') == '/some-endpoint'\n",
"def test_it_works_with_endpoints(self, app):...\n",
"assert url_for('some.endpoint')\n",
"app.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n",
"assert url_for('some.endpoint') == '/some-endpoint'\n",
"def test_it_works_with_controller_method_names(self, app):...\n",
"def about_us(self):...\n",
"app.add_url_rule('/about-us', endpoint='site_controller.about_us')\n",
"assert url_for('about_us', _cls=SiteController) == '/about-us'\n",
"def test_it_works_with_url_for_kwargs(self, app):...\n",
"def get(self, id):...\n",
"app.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n",
"assert url_for('get', id=1, _cls=SiteResource) == '/sites/1'\n",
"app.add_url_rule('/foo/<string:slug>', endpoint='some.endpoint')\n",
"assert url_for('some.endpoint', slug='hi') == '/foo/hi'\n",
"def test_it_falls_through_if_class_endpoint_not_found(self, app):...\n",
"def get(self, id):...\n",
"app.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n",
"url_for('delete', id=1, _cls=SiteResource)\n",
"def test_it_works_with_garbage(self):...\n",
"assert join(None) == '/'\n",
"assert join(None, None, '', 0) == '/'\n",
"def test_it_works_with_partially_valid_input(self):...\n",
"assert join('/', 'foo', None, 'bar', '', 'baz') == '/foo/bar/baz'\n",
"def test_it_strips_neighboring_slashes(self):...\n",
"assert join('/', '/foo', '/', '/bar') == '/foo/bar'\n",
"def test_it_doesnt_eat_single_slash(self):...\n",
"assert join('/', '/') == '/'\n",
"assert join(None, '/') == '/'\n",
"assert join('/', None) == '/'\n",
"def test_it_strips_trailing_slash(self):...\n",
"assert join('/foo/bar/') == '/foo/bar'\n",
"assert join('/foo/bar/', None) == '/foo/bar'\n",
"assert join('/foo/bar/', '/') == '/foo/bar'\n",
"assert join('/foo', 'bar/') == '/foo/bar'\n",
"def test_trailing_slash(self):...\n",
"assert join('/', trailing_slash=True) == '/'\n",
"assert join('/foo', 'baz', None, trailing_slash=True) == '/foo/baz/'\n",
"assert join('/foo', 'baz/', trailing_slash=True) == '/foo/baz/'\n",
"def test_it_works(self):...\n",
"assert method_name_to_url('fooBar') == '/foo-bar'\n",
"assert method_name_to_url('foo_bar') == '/foo-bar'\n",
"assert method_name_to_url('fooBar_baz') == '/foo-bar-baz'\n",
"assert method_name_to_url('_FooBar_baz-booFoo_') == '/foo-bar-baz-boo-foo'\n",
"def test_it_fails_on_garbage(self):...\n",
"assert _validate_redirect_url(None) is False\n",
"assert _validate_redirect_url(' ') is False\n",
"def test_it_fails_with_invalid_netloc(self, app, monkeypatch):...\n",
"monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n",
"assert _validate_redirect_url('http://fail.com') is False\n",
"monkeypatch.undo()\n",
"@pytest.mark.options(EXTERNAL_SERVER_NAME='works.com')...\n",
"monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n",
"assert _validate_redirect_url('http://works.com') is True\n",
"monkeypatch.undo()\n",
"def test_it_works_with_explicit_external_host(self, app, monkeypatch):...\n",
"monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n",
"result = _validate_redirect_url('http://works.com', _external_host='works.com')\n",
"assert result is True\n",
"monkeypatch.undo()\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
4,
0,
0
] | [
"Import'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assert'",
"Assert'",
"Assert'",
"Expr'",
"Assert'",
"Assert'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assign'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Expr'",
"Assert'",
"FunctionDef'",
"Expr'",
"Assert'",
"Expr'",
"Assert'",
"FunctionDef'",
"Assert'",
"Expr'",
"Assert'",
"FunctionDef'",
"FunctionDef'",
"Expr'",
"Assert'",
"FunctionDef'",
"FunctionDef'",
"Expr'",
"Assert'",
"Expr'",
"Assert'",
"FunctionDef'",
"FunctionDef'",
"Expr'",
"Expr'",
"FunctionDef'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"FunctionDef'",
"Assert'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"Assert'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"Assert'",
"Assert'",
"Assert'",
"FunctionDef'",
"Assert'",
"Assert'",
"FunctionDef'",
"Expr'",
"Assert'",
"Expr'",
"Condition",
"Expr'",
"Assert'",
"Expr'",
"FunctionDef'",
"Expr'",
"Assign'",
"Assert'",
"Expr'"
] |
[
"def FUNC_87(VAR_160, VAR_161, VAR_162):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_49[VAR_160].append(_('user \"{user}\" in row# {row}').format(VAR_19=_user,\n row=row_count))\n"
] | [
"def build_row_errors(key, _user, row_count):...\n",
"\"\"\"docstring\"\"\"\n",
"row_errors[key].append(_('user \"{user}\" in row# {row}').format(user=_user,\n row=row_count))\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'"
] |
[
"def FUNC_30(self, VAR_28, VAR_29, VAR_12):...\n",
"if VAR_29 == 'public':\n",
"VAR_103 = type(VAR_12) == bool and type(VAR_28) == int\n",
"if VAR_103:\n",
"return self.playlistdb.setPublic(VAR_15=self.getUserId(), VAR_28=plid,\n VAR_24=value)\n"
] | [
"def api_changeplaylist(self, plid, attribute, value):...\n",
"if attribute == 'public':\n",
"is_valid = type(value) == bool and type(plid) == int\n",
"if is_valid:\n",
"return self.playlistdb.setPublic(userid=self.getUserId(), plid=plid, public\n =value)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Condition",
"Return'"
] |
[
"from __future__ import annotations\n",
"from dataclasses import dataclass, field\n",
"from datetime import date, datetime\n",
"from typing import Any, Dict, List, Optional, Union, cast\n",
"from .an_enum import AnEnum\n",
"from .different_enum import DifferentEnum\n",
"\"\"\" A Model for testing all the ways custom objects can be used \"\"\"\n",
"VAR_1: AnEnum\n",
"VAR_5: Union[datetime, date]\n",
"VAR_2: date\n",
"VAR_6: Optional[List[List[DifferentEnum]]] = field(default_factory=lambda :\n cast(Optional[List[List[DifferentEnum]]], []))\n",
"VAR_3: Optional[Dict[Any, Any]] = field(default_factory=lambda : cast(\n Optional[Dict[Any, Any]], {}))\n",
"def FUNC_0(self) ->Dict[str, Any]:...\n",
"VAR_1 = self.an_enum_value.value\n",
"if isinstance(self.a_camel_date_time, datetime):\n",
"VAR_5 = self.a_camel_date_time.isoformat()\n",
"VAR_5 = self.a_camel_date_time.isoformat()\n",
"VAR_2 = self.a_date.isoformat()\n",
"if self.nested_list_of_enums is None:\n",
"VAR_6 = None\n",
"VAR_6 = []\n",
"VAR_3 = self.some_dict\n",
"for nested_list_of_enums_item_data in self.nested_list_of_enums:\n",
"return {'an_enum_value': VAR_1, 'aCamelDateTime': VAR_5, 'a_date': VAR_2,\n 'nested_list_of_enums': VAR_6, 'some_dict': VAR_3}\n",
"VAR_7 = []\n",
"for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n",
"VAR_8 = nested_list_of_enums_item_item_data.value\n",
"VAR_6.append(VAR_7)\n",
"VAR_7.append(VAR_8)\n"
] | [
"from __future__ import annotations\n",
"from dataclasses import dataclass, field\n",
"from datetime import date, datetime\n",
"from typing import Any, Dict, List, Optional, Union, cast\n",
"from .an_enum import AnEnum\n",
"from .different_enum import DifferentEnum\n",
"\"\"\" A Model for testing all the ways custom objects can be used \"\"\"\n",
"an_enum_value: AnEnum\n",
"a_camel_date_time: Union[datetime, date]\n",
"a_date: date\n",
"nested_list_of_enums: Optional[List[List[DifferentEnum]]] = field(\n default_factory=lambda : cast(Optional[List[List[DifferentEnum]]], []))\n",
"some_dict: Optional[Dict[Any, Any]] = field(default_factory=lambda : cast(\n Optional[Dict[Any, Any]], {}))\n",
"def to_dict(self) ->Dict[str, Any]:...\n",
"an_enum_value = self.an_enum_value.value\n",
"if isinstance(self.a_camel_date_time, datetime):\n",
"a_camel_date_time = self.a_camel_date_time.isoformat()\n",
"a_camel_date_time = self.a_camel_date_time.isoformat()\n",
"a_date = self.a_date.isoformat()\n",
"if self.nested_list_of_enums is None:\n",
"nested_list_of_enums = None\n",
"nested_list_of_enums = []\n",
"some_dict = self.some_dict\n",
"for nested_list_of_enums_item_data in self.nested_list_of_enums:\n",
"return {'an_enum_value': an_enum_value, 'aCamelDateTime': a_camel_date_time,\n 'a_date': a_date, 'nested_list_of_enums': nested_list_of_enums,\n 'some_dict': some_dict}\n",
"nested_list_of_enums_item = []\n",
"for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n",
"nested_list_of_enums_item_item = nested_list_of_enums_item_item_data.value\n",
"nested_list_of_enums.append(nested_list_of_enums_item)\n",
"nested_list_of_enums_item.append(nested_list_of_enums_item_item)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
5,
5,
0,
5,
0,
0,
5,
0,
0,
0,
0,
0,
0,
5,
0,
5,
0,
0,
0,
0,
0
] | [
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"ImportFrom'",
"Expr'",
"AnnAssign'",
"AnnAssign'",
"AnnAssign'",
"AnnAssign'",
"AnnAssign'",
"FunctionDef'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"For",
"Return'",
"Assign'",
"For",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_20(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return u'[%s]' % self.ticket_for_url\n"
] | [
"def _get_ticket(self):...\n",
"\"\"\"docstring\"\"\"\n",
"return u'[%s]' % self.ticket_for_url\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Return'"
] |
[
"def FUNC_77(self):...\n",
"VAR_78 = frappe.get_all('ToDo', fields=['owner'], filters={'reference_type':\n self.doctype, 'reference_name': self.name, 'status': ('!=', 'Cancelled')})\n",
"VAR_79 = set([assignment.owner for assignment in VAR_78])\n",
"return VAR_79\n"
] | [
"def get_assigned_users(self):...\n",
"assignments = frappe.get_all('ToDo', fields=['owner'], filters={\n 'reference_type': self.doctype, 'reference_name': self.name, 'status':\n ('!=', 'Cancelled')})\n",
"users = set([assignment.owner for assignment in assignments])\n",
"return users\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_14():...\n",
"VAR_23.component_instance.json_args = VAR_34\n",
"VAR_23.component_instance.special_args.extend(VAR_21)\n"
] | [
"def marshall_element_args():...\n",
"element.component_instance.json_args = serialized_json_args\n",
"element.component_instance.special_args.extend(special_args)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_23(self, VAR_46: str, VAR_47: str) ->bool:...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.instances or len(self.instances) == 1:\n",
"return True\n",
"return self.get_instance(VAR_47) == VAR_46\n"
] | [
"def should_handle(self, instance_name: str, key: str) ->bool:...\n",
"\"\"\"docstring\"\"\"\n",
"if not self.instances or len(self.instances) == 1:\n",
"return True\n",
"return self.get_instance(key) == instance_name\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Return'"
] |
[
"async def FUNC_19(self, VAR_5: str, VAR_11: str, VAR_27: str) ->EventBase:...\n",
"\"\"\"docstring\"\"\"\n",
"if get_domain_from_id(VAR_27) != VAR_5:\n",
"VAR_0.info(\n 'Got /make_leave request for user %r from different origin %s, ignoring',\n VAR_27, VAR_5)\n",
"VAR_28 = await self.store.get_room_version_id(VAR_11)\n",
"VAR_93 = self.event_builder_factory.new(VAR_28, {'type': VAR_188.Member,\n 'content': {'membership': Membership.LEAVE}, 'room_id': VAR_11,\n 'sender': VAR_27, 'state_key': VAR_27})\n",
"VAR_1, VAR_32 = await self.event_creation_handler.create_new_client_event(\n VAR_93=builder)\n",
"await self.auth.check_from_context(VAR_28, VAR_1, VAR_32, do_sig_check=False)\n",
"VAR_0.warning('Failed to create new leave %r because %s', VAR_1, VAR_170)\n",
"return VAR_1\n"
] | [
"async def on_make_leave_request(self, origin: str, room_id: str, user_id: str...\n",
"\"\"\"docstring\"\"\"\n",
"if get_domain_from_id(user_id) != origin:\n",
"logger.info(\n 'Got /make_leave request for user %r from different origin %s, ignoring',\n user_id, origin)\n",
"room_version = await self.store.get_room_version_id(room_id)\n",
"builder = self.event_builder_factory.new(room_version, {'type': EventTypes.\n Member, 'content': {'membership': Membership.LEAVE}, 'room_id': room_id,\n 'sender': user_id, 'state_key': user_id})\n",
"event, context = await self.event_creation_handler.create_new_client_event(\n builder=builder)\n",
"await self.auth.check_from_context(room_version, event, context,\n do_sig_check=False)\n",
"logger.warning('Failed to create new leave %r because %s', event, e)\n",
"return event\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"AsyncFunctionDef'",
"Docstring",
"Condition",
"Expr'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_3(self):...\n",
"\"\"\"Model with callable polymorphic functions specified.\"\"\"\n",
"@def_function.function...\n",
"if VAR_56:\n",
"return VAR_54 + VAR_55\n",
"return VAR_54 * VAR_55\n"
] | [
"def testShowAllWithFunctions(self):...\n",
"\"\"\"Model with callable polymorphic functions specified.\"\"\"\n",
"@def_function.function...\n",
"if c:\n",
"return a + b\n",
"return a * b\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Condition",
"Return'",
"Return'"
] |
[
"import time\n",
"import urllib2\n",
"from flask import Flask, render_template, session, request, json\n",
"from core.victim_objects import *\n",
"import core.stats\n",
"from core.utils import utils\n",
"from core.db import Database\n",
"VAR_0 = core.stats.trape\n",
"VAR_1 = core.stats.app\n",
"VAR_2 = Database()\n",
"@VAR_1.route('/' + VAR_0.victim_path)...\n",
"VAR_3 = urllib2.build_opener()\n",
"VAR_4 = victim_headers()\n",
"VAR_3.addheaders = VAR_4\n",
"VAR_5 = victim_inject_code(VAR_3.open(VAR_0.url_to_clone).read(), 'lure')\n",
"return VAR_5\n"
] | [
"import time\n",
"import urllib2\n",
"from flask import Flask, render_template, session, request, json\n",
"from core.victim_objects import *\n",
"import core.stats\n",
"from core.utils import utils\n",
"from core.db import Database\n",
"trape = core.stats.trape\n",
"app = core.stats.app\n",
"db = Database()\n",
"@app.route('/' + trape.victim_path)...\n",
"opener = urllib2.build_opener()\n",
"headers = victim_headers()\n",
"opener.addheaders = headers\n",
"html = victim_inject_code(opener.open(trape.url_to_clone).read(), 'lure')\n",
"return html\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Import'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Import'",
"ImportFrom'",
"ImportFrom'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_53 = {}\n",
"VAR_93 = getIntOrDefault(VAR_2, 'index', 0)\n",
"VAR_104 = BaseContainer(VAR_5, **{str(c_type): long(c_id)})\n",
"if VAR_25:\n",
"VAR_53['share'] = BaseShare(VAR_5, VAR_25)\n",
"if VAR_23 == 'well':\n",
"VAR_104.image = VAR_104.well.getImage(VAR_93)\n",
"VAR_118 = VAR_104.image.getAllRenderingDefs()\n",
"VAR_119 = {}\n",
"VAR_120 = VAR_104.image.getRenderingDefId()\n",
"for VAR_71 in VAR_118:\n",
"VAR_229 = VAR_71['owner']['id']\n",
"VAR_119 = VAR_119.values()\n",
"VAR_71['current'] = VAR_71['id'] == VAR_120\n",
"VAR_121 = []\n",
"if VAR_229 not in VAR_119 or VAR_119[VAR_229]['id'] < VAR_71['id']:\n",
"for VAR_71 in VAR_119:\n",
"VAR_119[VAR_229] = VAR_71\n",
"VAR_282 = []\n",
"VAR_122, VAR_123 = VAR_5.getMaxPlaneSize()\n",
"for VAR_318, VAR_383 in enumerate(VAR_71['c']):\n",
"VAR_124 = VAR_104.image.getSizeX()\n",
"VAR_348 = '-'\n",
"VAR_121.append({'id': VAR_71['id'], 'owner': VAR_71['owner'], 'c': ','.join\n (VAR_282), 'm': VAR_71['model'] == 'greyscale' and 'g' or 'c'})\n",
"VAR_125 = VAR_104.image.getSizeY()\n",
"if VAR_383['active']:\n",
"VAR_53['tiledImage'] = VAR_124 * VAR_125 > VAR_122 * VAR_123\n",
"VAR_348 = ''\n",
"VAR_349 = VAR_383['lut'] if 'lut' in VAR_383 else VAR_383['color']\n",
"VAR_53['manager'] = VAR_104\n",
"VAR_350 = 'r' if VAR_383['inverted'] else '-r'\n",
"VAR_53['rdefsJson'] = json.dumps(VAR_121)\n",
"VAR_282.append('%s%s|%s:%s%s$%s' % (VAR_348, VAR_318 + 1, VAR_383['start'],\n VAR_383['end'], VAR_350, VAR_349))\n",
"VAR_53['rdefs'] = VAR_119\n",
"VAR_53['template'] = 'webclient/annotations/metadata_preview.html'\n",
"return VAR_53\n"
] | [
"@login_required()...\n",
"\"\"\"docstring\"\"\"\n",
"context = {}\n",
"index = getIntOrDefault(request, 'index', 0)\n",
"manager = BaseContainer(conn, **{str(c_type): long(c_id)})\n",
"if share_id:\n",
"context['share'] = BaseShare(conn, share_id)\n",
"if c_type == 'well':\n",
"manager.image = manager.well.getImage(index)\n",
"allRdefs = manager.image.getAllRenderingDefs()\n",
"rdefs = {}\n",
"rdefId = manager.image.getRenderingDefId()\n",
"for r in allRdefs:\n",
"ownerId = r['owner']['id']\n",
"rdefs = rdefs.values()\n",
"r['current'] = r['id'] == rdefId\n",
"rdefQueries = []\n",
"if ownerId not in rdefs or rdefs[ownerId]['id'] < r['id']:\n",
"for r in rdefs:\n",
"rdefs[ownerId] = r\n",
"chs = []\n",
"max_w, max_h = conn.getMaxPlaneSize()\n",
"for i, c in enumerate(r['c']):\n",
"size_x = manager.image.getSizeX()\n",
"act = '-'\n",
"rdefQueries.append({'id': r['id'], 'owner': r['owner'], 'c': ','.join(chs),\n 'm': r['model'] == 'greyscale' and 'g' or 'c'})\n",
"size_y = manager.image.getSizeY()\n",
"if c['active']:\n",
"context['tiledImage'] = size_x * size_y > max_w * max_h\n",
"act = ''\n",
"color = c['lut'] if 'lut' in c else c['color']\n",
"context['manager'] = manager\n",
"reverse = 'r' if c['inverted'] else '-r'\n",
"context['rdefsJson'] = json.dumps(rdefQueries)\n",
"chs.append('%s%s|%s:%s%s$%s' % (act, i + 1, c['start'], c['end'], reverse,\n color))\n",
"context['rdefs'] = rdefs\n",
"context['template'] = 'webclient/annotations/metadata_preview.html'\n",
"return context\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"For",
"Assign'",
"Assign'",
"Assign'",
"For",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"@app.route('/tags')...\n",
"if not app.config['SEARCH_CONF']['engine'] == 'ripgrep' and not which('rg'):\n",
"flash('Ripgrep must be installed to view pages about embedded tags.', 'error')\n",
"VAR_10 = sorted(get_all_tags(force=True))\n",
"return redirect('/')\n",
"return render_template('tags/all.html', title='All Tags', VAR_10=tags)\n"
] | [
"@app.route('/tags')...\n",
"if not app.config['SEARCH_CONF']['engine'] == 'ripgrep' and not which('rg'):\n",
"flash('Ripgrep must be installed to view pages about embedded tags.', 'error')\n",
"tags = sorted(get_all_tags(force=True))\n",
"return redirect('/')\n",
"return render_template('tags/all.html', title='All Tags', tags=tags)\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Expr'",
"Assign'",
"Return'",
"Return'"
] |
[
"def FUNC_9():...\n",
"if session.pam_user:\n",
"session.flash = T('PAM authenticated user, cannot change password here')\n",
"VAR_26 = SQLFORM.factory(Field('current_admin_password', 'password'), Field\n ('new_admin_password', 'password', requires=IS_STRONG()), Field(\n 'new_admin_password_again', 'password'), VAR_144='span4 well')\n",
"redirect(URL('site'))\n",
"if VAR_26.accepts(request.vars):\n",
"if not verify_password(request.vars.current_admin_password):\n",
"return dict(VAR_26=form)\n",
"VAR_26.errors.current_admin_password = T('invalid password')\n",
"if VAR_26.vars.new_admin_password != VAR_26.vars.new_admin_password_again:\n",
"VAR_26.errors.new_admin_password_again = T('no match')\n",
"VAR_15 = abspath('parameters_%s.py' % request.env.server_port)\n",
"FUNC_4(VAR_15, 'password=\"%s\"' % CRYPT()(request.vars.new_admin_password)[0])\n",
"session.flash = T('password changed')\n",
"redirect(URL('site'))\n"
] | [
"def change_password():...\n",
"if session.pam_user:\n",
"session.flash = T('PAM authenticated user, cannot change password here')\n",
"form = SQLFORM.factory(Field('current_admin_password', 'password'), Field(\n 'new_admin_password', 'password', requires=IS_STRONG()), Field(\n 'new_admin_password_again', 'password'), _class='span4 well')\n",
"redirect(URL('site'))\n",
"if form.accepts(request.vars):\n",
"if not verify_password(request.vars.current_admin_password):\n",
"return dict(form=form)\n",
"form.errors.current_admin_password = T('invalid password')\n",
"if form.vars.new_admin_password != form.vars.new_admin_password_again:\n",
"form.errors.new_admin_password_again = T('no match')\n",
"path = abspath('parameters_%s.py' % request.env.server_port)\n",
"safe_write(path, 'password=\"%s\"' % CRYPT()(request.vars.new_admin_password)[0])\n",
"session.flash = T('password changed')\n",
"redirect(URL('site'))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Condition",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Expr'"
] |
[
"def FUNC_13():...\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if VAR_1:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist count=\"{:d}\">\\n'.format(VAR_39)\n",
"VAR_40 = 0\n",
"VAR_41 = 0\n",
"if VAR_26:\n",
"VAR_40 = int(VAR_25 * VAR_5)\n",
"VAR_42 = []\n",
"VAR_44 = cycle(VAR_17)\n",
"def FUNC_15():...\n",
"random.seed(VAR_24)\n",
"return list(map(lambda x: x + VAR_5 * VAR_41, sorted(random.sample(list(\n range(VAR_5)), VAR_40))))\n"
] | [
"def generate():...\n",
"yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n",
"if STYLE:\n",
"yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n",
"yield '<objectlist count=\"{:d}\">\\n'.format(total_entries)\n",
"mix_per_iteration = 0\n",
"iteration_count = 0\n",
"if make_cocktail:\n",
"mix_per_iteration = int(percentage * ITEMS_PER_ITERATION)\n",
"mix_indices = []\n",
"pool = cycle(mixer_list)\n",
"def generate_mix_indices():...\n",
"random.seed(seed)\n",
"return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"def FUNC_76(VAR_212):...\n",
"if VAR_178(VAR_212.getId()) == VAR_178(VAR_6):\n",
"return False\n",
"VAR_285 = VAR_212.getPrimaryPixels()\n",
"if VAR_285 is None or VAR_212.getPrimaryPixels().getPixelsType().getValue(\n",
"return False\n",
"VAR_286 = [VAR_30.getLabel() for VAR_30 in VAR_212.getChannels()]\n",
"VAR_286.sort()\n",
"if VAR_286 != VAR_220:\n",
"return False\n",
"return True\n"
] | [
"def compat(i):...\n",
"if long(i.getId()) == long(iid):\n",
"return False\n",
"pp = i.getPrimaryPixels()\n",
"if pp is None or i.getPrimaryPixels().getPixelsType().getValue(\n",
"return False\n",
"ew = [x.getLabel() for x in i.getChannels()]\n",
"ew.sort()\n",
"if ew != img_ew:\n",
"return False\n",
"return True\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Assign'",
"Condition",
"Return'",
"Assign'",
"Expr'",
"Condition",
"Return'",
"Return'"
] |
[
"def __init__(self, *VAR_0, **VAR_1):...\n",
"\"\"\"docstring\"\"\"\n",
"self.doctype = self.name = None\n",
"self._default_new_docs = {}\n",
"self.flags = frappe._dict()\n",
"if VAR_0 and VAR_0[0] and isinstance(VAR_0[0], string_types):\n",
"if len(VAR_0) == 1:\n",
"if VAR_0 and VAR_0[0] and isinstance(VAR_0[0], dict):\n",
"self.doctype = self.name = VAR_0[0]\n",
"self.doctype = VAR_0[0]\n",
"VAR_1 = VAR_0[0]\n",
"if VAR_1:\n",
"self.load_from_db()\n",
"if isinstance(VAR_0[1], dict):\n",
"super(CLASS_0, self).__init__(VAR_1)\n",
"return\n",
"self.name = frappe.db.get_value(VAR_0[0], VAR_0[1], 'name')\n",
"self.name = VAR_0[1]\n",
"self.init_valid_columns()\n",
"if self.name is None:\n",
"if 'for_update' in VAR_1:\n",
"frappe.throw(_('{0} {1} not found').format(_(VAR_0[0]), VAR_0[1]), frappe.\n DoesNotExistError)\n",
"self.flags.for_update = VAR_1.get('for_update')\n"
] | [
"def __init__(self, *args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"self.doctype = self.name = None\n",
"self._default_new_docs = {}\n",
"self.flags = frappe._dict()\n",
"if args and args[0] and isinstance(args[0], string_types):\n",
"if len(args) == 1:\n",
"if args and args[0] and isinstance(args[0], dict):\n",
"self.doctype = self.name = args[0]\n",
"self.doctype = args[0]\n",
"kwargs = args[0]\n",
"if kwargs:\n",
"self.load_from_db()\n",
"if isinstance(args[1], dict):\n",
"super(Document, self).__init__(kwargs)\n",
"return\n",
"self.name = frappe.db.get_value(args[0], args[1], 'name')\n",
"self.name = args[1]\n",
"self.init_valid_columns()\n",
"if self.name is None:\n",
"if 'for_update' in kwargs:\n",
"frappe.throw(_('{0} {1} not found').format(_(args[0]), args[1]), frappe.\n DoesNotExistError)\n",
"self.flags.for_update = kwargs.get('for_update')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Return'",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"For",
"Expr'",
"Assign'"
] |
[
"def __init__(self, VAR_7, VAR_8):...\n",
"self.errorstring = VAR_7\n",
"self.error = VAR_8\n",
"super().__init__(VAR_7)\n"
] | [
"def __init__(self, errorstring, error):...\n",
"self.errorstring = errorstring\n",
"self.error = error\n",
"super().__init__(errorstring)\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"def FUNC_11(self):...\n",
""
] | [
"def after_script_executed(self):...\n",
""
] | [
0,
0
] | [
"FunctionDef'",
"Condition"
] |
[
"@VAR_2.route('/advsearch', methods=['GET'])...\n",
"VAR_16 = get_cc_columns(filter_config_custom_read=True)\n",
"return FUNC_29(VAR_16)\n"
] | [
"@web.route('/advsearch', methods=['GET'])...\n",
"cc = get_cc_columns(filter_config_custom_read=True)\n",
"return render_prepare_search_form(cc)\n"
] | [
0,
0,
0
] | [
"Condition",
"Assign'",
"Return'"
] |
[
"@expose('/userinfo/')...\n",
"VAR_21 = self.datamodel.get(g.user.id, self._base_filters)\n",
"VAR_43 = self._get_show_widget(g.user.id, VAR_21, VAR_15=self.\n user_show_fieldsets)\n",
"self.update_redirect()\n",
"return self.render_template(self.show_template, VAR_37=self.user_info_title,\n VAR_43=widgets, appbuilder=self.appbuilder)\n"
] | [
"@expose('/userinfo/')...\n",
"item = self.datamodel.get(g.user.id, self._base_filters)\n",
"widgets = self._get_show_widget(g.user.id, item, show_fieldsets=self.\n user_show_fieldsets)\n",
"self.update_redirect()\n",
"return self.render_template(self.show_template, title=self.user_info_title,\n widgets=widgets, appbuilder=self.appbuilder)\n"
] | [
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Return'"
] |
[
"def FUNC_2(self, VAR_3, VAR_4):...\n",
"\"\"\"docstring\"\"\"\n",
"for basket_command_middleware in get_provide_objects(\n",
"if not issubclass(basket_command_middleware, BaseBasketCommandMiddleware):\n",
"return VAR_4\n",
"VAR_4 = dict(basket_command_middleware().preprocess_kwargs(VAR_2=self.\n basket, VAR_1=self.request, VAR_3=command, VAR_4=kwargs))\n"
] | [
"def preprocess_kwargs(self, command, kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"for basket_command_middleware in get_provide_objects(\n",
"if not issubclass(basket_command_middleware, BaseBasketCommandMiddleware):\n",
"return kwargs\n",
"kwargs = dict(basket_command_middleware().preprocess_kwargs(basket=self.\n basket, request=self.request, command=command, kwargs=kwargs))\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"For",
"Condition",
"Return'",
"Assign'"
] |
[
"def FUNC_14(self, VAR_6):...\n",
"\"\"\"docstring\"\"\"\n",
"self.helper.send_state(self.room_id, 'm.room.canonical_alias', VAR_6, tok=\n self.admin_user_tok)\n"
] | [
"def _set_canonical_alias(self, content):...\n",
"\"\"\"docstring\"\"\"\n",
"self.helper.send_state(self.room_id, 'm.room.canonical_alias', content, tok\n =self.admin_user_tok)\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Expr'"
] |
[
"@property...\n",
"return self._base_builder.room_id\n"
] | [
"@property...\n",
"return self._base_builder.room_id\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@VAR_2.route('/ajax/toggleread/<int:book_id>', methods=['POST'])...\n",
"if not config.config_read_column:\n",
"VAR_95 = ub.session.query(ub.ReadBook).filter(and_(ub.ReadBook.user_id ==\n int(VAR_87.id), ub.ReadBook.book_id == VAR_5)).first()\n",
"calibre_db.update_title_sort(config)\n",
"VAR_3.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n",
"return ''\n",
"if VAR_95:\n",
"VAR_95 = calibre_db.get_filtered_book(VAR_5)\n",
"return 'Custom Column No.{} is not existing in calibre database'.format(config\n .config_read_column), 400\n",
"if VAR_95.read_status == ub.ReadBook.STATUS_FINISHED:\n",
"VAR_123 = ub.ReadBook(VAR_11=current_user.id, VAR_5=book_id)\n",
"VAR_26 = getattr(VAR_95, 'custom_column_' + str(config.config_read_column))\n",
"VAR_95.read_status = ub.ReadBook.STATUS_UNREAD\n",
"VAR_95.read_status = ub.ReadBook.STATUS_FINISHED\n",
"VAR_123.read_status = ub.ReadBook.STATUS_FINISHED\n",
"if len(VAR_26):\n",
"if not VAR_95.kobo_reading_state:\n",
"VAR_95 = VAR_123\n",
"VAR_26[0].value = not VAR_26[0].value\n",
"VAR_143 = db.cc_classes[config.config_read_column]\n",
"VAR_124 = ub.KoboReadingState(VAR_11=current_user.id, VAR_5=book_id)\n",
"ub.session.merge(VAR_95)\n",
"calibre_db.session.commit()\n",
"VAR_144 = VAR_143(value=1, VAR_95=book_id)\n",
"VAR_124.current_bookmark = ub.KoboBookmark()\n",
"ub.session_commit('Book {} readbit toggled'.format(VAR_5))\n",
"calibre_db.session.add(VAR_144)\n",
"VAR_124.statistics = ub.KoboStatistics()\n",
"calibre_db.session.commit()\n",
"VAR_95.kobo_reading_state = VAR_124\n"
] | [
"@web.route('/ajax/toggleread/<int:book_id>', methods=['POST'])...\n",
"if not config.config_read_column:\n",
"book = ub.session.query(ub.ReadBook).filter(and_(ub.ReadBook.user_id == int\n (current_user.id), ub.ReadBook.book_id == book_id)).first()\n",
"calibre_db.update_title_sort(config)\n",
"log.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n",
"return ''\n",
"if book:\n",
"book = calibre_db.get_filtered_book(book_id)\n",
"return 'Custom Column No.{} is not existing in calibre database'.format(config\n .config_read_column), 400\n",
"if book.read_status == ub.ReadBook.STATUS_FINISHED:\n",
"readBook = ub.ReadBook(user_id=current_user.id, book_id=book_id)\n",
"read_status = getattr(book, 'custom_column_' + str(config.config_read_column))\n",
"book.read_status = ub.ReadBook.STATUS_UNREAD\n",
"book.read_status = ub.ReadBook.STATUS_FINISHED\n",
"readBook.read_status = ub.ReadBook.STATUS_FINISHED\n",
"if len(read_status):\n",
"if not book.kobo_reading_state:\n",
"book = readBook\n",
"read_status[0].value = not read_status[0].value\n",
"cc_class = db.cc_classes[config.config_read_column]\n",
"kobo_reading_state = ub.KoboReadingState(user_id=current_user.id, book_id=\n book_id)\n",
"ub.session.merge(book)\n",
"calibre_db.session.commit()\n",
"new_cc = cc_class(value=1, book=book_id)\n",
"kobo_reading_state.current_bookmark = ub.KoboBookmark()\n",
"ub.session_commit('Book {} readbit toggled'.format(book_id))\n",
"calibre_db.session.add(new_cc)\n",
"kobo_reading_state.statistics = ub.KoboStatistics()\n",
"calibre_db.session.commit()\n",
"book.kobo_reading_state = kobo_reading_state\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assign'",
"Expr'",
"Expr'",
"Return'",
"Condition",
"Assign'",
"Return'",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Assign'",
"Expr'",
"Assign'"
] |
[
"@FUNC_0...\n",
"return MediaRepository(self)\n"
] | [
"@cache_in_self...\n",
"return MediaRepository(self)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_9(VAR_22):...\n",
"\"\"\"docstring\"\"\"\n",
"if len(VAR_22) > 13:\n",
"VAR_30 = int(VAR_22, 36)\n",
"if not six.PY3 and VAR_30 > sys.maxint:\n",
"return VAR_30\n"
] | [
"def base36_to_int(s):...\n",
"\"\"\"docstring\"\"\"\n",
"if len(s) > 13:\n",
"value = int(s, 36)\n",
"if not six.PY3 and value > sys.maxint:\n",
"return value\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Assign'",
"Condition",
"Return'"
] |
[
"@VAR_25.whitelist(allow_guest=True)...\n",
"VAR_25.local.login_manager.logout()\n",
"VAR_25.db.commit()\n",
"VAR_25.respond_as_web_page(_('Logged Out'), _(\n 'You have been successfully logged out'), indicator_color='green')\n"
] | [
"@frappe.whitelist(allow_guest=True)...\n",
"frappe.local.login_manager.logout()\n",
"frappe.db.commit()\n",
"frappe.respond_as_web_page(_('Logged Out'), _(\n 'You have been successfully logged out'), indicator_color='green')\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_2(VAR_2, VAR_3, VAR_4):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_24 = FUNC_1(VAR_2)\n",
"for port in (VAR_3, VAR_4):\n",
"VAR_37 = socket.socket(VAR_24, socket.SOCK_STREAM)\n",
"return True\n",
"VAR_37.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n",
"VAR_32 = 'Unable to bind socket {0}:{1}'.format(VAR_2, port)\n",
"VAR_37.close()\n",
"VAR_37.bind((VAR_2, int(port)))\n",
"if exc.args:\n",
"VAR_32 = '{0}, error: {1}'.format(VAR_32, str(exc))\n",
"VAR_32 = '{0}, this might not be a problem.'.format(VAR_32)\n",
"VAR_32 += '; Is there another salt-master running?'\n",
"if is_console_configured():\n",
"VAR_0.warning(VAR_32)\n",
"sys.stderr.write('WARNING: {0}\\n'.format(VAR_32))\n",
"return False\n"
] | [
"def verify_socket(interface, pub_port, ret_port):...\n",
"\"\"\"docstring\"\"\"\n",
"addr_family = lookup_family(interface)\n",
"for port in (pub_port, ret_port):\n",
"sock = socket.socket(addr_family, socket.SOCK_STREAM)\n",
"return True\n",
"sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n",
"msg = 'Unable to bind socket {0}:{1}'.format(interface, port)\n",
"sock.close()\n",
"sock.bind((interface, int(port)))\n",
"if exc.args:\n",
"msg = '{0}, error: {1}'.format(msg, str(exc))\n",
"msg = '{0}, this might not be a problem.'.format(msg)\n",
"msg += '; Is there another salt-master running?'\n",
"if is_console_configured():\n",
"log.warning(msg)\n",
"sys.stderr.write('WARNING: {0}\\n'.format(msg))\n",
"return False\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Assign'",
"Return'",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"AugAssign'",
"Condition",
"Expr'",
"Expr'",
"Return'"
] |
[
"def __init__(self, VAR_8):...\n",
"self.store = VAR_8.get_datastore()\n",
"self.config = VAR_8.get_config()\n"
] | [
"def __init__(self, hs):...\n",
"self.store = hs.get_datastore()\n",
"self.config = hs.get_config()\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'"
] |
[
"@FUNC_2.__func__...\n",
"\"\"\"docstring\"\"\"\n",
"self._cancel()\n"
] | [
"@whitelist.__func__...\n",
"\"\"\"docstring\"\"\"\n",
"self._cancel()\n"
] | [
0,
0,
0
] | [
"Condition",
"Docstring",
"Expr'"
] |
[
"@property...\n",
"return self._base_builder.type\n"
] | [
"@property...\n",
"return self._base_builder.type\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"def FUNC_34(self):...\n",
"return self.db[self.settings.table_cas_name]\n"
] | [
"def table_cas(self):...\n",
"return self.db[self.settings.table_cas_name]\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
[
"@login_required(setGroupContext=True)...\n",
"\"\"\"docstring\"\"\"\n",
"if VAR_35 is None or VAR_36 is None or VAR_37 is None:\n",
"VAR_54 = 'webclient/history/history_details.html'\n",
"VAR_87 = int(VAR_2.GET.get('page', 1))\n",
"VAR_192 = VAR_2.session.get('user_id')\n",
"VAR_193 = BaseCalendar(VAR_5=conn, VAR_35=year, VAR_36=month, VAR_37=day,\n eid=filter_user_id)\n",
"VAR_193.get_items(VAR_87)\n",
"VAR_53 = {'controller': VAR_193}\n",
"VAR_53['template'] = VAR_54\n",
"return VAR_53\n"
] | [
"@login_required(setGroupContext=True)...\n",
"\"\"\"docstring\"\"\"\n",
"if year is None or month is None or day is None:\n",
"template = 'webclient/history/history_details.html'\n",
"page = int(request.GET.get('page', 1))\n",
"filter_user_id = request.session.get('user_id')\n",
"controller = BaseCalendar(conn=conn, year=year, month=month, day=day, eid=\n filter_user_id)\n",
"controller.get_items(page)\n",
"context = {'controller': controller}\n",
"context['template'] = template\n",
"return context\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Condition",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Expr'",
"Assign'",
"Assign'",
"Return'"
] |
[
"def __init__(self):...\n",
"self.var = variables.Variable(1.0, name='my_var')\n",
"self.write_var = variables.Variable(1.0, name='write_var')\n"
] | [
"def __init__(self):...\n",
"self.var = variables.Variable(1.0, name='my_var')\n",
"self.write_var = variables.Variable(1.0, name='write_var')\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'"
] |
[
"def FUNC_16(VAR_19):...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_41 = example_pb2.Example()\n",
"for feature_name, feature_list in VAR_19.items():\n",
"if not isinstance(feature_list, list):\n",
"return VAR_41.SerializeToString()\n",
"if isinstance(feature_list[0], float):\n",
"VAR_41.features.feature[feature_name].float_list.value.extend(feature_list)\n",
"if isinstance(feature_list[0], str):\n",
"VAR_41.features.feature[feature_name].bytes_list.value.extend([f.encode(\n 'utf8') for f in feature_list])\n",
"if isinstance(feature_list[0], bytes):\n",
"VAR_41.features.feature[feature_name].bytes_list.value.extend(feature_list)\n",
"if isinstance(feature_list[0], six.integer_types):\n",
"VAR_41.features.feature[feature_name].int64_list.value.extend(feature_list)\n"
] | [
"def _create_example_string(example_dict):...\n",
"\"\"\"docstring\"\"\"\n",
"example = example_pb2.Example()\n",
"for feature_name, feature_list in example_dict.items():\n",
"if not isinstance(feature_list, list):\n",
"return example.SerializeToString()\n",
"if isinstance(feature_list[0], float):\n",
"example.features.feature[feature_name].float_list.value.extend(feature_list)\n",
"if isinstance(feature_list[0], str):\n",
"example.features.feature[feature_name].bytes_list.value.extend([f.encode(\n 'utf8') for f in feature_list])\n",
"if isinstance(feature_list[0], bytes):\n",
"example.features.feature[feature_name].bytes_list.value.extend(feature_list)\n",
"if isinstance(feature_list[0], six.integer_types):\n",
"example.features.feature[feature_name].int64_list.value.extend(feature_list)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"For",
"Condition",
"Return'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Condition",
"Expr'"
] |
[
"def FUNC_33(VAR_39, VAR_36):...\n",
"VAR_37 = VAR_36.title\n",
"VAR_38 = VAR_36.author\n",
"VAR_65, VAR_62, VAR_69 = FUNC_32(VAR_37, VAR_38)\n",
"VAR_70 = helper.get_valid_filename(VAR_37)\n",
"VAR_71 = helper.get_valid_filename(VAR_69.name)\n",
"VAR_72 = os.path.join(VAR_71, VAR_70).replace('\\\\', '/')\n",
"VAR_41 = db.Books(VAR_37, '', VAR_65, datetime.utcnow(), datetime(101, 1, 1\n ), '1', datetime.utcnow(), VAR_72, VAR_36.cover, VAR_69, [], '')\n",
"VAR_39 |= FUNC_7(VAR_62, VAR_41.authors, db.Authors, calibre_db.session,\n 'author')\n",
"VAR_39 |= FUNC_18(VAR_36.series_id, VAR_41)\n",
"VAR_39 |= FUNC_20(VAR_36.languages, VAR_41, VAR_26=True)\n",
"VAR_39 |= FUNC_16(VAR_36.tags, VAR_41)\n",
"VAR_39 |= FUNC_21(VAR_36.publisher, VAR_41)\n",
"VAR_39 |= FUNC_17(VAR_36.series, VAR_41)\n",
"VAR_73 = os.path.getsize(VAR_36.file_path)\n",
"VAR_74 = db.Data(VAR_41, VAR_36.extension.upper()[1:], VAR_73, VAR_70)\n",
"VAR_41.data.append(VAR_74)\n",
"calibre_db.session.add(VAR_41)\n",
"calibre_db.session.flush()\n",
"return VAR_41, VAR_62, VAR_70\n"
] | [
"def create_book_on_upload(modif_date, meta):...\n",
"title = meta.title\n",
"authr = meta.author\n",
"sort_authors, input_authors, db_author = prepare_authors_on_upload(title, authr\n )\n",
"title_dir = helper.get_valid_filename(title)\n",
"author_dir = helper.get_valid_filename(db_author.name)\n",
"path = os.path.join(author_dir, title_dir).replace('\\\\', '/')\n",
"db_book = db.Books(title, '', sort_authors, datetime.utcnow(), datetime(101,\n 1, 1), '1', datetime.utcnow(), path, meta.cover, db_author, [], '')\n",
"modif_date |= modify_database_object(input_authors, db_book.authors, db.\n Authors, calibre_db.session, 'author')\n",
"modif_date |= edit_book_series_index(meta.series_id, db_book)\n",
"modif_date |= edit_book_languages(meta.languages, db_book, upload=True)\n",
"modif_date |= edit_book_tags(meta.tags, db_book)\n",
"modif_date |= edit_book_publisher(meta.publisher, db_book)\n",
"modif_date |= edit_book_series(meta.series, db_book)\n",
"file_size = os.path.getsize(meta.file_path)\n",
"db_data = db.Data(db_book, meta.extension.upper()[1:], file_size, title_dir)\n",
"db_book.data.append(db_data)\n",
"calibre_db.session.add(db_book)\n",
"calibre_db.session.flush()\n",
"return db_book, input_authors, title_dir\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"Assign'",
"AugAssign'",
"AugAssign'",
"AugAssign'",
"AugAssign'",
"AugAssign'",
"AugAssign'",
"Assign'",
"Assign'",
"Expr'",
"Expr'",
"Expr'",
"Return'"
] |
[
"def FUNC_20(self, *VAR_12, **VAR_13):...\n",
"\"\"\"docstring\"\"\"\n",
"if self._backend is not None and objects.backend != self._backend:\n",
"return self.wrong_backend_handler(*VAR_12, **kwargs)\n",
"return self._function(*VAR_12, **kwargs)\n"
] | [
"def wrapper(self, *args, **kwargs):...\n",
"\"\"\"docstring\"\"\"\n",
"if self._backend is not None and objects.backend != self._backend:\n",
"return self.wrong_backend_handler(*args, **kwargs)\n",
"return self._function(*args, **kwargs)\n"
] | [
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Condition",
"Return'",
"Return'"
] |
[
"def __init__(self):...\n",
"self.reqparse = reqparse.RequestParser()\n",
"self.req = CLASS_2.prepare_from_flask_request(request)\n",
"super(CLASS_2, self).__init__()\n"
] | [
"def __init__(self):...\n",
"self.reqparse = reqparse.RequestParser()\n",
"self.req = OneLogin.prepare_from_flask_request(request)\n",
"super(OneLogin, self).__init__()\n"
] | [
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Expr'"
] |
[
"@apfell.route('/apiui/command_help')...\n",
"VAR_4 = VAR_0.get_template('apiui_command_help.html')\n",
"if len(VAR_1.query_args) != 0:\n",
"VAR_5 = urllib.parse.unquote(VAR_1.query_args[0][1])\n",
"VAR_5 = ''\n",
"print(VAR_5)\n",
"if use_ssl:\n",
"VAR_6 = VAR_4.render(VAR_3=await respect_pivot(links, request), name=user[\n 'username'], http='https', ws='wss', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n",
"VAR_6 = VAR_4.render(VAR_3=await respect_pivot(links, request), name=user[\n 'username'], http='http', ws='ws', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n",
"return response.html(VAR_6)\n"
] | [
"@apfell.route('/apiui/command_help')...\n",
"template = env.get_template('apiui_command_help.html')\n",
"if len(request.query_args) != 0:\n",
"data = urllib.parse.unquote(request.query_args[0][1])\n",
"data = ''\n",
"print(data)\n",
"if use_ssl:\n",
"content = template.render(links=await respect_pivot(links, request), name=\n user['username'], http='https', ws='wss', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n",
"content = template.render(links=await respect_pivot(links, request), name=\n user['username'], http='http', ws='ws', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n",
"return response.html(content)\n"
] | [
0,
0,
0,
0,
0,
2,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"Expr'",
"Condition",
"Assign'",
"Assign'",
"Return'"
] |
[
"def FUNC_11(self, VAR_50: 'APIRouter', *, VAR_41: str='', VAR_30: Optional[...\n",
"if VAR_41:\n",
"assert VAR_41.startswith('/'), \"A path prefix must start with '/'\"\n",
"for r in VAR_50.routes:\n",
"assert not VAR_41.endswith('/'\n ), \"A path prefix must not end with '/', as the routes will start with '/'\"\n",
"VAR_26 = getattr(r, 'path')\n",
"if VAR_35 is None:\n",
"VAR_28 = getattr(r, 'name', 'unknown')\n",
"VAR_35 = {}\n",
"for VAR_65 in VAR_50.routes:\n",
"if VAR_26 is not None and not VAR_26:\n",
"if isinstance(VAR_65, CLASS_1):\n",
"for handler in VAR_50.on_startup:\n",
"VAR_60 = {**VAR_35, **VAR_65.responses}\n",
"if isinstance(VAR_65, routing.Route):\n",
"self.add_event_handler('startup', handler)\n",
"for handler in VAR_50.on_shutdown:\n",
"VAR_76 = get_value_or_default(VAR_65.response_class, VAR_50.\n default_response_class, VAR_42, self.default_response_class)\n",
"VAR_37 = list(VAR_65.methods or [])\n",
"if isinstance(VAR_65, CLASS_0):\n",
"self.add_event_handler('shutdown', handler)\n",
"VAR_62 = []\n",
"self.add_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_37=methods,\n VAR_39=route.include_in_schema, VAR_28=route.name)\n",
"self.add_api_websocket_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_28=\n route.name)\n",
"if isinstance(VAR_65, routing.WebSocketRoute):\n",
"if VAR_30:\n",
"self.add_websocket_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_28=\n route.name)\n",
"VAR_62.extend(VAR_30)\n",
"if VAR_65.tags:\n",
"VAR_62.extend(VAR_65.tags)\n",
"VAR_63: List[params.Depends] = []\n",
"if VAR_31:\n",
"VAR_63.extend(VAR_31)\n",
"if VAR_65.dependencies:\n",
"VAR_63.extend(VAR_65.dependencies)\n",
"VAR_64 = []\n",
"if VAR_40:\n",
"VAR_64.extend(VAR_40)\n",
"if VAR_65.callbacks:\n",
"VAR_64.extend(VAR_65.callbacks)\n",
"self.add_api_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_29=route.\n response_model, VAR_13=route.status_code, VAR_30=current_tags, VAR_31=\n current_dependencies, VAR_32=route.summary, VAR_33=route.description,\n VAR_34=route.response_description, VAR_35=combined_responses, VAR_36=\n route.deprecated or deprecated or self.deprecated, VAR_37=route.methods,\n VAR_38=route.operation_id, VAR_16=route.response_model_include, VAR_17=\n route.response_model_exclude, VAR_18=route.response_model_by_alias,\n VAR_19=route.response_model_exclude_unset, VAR_20=route.\n response_model_exclude_defaults, VAR_21=route.\n response_model_exclude_none, VAR_39=route.include_in_schema and self.\n include_in_schema and include_in_schema, VAR_14=use_response_class,\n VAR_28=route.name, VAR_49=type(route), VAR_40=current_callbacks)\n"
] | [
"def include_router(self, router: 'APIRouter', *, prefix: str='', tags:...\n",
"if prefix:\n",
"assert prefix.startswith('/'), \"A path prefix must start with '/'\"\n",
"for r in router.routes:\n",
"assert not prefix.endswith('/'\n ), \"A path prefix must not end with '/', as the routes will start with '/'\"\n",
"path = getattr(r, 'path')\n",
"if responses is None:\n",
"name = getattr(r, 'name', 'unknown')\n",
"responses = {}\n",
"for route in router.routes:\n",
"if path is not None and not path:\n",
"if isinstance(route, APIRoute):\n",
"for handler in router.on_startup:\n",
"combined_responses = {**responses, **route.responses}\n",
"if isinstance(route, routing.Route):\n",
"self.add_event_handler('startup', handler)\n",
"for handler in router.on_shutdown:\n",
"use_response_class = get_value_or_default(route.response_class, router.\n default_response_class, default_response_class, self.default_response_class\n )\n",
"methods = list(route.methods or [])\n",
"if isinstance(route, APIWebSocketRoute):\n",
"self.add_event_handler('shutdown', handler)\n",
"current_tags = []\n",
"self.add_route(prefix + route.path, route.endpoint, methods=methods,\n include_in_schema=route.include_in_schema, name=route.name)\n",
"self.add_api_websocket_route(prefix + route.path, route.endpoint, name=\n route.name)\n",
"if isinstance(route, routing.WebSocketRoute):\n",
"if tags:\n",
"self.add_websocket_route(prefix + route.path, route.endpoint, name=route.name)\n",
"current_tags.extend(tags)\n",
"if route.tags:\n",
"current_tags.extend(route.tags)\n",
"current_dependencies: List[params.Depends] = []\n",
"if dependencies:\n",
"current_dependencies.extend(dependencies)\n",
"if route.dependencies:\n",
"current_dependencies.extend(route.dependencies)\n",
"current_callbacks = []\n",
"if callbacks:\n",
"current_callbacks.extend(callbacks)\n",
"if route.callbacks:\n",
"current_callbacks.extend(route.callbacks)\n",
"self.add_api_route(prefix + route.path, route.endpoint, response_model=\n route.response_model, status_code=route.status_code, tags=current_tags,\n dependencies=current_dependencies, summary=route.summary, description=\n route.description, response_description=route.response_description,\n responses=combined_responses, deprecated=route.deprecated or deprecated or\n self.deprecated, methods=route.methods, operation_id=route.operation_id,\n response_model_include=route.response_model_include,\n response_model_exclude=route.response_model_exclude,\n response_model_by_alias=route.response_model_by_alias,\n response_model_exclude_unset=route.response_model_exclude_unset,\n response_model_exclude_defaults=route.response_model_exclude_defaults,\n response_model_exclude_none=route.response_model_exclude_none,\n include_in_schema=route.include_in_schema and self.include_in_schema and\n include_in_schema, response_class=use_response_class, name=route.name,\n route_class_override=type(route), callbacks=current_callbacks)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Assert'",
"For",
"Assert'",
"Assign'",
"Condition",
"Assign'",
"Assign'",
"For",
"Condition",
"Condition",
"For",
"Assign'",
"Condition",
"Expr'",
"For",
"Assign'",
"Assign'",
"Condition",
"Expr'",
"Assign'",
"Expr'",
"Expr'",
"Condition",
"Condition",
"Expr'",
"Expr'",
"Condition",
"Expr'",
"AnnAssign'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Assign'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Expr'"
] |
[
"def __init__(self, VAR_4):...\n",
"self.response = VAR_4\n",
"super(CLASS_1, self).__init__(force_text(VAR_4))\n"
] | [
"def __init__(self, response):...\n",
"self.response = response\n",
"super(ExceptionalResponse, self).__init__(force_text(response))\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Expr'"
] |
[
"def FUNC_31(VAR_37, VAR_38):...\n",
"if VAR_37[1].get('_key', None) == 'plugin_corewizard_acl':\n",
"return f'0:{to_unicode(VAR_37[0])}'\n",
"if VAR_37[1].get('mandatory', False):\n",
"return f'1:{to_unicode(VAR_37[0])}'\n",
"return f'2:{to_unicode(VAR_37[0])}'\n"
] | [
"def wizard_key_extractor(d, k):...\n",
"if d[1].get('_key', None) == 'plugin_corewizard_acl':\n",
"return f'0:{to_unicode(d[0])}'\n",
"if d[1].get('mandatory', False):\n",
"return f'1:{to_unicode(d[0])}'\n",
"return f'2:{to_unicode(d[0])}'\n"
] | [
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Condition",
"Return'",
"Condition",
"Return'",
"Return'"
] |
[
"@VAR_1.route('/ajax/mergebooks', methods=['POST'])...\n",
"VAR_79 = VAR_32.get_json().get('Merge_books')\n",
"VAR_81 = list()\n",
"if VAR_79:\n",
"VAR_100 = calibre_db.get_book(VAR_79[0])\n",
"return ''\n",
"VAR_79.pop(0)\n",
"if VAR_100:\n",
"for file in VAR_100.data:\n",
"VAR_81.append(file.format)\n",
"VAR_114 = helper.get_valid_filename(VAR_100.title\n ) + ' - ' + helper.get_valid_filename(VAR_100.authors[0].name)\n",
"for VAR_15 in VAR_79:\n",
"VAR_118 = calibre_db.get_book(VAR_15)\n",
"if VAR_118:\n",
"for element in VAR_118.data:\n",
"if element.format not in VAR_81:\n",
"FUNC_13(VAR_118.id, '', True)\n",
"VAR_122 = os.path.normpath(os.path.join(config.config_calibre_dir, VAR_100.\n path, VAR_114 + '.' + element.format.lower()))\n",
"return json.dumps({'success': True})\n",
"VAR_123 = os.path.normpath(os.path.join(config.config_calibre_dir, VAR_118.\n path, element.name + '.' + element.format.lower()))\n",
"copyfile(VAR_123, VAR_122)\n",
"VAR_100.data.append(db.Data(VAR_100.id, element.format, element.\n uncompressed_size, VAR_114))\n"
] | [
"@editbook.route('/ajax/mergebooks', methods=['POST'])...\n",
"vals = request.get_json().get('Merge_books')\n",
"to_file = list()\n",
"if vals:\n",
"to_book = calibre_db.get_book(vals[0])\n",
"return ''\n",
"vals.pop(0)\n",
"if to_book:\n",
"for file in to_book.data:\n",
"to_file.append(file.format)\n",
"to_name = helper.get_valid_filename(to_book.title\n ) + ' - ' + helper.get_valid_filename(to_book.authors[0].name)\n",
"for book_id in vals:\n",
"from_book = calibre_db.get_book(book_id)\n",
"if from_book:\n",
"for element in from_book.data:\n",
"if element.format not in to_file:\n",
"delete_book(from_book.id, '', True)\n",
"filepath_new = os.path.normpath(os.path.join(config.config_calibre_dir,\n to_book.path, to_name + '.' + element.format.lower()))\n",
"return json.dumps({'success': True})\n",
"filepath_old = os.path.normpath(os.path.join(config.config_calibre_dir,\n from_book.path, element.name + '.' + element.format.lower()))\n",
"copyfile(filepath_old, filepath_new)\n",
"to_book.data.append(db.Data(to_book.id, element.format, element.\n uncompressed_size, to_name))\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Return'",
"Expr'",
"Condition",
"For",
"Expr'",
"Assign'",
"For",
"Assign'",
"Condition",
"For",
"For",
"Expr'",
"Assign'",
"Return'",
"Assign'",
"Expr'",
"Expr'"
] |
[
"def FUNC_15():...\n",
"random.seed(VAR_24)\n",
"return list(map(lambda x: x + VAR_5 * VAR_45, sorted(random.sample(list(\n range(VAR_5)), VAR_44))))\n"
] | [
"def generate_mix_indices():...\n",
"random.seed(seed)\n",
"return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n"
] | [
0,
0,
0
] | [
"FunctionDef'",
"Expr'",
"Return'"
] |
[
"@VAR_2.route('/ajax/editlistusers/<param>', methods=['POST'])...\n",
"VAR_61 = request.form.to_dict(flat=False)\n",
"VAR_47 = ub.session.query(ub.User)\n",
"if not config.config_anonbrowse:\n",
"VAR_47 = VAR_47.filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS)\n",
"if 'pk' in VAR_61:\n",
"VAR_54 = [VAR_47.filter(ub.User.id == VAR_61['pk'][0]).one_or_none()]\n",
"if 'pk[]' in VAR_61:\n",
"if 'field_index' in VAR_61:\n",
"VAR_54 = VAR_47.filter(ub.User.id.in_(VAR_61['pk[]'])).all()\n",
"return _('Malformed request'), 400\n",
"VAR_61['field_index'] = VAR_61['field_index'][0]\n",
"if 'value' in VAR_61:\n",
"VAR_61['value'] = VAR_61['value'][0]\n",
"if not 'value[]' in VAR_61:\n",
"for VAR_12 in VAR_54:\n",
"return _('Malformed request'), 400\n",
"ub.session_commit()\n",
"if VAR_4 in ['denied_tags', 'allowed_tags', 'allowed_column_value',\n",
"VAR_0.debug_or_exception(ex)\n",
"return ''\n",
"if 'value[]' in VAR_61:\n",
"VAR_61['value'] = VAR_61['value'].strip()\n",
"return str(ex), 400\n",
"setattr(VAR_12, VAR_4, FUNC_29(VAR_12, VAR_61['action'][0], VAR_4, VAR_61[\n 'value[]']))\n",
"setattr(VAR_12, VAR_4, VAR_61['value'].strip())\n",
"if VAR_4 == 'name':\n",
"if VAR_12.name == 'Guest':\n",
"if VAR_4 == 'email':\n",
"VAR_12.name = check_username(VAR_61['value'])\n",
"VAR_12.email = check_email(VAR_61['value'])\n",
"if VAR_4 == 'kobo_only_shelves_sync':\n",
"VAR_12.kobo_only_shelves_sync = VAR_119(VAR_61['value'] == 'true')\n",
"if VAR_4 == 'kindle_mail':\n",
"VAR_12.kindle_mail = valid_email(VAR_61['value']) if VAR_61['value'] else ''\n",
"if VAR_4.endswith('role'):\n",
"VAR_125 = VAR_119(VAR_61['field_index'])\n",
"if VAR_4.startswith('sidebar'):\n",
"if VAR_12.name == 'Guest' and VAR_125 in [constants.ROLE_ADMIN, constants.\n",
"VAR_125 = VAR_119(VAR_61['field_index'])\n",
"if VAR_4 == 'locale':\n",
"if VAR_125 > 0 and VAR_125 <= constants.ROLE_VIEWER and (VAR_125 & VAR_125 -\n",
"if VAR_12.name == 'Guest' and VAR_125 == constants.SIDEBAR_READ_AND_UNREAD:\n",
"if VAR_12.name == 'Guest':\n",
"if VAR_4 == 'default_language':\n",
"if VAR_61['value'] == 'true':\n",
"if VAR_125 > 0 and VAR_125 <= constants.SIDEBAR_LIST and (VAR_125 & VAR_125 -\n",
"if VAR_61['value'] in _BABEL_TRANSLATIONS:\n",
"VAR_23 = calibre_db.session.query(db.Languages).join(db.books_languages_link\n ).join(db.Books).filter(calibre_db.common_filters()).group_by(VAR_130(\n 'books_languages_link.lang_code')).all()\n",
"return _('Parameter not found'), 400\n",
"VAR_12.role |= VAR_125\n",
"if VAR_61['value'] == 'false':\n",
"if VAR_61['value'] == 'true':\n",
"VAR_12.locale = VAR_61['value']\n",
"VAR_131 = [lang.lang_code for lang in VAR_23] + ['all']\n",
"if VAR_125 == constants.ROLE_ADMIN:\n",
"VAR_12.sidebar_view |= VAR_125\n",
"if VAR_61['value'] == 'false':\n",
"if VAR_61['value'] in VAR_131:\n",
"if not ub.session.query(ub.User).filter(ub.User.role.op('&')(constants.\n",
"VAR_12.role &= ~VAR_125\n",
"VAR_12.sidebar_view &= ~VAR_125\n",
"VAR_12.default_language = VAR_61['value']\n",
"return Response(json.dumps([{'type': 'danger', 'message': _(\n u\"No admin user remaining, can't remove admin role\", nick=user.name)}]),\n mimetype='application/json')\n"
] | [
"@admi.route('/ajax/editlistusers/<param>', methods=['POST'])...\n",
"vals = request.form.to_dict(flat=False)\n",
"all_user = ub.session.query(ub.User)\n",
"if not config.config_anonbrowse:\n",
"all_user = all_user.filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS)\n",
"if 'pk' in vals:\n",
"users = [all_user.filter(ub.User.id == vals['pk'][0]).one_or_none()]\n",
"if 'pk[]' in vals:\n",
"if 'field_index' in vals:\n",
"users = all_user.filter(ub.User.id.in_(vals['pk[]'])).all()\n",
"return _('Malformed request'), 400\n",
"vals['field_index'] = vals['field_index'][0]\n",
"if 'value' in vals:\n",
"vals['value'] = vals['value'][0]\n",
"if not 'value[]' in vals:\n",
"for user in users:\n",
"return _('Malformed request'), 400\n",
"ub.session_commit()\n",
"if param in ['denied_tags', 'allowed_tags', 'allowed_column_value',\n",
"log.debug_or_exception(ex)\n",
"return ''\n",
"if 'value[]' in vals:\n",
"vals['value'] = vals['value'].strip()\n",
"return str(ex), 400\n",
"setattr(user, param, prepare_tags(user, vals['action'][0], param, vals[\n 'value[]']))\n",
"setattr(user, param, vals['value'].strip())\n",
"if param == 'name':\n",
"if user.name == 'Guest':\n",
"if param == 'email':\n",
"user.name = check_username(vals['value'])\n",
"user.email = check_email(vals['value'])\n",
"if param == 'kobo_only_shelves_sync':\n",
"user.kobo_only_shelves_sync = int(vals['value'] == 'true')\n",
"if param == 'kindle_mail':\n",
"user.kindle_mail = valid_email(vals['value']) if vals['value'] else ''\n",
"if param.endswith('role'):\n",
"value = int(vals['field_index'])\n",
"if param.startswith('sidebar'):\n",
"if user.name == 'Guest' and value in [constants.ROLE_ADMIN, constants.\n",
"value = int(vals['field_index'])\n",
"if param == 'locale':\n",
"if value > 0 and value <= constants.ROLE_VIEWER and (value & value - 1 == 0 or\n",
"if user.name == 'Guest' and value == constants.SIDEBAR_READ_AND_UNREAD:\n",
"if user.name == 'Guest':\n",
"if param == 'default_language':\n",
"if vals['value'] == 'true':\n",
"if value > 0 and value <= constants.SIDEBAR_LIST and (value & value - 1 == \n",
"if vals['value'] in _BABEL_TRANSLATIONS:\n",
"languages = calibre_db.session.query(db.Languages).join(db.books_languages_link\n ).join(db.Books).filter(calibre_db.common_filters()).group_by(text(\n 'books_languages_link.lang_code')).all()\n",
"return _('Parameter not found'), 400\n",
"user.role |= value\n",
"if vals['value'] == 'false':\n",
"if vals['value'] == 'true':\n",
"user.locale = vals['value']\n",
"lang_codes = [lang.lang_code for lang in languages] + ['all']\n",
"if value == constants.ROLE_ADMIN:\n",
"user.sidebar_view |= value\n",
"if vals['value'] == 'false':\n",
"if vals['value'] in lang_codes:\n",
"if not ub.session.query(ub.User).filter(ub.User.role.op('&')(constants.\n",
"user.role &= ~value\n",
"user.sidebar_view &= ~value\n",
"user.default_language = vals['value']\n",
"return Response(json.dumps([{'type': 'danger', 'message': _(\n u\"No admin user remaining, can't remove admin role\", nick=user.name)}]),\n mimetype='application/json')\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Return'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"For",
"Return'",
"Expr'",
"Condition",
"Expr'",
"Return'",
"Condition",
"Assign'",
"Return'",
"Expr'",
"Expr'",
"Condition",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Condition",
"Assign'",
"Condition",
"Condition",
"Condition",
"Condition",
"Condition",
"Condition",
"Condition",
"Condition",
"Assign'",
"Return'",
"AugAssign'",
"Condition",
"Condition",
"Assign'",
"Assign'",
"Condition",
"AugAssign'",
"Condition",
"Condition",
"Condition",
"AugAssign'",
"AugAssign'",
"Assign'",
"Return'"
] |
[
"def FUNC_10(self, VAR_0):...\n",
"VAR_13 = VAR_0.patch('httpx.get')\n",
"VAR_14 = VAR_0.patch('openapi_python_client.Path')\n",
"VAR_15 = VAR_0.patch('yaml.safe_load')\n",
"from openapi_python_client import _get_document\n",
"VAR_11 = VAR_2(VAR_6=mocker.MagicMock(), VAR_7=mocker.MagicMock())\n",
"assert VAR_11 == GeneratorError(header='Provide URL or Path, not both.')\n",
"VAR_13.assert_not_called()\n",
"VAR_14.assert_not_called()\n",
"VAR_15.assert_not_called()\n"
] | [
"def test__get_document_url_and_path(self, mocker):...\n",
"get = mocker.patch('httpx.get')\n",
"Path = mocker.patch('openapi_python_client.Path')\n",
"loads = mocker.patch('yaml.safe_load')\n",
"from openapi_python_client import _get_document\n",
"result = _get_document(url=mocker.MagicMock(), path=mocker.MagicMock())\n",
"assert result == GeneratorError(header='Provide URL or Path, not both.')\n",
"get.assert_not_called()\n",
"Path.assert_not_called()\n",
"loads.assert_not_called()\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Assign'",
"Assign'",
"Assign'",
"ImportFrom'",
"Assign'",
"Assert'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"@VAR_0.route('/simpleshelf/<int:shelf_id>')...\n",
"return FUNC_15(2, VAR_3, 1, None)\n"
] | [
"@shelf.route('/simpleshelf/<int:shelf_id>')...\n",
"return render_show_shelf(2, shelf_id, 1, None)\n"
] | [
0,
0
] | [
"Condition",
"Return'"
] |
[
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_2 = FUNC_1('/groups/%s/roles', VAR_30)\n",
"return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n"
] | [
"@log_function...\n",
"\"\"\"docstring\"\"\"\n",
"path = _create_v1_path('/groups/%s/roles', group_id)\n",
"return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Docstring",
"Assign'",
"Return'"
] |
[
"@wraps(VAR_4)...\n",
"if VAR_87.role_download():\n",
"return VAR_4(*VAR_49, **kwargs)\n",
"abort(403)\n"
] | [
"@wraps(f)...\n",
"if current_user.role_download():\n",
"return f(*args, **kwargs)\n",
"abort(403)\n"
] | [
0,
0,
0,
0
] | [
"Condition",
"Condition",
"Return'",
"Expr'"
] |
[
"def FUNC_4(VAR_21: Text) ->Tuple[Optional[Text], Optional[Text]]:...\n",
"\"\"\"docstring\"\"\"\n",
"VAR_50 = os.path.join(VAR_21, DEFAULT_CORE_SUBDIRECTORY_NAME)\n",
"VAR_51 = os.path.join(VAR_21, DEFAULT_NLU_SUBDIRECTORY_NAME)\n",
"if not os.path.isdir(VAR_50):\n",
"VAR_50 = None\n",
"if not os.path.isdir(VAR_51):\n",
"VAR_51 = None\n",
"if not VAR_50 and not VAR_51:\n",
"return VAR_50, VAR_51\n"
] | [
"def get_model_subdirectories(unpacked_model_path: Text) ->Tuple[Optional[...\n",
"\"\"\"docstring\"\"\"\n",
"core_path = os.path.join(unpacked_model_path, DEFAULT_CORE_SUBDIRECTORY_NAME)\n",
"nlu_path = os.path.join(unpacked_model_path, DEFAULT_NLU_SUBDIRECTORY_NAME)\n",
"if not os.path.isdir(core_path):\n",
"core_path = None\n",
"if not os.path.isdir(nlu_path):\n",
"nlu_path = None\n",
"if not core_path and not nlu_path:\n",
"return core_path, nlu_path\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"FunctionDef'",
"Docstring",
"Assign'",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Assign'",
"Condition",
"Return'"
] |
[
"async def FUNC_47(VAR_82):...\n",
"for VAR_196 in VAR_82:\n",
"return False\n",
"await self.backfill(VAR_196, VAR_11, VAR_17=100, VAR_18=extremities)\n",
"VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n",
"return True\n",
"if 400 <= VAR_170.code < 500:\n",
"VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n",
"if 400 <= VAR_170.code < 500:\n",
"VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n",
"VAR_0.info(str(VAR_170))\n",
"VAR_0.info('Failed to get backfill from %s because %s', VAR_196, VAR_170)\n",
"VAR_0.info(VAR_170)\n",
"VAR_0.exception('Failed to backfill from %s because %s', VAR_196, VAR_170)\n"
] | [
"async def try_backfill(domains):...\n",
"for dom in domains:\n",
"return False\n",
"await self.backfill(dom, room_id, limit=100, extremities=extremities)\n",
"logger.info('Failed to backfill from %s because %s', dom, e)\n",
"return True\n",
"if 400 <= e.code < 500:\n",
"logger.info('Failed to backfill from %s because %s', dom, e)\n",
"if 400 <= e.code < 500:\n",
"logger.info('Failed to backfill from %s because %s', dom, e)\n",
"logger.info(str(e))\n",
"logger.info('Failed to get backfill from %s because %s', dom, e)\n",
"logger.info(e)\n",
"logger.exception('Failed to backfill from %s because %s', dom, e)\n"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] | [
"AsyncFunctionDef'",
"For",
"Return'",
"Expr'",
"Expr'",
"Return'",
"Condition",
"Expr'",
"Condition",
"Expr'",
"Expr'",
"Expr'",
"Expr'",
"Expr'"
] |
[
"def FUNC_23(self):...\n",
"return False\n"
] | [
"def role_admin(self):...\n",
"return False\n"
] | [
0,
0
] | [
"FunctionDef'",
"Return'"
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.