lines
sequencelengths
1
383
raw_lines
sequencelengths
1
383
label
sequencelengths
1
383
type
sequencelengths
1
383
[ "@override_settings(LOGIN_REDIRECT_URL='/custom/')...\n", "self.assertLoginRedirectURLEqual('/custom/')\n" ]
[ "@override_settings(LOGIN_REDIRECT_URL='/custom/')...\n", "self.assertLoginRedirectURLEqual('/custom/')\n" ]
[ 0, 0 ]
[ "Condition", "Expr'" ]
[ "def FUNC_44(VAR_15):...\n", "\"\"\"docstring\"\"\"\n", "VAR_108 = VAR_15.replace('\\\\', '/')\n", "if os.path.isabs(VAR_108) and os.path.isfile(VAR_108):\n", "VAR_168, VAR_5 = os.path.split(VAR_108)\n", "return ''\n", "VAR_11, VAR_169 = os.path.splitext(VAR_5)\n", "VAR_3 = FUNC_5()\n", "VAR_170 = {'controllers': '.py', 'models': '.py', 'views': '.html'}\n", "for VAR_143 in VAR_170.keys():\n", "VAR_194 = VAR_168.endswith('%s/%s' % (VAR_3, VAR_143))\n", "if VAR_169.lower() == VAR_170[VAR_143] and VAR_194:\n", "return to_native(A('\"' + VAR_108 + '\"', _href=URL(r=request, f=\n 'edit/%s/%s/%s' % (app, key, filename))).xml())\n" ]
[ "def make_link(path):...\n", "\"\"\"docstring\"\"\"\n", "tryFile = path.replace('\\\\', '/')\n", "if os.path.isabs(tryFile) and os.path.isfile(tryFile):\n", "folder, filename = os.path.split(tryFile)\n", "return ''\n", "base, ext = os.path.splitext(filename)\n", "app = get_app()\n", "editable = {'controllers': '.py', 'models': '.py', 'views': '.html'}\n", "for key in editable.keys():\n", "check_extension = folder.endswith('%s/%s' % (app, key))\n", "if ext.lower() == editable[key] and check_extension:\n", "return to_native(A('\"' + tryFile + '\"', _href=URL(r=request, f=\n 'edit/%s/%s/%s' % (app, key, filename))).xml())\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Condition", "Return'" ]
[ "def FUNC_0(VAR_2, VAR_3):...\n", "\"\"\"docstring\"\"\"\n", "for i in range(min(5, VAR_3)):\n", "yield VAR_2 + i\n", "for i in range(VAR_3 - 5):\n", "yield max(1, VAR_2 + random.randint(-2 * VAR_3, 2 * VAR_3))\n" ]
[ "def random_ports(port, n):...\n", "\"\"\"docstring\"\"\"\n", "for i in range(min(5, n)):\n", "yield port + i\n", "for i in range(n - 5):\n", "yield max(1, port + random.randint(-2 * n, 2 * n))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "For", "Expr'", "For", "Expr'" ]
[ "async def FUNC_45(self, VAR_50: List[str], VAR_11: str) ->Optional[dict]:...\n", "\"\"\"docstring\"\"\"\n", "for host in VAR_50:\n", "VAR_168 = await self.federation_client.get_room_complexity(host, VAR_11)\n", "return None\n", "if VAR_168:\n", "return VAR_168\n" ]
[ "async def get_room_complexity(self, remote_room_hosts: List[str], room_id: str...\n", "\"\"\"docstring\"\"\"\n", "for host in remote_room_hosts:\n", "res = await self.federation_client.get_room_complexity(host, room_id)\n", "return None\n", "if res:\n", "return res\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Docstring", "For", "Assign'", "Return'", "Condition", "Return'" ]
[ "def FUNC_4(VAR_14, VAR_18=4):...\n", "VAR_14 = VAR_3.sub('\\n', VAR_14)\n", "VAR_14 = VAR_14.expandtabs(VAR_18)\n", "VAR_14 = VAR_14.replace('␤', '\\n')\n", "VAR_21 = re.compile('^ +$', re.M)\n", "return VAR_21.sub('', VAR_14)\n" ]
[ "def preprocessing(text, tab=4):...\n", "text = _newline_pattern.sub('\\n', text)\n", "text = text.expandtabs(tab)\n", "text = text.replace('␤', '\\n')\n", "pattern = re.compile('^ +$', re.M)\n", "return pattern.sub('', text)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_2(self, VAR_5, VAR_6, *VAR_21, **VAR_22):...\n", "VAR_31 = self.rel.to\n", "VAR_36 = VAR_31._meta.app_label, VAR_31._meta.model_name\n", "self.widget.choices = self.choices\n", "VAR_27 = [self.widget.render(VAR_5, VAR_6, *VAR_21, **kwargs)]\n", "if self.can_add_related:\n", "VAR_38 = reverse('admin:%s_%s_add' % VAR_36, current_app=self.admin_site.name)\n", "return mark_safe(''.join(VAR_27))\n", "VAR_27.append(\n '<a href=\"%s\" class=\"add-another\" id=\"add_id_%s\" onclick=\"return showAddAnotherPopup(this);\"> '\n % (VAR_38, VAR_5))\n", "VAR_27.append('<img src=\"%s\" width=\"10\" height=\"10\" alt=\"%s\"/></a>' % (\n static('admin/img/icon_addlink.gif'), _('Add Another')))\n" ]
[ "def render(self, name, value, *args, **kwargs):...\n", "rel_to = self.rel.to\n", "info = rel_to._meta.app_label, rel_to._meta.model_name\n", "self.widget.choices = self.choices\n", "output = [self.widget.render(name, value, *args, **kwargs)]\n", "if self.can_add_related:\n", "related_url = reverse('admin:%s_%s_add' % info, current_app=self.admin_site\n .name)\n", "return mark_safe(''.join(output))\n", "output.append(\n '<a href=\"%s\" class=\"add-another\" id=\"add_id_%s\" onclick=\"return showAddAnotherPopup(this);\"> '\n % (related_url, name))\n", "output.append('<img src=\"%s\" width=\"10\" height=\"10\" alt=\"%s\"/></a>' % (\n static('admin/img/icon_addlink.gif'), _('Add Another')))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Expr'", "Expr'" ]
[ "def FUNC_1(self, VAR_4, VAR_5, VAR_6):...\n", "self.room_id = self.helper.create_room_as(self.user_id)\n", "self.helper.join(self.room_id, VAR_2='@jim:red')\n" ]
[ "def prepare(self, reactor, clock, hs):...\n", "self.room_id = self.helper.create_room_as(self.user_id)\n", "self.helper.join(self.room_id, user='@jim:red')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_27(self):...\n", "self.assert_expected(self.folder.t, 'InterpolationInContent.html')\n" ]
[ "def testInterpolationInContent(self):...\n", "self.assert_expected(self.folder.t, 'InterpolationInContent.html')\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Expr'" ]
[ "def FUNC_4(self):...\n", "from openapi_python_client.parser.properties import EnumProperty\n", "VAR_23 = ['abc', '123', 'a23', '1bc']\n", "VAR_24 = VAR_27.values_from_list(VAR_23)\n", "assert VAR_24 == {'ABC': 'abc', 'VALUE_1': '123', 'A23': 'a23', 'VALUE_3':\n '1bc'}\n" ]
[ "def test_values_from_list(self):...\n", "from openapi_python_client.parser.properties import EnumProperty\n", "data = ['abc', '123', 'a23', '1bc']\n", "result = EnumProperty.values_from_list(data)\n", "assert result == {'ABC': 'abc', 'VALUE_1': '123', 'A23': 'a23', 'VALUE_3':\n '1bc'}\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "ImportFrom'", "Assign'", "Assign'", "Assert'" ]
[ "@pytest.mark.posix...\n", "assert filescheme.is_root(VAR_9) == VAR_10\n" ]
[ "@pytest.mark.posix...\n", "assert filescheme.is_root(directory) == is_root\n" ]
[ 0, 0 ]
[ "Condition", "Assert'" ]
[ "def FUNC_73(self, VAR_2, VAR_8, VAR_175):...\n", "\"\"\"docstring\"\"\"\n", "VAR_203 = VAR_8.getEventContext()\n", "VAR_24 = {}\n", "for VAR_167 in ['sessionId', 'sessionUuid', 'userId', 'userName', 'groupId',\n", "if hasattr(VAR_203, VAR_167):\n", "return JsonResponse({'success': True, 'eventContext': VAR_24})\n", "VAR_24[VAR_167] = getattr(VAR_203, VAR_167)\n" ]
[ "def handle_logged_in(self, request, conn, connector):...\n", "\"\"\"docstring\"\"\"\n", "c = conn.getEventContext()\n", "ctx = {}\n", "for a in ['sessionId', 'sessionUuid', 'userId', 'userName', 'groupId',\n", "if hasattr(c, a):\n", "return JsonResponse({'success': True, 'eventContext': ctx})\n", "ctx[a] = getattr(c, a)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "For", "Condition", "Return'", "Assign'" ]
[ "def FUNC_57(self):...\n", "\"\"\"docstring\"\"\"\n", "self.loggedIn = True\n" ]
[ "def logIn(self):...\n", "\"\"\"docstring\"\"\"\n", "self.loggedIn = True\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'" ]
[ "def FUNC_1(VAR_7, **VAR_8):...\n", "if not VAR_7.user.is_authenticated:\n", "return redirect(VAR_7.GET.get('next', reverse(settings.LOGIN_URL)))\n", "if VAR_7.method == 'POST':\n", "return VAR_2(VAR_7, **kwargs)\n", "return render(VAR_7, 'spirit/user/auth/logout.html')\n" ]
[ "def custom_logout(request, **kwargs):...\n", "if not request.user.is_authenticated:\n", "return redirect(request.GET.get('next', reverse(settings.LOGIN_URL)))\n", "if request.method == 'POST':\n", "return _logout_view(request, **kwargs)\n", "return render(request, 'spirit/user/auth/logout.html')\n" ]
[ 0, 0, 4, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "def FUNC_11(self, VAR_1):...\n", "VAR_29 = VAR_1.patch(f'{VAR_0}._string_based_property')\n", "VAR_5 = VAR_1.MagicMock()\n", "VAR_26 = VAR_1.MagicMock()\n", "VAR_23 = oai.Schema.construct(type='string')\n", "from openapi_python_client.parser.properties import property_from_data\n", "VAR_4 = property_from_data(VAR_5=name, VAR_26=required, VAR_23=data)\n", "assert VAR_4 == VAR_29.return_value\n", "VAR_29.assert_called_once_with(VAR_5=name, VAR_26=required, VAR_23=data)\n" ]
[ "def test_property_from_data_string(self, mocker):...\n", "_string_based_property = mocker.patch(f'{MODULE_NAME}._string_based_property')\n", "name = mocker.MagicMock()\n", "required = mocker.MagicMock()\n", "data = oai.Schema.construct(type='string')\n", "from openapi_python_client.parser.properties import property_from_data\n", "p = property_from_data(name=name, required=required, data=data)\n", "assert p == _string_based_property.return_value\n", "_string_based_property.assert_called_once_with(name=name, required=required,\n data=data)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assert'", "Expr'" ]
[ "def FUNC_11(self, VAR_55):...\n", "\"\"\"docstring\"\"\"\n", "VAR_81 = []\n", "self._kill_elements(VAR_55, lambda VAR_15: VAR_6.search(VAR_15.text), etree\n .Comment)\n" ]
[ "def kill_conditional_comments(self, doc):...\n", "\"\"\"docstring\"\"\"\n", "bad = []\n", "self._kill_elements(doc, lambda el: _conditional_comment_re.search(el.text),\n etree.Comment)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'" ]
[ "def FUNC_5(VAR_11):...\n", "return os.path.join(VAR_3, VAR_11)\n" ]
[ "def _get_obj_absolute_path(obj_path):...\n", "return os.path.join(DATAROOT, obj_path)\n" ]
[ 0, 1 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_2(VAR_0, VAR_2, VAR_3):...\n", "VAR_7 = Topic.objects.get_public_or_404(VAR_2, VAR_0.user)\n", "if VAR_7.slug != VAR_3:\n", "return HttpResponsePermanentRedirect(VAR_7.get_absolute_url())\n", "utils.topic_viewed(VAR_0=request, VAR_7=topic)\n", "VAR_8 = Comment.objects.for_topic(VAR_7=topic).with_likes(VAR_4=request.user\n ).with_polls(VAR_4=request.user).order_by('date')\n", "VAR_8 = paginate(VAR_8, per_page=config.comments_per_page, page_number=\n request.GET.get('page', 1))\n", "return render(VAR_0=request, template_name='spirit/topic/detail.html',\n context={'topic': topic, 'comments': comments})\n" ]
[ "def detail(request, pk, slug):...\n", "topic = Topic.objects.get_public_or_404(pk, request.user)\n", "if topic.slug != slug:\n", "return HttpResponsePermanentRedirect(topic.get_absolute_url())\n", "utils.topic_viewed(request=request, topic=topic)\n", "comments = Comment.objects.for_topic(topic=topic).with_likes(user=request.user\n ).with_polls(user=request.user).order_by('date')\n", "comments = paginate(comments, per_page=config.comments_per_page,\n page_number=request.GET.get('page', 1))\n", "return render(request=request, template_name='spirit/topic/detail.html',\n context={'topic': topic, 'comments': comments})\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_88(*VAR_79, **VAR_42):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.utils.formatters\n", "return frappe.utils.formatters.format_value(*VAR_79, **kwargs)\n" ]
[ "def format_value(*args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "import frappe.utils.formatters\n", "return frappe.utils.formatters.format_value(*args, **kwargs)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Import'", "Return'" ]
[ "@VAR_1.route('/ajax/editbooks/<param>', methods=['POST'])...\n", "VAR_79 = VAR_32.form.to_dict()\n", "VAR_17 = calibre_db.get_book(VAR_79['pk'])\n", "VAR_80 = ''\n", "if VAR_43 == 'series_index':\n", "FUNC_18(VAR_79['value'], VAR_17)\n", "if VAR_43 == 'tags':\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.\n series_index}), mimetype='application/json')\n", "FUNC_16(VAR_79['value'], VAR_17)\n", "if VAR_43 == 'series':\n", "VAR_17.last_modified = datetime.utcnow()\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([tag.\n name for tag in VAR_17.tags])}), mimetype='application/json')\n", "FUNC_17(VAR_79['value'], VAR_17)\n", "if VAR_43 == 'publishers':\n", "calibre_db.session.commit()\n", "calibre_db.session.rollback()\n", "return VAR_80\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([serie\n .name for serie in VAR_17.series])}), mimetype='application/json')\n", "FUNC_21(VAR_79['value'], VAR_17)\n", "if VAR_43 == 'languages':\n", "if VAR_43 == 'title' and VAR_79.get('checkT') == 'false':\n", "VAR_2.error('Database error: %s', e)\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join([\n VAR_86.name for VAR_86 in VAR_17.publishers])}), mimetype=\n 'application/json')\n", "VAR_27 = list()\n", "if VAR_43 == 'author_sort':\n", "VAR_17.sort = VAR_124\n", "FUNC_20(VAR_79['value'], VAR_17, VAR_27=invalid)\n", "VAR_17.author_sort = VAR_79['value']\n", "if VAR_43 == 'title':\n", "calibre_db.session.commit()\n", "if VAR_27:\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.\n author_sort}), mimetype='application/json')\n", "VAR_124 = VAR_17.sort\n", "if VAR_43 == 'sort':\n", "VAR_80 = Response(json.dumps({'success': False, 'msg':\n 'Invalid languages in request: {}'.format(','.join(VAR_27))}), mimetype\n ='application/json')\n", "VAR_121 = list()\n", "FUNC_27(VAR_17, VAR_79.get('value', ''))\n", "VAR_17.sort = VAR_79['value']\n", "if VAR_43 == 'authors':\n", "for lang in VAR_17.languages:\n", "helper.update_dir_stucture(VAR_17.id, config.config_calibre_dir)\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.sort}),\n mimetype='application/json')\n", "VAR_62, VAR_125 = FUNC_28(VAR_17, VAR_79['value'], VAR_79.get('checkA',\n None) == 'true')\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': ', '.join(\n VAR_121)}), mimetype='application/json')\n", "VAR_121.append(LC.parse(lang.lang_code).get_language_name(get_locale()))\n", "VAR_121.append(_(isoLanguages.get(part3=lang.lang_code).name))\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': VAR_17.title}),\n mimetype='application/json')\n", "helper.update_dir_stucture(VAR_17.id, config.config_calibre_dir, VAR_62[0])\n", "VAR_80 = Response(json.dumps({'success': True, 'newValue': ' & '.join([\n author.replace('|', ',') for author in VAR_62])}), mimetype=\n 'application/json')\n" ]
[ "@editbook.route('/ajax/editbooks/<param>', methods=['POST'])...\n", "vals = request.form.to_dict()\n", "book = calibre_db.get_book(vals['pk'])\n", "ret = ''\n", "if param == 'series_index':\n", "edit_book_series_index(vals['value'], book)\n", "if param == 'tags':\n", "ret = Response(json.dumps({'success': True, 'newValue': book.series_index}),\n mimetype='application/json')\n", "edit_book_tags(vals['value'], book)\n", "if param == 'series':\n", "book.last_modified = datetime.utcnow()\n", "ret = Response(json.dumps({'success': True, 'newValue': ', '.join([tag.name for\n tag in book.tags])}), mimetype='application/json')\n", "edit_book_series(vals['value'], book)\n", "if param == 'publishers':\n", "calibre_db.session.commit()\n", "calibre_db.session.rollback()\n", "return ret\n", "ret = Response(json.dumps({'success': True, 'newValue': ', '.join([serie.\n name for serie in book.series])}), mimetype='application/json')\n", "edit_book_publisher(vals['value'], book)\n", "if param == 'languages':\n", "if param == 'title' and vals.get('checkT') == 'false':\n", "log.error('Database error: %s', e)\n", "ret = Response(json.dumps({'success': True, 'newValue': ', '.join([\n publisher.name for publisher in book.publishers])}), mimetype=\n 'application/json')\n", "invalid = list()\n", "if param == 'author_sort':\n", "book.sort = sort\n", "edit_book_languages(vals['value'], book, invalid=invalid)\n", "book.author_sort = vals['value']\n", "if param == 'title':\n", "calibre_db.session.commit()\n", "if invalid:\n", "ret = Response(json.dumps({'success': True, 'newValue': book.author_sort}),\n mimetype='application/json')\n", "sort = book.sort\n", "if param == 'sort':\n", "ret = Response(json.dumps({'success': False, 'msg':\n 'Invalid languages in request: {}'.format(','.join(invalid))}),\n mimetype='application/json')\n", "lang_names = list()\n", "handle_title_on_edit(book, vals.get('value', ''))\n", "book.sort = vals['value']\n", "if param == 'authors':\n", "for lang in book.languages:\n", "helper.update_dir_stucture(book.id, config.config_calibre_dir)\n", "ret = Response(json.dumps({'success': True, 'newValue': book.sort}),\n mimetype='application/json')\n", "input_authors, __ = handle_author_on_edit(book, vals['value'], vals.get(\n 'checkA', None) == 'true')\n", "ret = Response(json.dumps({'success': True, 'newValue': ', '.join(\n lang_names)}), mimetype='application/json')\n", "lang_names.append(LC.parse(lang.lang_code).get_language_name(get_locale()))\n", "lang_names.append(_(isoLanguages.get(part3=lang.lang_code).name))\n", "ret = Response(json.dumps({'success': True, 'newValue': book.title}),\n mimetype='application/json')\n", "helper.update_dir_stucture(book.id, config.config_calibre_dir, input_authors[0]\n )\n", "ret = Response(json.dumps({'success': True, 'newValue': ' & '.join([author.\n replace('|', ',') for author in input_authors])}), mimetype=\n 'application/json')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Assign'", "Expr'", "Condition", "Expr'", "Expr'", "Return'", "Assign'", "Expr'", "Condition", "Condition", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Expr'", "Assign'", "Condition", "Expr'", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "For", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'" ]
[ "def __post_init__(self) ->None:...\n", "super().__post_init__()\n", "if self.default is not None:\n", "self.default = (\n f'field(default_factory=lambda: cast({self.get_type_string()}, {self.default}))'\n )\n" ]
[ "def __post_init__(self) ->None:...\n", "super().__post_init__()\n", "if self.default is not None:\n", "self.default = (\n f'field(default_factory=lambda: cast({self.get_type_string()}, {self.default}))'\n )\n" ]
[ 0, 5, 5, 5 ]
[ "FunctionDef'", "Expr'", "Condition", "Assign'" ]
[ "def FUNC_13(VAR_13):...\n", "\"\"\"docstring\"\"\"\n", "VAR_31 = LOG_LEVELS.get(str(VAR_13.get('log_level')).lower(), logging.NOTSET)\n", "if VAR_31 < logging.INFO:\n", "VAR_0.warning(\n 'Insecure logging configuration detected! Sensitive data may be logged.')\n" ]
[ "def verify_log(opts):...\n", "\"\"\"docstring\"\"\"\n", "level = LOG_LEVELS.get(str(opts.get('log_level')).lower(), logging.NOTSET)\n", "if level < logging.INFO:\n", "log.warning(\n 'Insecure logging configuration detected! Sensitive data may be logged.')\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Expr'" ]
[ "@defer.inlineCallbacks...\n", "VAR_7 = {'room': {'state': {'types': ['m.*']}}}\n", "VAR_8 = yield defer.ensureDeferred(self.filtering.add_user_filter(VAR_0=\n user_localpart, VAR_10=user_filter_json))\n", "self.assertEquals(VAR_8, 0)\n", "self.assertEquals(VAR_7, (yield defer.ensureDeferred(self.datastore.\n get_user_filter(VAR_0=user_localpart, VAR_8=0))))\n" ]
[ "@defer.inlineCallbacks...\n", "user_filter_json = {'room': {'state': {'types': ['m.*']}}}\n", "filter_id = yield defer.ensureDeferred(self.filtering.add_user_filter(\n user_localpart=user_localpart, user_filter=user_filter_json))\n", "self.assertEquals(filter_id, 0)\n", "self.assertEquals(user_filter_json, (yield defer.ensureDeferred(self.\n datastore.get_user_filter(user_localpart=user_localpart, filter_id=0))))\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_11(self, VAR_11):...\n", "VAR_60 = self.download_check_files(VAR_11)\n", "if VAR_60 == 'not_permitted':\n", "return 'You are not allowed to download files.'\n", "if VAR_60 == 'invalid_file':\n", "return 'Error: invalid filename found in {list}'.format(list=filelist)\n", "if VAR_60 == 'too_big':\n", "VAR_59 = cherry.config['media.maximum_download_size']\n", "if VAR_60 == 'ok':\n", "return 'string'.format(maxsize=size_limit / 1024 / 1024)\n", "return VAR_60\n", "VAR_125 = 'Error status check for download: {status!r}'.format(VAR_60=status)\n", "log.e(VAR_125)\n", "return VAR_125\n" ]
[ "def api_downloadcheck(self, filelist):...\n", "status = self.download_check_files(filelist)\n", "if status == 'not_permitted':\n", "return 'You are not allowed to download files.'\n", "if status == 'invalid_file':\n", "return 'Error: invalid filename found in {list}'.format(list=filelist)\n", "if status == 'too_big':\n", "size_limit = cherry.config['media.maximum_download_size']\n", "if status == 'ok':\n", "return \"\"\"Can't download: Playlist is bigger than {maxsize} mB.\n The server administrator can change this configuration.\n \"\"\".format(\n maxsize=size_limit / 1024 / 1024)\n", "return status\n", "message = 'Error status check for download: {status!r}'.format(status=status)\n", "log.e(message)\n", "return message\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'", "Condition", "Return'", "Condition", "Assign'", "Condition", "Return'", "Return'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_52(self):...\n", "\"\"\"docstring\"\"\"\n", "VAR_69 = self.get_doc_before_save()\n", "if self._action == 'save':\n", "self.run_method('on_update')\n", "if self._action == 'submit':\n", "self.clear_cache()\n", "self.run_method('on_update')\n", "if self._action == 'cancel':\n", "self.notify_update()\n", "self.run_method('on_submit')\n", "self.run_method('on_cancel')\n", "if self._action == 'update_after_submit':\n", "update_global_search(self)\n", "self.check_no_back_links_exist()\n", "self.run_method('on_update_after_submit')\n", "self.save_version()\n", "self.run_method('on_change')\n", "if (self.doctype, self.name) in frappe.flags.currently_saving:\n", "frappe.flags.currently_saving.remove((self.doctype, self.name))\n", "self.latest = None\n" ]
[ "def run_post_save_methods(self):...\n", "\"\"\"docstring\"\"\"\n", "doc_before_save = self.get_doc_before_save()\n", "if self._action == 'save':\n", "self.run_method('on_update')\n", "if self._action == 'submit':\n", "self.clear_cache()\n", "self.run_method('on_update')\n", "if self._action == 'cancel':\n", "self.notify_update()\n", "self.run_method('on_submit')\n", "self.run_method('on_cancel')\n", "if self._action == 'update_after_submit':\n", "update_global_search(self)\n", "self.check_no_back_links_exist()\n", "self.run_method('on_update_after_submit')\n", "self.save_version()\n", "self.run_method('on_change')\n", "if (self.doctype, self.name) in frappe.flags.currently_saving:\n", "frappe.flags.currently_saving.remove((self.doctype, self.name))\n", "self.latest = None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'", "Condition", "Expr'", "Assign'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_87 = FUNC_0(VAR_2, 'page', 1)\n", "return HttpResponseBadRequest('Invalid parameter value')\n", "if not VAR_5.isValidGroup(VAR_20):\n", "VAR_88 = FUNC_0(VAR_2, 'limit', settings.PAGE)\n", "return HttpResponseForbidden('Not a member of Group: %s' % VAR_20)\n", "VAR_25 = 'share_id' in VAR_6 and VAR_241(VAR_6['share_id']) or None\n", "VAR_20 = FUNC_0(VAR_2, 'group', -1)\n", "VAR_106 = tree.marshal_images(VAR_5=conn, VAR_247=orphaned, VAR_11=\n experimenter_id, VAR_246=dataset_id, VAR_25=share_id, VAR_248=\n load_pixels, VAR_20=group_id, VAR_87=page, VAR_250=date, VAR_249=\n thumb_version, VAR_88=limit)\n", "return HttpResponseBadRequest(e.serverStackTrace)\n", "return JsonResponse({'images': VAR_106})\n", "VAR_246 = FUNC_0(VAR_2, 'id', None)\n", "VAR_247 = FUNC_3(VAR_2, 'orphaned', False)\n", "VAR_248 = FUNC_3(VAR_2, 'sizeXYZ', False)\n", "VAR_249 = FUNC_3(VAR_2, 'thumbVersion', False)\n", "VAR_250 = FUNC_3(VAR_2, 'date', False)\n", "VAR_11 = FUNC_0(VAR_2, 'experimenter_id', -1)\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "page = get_long_or_default(request, 'page', 1)\n", "return HttpResponseBadRequest('Invalid parameter value')\n", "if not conn.isValidGroup(group_id):\n", "limit = get_long_or_default(request, 'limit', settings.PAGE)\n", "return HttpResponseForbidden('Not a member of Group: %s' % group_id)\n", "share_id = 'share_id' in kwargs and long(kwargs['share_id']) or None\n", "group_id = get_long_or_default(request, 'group', -1)\n", "images = tree.marshal_images(conn=conn, orphaned=orphaned, experimenter_id=\n experimenter_id, dataset_id=dataset_id, share_id=share_id, load_pixels=\n load_pixels, group_id=group_id, page=page, date=date, thumb_version=\n thumb_version, limit=limit)\n", "return HttpResponseBadRequest(e.serverStackTrace)\n", "return JsonResponse({'images': images})\n", "dataset_id = get_long_or_default(request, 'id', None)\n", "orphaned = get_bool_or_default(request, 'orphaned', False)\n", "load_pixels = get_bool_or_default(request, 'sizeXYZ', False)\n", "thumb_version = get_bool_or_default(request, 'thumbVersion', False)\n", "date = get_bool_or_default(request, 'date', False)\n", "experimenter_id = get_long_or_default(request, 'experimenter_id', -1)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'", "Condition", "Assign'", "Return'", "Assign'", "Assign'", "Assign'", "Return'", "Return'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'" ]
[ "@staticmethod...\n", "VAR_26 = VAR_13(VAR_24.text)\n", "if not VAR_26:\n" ]
[ "@staticmethod...\n", "is_valid = function(document.text)\n", "if not is_valid:\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Condition" ]
[ "def FUNC_36(self):...\n", "return self._queryset\n" ]
[ "def _get_queryset(self):...\n", "return self._queryset\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_11(VAR_37):...\n", "VAR_4, VAR_39 = VAR_37\n", "VAR_25.write(VAR_4)\n", "return VAR_39\n" ]
[ "def write_to(r):...\n", "data, response = r\n", "output_stream.write(data)\n", "return response\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_4():...\n", "return VAR_7\n" ]
[ "def getTrustedEngine():...\n", "return _trusted_engine\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@FUNC_0...\n", "return EventSources(self)\n" ]
[ "@cache_in_self...\n", "return EventSources(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "from security_monkey import app, db\n", "from flask_wtf.csrf import generate_csrf\n", "from security_monkey.auth.models import RBACRole\n", "from security_monkey.decorators import crossdomain\n", "from flask_restful import fields, marshal, Resource, reqparse\n", "from flask_login import current_user\n", "VAR_0 = ['https://{}:{}'.format(app.config.get('FQDN'), app.config.get(\n 'WEB_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.config\n .get('API_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.\n config.get('NGINX_PORT')), 'https://{}:80'.format(app.config.get('FQDN'))]\n", "VAR_1 = {'id': fields.Integer, 'date_created': fields.String,\n 'date_last_ephemeral_change': fields.String, 'active': fields.Boolean,\n 'item_id': fields.Integer}\n", "VAR_2 = {'id': fields.Integer, 'region': fields.String, 'name': fields.String}\n", "VAR_3 = {'id': fields.Integer, 'score': fields.Integer, 'issue': fields.\n String, 'notes': fields.String, 'justified': fields.Boolean,\n 'justification': fields.String, 'justified_date': fields.String,\n 'item_id': fields.Integer}\n", "VAR_4 = {'id': fields.Integer, 'revision_id': fields.Integer,\n 'date_created': fields.String, 'text': fields.String}\n", "VAR_5 = {'id': fields.Integer, 'date_created': fields.String, 'text':\n fields.String, 'item_id': fields.Integer}\n", "VAR_6 = {'daily_audit_email': fields.Boolean, 'change_reports': fields.String}\n", "VAR_7 = {'id': fields.Integer, 'name': fields.String, 'identifier': fields.\n String, 'notes': fields.String, 'active': fields.Boolean, 'third_party':\n fields.Boolean, 'account_type': fields.String}\n", "VAR_8 = {'id': fields.Integer, 'active': fields.Boolean, 'email': fields.\n String, 'role': fields.String, 'confirmed_at': fields.String,\n 'daily_audit_email': fields.Boolean, 'change_reports': fields.String,\n 'last_login_at': fields.String, 'current_login_at': fields.String,\n 'login_count': fields.Integer, 'last_login_ip': fields.String,\n 'current_login_ip': fields.String}\n", "VAR_9 = {'id': fields.Integer, 'name': fields.String, 'description': fields\n .String}\n", "VAR_10 = {'id': fields.Integer, 'name': fields.String, 'notes': fields.\n String, 'cidr': fields.String}\n", "VAR_11 = {'id': fields.Integer, 'prefix': fields.String, 'notes': fields.String\n }\n", "VAR_12 = {'id': fields.Integer, 'disabled': fields.Boolean, 'issue_text':\n fields.String}\n", "VAR_13 = {'id': fields.Integer, 'name': fields.String}\n", "def __init__(self):...\n", "self.reqparse = reqparse.RequestParser()\n", "super(CLASS_0, self).__init__()\n", "self.auth_dict = dict()\n", "if current_user.is_authenticated():\n", "VAR_15 = []\n", "if app.config.get('FRONTED_BY_NGINX'):\n", "for role in current_user.roles:\n", "VAR_16 = 'https://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'NGINX_PORT'), '/login')\n", "VAR_16 = 'http://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'API_PORT'), '/login')\n", "VAR_15.append(marshal(role.__dict__, VAR_9))\n", "VAR_15.append({'name': current_user.role})\n", "self.auth_dict = {'authenticated': False, 'user': None, 'url': VAR_16}\n", "for role in RBACRole.roles[current_user.role].get_parents():\n", "@app.after_request...\n", "VAR_15.append({'name': role.name})\n", "self.auth_dict = {'authenticated': True, 'user': current_user.email,\n 'roles': VAR_15}\n", "VAR_14.set_cookie('XSRF-COOKIE', generate_csrf())\n", "return VAR_14\n" ]
[ "from security_monkey import app, db\n", "from flask_wtf.csrf import generate_csrf\n", "from security_monkey.auth.models import RBACRole\n", "from security_monkey.decorators import crossdomain\n", "from flask_restful import fields, marshal, Resource, reqparse\n", "from flask_login import current_user\n", "ORIGINS = ['https://{}:{}'.format(app.config.get('FQDN'), app.config.get(\n 'WEB_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.config\n .get('API_PORT')), 'https://{}:{}'.format(app.config.get('FQDN'), app.\n config.get('NGINX_PORT')), 'https://{}:80'.format(app.config.get('FQDN'))]\n", "REVISION_FIELDS = {'id': fields.Integer, 'date_created': fields.String,\n 'date_last_ephemeral_change': fields.String, 'active': fields.Boolean,\n 'item_id': fields.Integer}\n", "ITEM_FIELDS = {'id': fields.Integer, 'region': fields.String, 'name':\n fields.String}\n", "AUDIT_FIELDS = {'id': fields.Integer, 'score': fields.Integer, 'issue':\n fields.String, 'notes': fields.String, 'justified': fields.Boolean,\n 'justification': fields.String, 'justified_date': fields.String,\n 'item_id': fields.Integer}\n", "REVISION_COMMENT_FIELDS = {'id': fields.Integer, 'revision_id': fields.\n Integer, 'date_created': fields.String, 'text': fields.String}\n", "ITEM_COMMENT_FIELDS = {'id': fields.Integer, 'date_created': fields.String,\n 'text': fields.String, 'item_id': fields.Integer}\n", "USER_SETTINGS_FIELDS = {'daily_audit_email': fields.Boolean,\n 'change_reports': fields.String}\n", "ACCOUNT_FIELDS = {'id': fields.Integer, 'name': fields.String, 'identifier':\n fields.String, 'notes': fields.String, 'active': fields.Boolean,\n 'third_party': fields.Boolean, 'account_type': fields.String}\n", "USER_FIELDS = {'id': fields.Integer, 'active': fields.Boolean, 'email':\n fields.String, 'role': fields.String, 'confirmed_at': fields.String,\n 'daily_audit_email': fields.Boolean, 'change_reports': fields.String,\n 'last_login_at': fields.String, 'current_login_at': fields.String,\n 'login_count': fields.Integer, 'last_login_ip': fields.String,\n 'current_login_ip': fields.String}\n", "ROLE_FIELDS = {'id': fields.Integer, 'name': fields.String, 'description':\n fields.String}\n", "WHITELIST_FIELDS = {'id': fields.Integer, 'name': fields.String, 'notes':\n fields.String, 'cidr': fields.String}\n", "IGNORELIST_FIELDS = {'id': fields.Integer, 'prefix': fields.String, 'notes':\n fields.String}\n", "AUDITORSETTING_FIELDS = {'id': fields.Integer, 'disabled': fields.Boolean,\n 'issue_text': fields.String}\n", "ITEM_LINK_FIELDS = {'id': fields.Integer, 'name': fields.String}\n", "def __init__(self):...\n", "self.reqparse = reqparse.RequestParser()\n", "super(AuthenticatedService, self).__init__()\n", "self.auth_dict = dict()\n", "if current_user.is_authenticated():\n", "roles_marshal = []\n", "if app.config.get('FRONTED_BY_NGINX'):\n", "for role in current_user.roles:\n", "url = 'https://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'NGINX_PORT'), '/login')\n", "url = 'http://{}:{}{}'.format(app.config.get('FQDN'), app.config.get(\n 'API_PORT'), '/login')\n", "roles_marshal.append(marshal(role.__dict__, ROLE_FIELDS))\n", "roles_marshal.append({'name': current_user.role})\n", "self.auth_dict = {'authenticated': False, 'user': None, 'url': url}\n", "for role in RBACRole.roles[current_user.role].get_parents():\n", "@app.after_request...\n", "roles_marshal.append({'name': role.name})\n", "self.auth_dict = {'authenticated': True, 'user': current_user.email,\n 'roles': roles_marshal}\n", "response.set_cookie('XSRF-COOKIE', generate_csrf())\n", "return response\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Condition", "For", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "For", "Condition", "Expr'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_23(VAR_48, VAR_49=VAR_29):...\n", "return VAR_49 is not None and (len(VAR_48) > VAR_49 and '%s...' % VAR_48[:\n max(0, VAR_49 - 3)]) or VAR_48\n" ]
[ "def trim_url(x, limit=trim_url_limit):...\n", "return limit is not None and (len(x) > limit and '%s...' % x[:max(0, limit -\n 3)]) or x\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_6(self):...\n", "return self.kw_component\n" ]
[ "def get_kwargs(self):...\n", "return self.kw_component\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_14(VAR_7, VAR_8):...\n", "VAR_10 = [db.Books.timestamp.desc()]\n", "if VAR_7 == 'stored':\n", "VAR_7 = VAR_87.get_view_property(VAR_8, 'stored')\n", "VAR_87.set_view_property(VAR_8, 'stored', VAR_7)\n", "if VAR_7 == 'pubnew':\n", "VAR_10 = [db.Books.pubdate.desc()]\n", "if VAR_7 == 'pubold':\n", "VAR_10 = [db.Books.pubdate]\n", "if VAR_7 == 'abc':\n", "VAR_10 = [db.Books.sort]\n", "if VAR_7 == 'zyx':\n", "VAR_10 = [db.Books.sort.desc()]\n", "if VAR_7 == 'new':\n", "VAR_10 = [db.Books.timestamp.desc()]\n", "if VAR_7 == 'old':\n", "VAR_10 = [db.Books.timestamp]\n", "if VAR_7 == 'authaz':\n", "VAR_10 = [db.Books.author_sort.asc(), db.Series.name, db.Books.series_index]\n", "if VAR_7 == 'authza':\n", "VAR_10 = [db.Books.author_sort.desc(), db.Series.name.desc(), db.Books.\n series_index.desc()]\n", "if VAR_7 == 'seriesasc':\n", "VAR_10 = [db.Books.series_index.asc()]\n", "if VAR_7 == 'seriesdesc':\n", "VAR_10 = [db.Books.series_index.desc()]\n", "if VAR_7 == 'hotdesc':\n", "VAR_10 = [func.count(ub.Downloads.book_id).desc()]\n", "if VAR_7 == 'hotasc':\n", "VAR_10 = [func.count(ub.Downloads.book_id).asc()]\n", "if VAR_7 is None:\n", "VAR_7 = 'new'\n", "return VAR_10, VAR_7\n" ]
[ "def get_sort_function(sort, data):...\n", "order = [db.Books.timestamp.desc()]\n", "if sort == 'stored':\n", "sort = current_user.get_view_property(data, 'stored')\n", "current_user.set_view_property(data, 'stored', sort)\n", "if sort == 'pubnew':\n", "order = [db.Books.pubdate.desc()]\n", "if sort == 'pubold':\n", "order = [db.Books.pubdate]\n", "if sort == 'abc':\n", "order = [db.Books.sort]\n", "if sort == 'zyx':\n", "order = [db.Books.sort.desc()]\n", "if sort == 'new':\n", "order = [db.Books.timestamp.desc()]\n", "if sort == 'old':\n", "order = [db.Books.timestamp]\n", "if sort == 'authaz':\n", "order = [db.Books.author_sort.asc(), db.Series.name, db.Books.series_index]\n", "if sort == 'authza':\n", "order = [db.Books.author_sort.desc(), db.Series.name.desc(), db.Books.\n series_index.desc()]\n", "if sort == 'seriesasc':\n", "order = [db.Books.series_index.asc()]\n", "if sort == 'seriesdesc':\n", "order = [db.Books.series_index.desc()]\n", "if sort == 'hotdesc':\n", "order = [func.count(ub.Downloads.book_id).desc()]\n", "if sort == 'hotasc':\n", "order = [func.count(ub.Downloads.book_id).asc()]\n", "if sort is None:\n", "sort = 'new'\n", "return order, sort\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Assign'", "Expr'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Return'" ]
[ "def FUNC_0(self):...\n", "return self.post()\n" ]
[ "def get(self):...\n", "return self.post()\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@wraps(VAR_4)...\n", "if VAR_87.role_viewer():\n", "return VAR_4(*VAR_49, **kwargs)\n", "abort(403)\n" ]
[ "@wraps(f)...\n", "if current_user.role_viewer():\n", "return f(*args, **kwargs)\n", "abort(403)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Expr'" ]
[ "def FUNC_69(VAR_12, VAR_29):...\n", "VAR_94 = re.search('([a-zA-Z0-9-]+)=%s', VAR_29, re.IGNORECASE | re.UNICODE)\n", "if VAR_94:\n", "return VAR_94.group(1)\n" ]
[ "def extract_dynamic_field_from_filter(user, filtr):...\n", "match = re.search('([a-zA-Z0-9-]+)=%s', filtr, re.IGNORECASE | re.UNICODE)\n", "if match:\n", "return match.group(1)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Condition", "Return'" ]
[ "def FUNC_1(self, VAR_5, VAR_7, VAR_8=None):...\n", "\"\"\"docstring\"\"\"\n", "if self.ignore_login_fail:\n", "return HttpResponse('Connection Failed')\n", "if self.login_redirect is not None:\n", "return super(CLASS_0, self).on_not_logged_in(VAR_5, VAR_7, VAR_8)\n", "VAR_7 = reverse(self.login_redirect)\n" ]
[ "def on_not_logged_in(self, request, url, error=None):...\n", "\"\"\"docstring\"\"\"\n", "if self.ignore_login_fail:\n", "return HttpResponse('Connection Failed')\n", "if self.login_redirect is not None:\n", "return super(login_required, self).on_not_logged_in(request, url, error)\n", "url = reverse(self.login_redirect)\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Condition", "Return'", "Assign'" ]
[ "def FUNC_0(self, VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "return yaml.load(self.vault.decrypt(VAR_1))\n" ]
[ "def load(self, stream):...\n", "\"\"\"docstring\"\"\"\n", "return yaml.load(self.vault.decrypt(stream))\n" ]
[ 0, 0, 5 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_9(self, VAR_5):...\n", "for VAR_20 in self._subexprs:\n", "return 0\n", "VAR_20(VAR_5)\n", "return 1\n" ]
[ "def _exists(self, econtext):...\n", "for expr in self._subexprs:\n", "return 0\n", "expr(econtext)\n", "return 1\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "For", "Return'", "Expr'", "Return'" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "VAR_10 = SlashSeparatedCourseKey.from_deprecated_string(VAR_10)\n", "VAR_121 = RegistrationCodeRedemption.objects.select_related('registration_code'\n ).filter(registration_code__course_id=course_id)\n", "VAR_122 = []\n", "if VAR_121.exists():\n", "VAR_184 = [VAR_82.registration_code.code for VAR_82 in VAR_121]\n", "VAR_34 = 'spent'\n", "VAR_122 = CourseRegistrationCode.objects.filter(VAR_10=course_id, code__in=\n redeemed_registration_codes).order_by('invoice_item__invoice__company_name'\n ).select_related('invoice_item__invoice')\n", "return FUNC_34('Spent_Registration_Codes.csv', VAR_122, VAR_34)\n", "VAR_88 = VAR_9.POST['spent_company_name']\n", "if VAR_88:\n", "VAR_122 = VAR_122.filter(invoice_item__invoice__company_name=company_name)\n" ]
[ "@ensure_csrf_cookie...\n", "\"\"\"docstring\"\"\"\n", "course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)\n", "code_redemption_set = RegistrationCodeRedemption.objects.select_related(\n 'registration_code').filter(registration_code__course_id=course_id)\n", "spent_codes_list = []\n", "if code_redemption_set.exists():\n", "redeemed_registration_codes = [code.registration_code.code for code in\n code_redemption_set]\n", "csv_type = 'spent'\n", "spent_codes_list = CourseRegistrationCode.objects.filter(course_id=\n course_id, code__in=redeemed_registration_codes).order_by(\n 'invoice_item__invoice__company_name').select_related(\n 'invoice_item__invoice')\n", "return registration_codes_csv('Spent_Registration_Codes.csv',\n spent_codes_list, csv_type)\n", "company_name = request.POST['spent_company_name']\n", "if company_name:\n", "spent_codes_list = spent_codes_list.filter(invoice_item__invoice__company_name\n =company_name)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Return'", "Assign'", "Condition", "Assign'" ]
[ "def FUNC_12(self, VAR_0, VAR_13):...\n", "VAR_33 = 'C:\\\\' if utils.is_windows else '/'\n", "VAR_25 = VAR_13(VAR_33)\n", "assert not VAR_25.parent\n" ]
[ "def test_root_dir(self, tmpdir, parser):...\n", "root_dir = 'C:\\\\' if utils.is_windows else '/'\n", "parsed = parser(root_dir)\n", "assert not parsed.parent\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assert'" ]
[ "def FUNC_15(self, VAR_31, VAR_32):...\n", "return CLASS_7(VAR_31, VAR_32)\n" ]
[ "def createErrorInfo(self, err, position):...\n", "return ErrorInfo(err, position)\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "def FUNC_5(self, VAR_7):...\n", "\"\"\"docstring\"\"\"\n", "VAR_12, VAR_13 = self.make_request('GET', '/joined_groups'.encode('ascii'),\n VAR_7=access_token)\n", "self.assertEqual(200, int(VAR_13.result['code']), msg=channel.result['body'])\n", "return VAR_13.json_body['groups']\n" ]
[ "def _get_groups_user_is_in(self, access_token):...\n", "\"\"\"docstring\"\"\"\n", "request, channel = self.make_request('GET', '/joined_groups'.encode('ascii'\n ), access_token=access_token)\n", "self.assertEqual(200, int(channel.result['code']), msg=channel.result['body'])\n", "return channel.json_body['groups']\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Expr'", "Return'" ]
[ "async def FUNC_18(VAR_3, **VAR_4):...\n", "VAR_17 = RedirectException(b'/no/over/there', 304)\n", "VAR_17.cookies.append(b'session=yespls')\n" ]
[ "async def callback(request, **kwargs):...\n", "e = RedirectException(b'/no/over/there', 304)\n", "e.cookies.append(b'session=yespls')\n" ]
[ 0, 0, 0 ]
[ "AsyncFunctionDef'", "Assign'", "Expr'" ]
[ "@login_required...\n", "if not is_ajax(VAR_0):\n", "return Http404()\n", "VAR_6 = TopicNotification.objects.for_access(VAR_0.user).order_by('is_read',\n '-date').with_related_data()\n", "VAR_6 = VAR_6[:settings.ST_NOTIFICATIONS_PER_PAGE]\n", "VAR_6 = [{'user': escape(n.comment.user.st.nickname), 'action': n.action,\n 'title': escape(n.topic.title), 'url': n.get_absolute_url(), 'is_read':\n n.is_read} for n in VAR_6]\n", "return HttpResponse(json.dumps({'n': VAR_6}), content_type='application/json')\n" ]
[ "@login_required...\n", "if not is_ajax(request):\n", "return Http404()\n", "notifications = TopicNotification.objects.for_access(request.user).order_by(\n 'is_read', '-date').with_related_data()\n", "notifications = notifications[:settings.ST_NOTIFICATIONS_PER_PAGE]\n", "notifications = [{'user': escape(n.comment.user.st.nickname), 'action': n.\n action, 'title': escape(n.topic.title), 'url': n.get_absolute_url(),\n 'is_read': n.is_read} for n in notifications]\n", "return HttpResponse(json.dumps({'n': notifications}), content_type=\n 'application/json')\n" ]
[ 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_12(VAR_2, VAR_3, VAR_5, VAR_10, VAR_11, VAR_12, VAR_13=None,...\n", "\"\"\"docstring\"\"\"\n", "VAR_4 = saved_model_utils.get_meta_graph_def(VAR_2, VAR_3)\n", "VAR_24 = FUNC_2(VAR_4, VAR_5)\n", "for input_key_name in VAR_10.keys():\n", "if input_key_name not in VAR_24:\n", "VAR_36 = {VAR_24[key].name: tensor for key, tensor in VAR_10.items()}\n", "VAR_25 = FUNC_3(VAR_4, VAR_5)\n", "VAR_37 = sorted(VAR_25.keys())\n", "VAR_38 = [VAR_25[tensor_key].name for tensor_key in VAR_37]\n", "if VAR_14:\n", "print('Initializing TPU System ...')\n", "loader.load(VAR_77, VAR_3.split(','), VAR_2)\n", "VAR_77.run(tpu.initialize_system())\n", "if VAR_15:\n", "VAR_77 = local_cli_wrapper.LocalCLIDebugWrapperSession(VAR_77)\n", "VAR_66 = VAR_77.run(VAR_38, feed_dict=inputs_feed_dict)\n", "for i, output in enumerate(VAR_66):\n", "VAR_78 = VAR_37[i]\n", "print(\"\"\"Result for output key %s:\n%s\"\"\" % (VAR_78, output))\n", "if VAR_11:\n", "if not os.path.isdir(VAR_11):\n", "os.makedirs(VAR_11)\n", "VAR_79 = os.path.join(VAR_11, VAR_78 + '.npy')\n", "if not VAR_12 and os.path.exists(VAR_79):\n", "np.save(VAR_79, output)\n", "print('Output %s is saved to %s' % (VAR_78, VAR_79))\n" ]
[ "def run_saved_model_with_feed_dict(saved_model_dir, tag_set,...\n", "\"\"\"docstring\"\"\"\n", "meta_graph_def = saved_model_utils.get_meta_graph_def(saved_model_dir, tag_set)\n", "inputs_tensor_info = _get_inputs_tensor_info_from_meta_graph_def(meta_graph_def\n , signature_def_key)\n", "for input_key_name in input_tensor_key_feed_dict.keys():\n", "if input_key_name not in inputs_tensor_info:\n", "inputs_feed_dict = {inputs_tensor_info[key].name: tensor for key, tensor in\n input_tensor_key_feed_dict.items()}\n", "outputs_tensor_info = _get_outputs_tensor_info_from_meta_graph_def(\n meta_graph_def, signature_def_key)\n", "output_tensor_keys_sorted = sorted(outputs_tensor_info.keys())\n", "output_tensor_names_sorted = [outputs_tensor_info[tensor_key].name for\n tensor_key in output_tensor_keys_sorted]\n", "if init_tpu:\n", "print('Initializing TPU System ...')\n", "loader.load(sess, tag_set.split(','), saved_model_dir)\n", "sess.run(tpu.initialize_system())\n", "if tf_debug:\n", "sess = local_cli_wrapper.LocalCLIDebugWrapperSession(sess)\n", "outputs = sess.run(output_tensor_names_sorted, feed_dict=inputs_feed_dict)\n", "for i, output in enumerate(outputs):\n", "output_tensor_key = output_tensor_keys_sorted[i]\n", "print(\"\"\"Result for output key %s:\n%s\"\"\" % (output_tensor_key, output))\n", "if outdir:\n", "if not os.path.isdir(outdir):\n", "os.makedirs(outdir)\n", "output_full_path = os.path.join(outdir, output_tensor_key + '.npy')\n", "if not overwrite_flag and os.path.exists(output_full_path):\n", "np.save(output_full_path, output)\n", "print('Output %s is saved to %s' % (output_tensor_key, output_full_path))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "For", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "For", "Assign'", "Expr'", "Condition", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Expr'" ]
[ "import pytest\n", "from werkzeug.routing import BuildError\n", "from flask_unchained.bundles.controller import Controller, Resource\n", "from flask_unchained.bundles.controller.utils import controller_name, get_param_tuples, join, method_name_to_url, url_for, _validate_redirect_url\n", "from py_meta_utils import deep_getattr\n", "def FUNC_0():...\n", "VAR_0 = {'a': 'clsdict'}\n", "VAR_4 = 'first'\n", "VAR_5 = 'first'\n", "VAR_5 = 'second'\n", "VAR_6 = 'second'\n", "VAR_1 = CLASS_6, CLASS_7\n", "assert deep_getattr(VAR_0, VAR_1, 'a') == 'clsdict'\n", "assert deep_getattr(VAR_0, VAR_1, 'b') == 'first'\n", "assert deep_getattr(VAR_0, VAR_1, 'c') == 'second'\n", "deep_getattr(VAR_0, VAR_1, 'd')\n", "assert deep_getattr(VAR_0, VAR_1, 'a', 'default') == 'clsdict'\n", "assert deep_getattr(VAR_0, VAR_1, 'b', 'default') == 'first'\n", "assert deep_getattr(VAR_0, VAR_1, 'c', 'default') == 'second'\n", "assert deep_getattr(VAR_0, VAR_1, 'd', 'default') == 'default'\n", "def FUNC_1(self):...\n", "assert controller_name(CLASS_8) == 'user'\n", "def FUNC_2(self):...\n", "assert controller_name(CLASS_9) == 'api'\n", "def FUNC_3(self):...\n", "assert controller_name(CLASS_10) == 'some'\n", "def FUNC_4(self):...\n", "assert controller_name(CLASS_11) == 'more_than_one_word'\n", "def FUNC_5(self):...\n", "assert controller_name(CLASS_12) == 'user'\n", "def FUNC_6(self):...\n", "assert controller_name(CLASS_13) == 'role'\n", "def FUNC_7(self):...\n", "assert controller_name(CLASS_14) == 'role_view_controller'\n", "def FUNC_8(self):...\n", "assert controller_name(CLASS_15) == 'some_ctrl'\n", "def FUNC_9(self):...\n", "assert get_param_tuples('<int:id>') == [('int', 'id')]\n", "def FUNC_10(self):...\n", "assert get_param_tuples(None) == []\n", "def FUNC_11(self):...\n", "VAR_7 = '/users/<int:user_id>/roles/<string:slug>'\n", "assert get_param_tuples(VAR_7) == [('int', 'user_id'), ('string', 'slug')]\n", "def FUNC_12(self):...\n", "assert url_for('/foobar') == '/foobar'\n", "def FUNC_13(self):...\n", "assert url_for(None) is None\n", "def FUNC_14(self, VAR_2):...\n", "VAR_2.config.from_mapping({'MY_KEY': '/my-key'})\n", "assert url_for('MY_KEY') == '/my-key'\n", "def FUNC_15(self, VAR_2):...\n", "VAR_2.config.from_mapping({'MY_KEY': 'some.endpoint'})\n", "assert url_for('MY_KEY')\n", "VAR_2.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('MY_KEY') == '/some-endpoint'\n", "def FUNC_16(self, VAR_2):...\n", "assert url_for('some.endpoint')\n", "VAR_2.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('some.endpoint') == '/some-endpoint'\n", "def FUNC_17(self, VAR_2):...\n", "def FUNC_29(self):...\n", "VAR_2.add_url_rule('/about-us', endpoint='site_controller.about_us')\n", "assert url_for('about_us', _cls=SiteController) == '/about-us'\n", "def FUNC_18(self, VAR_2):...\n", "def FUNC_30(self, VAR_8):...\n", "VAR_2.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "assert url_for('get', VAR_8=1, _cls=SiteResource) == '/sites/1'\n", "VAR_2.add_url_rule('/foo/<string:slug>', endpoint='some.endpoint')\n", "assert url_for('some.endpoint', slug='hi') == '/foo/hi'\n", "def FUNC_19(self, VAR_2):...\n", "def FUNC_30(self, VAR_8):...\n", "VAR_2.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "url_for('delete', VAR_8=1, _cls=SiteResource)\n", "def FUNC_13(self):...\n", "assert join(None) == '/'\n", "assert join(None, None, '', 0) == '/'\n", "def FUNC_20(self):...\n", "assert join('/', 'foo', None, 'bar', '', 'baz') == '/foo/bar/baz'\n", "def FUNC_21(self):...\n", "assert join('/', '/foo', '/', '/bar') == '/foo/bar'\n", "def FUNC_22(self):...\n", "assert join('/', '/') == '/'\n", "assert join(None, '/') == '/'\n", "assert join('/', None) == '/'\n", "def FUNC_23(self):...\n", "assert join('/foo/bar/') == '/foo/bar'\n", "assert join('/foo/bar/', None) == '/foo/bar'\n", "assert join('/foo/bar/', '/') == '/foo/bar'\n", "assert join('/foo', 'bar/') == '/foo/bar'\n", "def FUNC_24(self):...\n", "assert join('/', trailing_slash=True) == '/'\n", "assert join('/foo', 'baz', None, trailing_slash=True) == '/foo/baz/'\n", "assert join('/foo', 'baz/', trailing_slash=True) == '/foo/baz/'\n", "def FUNC_9(self):...\n", "assert method_name_to_url('fooBar') == '/foo-bar'\n", "assert method_name_to_url('foo_bar') == '/foo-bar'\n", "assert method_name_to_url('fooBar_baz') == '/foo-bar-baz'\n", "assert method_name_to_url('_FooBar_baz-booFoo_') == '/foo-bar-baz-boo-foo'\n", "def FUNC_25(self):...\n", "assert _validate_redirect_url(None) is False\n", "assert _validate_redirect_url(' ') is False\n", "def FUNC_26(self, VAR_2, VAR_3):...\n", "VAR_3.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://fail.com') is False\n", "VAR_3.undo()\n", "@pytest.mark.options(EXTERNAL_SERVER_NAME='works.com')...\n", "VAR_3.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://works.com') is True\n", "VAR_3.undo()\n", "def FUNC_28(self, VAR_2, VAR_3):...\n", "VAR_3.setattr('flask.request.host_url', 'http://example.com')\n", "VAR_9 = _validate_redirect_url('http://works.com', _external_host='works.com')\n", "assert VAR_9 is True\n", "VAR_3.undo()\n" ]
[ "import pytest\n", "from werkzeug.routing import BuildError\n", "from flask_unchained.bundles.controller import Controller, Resource\n", "from flask_unchained.bundles.controller.utils import controller_name, get_param_tuples, join, method_name_to_url, url_for, _validate_redirect_url\n", "from py_meta_utils import deep_getattr\n", "def test_deep_getattr():...\n", "clsdict = {'a': 'clsdict'}\n", "a = 'first'\n", "b = 'first'\n", "b = 'second'\n", "c = 'second'\n", "bases = First, Second\n", "assert deep_getattr(clsdict, bases, 'a') == 'clsdict'\n", "assert deep_getattr(clsdict, bases, 'b') == 'first'\n", "assert deep_getattr(clsdict, bases, 'c') == 'second'\n", "deep_getattr(clsdict, bases, 'd')\n", "assert deep_getattr(clsdict, bases, 'a', 'default') == 'clsdict'\n", "assert deep_getattr(clsdict, bases, 'b', 'default') == 'first'\n", "assert deep_getattr(clsdict, bases, 'c', 'default') == 'second'\n", "assert deep_getattr(clsdict, bases, 'd', 'default') == 'default'\n", "def test_it_strips_controller(self):...\n", "assert controller_name(UserController) == 'user'\n", "def test_it_handles_acronyms(self):...\n", "assert controller_name(APIController) == 'api'\n", "def test_it_strips_view(self):...\n", "assert controller_name(SomeView) == 'some'\n", "def test_it_works_with_more_than_one_word(self):...\n", "assert controller_name(MoreThanOneWordController) == 'more_than_one_word'\n", "def test_it_strips_resource(self):...\n", "assert controller_name(UserResource) == 'user'\n", "def test_it_strips_method_view(self):...\n", "assert controller_name(RoleMethodView) == 'role'\n", "def test_it_only_strips_one_suffix(self):...\n", "assert controller_name(RoleViewControllerResource) == 'role_view_controller'\n", "def test_it_works_without_stripping_any_suffixes(self):...\n", "assert controller_name(SomeCtrl) == 'some_ctrl'\n", "def test_it_works(self):...\n", "assert get_param_tuples('<int:id>') == [('int', 'id')]\n", "def test_it_works_on_garbage(self):...\n", "assert get_param_tuples(None) == []\n", "def test_multiple(self):...\n", "path = '/users/<int:user_id>/roles/<string:slug>'\n", "assert get_param_tuples(path) == [('int', 'user_id'), ('string', 'slug')]\n", "def test_it_works_with_already_formed_path(self):...\n", "assert url_for('/foobar') == '/foobar'\n", "def test_it_works_with_garbage(self):...\n", "assert url_for(None) is None\n", "def test_it_works_with_config_keys_returning_path(self, app):...\n", "app.config.from_mapping({'MY_KEY': '/my-key'})\n", "assert url_for('MY_KEY') == '/my-key'\n", "def test_it_works_with_config_keys_returning_endpoints(self, app):...\n", "app.config.from_mapping({'MY_KEY': 'some.endpoint'})\n", "assert url_for('MY_KEY')\n", "app.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('MY_KEY') == '/some-endpoint'\n", "def test_it_works_with_endpoints(self, app):...\n", "assert url_for('some.endpoint')\n", "app.add_url_rule('/some-endpoint', endpoint='some.endpoint')\n", "assert url_for('some.endpoint') == '/some-endpoint'\n", "def test_it_works_with_controller_method_names(self, app):...\n", "def about_us(self):...\n", "app.add_url_rule('/about-us', endpoint='site_controller.about_us')\n", "assert url_for('about_us', _cls=SiteController) == '/about-us'\n", "def test_it_works_with_url_for_kwargs(self, app):...\n", "def get(self, id):...\n", "app.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "assert url_for('get', id=1, _cls=SiteResource) == '/sites/1'\n", "app.add_url_rule('/foo/<string:slug>', endpoint='some.endpoint')\n", "assert url_for('some.endpoint', slug='hi') == '/foo/hi'\n", "def test_it_falls_through_if_class_endpoint_not_found(self, app):...\n", "def get(self, id):...\n", "app.add_url_rule('/sites/<int:id>', endpoint='site_resource.get')\n", "url_for('delete', id=1, _cls=SiteResource)\n", "def test_it_works_with_garbage(self):...\n", "assert join(None) == '/'\n", "assert join(None, None, '', 0) == '/'\n", "def test_it_works_with_partially_valid_input(self):...\n", "assert join('/', 'foo', None, 'bar', '', 'baz') == '/foo/bar/baz'\n", "def test_it_strips_neighboring_slashes(self):...\n", "assert join('/', '/foo', '/', '/bar') == '/foo/bar'\n", "def test_it_doesnt_eat_single_slash(self):...\n", "assert join('/', '/') == '/'\n", "assert join(None, '/') == '/'\n", "assert join('/', None) == '/'\n", "def test_it_strips_trailing_slash(self):...\n", "assert join('/foo/bar/') == '/foo/bar'\n", "assert join('/foo/bar/', None) == '/foo/bar'\n", "assert join('/foo/bar/', '/') == '/foo/bar'\n", "assert join('/foo', 'bar/') == '/foo/bar'\n", "def test_trailing_slash(self):...\n", "assert join('/', trailing_slash=True) == '/'\n", "assert join('/foo', 'baz', None, trailing_slash=True) == '/foo/baz/'\n", "assert join('/foo', 'baz/', trailing_slash=True) == '/foo/baz/'\n", "def test_it_works(self):...\n", "assert method_name_to_url('fooBar') == '/foo-bar'\n", "assert method_name_to_url('foo_bar') == '/foo-bar'\n", "assert method_name_to_url('fooBar_baz') == '/foo-bar-baz'\n", "assert method_name_to_url('_FooBar_baz-booFoo_') == '/foo-bar-baz-boo-foo'\n", "def test_it_fails_on_garbage(self):...\n", "assert _validate_redirect_url(None) is False\n", "assert _validate_redirect_url(' ') is False\n", "def test_it_fails_with_invalid_netloc(self, app, monkeypatch):...\n", "monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://fail.com') is False\n", "monkeypatch.undo()\n", "@pytest.mark.options(EXTERNAL_SERVER_NAME='works.com')...\n", "monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n", "assert _validate_redirect_url('http://works.com') is True\n", "monkeypatch.undo()\n", "def test_it_works_with_explicit_external_host(self, app, monkeypatch):...\n", "monkeypatch.setattr('flask.request.host_url', 'http://example.com')\n", "result = _validate_redirect_url('http://works.com', _external_host='works.com')\n", "assert result is True\n", "monkeypatch.undo()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0 ]
[ "Import'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assert'", "Assert'", "Assert'", "Expr'", "Assert'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assign'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Expr'", "Assert'", "FunctionDef'", "Expr'", "Assert'", "Expr'", "Assert'", "FunctionDef'", "Assert'", "Expr'", "Assert'", "FunctionDef'", "FunctionDef'", "Expr'", "Assert'", "FunctionDef'", "FunctionDef'", "Expr'", "Assert'", "Expr'", "Assert'", "FunctionDef'", "FunctionDef'", "Expr'", "Expr'", "FunctionDef'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "FunctionDef'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "Assert'", "Assert'", "Assert'", "FunctionDef'", "Assert'", "Assert'", "FunctionDef'", "Expr'", "Assert'", "Expr'", "Condition", "Expr'", "Assert'", "Expr'", "FunctionDef'", "Expr'", "Assign'", "Assert'", "Expr'" ]
[ "def FUNC_87(VAR_160, VAR_161, VAR_162):...\n", "\"\"\"docstring\"\"\"\n", "VAR_49[VAR_160].append(_('user \"{user}\" in row# {row}').format(VAR_19=_user,\n row=row_count))\n" ]
[ "def build_row_errors(key, _user, row_count):...\n", "\"\"\"docstring\"\"\"\n", "row_errors[key].append(_('user \"{user}\" in row# {row}').format(user=_user,\n row=row_count))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "def FUNC_30(self, VAR_28, VAR_29, VAR_12):...\n", "if VAR_29 == 'public':\n", "VAR_103 = type(VAR_12) == bool and type(VAR_28) == int\n", "if VAR_103:\n", "return self.playlistdb.setPublic(VAR_15=self.getUserId(), VAR_28=plid,\n VAR_24=value)\n" ]
[ "def api_changeplaylist(self, plid, attribute, value):...\n", "if attribute == 'public':\n", "is_valid = type(value) == bool and type(plid) == int\n", "if is_valid:\n", "return self.playlistdb.setPublic(userid=self.getUserId(), plid=plid, public\n =value)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Condition", "Return'" ]
[ "from __future__ import annotations\n", "from dataclasses import dataclass, field\n", "from datetime import date, datetime\n", "from typing import Any, Dict, List, Optional, Union, cast\n", "from .an_enum import AnEnum\n", "from .different_enum import DifferentEnum\n", "\"\"\" A Model for testing all the ways custom objects can be used \"\"\"\n", "VAR_1: AnEnum\n", "VAR_5: Union[datetime, date]\n", "VAR_2: date\n", "VAR_6: Optional[List[List[DifferentEnum]]] = field(default_factory=lambda :\n cast(Optional[List[List[DifferentEnum]]], []))\n", "VAR_3: Optional[Dict[Any, Any]] = field(default_factory=lambda : cast(\n Optional[Dict[Any, Any]], {}))\n", "def FUNC_0(self) ->Dict[str, Any]:...\n", "VAR_1 = self.an_enum_value.value\n", "if isinstance(self.a_camel_date_time, datetime):\n", "VAR_5 = self.a_camel_date_time.isoformat()\n", "VAR_5 = self.a_camel_date_time.isoformat()\n", "VAR_2 = self.a_date.isoformat()\n", "if self.nested_list_of_enums is None:\n", "VAR_6 = None\n", "VAR_6 = []\n", "VAR_3 = self.some_dict\n", "for nested_list_of_enums_item_data in self.nested_list_of_enums:\n", "return {'an_enum_value': VAR_1, 'aCamelDateTime': VAR_5, 'a_date': VAR_2,\n 'nested_list_of_enums': VAR_6, 'some_dict': VAR_3}\n", "VAR_7 = []\n", "for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n", "VAR_8 = nested_list_of_enums_item_item_data.value\n", "VAR_6.append(VAR_7)\n", "VAR_7.append(VAR_8)\n" ]
[ "from __future__ import annotations\n", "from dataclasses import dataclass, field\n", "from datetime import date, datetime\n", "from typing import Any, Dict, List, Optional, Union, cast\n", "from .an_enum import AnEnum\n", "from .different_enum import DifferentEnum\n", "\"\"\" A Model for testing all the ways custom objects can be used \"\"\"\n", "an_enum_value: AnEnum\n", "a_camel_date_time: Union[datetime, date]\n", "a_date: date\n", "nested_list_of_enums: Optional[List[List[DifferentEnum]]] = field(\n default_factory=lambda : cast(Optional[List[List[DifferentEnum]]], []))\n", "some_dict: Optional[Dict[Any, Any]] = field(default_factory=lambda : cast(\n Optional[Dict[Any, Any]], {}))\n", "def to_dict(self) ->Dict[str, Any]:...\n", "an_enum_value = self.an_enum_value.value\n", "if isinstance(self.a_camel_date_time, datetime):\n", "a_camel_date_time = self.a_camel_date_time.isoformat()\n", "a_camel_date_time = self.a_camel_date_time.isoformat()\n", "a_date = self.a_date.isoformat()\n", "if self.nested_list_of_enums is None:\n", "nested_list_of_enums = None\n", "nested_list_of_enums = []\n", "some_dict = self.some_dict\n", "for nested_list_of_enums_item_data in self.nested_list_of_enums:\n", "return {'an_enum_value': an_enum_value, 'aCamelDateTime': a_camel_date_time,\n 'a_date': a_date, 'nested_list_of_enums': nested_list_of_enums,\n 'some_dict': some_dict}\n", "nested_list_of_enums_item = []\n", "for nested_list_of_enums_item_item_data in nested_list_of_enums_item_data:\n", "nested_list_of_enums_item_item = nested_list_of_enums_item_item_data.value\n", "nested_list_of_enums.append(nested_list_of_enums_item)\n", "nested_list_of_enums_item.append(nested_list_of_enums_item_item)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 5, 5, 0, 5, 0, 0, 5, 0, 0, 0, 0, 0, 0, 5, 0, 5, 0, 0, 0, 0, 0 ]
[ "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "ImportFrom'", "Expr'", "AnnAssign'", "AnnAssign'", "AnnAssign'", "AnnAssign'", "AnnAssign'", "FunctionDef'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "For", "Return'", "Assign'", "For", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_20(self):...\n", "\"\"\"docstring\"\"\"\n", "return u'[%s]' % self.ticket_for_url\n" ]
[ "def _get_ticket(self):...\n", "\"\"\"docstring\"\"\"\n", "return u'[%s]' % self.ticket_for_url\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Return'" ]
[ "def FUNC_77(self):...\n", "VAR_78 = frappe.get_all('ToDo', fields=['owner'], filters={'reference_type':\n self.doctype, 'reference_name': self.name, 'status': ('!=', 'Cancelled')})\n", "VAR_79 = set([assignment.owner for assignment in VAR_78])\n", "return VAR_79\n" ]
[ "def get_assigned_users(self):...\n", "assignments = frappe.get_all('ToDo', fields=['owner'], filters={\n 'reference_type': self.doctype, 'reference_name': self.name, 'status':\n ('!=', 'Cancelled')})\n", "users = set([assignment.owner for assignment in assignments])\n", "return users\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_14():...\n", "VAR_23.component_instance.json_args = VAR_34\n", "VAR_23.component_instance.special_args.extend(VAR_21)\n" ]
[ "def marshall_element_args():...\n", "element.component_instance.json_args = serialized_json_args\n", "element.component_instance.special_args.extend(special_args)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_23(self, VAR_46: str, VAR_47: str) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "if not self.instances or len(self.instances) == 1:\n", "return True\n", "return self.get_instance(VAR_47) == VAR_46\n" ]
[ "def should_handle(self, instance_name: str, key: str) ->bool:...\n", "\"\"\"docstring\"\"\"\n", "if not self.instances or len(self.instances) == 1:\n", "return True\n", "return self.get_instance(key) == instance_name\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Return'" ]
[ "async def FUNC_19(self, VAR_5: str, VAR_11: str, VAR_27: str) ->EventBase:...\n", "\"\"\"docstring\"\"\"\n", "if get_domain_from_id(VAR_27) != VAR_5:\n", "VAR_0.info(\n 'Got /make_leave request for user %r from different origin %s, ignoring',\n VAR_27, VAR_5)\n", "VAR_28 = await self.store.get_room_version_id(VAR_11)\n", "VAR_93 = self.event_builder_factory.new(VAR_28, {'type': VAR_188.Member,\n 'content': {'membership': Membership.LEAVE}, 'room_id': VAR_11,\n 'sender': VAR_27, 'state_key': VAR_27})\n", "VAR_1, VAR_32 = await self.event_creation_handler.create_new_client_event(\n VAR_93=builder)\n", "await self.auth.check_from_context(VAR_28, VAR_1, VAR_32, do_sig_check=False)\n", "VAR_0.warning('Failed to create new leave %r because %s', VAR_1, VAR_170)\n", "return VAR_1\n" ]
[ "async def on_make_leave_request(self, origin: str, room_id: str, user_id: str...\n", "\"\"\"docstring\"\"\"\n", "if get_domain_from_id(user_id) != origin:\n", "logger.info(\n 'Got /make_leave request for user %r from different origin %s, ignoring',\n user_id, origin)\n", "room_version = await self.store.get_room_version_id(room_id)\n", "builder = self.event_builder_factory.new(room_version, {'type': EventTypes.\n Member, 'content': {'membership': Membership.LEAVE}, 'room_id': room_id,\n 'sender': user_id, 'state_key': user_id})\n", "event, context = await self.event_creation_handler.create_new_client_event(\n builder=builder)\n", "await self.auth.check_from_context(room_version, event, context,\n do_sig_check=False)\n", "logger.warning('Failed to create new leave %r because %s', event, e)\n", "return event\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "Docstring", "Condition", "Expr'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_3(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if VAR_56:\n", "return VAR_54 + VAR_55\n", "return VAR_54 * VAR_55\n" ]
[ "def testShowAllWithFunctions(self):...\n", "\"\"\"Model with callable polymorphic functions specified.\"\"\"\n", "@def_function.function...\n", "if c:\n", "return a + b\n", "return a * b\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Condition", "Return'", "Return'" ]
[ "import time\n", "import urllib2\n", "from flask import Flask, render_template, session, request, json\n", "from core.victim_objects import *\n", "import core.stats\n", "from core.utils import utils\n", "from core.db import Database\n", "VAR_0 = core.stats.trape\n", "VAR_1 = core.stats.app\n", "VAR_2 = Database()\n", "@VAR_1.route('/' + VAR_0.victim_path)...\n", "VAR_3 = urllib2.build_opener()\n", "VAR_4 = victim_headers()\n", "VAR_3.addheaders = VAR_4\n", "VAR_5 = victim_inject_code(VAR_3.open(VAR_0.url_to_clone).read(), 'lure')\n", "return VAR_5\n" ]
[ "import time\n", "import urllib2\n", "from flask import Flask, render_template, session, request, json\n", "from core.victim_objects import *\n", "import core.stats\n", "from core.utils import utils\n", "from core.db import Database\n", "trape = core.stats.trape\n", "app = core.stats.app\n", "db = Database()\n", "@app.route('/' + trape.victim_path)...\n", "opener = urllib2.build_opener()\n", "headers = victim_headers()\n", "opener.addheaders = headers\n", "html = victim_inject_code(opener.open(trape.url_to_clone).read(), 'lure')\n", "return html\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Import'", "Import'", "ImportFrom'", "ImportFrom'", "Import'", "ImportFrom'", "ImportFrom'", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Return'" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "VAR_53 = {}\n", "VAR_93 = getIntOrDefault(VAR_2, 'index', 0)\n", "VAR_104 = BaseContainer(VAR_5, **{str(c_type): long(c_id)})\n", "if VAR_25:\n", "VAR_53['share'] = BaseShare(VAR_5, VAR_25)\n", "if VAR_23 == 'well':\n", "VAR_104.image = VAR_104.well.getImage(VAR_93)\n", "VAR_118 = VAR_104.image.getAllRenderingDefs()\n", "VAR_119 = {}\n", "VAR_120 = VAR_104.image.getRenderingDefId()\n", "for VAR_71 in VAR_118:\n", "VAR_229 = VAR_71['owner']['id']\n", "VAR_119 = VAR_119.values()\n", "VAR_71['current'] = VAR_71['id'] == VAR_120\n", "VAR_121 = []\n", "if VAR_229 not in VAR_119 or VAR_119[VAR_229]['id'] < VAR_71['id']:\n", "for VAR_71 in VAR_119:\n", "VAR_119[VAR_229] = VAR_71\n", "VAR_282 = []\n", "VAR_122, VAR_123 = VAR_5.getMaxPlaneSize()\n", "for VAR_318, VAR_383 in enumerate(VAR_71['c']):\n", "VAR_124 = VAR_104.image.getSizeX()\n", "VAR_348 = '-'\n", "VAR_121.append({'id': VAR_71['id'], 'owner': VAR_71['owner'], 'c': ','.join\n (VAR_282), 'm': VAR_71['model'] == 'greyscale' and 'g' or 'c'})\n", "VAR_125 = VAR_104.image.getSizeY()\n", "if VAR_383['active']:\n", "VAR_53['tiledImage'] = VAR_124 * VAR_125 > VAR_122 * VAR_123\n", "VAR_348 = ''\n", "VAR_349 = VAR_383['lut'] if 'lut' in VAR_383 else VAR_383['color']\n", "VAR_53['manager'] = VAR_104\n", "VAR_350 = 'r' if VAR_383['inverted'] else '-r'\n", "VAR_53['rdefsJson'] = json.dumps(VAR_121)\n", "VAR_282.append('%s%s|%s:%s%s$%s' % (VAR_348, VAR_318 + 1, VAR_383['start'],\n VAR_383['end'], VAR_350, VAR_349))\n", "VAR_53['rdefs'] = VAR_119\n", "VAR_53['template'] = 'webclient/annotations/metadata_preview.html'\n", "return VAR_53\n" ]
[ "@login_required()...\n", "\"\"\"docstring\"\"\"\n", "context = {}\n", "index = getIntOrDefault(request, 'index', 0)\n", "manager = BaseContainer(conn, **{str(c_type): long(c_id)})\n", "if share_id:\n", "context['share'] = BaseShare(conn, share_id)\n", "if c_type == 'well':\n", "manager.image = manager.well.getImage(index)\n", "allRdefs = manager.image.getAllRenderingDefs()\n", "rdefs = {}\n", "rdefId = manager.image.getRenderingDefId()\n", "for r in allRdefs:\n", "ownerId = r['owner']['id']\n", "rdefs = rdefs.values()\n", "r['current'] = r['id'] == rdefId\n", "rdefQueries = []\n", "if ownerId not in rdefs or rdefs[ownerId]['id'] < r['id']:\n", "for r in rdefs:\n", "rdefs[ownerId] = r\n", "chs = []\n", "max_w, max_h = conn.getMaxPlaneSize()\n", "for i, c in enumerate(r['c']):\n", "size_x = manager.image.getSizeX()\n", "act = '-'\n", "rdefQueries.append({'id': r['id'], 'owner': r['owner'], 'c': ','.join(chs),\n 'm': r['model'] == 'greyscale' and 'g' or 'c'})\n", "size_y = manager.image.getSizeY()\n", "if c['active']:\n", "context['tiledImage'] = size_x * size_y > max_w * max_h\n", "act = ''\n", "color = c['lut'] if 'lut' in c else c['color']\n", "context['manager'] = manager\n", "reverse = 'r' if c['inverted'] else '-r'\n", "context['rdefsJson'] = json.dumps(rdefQueries)\n", "chs.append('%s%s|%s:%s%s$%s' % (act, i + 1, c['start'], c['end'], reverse,\n color))\n", "context['rdefs'] = rdefs\n", "context['template'] = 'webclient/annotations/metadata_preview.html'\n", "return context\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "For", "Assign'", "Assign'", "Assign'", "For", "Assign'", "Assign'", "Expr'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "@app.route('/tags')...\n", "if not app.config['SEARCH_CONF']['engine'] == 'ripgrep' and not which('rg'):\n", "flash('Ripgrep must be installed to view pages about embedded tags.', 'error')\n", "VAR_10 = sorted(get_all_tags(force=True))\n", "return redirect('/')\n", "return render_template('tags/all.html', title='All Tags', VAR_10=tags)\n" ]
[ "@app.route('/tags')...\n", "if not app.config['SEARCH_CONF']['engine'] == 'ripgrep' and not which('rg'):\n", "flash('Ripgrep must be installed to view pages about embedded tags.', 'error')\n", "tags = sorted(get_all_tags(force=True))\n", "return redirect('/')\n", "return render_template('tags/all.html', title='All Tags', tags=tags)\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Expr'", "Assign'", "Return'", "Return'" ]
[ "def FUNC_9():...\n", "if session.pam_user:\n", "session.flash = T('PAM authenticated user, cannot change password here')\n", "VAR_26 = SQLFORM.factory(Field('current_admin_password', 'password'), Field\n ('new_admin_password', 'password', requires=IS_STRONG()), Field(\n 'new_admin_password_again', 'password'), VAR_144='span4 well')\n", "redirect(URL('site'))\n", "if VAR_26.accepts(request.vars):\n", "if not verify_password(request.vars.current_admin_password):\n", "return dict(VAR_26=form)\n", "VAR_26.errors.current_admin_password = T('invalid password')\n", "if VAR_26.vars.new_admin_password != VAR_26.vars.new_admin_password_again:\n", "VAR_26.errors.new_admin_password_again = T('no match')\n", "VAR_15 = abspath('parameters_%s.py' % request.env.server_port)\n", "FUNC_4(VAR_15, 'password=\"%s\"' % CRYPT()(request.vars.new_admin_password)[0])\n", "session.flash = T('password changed')\n", "redirect(URL('site'))\n" ]
[ "def change_password():...\n", "if session.pam_user:\n", "session.flash = T('PAM authenticated user, cannot change password here')\n", "form = SQLFORM.factory(Field('current_admin_password', 'password'), Field(\n 'new_admin_password', 'password', requires=IS_STRONG()), Field(\n 'new_admin_password_again', 'password'), _class='span4 well')\n", "redirect(URL('site'))\n", "if form.accepts(request.vars):\n", "if not verify_password(request.vars.current_admin_password):\n", "return dict(form=form)\n", "form.errors.current_admin_password = T('invalid password')\n", "if form.vars.new_admin_password != form.vars.new_admin_password_again:\n", "form.errors.new_admin_password_again = T('no match')\n", "path = abspath('parameters_%s.py' % request.env.server_port)\n", "safe_write(path, 'password=\"%s\"' % CRYPT()(request.vars.new_admin_password)[0])\n", "session.flash = T('password changed')\n", "redirect(URL('site'))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Assign'", "Assign'", "Expr'", "Condition", "Condition", "Return'", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Assign'", "Expr'" ]
[ "def FUNC_13():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if VAR_1:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(VAR_39)\n", "VAR_40 = 0\n", "VAR_41 = 0\n", "if VAR_26:\n", "VAR_40 = int(VAR_25 * VAR_5)\n", "VAR_42 = []\n", "VAR_44 = cycle(VAR_17)\n", "def FUNC_15():...\n", "random.seed(VAR_24)\n", "return list(map(lambda x: x + VAR_5 * VAR_41, sorted(random.sample(list(\n range(VAR_5)), VAR_40))))\n" ]
[ "def generate():...\n", "yield '<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\\n'\n", "if STYLE:\n", "yield '<?xml-stylesheet type=\"text/xsl\" href=\"/scopelist.xsl\" ?>\\n'\n", "yield '<objectlist count=\"{:d}\">\\n'.format(total_entries)\n", "mix_per_iteration = 0\n", "iteration_count = 0\n", "if make_cocktail:\n", "mix_per_iteration = int(percentage * ITEMS_PER_ITERATION)\n", "mix_indices = []\n", "pool = cycle(mixer_list)\n", "def generate_mix_indices():...\n", "random.seed(seed)\n", "return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Condition", "Expr'", "Expr'", "Assign'", "Assign'", "Condition", "Assign'", "Assign'", "Assign'", "FunctionDef'", "Expr'", "Return'" ]
[ "def FUNC_76(VAR_212):...\n", "if VAR_178(VAR_212.getId()) == VAR_178(VAR_6):\n", "return False\n", "VAR_285 = VAR_212.getPrimaryPixels()\n", "if VAR_285 is None or VAR_212.getPrimaryPixels().getPixelsType().getValue(\n", "return False\n", "VAR_286 = [VAR_30.getLabel() for VAR_30 in VAR_212.getChannels()]\n", "VAR_286.sort()\n", "if VAR_286 != VAR_220:\n", "return False\n", "return True\n" ]
[ "def compat(i):...\n", "if long(i.getId()) == long(iid):\n", "return False\n", "pp = i.getPrimaryPixels()\n", "if pp is None or i.getPrimaryPixels().getPixelsType().getValue(\n", "return False\n", "ew = [x.getLabel() for x in i.getChannels()]\n", "ew.sort()\n", "if ew != img_ew:\n", "return False\n", "return True\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Assign'", "Condition", "Return'", "Assign'", "Expr'", "Condition", "Return'", "Return'" ]
[ "def __init__(self, *VAR_0, **VAR_1):...\n", "\"\"\"docstring\"\"\"\n", "self.doctype = self.name = None\n", "self._default_new_docs = {}\n", "self.flags = frappe._dict()\n", "if VAR_0 and VAR_0[0] and isinstance(VAR_0[0], string_types):\n", "if len(VAR_0) == 1:\n", "if VAR_0 and VAR_0[0] and isinstance(VAR_0[0], dict):\n", "self.doctype = self.name = VAR_0[0]\n", "self.doctype = VAR_0[0]\n", "VAR_1 = VAR_0[0]\n", "if VAR_1:\n", "self.load_from_db()\n", "if isinstance(VAR_0[1], dict):\n", "super(CLASS_0, self).__init__(VAR_1)\n", "return\n", "self.name = frappe.db.get_value(VAR_0[0], VAR_0[1], 'name')\n", "self.name = VAR_0[1]\n", "self.init_valid_columns()\n", "if self.name is None:\n", "if 'for_update' in VAR_1:\n", "frappe.throw(_('{0} {1} not found').format(_(VAR_0[0]), VAR_0[1]), frappe.\n DoesNotExistError)\n", "self.flags.for_update = VAR_1.get('for_update')\n" ]
[ "def __init__(self, *args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "self.doctype = self.name = None\n", "self._default_new_docs = {}\n", "self.flags = frappe._dict()\n", "if args and args[0] and isinstance(args[0], string_types):\n", "if len(args) == 1:\n", "if args and args[0] and isinstance(args[0], dict):\n", "self.doctype = self.name = args[0]\n", "self.doctype = args[0]\n", "kwargs = args[0]\n", "if kwargs:\n", "self.load_from_db()\n", "if isinstance(args[1], dict):\n", "super(Document, self).__init__(kwargs)\n", "return\n", "self.name = frappe.db.get_value(args[0], args[1], 'name')\n", "self.name = args[1]\n", "self.init_valid_columns()\n", "if self.name is None:\n", "if 'for_update' in kwargs:\n", "frappe.throw(_('{0} {1} not found').format(_(args[0]), args[1]), frappe.\n DoesNotExistError)\n", "self.flags.for_update = kwargs.get('for_update')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Return'", "Assign'", "Assign'", "Expr'", "Condition", "For", "Expr'", "Assign'" ]
[ "def __init__(self, VAR_7, VAR_8):...\n", "self.errorstring = VAR_7\n", "self.error = VAR_8\n", "super().__init__(VAR_7)\n" ]
[ "def __init__(self, errorstring, error):...\n", "self.errorstring = errorstring\n", "self.error = error\n", "super().__init__(errorstring)\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "def FUNC_11(self):...\n", "" ]
[ "def after_script_executed(self):...\n", "" ]
[ 0, 0 ]
[ "FunctionDef'", "Condition" ]
[ "@VAR_2.route('/advsearch', methods=['GET'])...\n", "VAR_16 = get_cc_columns(filter_config_custom_read=True)\n", "return FUNC_29(VAR_16)\n" ]
[ "@web.route('/advsearch', methods=['GET'])...\n", "cc = get_cc_columns(filter_config_custom_read=True)\n", "return render_prepare_search_form(cc)\n" ]
[ 0, 0, 0 ]
[ "Condition", "Assign'", "Return'" ]
[ "@expose('/userinfo/')...\n", "VAR_21 = self.datamodel.get(g.user.id, self._base_filters)\n", "VAR_43 = self._get_show_widget(g.user.id, VAR_21, VAR_15=self.\n user_show_fieldsets)\n", "self.update_redirect()\n", "return self.render_template(self.show_template, VAR_37=self.user_info_title,\n VAR_43=widgets, appbuilder=self.appbuilder)\n" ]
[ "@expose('/userinfo/')...\n", "item = self.datamodel.get(g.user.id, self._base_filters)\n", "widgets = self._get_show_widget(g.user.id, item, show_fieldsets=self.\n user_show_fieldsets)\n", "self.update_redirect()\n", "return self.render_template(self.show_template, title=self.user_info_title,\n widgets=widgets, appbuilder=self.appbuilder)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Expr'", "Return'" ]
[ "def FUNC_2(self, VAR_3, VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "for basket_command_middleware in get_provide_objects(\n", "if not issubclass(basket_command_middleware, BaseBasketCommandMiddleware):\n", "return VAR_4\n", "VAR_4 = dict(basket_command_middleware().preprocess_kwargs(VAR_2=self.\n basket, VAR_1=self.request, VAR_3=command, VAR_4=kwargs))\n" ]
[ "def preprocess_kwargs(self, command, kwargs):...\n", "\"\"\"docstring\"\"\"\n", "for basket_command_middleware in get_provide_objects(\n", "if not issubclass(basket_command_middleware, BaseBasketCommandMiddleware):\n", "return kwargs\n", "kwargs = dict(basket_command_middleware().preprocess_kwargs(basket=self.\n basket, request=self.request, command=command, kwargs=kwargs))\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "For", "Condition", "Return'", "Assign'" ]
[ "def FUNC_14(self, VAR_6):...\n", "\"\"\"docstring\"\"\"\n", "self.helper.send_state(self.room_id, 'm.room.canonical_alias', VAR_6, tok=\n self.admin_user_tok)\n" ]
[ "def _set_canonical_alias(self, content):...\n", "\"\"\"docstring\"\"\"\n", "self.helper.send_state(self.room_id, 'm.room.canonical_alias', content, tok\n =self.admin_user_tok)\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Expr'" ]
[ "@property...\n", "return self._base_builder.room_id\n" ]
[ "@property...\n", "return self._base_builder.room_id\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@VAR_2.route('/ajax/toggleread/<int:book_id>', methods=['POST'])...\n", "if not config.config_read_column:\n", "VAR_95 = ub.session.query(ub.ReadBook).filter(and_(ub.ReadBook.user_id ==\n int(VAR_87.id), ub.ReadBook.book_id == VAR_5)).first()\n", "calibre_db.update_title_sort(config)\n", "VAR_3.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n", "return ''\n", "if VAR_95:\n", "VAR_95 = calibre_db.get_filtered_book(VAR_5)\n", "return 'Custom Column No.{} is not existing in calibre database'.format(config\n .config_read_column), 400\n", "if VAR_95.read_status == ub.ReadBook.STATUS_FINISHED:\n", "VAR_123 = ub.ReadBook(VAR_11=current_user.id, VAR_5=book_id)\n", "VAR_26 = getattr(VAR_95, 'custom_column_' + str(config.config_read_column))\n", "VAR_95.read_status = ub.ReadBook.STATUS_UNREAD\n", "VAR_95.read_status = ub.ReadBook.STATUS_FINISHED\n", "VAR_123.read_status = ub.ReadBook.STATUS_FINISHED\n", "if len(VAR_26):\n", "if not VAR_95.kobo_reading_state:\n", "VAR_95 = VAR_123\n", "VAR_26[0].value = not VAR_26[0].value\n", "VAR_143 = db.cc_classes[config.config_read_column]\n", "VAR_124 = ub.KoboReadingState(VAR_11=current_user.id, VAR_5=book_id)\n", "ub.session.merge(VAR_95)\n", "calibre_db.session.commit()\n", "VAR_144 = VAR_143(value=1, VAR_95=book_id)\n", "VAR_124.current_bookmark = ub.KoboBookmark()\n", "ub.session_commit('Book {} readbit toggled'.format(VAR_5))\n", "calibre_db.session.add(VAR_144)\n", "VAR_124.statistics = ub.KoboStatistics()\n", "calibre_db.session.commit()\n", "VAR_95.kobo_reading_state = VAR_124\n" ]
[ "@web.route('/ajax/toggleread/<int:book_id>', methods=['POST'])...\n", "if not config.config_read_column:\n", "book = ub.session.query(ub.ReadBook).filter(and_(ub.ReadBook.user_id == int\n (current_user.id), ub.ReadBook.book_id == book_id)).first()\n", "calibre_db.update_title_sort(config)\n", "log.error(u'Custom Column No.%d is not existing in calibre database',\n config.config_read_column)\n", "return ''\n", "if book:\n", "book = calibre_db.get_filtered_book(book_id)\n", "return 'Custom Column No.{} is not existing in calibre database'.format(config\n .config_read_column), 400\n", "if book.read_status == ub.ReadBook.STATUS_FINISHED:\n", "readBook = ub.ReadBook(user_id=current_user.id, book_id=book_id)\n", "read_status = getattr(book, 'custom_column_' + str(config.config_read_column))\n", "book.read_status = ub.ReadBook.STATUS_UNREAD\n", "book.read_status = ub.ReadBook.STATUS_FINISHED\n", "readBook.read_status = ub.ReadBook.STATUS_FINISHED\n", "if len(read_status):\n", "if not book.kobo_reading_state:\n", "book = readBook\n", "read_status[0].value = not read_status[0].value\n", "cc_class = db.cc_classes[config.config_read_column]\n", "kobo_reading_state = ub.KoboReadingState(user_id=current_user.id, book_id=\n book_id)\n", "ub.session.merge(book)\n", "calibre_db.session.commit()\n", "new_cc = cc_class(value=1, book=book_id)\n", "kobo_reading_state.current_bookmark = ub.KoboBookmark()\n", "ub.session_commit('Book {} readbit toggled'.format(book_id))\n", "calibre_db.session.add(new_cc)\n", "kobo_reading_state.statistics = ub.KoboStatistics()\n", "calibre_db.session.commit()\n", "book.kobo_reading_state = kobo_reading_state\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assign'", "Expr'", "Expr'", "Return'", "Condition", "Assign'", "Return'", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Condition", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Assign'", "Expr'", "Expr'", "Assign'", "Expr'", "Assign'" ]
[ "@FUNC_0...\n", "return MediaRepository(self)\n" ]
[ "@cache_in_self...\n", "return MediaRepository(self)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_9(VAR_22):...\n", "\"\"\"docstring\"\"\"\n", "if len(VAR_22) > 13:\n", "VAR_30 = int(VAR_22, 36)\n", "if not six.PY3 and VAR_30 > sys.maxint:\n", "return VAR_30\n" ]
[ "def base36_to_int(s):...\n", "\"\"\"docstring\"\"\"\n", "if len(s) > 13:\n", "value = int(s, 36)\n", "if not six.PY3 and value > sys.maxint:\n", "return value\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Assign'", "Condition", "Return'" ]
[ "@VAR_25.whitelist(allow_guest=True)...\n", "VAR_25.local.login_manager.logout()\n", "VAR_25.db.commit()\n", "VAR_25.respond_as_web_page(_('Logged Out'), _(\n 'You have been successfully logged out'), indicator_color='green')\n" ]
[ "@frappe.whitelist(allow_guest=True)...\n", "frappe.local.login_manager.logout()\n", "frappe.db.commit()\n", "frappe.respond_as_web_page(_('Logged Out'), _(\n 'You have been successfully logged out'), indicator_color='green')\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_2(VAR_2, VAR_3, VAR_4):...\n", "\"\"\"docstring\"\"\"\n", "VAR_24 = FUNC_1(VAR_2)\n", "for port in (VAR_3, VAR_4):\n", "VAR_37 = socket.socket(VAR_24, socket.SOCK_STREAM)\n", "return True\n", "VAR_37.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n", "VAR_32 = 'Unable to bind socket {0}:{1}'.format(VAR_2, port)\n", "VAR_37.close()\n", "VAR_37.bind((VAR_2, int(port)))\n", "if exc.args:\n", "VAR_32 = '{0}, error: {1}'.format(VAR_32, str(exc))\n", "VAR_32 = '{0}, this might not be a problem.'.format(VAR_32)\n", "VAR_32 += '; Is there another salt-master running?'\n", "if is_console_configured():\n", "VAR_0.warning(VAR_32)\n", "sys.stderr.write('WARNING: {0}\\n'.format(VAR_32))\n", "return False\n" ]
[ "def verify_socket(interface, pub_port, ret_port):...\n", "\"\"\"docstring\"\"\"\n", "addr_family = lookup_family(interface)\n", "for port in (pub_port, ret_port):\n", "sock = socket.socket(addr_family, socket.SOCK_STREAM)\n", "return True\n", "sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n", "msg = 'Unable to bind socket {0}:{1}'.format(interface, port)\n", "sock.close()\n", "sock.bind((interface, int(port)))\n", "if exc.args:\n", "msg = '{0}, error: {1}'.format(msg, str(exc))\n", "msg = '{0}, this might not be a problem.'.format(msg)\n", "msg += '; Is there another salt-master running?'\n", "if is_console_configured():\n", "log.warning(msg)\n", "sys.stderr.write('WARNING: {0}\\n'.format(msg))\n", "return False\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Assign'", "Return'", "Expr'", "Assign'", "Expr'", "Expr'", "Condition", "Assign'", "Assign'", "AugAssign'", "Condition", "Expr'", "Expr'", "Return'" ]
[ "def __init__(self, VAR_8):...\n", "self.store = VAR_8.get_datastore()\n", "self.config = VAR_8.get_config()\n" ]
[ "def __init__(self, hs):...\n", "self.store = hs.get_datastore()\n", "self.config = hs.get_config()\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'" ]
[ "@FUNC_2.__func__...\n", "\"\"\"docstring\"\"\"\n", "self._cancel()\n" ]
[ "@whitelist.__func__...\n", "\"\"\"docstring\"\"\"\n", "self._cancel()\n" ]
[ 0, 0, 0 ]
[ "Condition", "Docstring", "Expr'" ]
[ "@property...\n", "return self._base_builder.type\n" ]
[ "@property...\n", "return self._base_builder.type\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "def FUNC_34(self):...\n", "return self.db[self.settings.table_cas_name]\n" ]
[ "def table_cas(self):...\n", "return self.db[self.settings.table_cas_name]\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]
[ "@login_required(setGroupContext=True)...\n", "\"\"\"docstring\"\"\"\n", "if VAR_35 is None or VAR_36 is None or VAR_37 is None:\n", "VAR_54 = 'webclient/history/history_details.html'\n", "VAR_87 = int(VAR_2.GET.get('page', 1))\n", "VAR_192 = VAR_2.session.get('user_id')\n", "VAR_193 = BaseCalendar(VAR_5=conn, VAR_35=year, VAR_36=month, VAR_37=day,\n eid=filter_user_id)\n", "VAR_193.get_items(VAR_87)\n", "VAR_53 = {'controller': VAR_193}\n", "VAR_53['template'] = VAR_54\n", "return VAR_53\n" ]
[ "@login_required(setGroupContext=True)...\n", "\"\"\"docstring\"\"\"\n", "if year is None or month is None or day is None:\n", "template = 'webclient/history/history_details.html'\n", "page = int(request.GET.get('page', 1))\n", "filter_user_id = request.session.get('user_id')\n", "controller = BaseCalendar(conn=conn, year=year, month=month, day=day, eid=\n filter_user_id)\n", "controller.get_items(page)\n", "context = {'controller': controller}\n", "context['template'] = template\n", "return context\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Condition", "Assign'", "Assign'", "Assign'", "Assign'", "Expr'", "Assign'", "Assign'", "Return'" ]
[ "def __init__(self):...\n", "self.var = variables.Variable(1.0, name='my_var')\n", "self.write_var = variables.Variable(1.0, name='write_var')\n" ]
[ "def __init__(self):...\n", "self.var = variables.Variable(1.0, name='my_var')\n", "self.write_var = variables.Variable(1.0, name='write_var')\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'" ]
[ "def FUNC_16(VAR_19):...\n", "\"\"\"docstring\"\"\"\n", "VAR_41 = example_pb2.Example()\n", "for feature_name, feature_list in VAR_19.items():\n", "if not isinstance(feature_list, list):\n", "return VAR_41.SerializeToString()\n", "if isinstance(feature_list[0], float):\n", "VAR_41.features.feature[feature_name].float_list.value.extend(feature_list)\n", "if isinstance(feature_list[0], str):\n", "VAR_41.features.feature[feature_name].bytes_list.value.extend([f.encode(\n 'utf8') for f in feature_list])\n", "if isinstance(feature_list[0], bytes):\n", "VAR_41.features.feature[feature_name].bytes_list.value.extend(feature_list)\n", "if isinstance(feature_list[0], six.integer_types):\n", "VAR_41.features.feature[feature_name].int64_list.value.extend(feature_list)\n" ]
[ "def _create_example_string(example_dict):...\n", "\"\"\"docstring\"\"\"\n", "example = example_pb2.Example()\n", "for feature_name, feature_list in example_dict.items():\n", "if not isinstance(feature_list, list):\n", "return example.SerializeToString()\n", "if isinstance(feature_list[0], float):\n", "example.features.feature[feature_name].float_list.value.extend(feature_list)\n", "if isinstance(feature_list[0], str):\n", "example.features.feature[feature_name].bytes_list.value.extend([f.encode(\n 'utf8') for f in feature_list])\n", "if isinstance(feature_list[0], bytes):\n", "example.features.feature[feature_name].bytes_list.value.extend(feature_list)\n", "if isinstance(feature_list[0], six.integer_types):\n", "example.features.feature[feature_name].int64_list.value.extend(feature_list)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "For", "Condition", "Return'", "Condition", "Expr'", "Condition", "Expr'", "Condition", "Expr'", "Condition", "Expr'" ]
[ "def FUNC_33(VAR_39, VAR_36):...\n", "VAR_37 = VAR_36.title\n", "VAR_38 = VAR_36.author\n", "VAR_65, VAR_62, VAR_69 = FUNC_32(VAR_37, VAR_38)\n", "VAR_70 = helper.get_valid_filename(VAR_37)\n", "VAR_71 = helper.get_valid_filename(VAR_69.name)\n", "VAR_72 = os.path.join(VAR_71, VAR_70).replace('\\\\', '/')\n", "VAR_41 = db.Books(VAR_37, '', VAR_65, datetime.utcnow(), datetime(101, 1, 1\n ), '1', datetime.utcnow(), VAR_72, VAR_36.cover, VAR_69, [], '')\n", "VAR_39 |= FUNC_7(VAR_62, VAR_41.authors, db.Authors, calibre_db.session,\n 'author')\n", "VAR_39 |= FUNC_18(VAR_36.series_id, VAR_41)\n", "VAR_39 |= FUNC_20(VAR_36.languages, VAR_41, VAR_26=True)\n", "VAR_39 |= FUNC_16(VAR_36.tags, VAR_41)\n", "VAR_39 |= FUNC_21(VAR_36.publisher, VAR_41)\n", "VAR_39 |= FUNC_17(VAR_36.series, VAR_41)\n", "VAR_73 = os.path.getsize(VAR_36.file_path)\n", "VAR_74 = db.Data(VAR_41, VAR_36.extension.upper()[1:], VAR_73, VAR_70)\n", "VAR_41.data.append(VAR_74)\n", "calibre_db.session.add(VAR_41)\n", "calibre_db.session.flush()\n", "return VAR_41, VAR_62, VAR_70\n" ]
[ "def create_book_on_upload(modif_date, meta):...\n", "title = meta.title\n", "authr = meta.author\n", "sort_authors, input_authors, db_author = prepare_authors_on_upload(title, authr\n )\n", "title_dir = helper.get_valid_filename(title)\n", "author_dir = helper.get_valid_filename(db_author.name)\n", "path = os.path.join(author_dir, title_dir).replace('\\\\', '/')\n", "db_book = db.Books(title, '', sort_authors, datetime.utcnow(), datetime(101,\n 1, 1), '1', datetime.utcnow(), path, meta.cover, db_author, [], '')\n", "modif_date |= modify_database_object(input_authors, db_book.authors, db.\n Authors, calibre_db.session, 'author')\n", "modif_date |= edit_book_series_index(meta.series_id, db_book)\n", "modif_date |= edit_book_languages(meta.languages, db_book, upload=True)\n", "modif_date |= edit_book_tags(meta.tags, db_book)\n", "modif_date |= edit_book_publisher(meta.publisher, db_book)\n", "modif_date |= edit_book_series(meta.series, db_book)\n", "file_size = os.path.getsize(meta.file_path)\n", "db_data = db.Data(db_book, meta.extension.upper()[1:], file_size, title_dir)\n", "db_book.data.append(db_data)\n", "calibre_db.session.add(db_book)\n", "calibre_db.session.flush()\n", "return db_book, input_authors, title_dir\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "Assign'", "AugAssign'", "AugAssign'", "AugAssign'", "AugAssign'", "AugAssign'", "AugAssign'", "Assign'", "Assign'", "Expr'", "Expr'", "Expr'", "Return'" ]
[ "def FUNC_20(self, *VAR_12, **VAR_13):...\n", "\"\"\"docstring\"\"\"\n", "if self._backend is not None and objects.backend != self._backend:\n", "return self.wrong_backend_handler(*VAR_12, **kwargs)\n", "return self._function(*VAR_12, **kwargs)\n" ]
[ "def wrapper(self, *args, **kwargs):...\n", "\"\"\"docstring\"\"\"\n", "if self._backend is not None and objects.backend != self._backend:\n", "return self.wrong_backend_handler(*args, **kwargs)\n", "return self._function(*args, **kwargs)\n" ]
[ 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Condition", "Return'", "Return'" ]
[ "def __init__(self):...\n", "self.reqparse = reqparse.RequestParser()\n", "self.req = CLASS_2.prepare_from_flask_request(request)\n", "super(CLASS_2, self).__init__()\n" ]
[ "def __init__(self):...\n", "self.reqparse = reqparse.RequestParser()\n", "self.req = OneLogin.prepare_from_flask_request(request)\n", "super(OneLogin, self).__init__()\n" ]
[ 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Expr'" ]
[ "@apfell.route('/apiui/command_help')...\n", "VAR_4 = VAR_0.get_template('apiui_command_help.html')\n", "if len(VAR_1.query_args) != 0:\n", "VAR_5 = urllib.parse.unquote(VAR_1.query_args[0][1])\n", "VAR_5 = ''\n", "print(VAR_5)\n", "if use_ssl:\n", "VAR_6 = VAR_4.render(VAR_3=await respect_pivot(links, request), name=user[\n 'username'], http='https', ws='wss', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n", "VAR_6 = VAR_4.render(VAR_3=await respect_pivot(links, request), name=user[\n 'username'], http='http', ws='ws', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n", "return response.html(VAR_6)\n" ]
[ "@apfell.route('/apiui/command_help')...\n", "template = env.get_template('apiui_command_help.html')\n", "if len(request.query_args) != 0:\n", "data = urllib.parse.unquote(request.query_args[0][1])\n", "data = ''\n", "print(data)\n", "if use_ssl:\n", "content = template.render(links=await respect_pivot(links, request), name=\n user['username'], http='https', ws='wss', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n", "content = template.render(links=await respect_pivot(links, request), name=\n user['username'], http='http', ws='ws', config=user['ui_config'],\n view_utc_time=user['view_utc_time'], agent=data)\n", "return response.html(content)\n" ]
[ 0, 0, 0, 0, 0, 2, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Condition", "Assign'", "Assign'", "Expr'", "Condition", "Assign'", "Assign'", "Return'" ]
[ "def FUNC_11(self, VAR_50: 'APIRouter', *, VAR_41: str='', VAR_30: Optional[...\n", "if VAR_41:\n", "assert VAR_41.startswith('/'), \"A path prefix must start with '/'\"\n", "for r in VAR_50.routes:\n", "assert not VAR_41.endswith('/'\n ), \"A path prefix must not end with '/', as the routes will start with '/'\"\n", "VAR_26 = getattr(r, 'path')\n", "if VAR_35 is None:\n", "VAR_28 = getattr(r, 'name', 'unknown')\n", "VAR_35 = {}\n", "for VAR_65 in VAR_50.routes:\n", "if VAR_26 is not None and not VAR_26:\n", "if isinstance(VAR_65, CLASS_1):\n", "for handler in VAR_50.on_startup:\n", "VAR_60 = {**VAR_35, **VAR_65.responses}\n", "if isinstance(VAR_65, routing.Route):\n", "self.add_event_handler('startup', handler)\n", "for handler in VAR_50.on_shutdown:\n", "VAR_76 = get_value_or_default(VAR_65.response_class, VAR_50.\n default_response_class, VAR_42, self.default_response_class)\n", "VAR_37 = list(VAR_65.methods or [])\n", "if isinstance(VAR_65, CLASS_0):\n", "self.add_event_handler('shutdown', handler)\n", "VAR_62 = []\n", "self.add_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_37=methods,\n VAR_39=route.include_in_schema, VAR_28=route.name)\n", "self.add_api_websocket_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_28=\n route.name)\n", "if isinstance(VAR_65, routing.WebSocketRoute):\n", "if VAR_30:\n", "self.add_websocket_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_28=\n route.name)\n", "VAR_62.extend(VAR_30)\n", "if VAR_65.tags:\n", "VAR_62.extend(VAR_65.tags)\n", "VAR_63: List[params.Depends] = []\n", "if VAR_31:\n", "VAR_63.extend(VAR_31)\n", "if VAR_65.dependencies:\n", "VAR_63.extend(VAR_65.dependencies)\n", "VAR_64 = []\n", "if VAR_40:\n", "VAR_64.extend(VAR_40)\n", "if VAR_65.callbacks:\n", "VAR_64.extend(VAR_65.callbacks)\n", "self.add_api_route(VAR_41 + VAR_65.path, VAR_65.endpoint, VAR_29=route.\n response_model, VAR_13=route.status_code, VAR_30=current_tags, VAR_31=\n current_dependencies, VAR_32=route.summary, VAR_33=route.description,\n VAR_34=route.response_description, VAR_35=combined_responses, VAR_36=\n route.deprecated or deprecated or self.deprecated, VAR_37=route.methods,\n VAR_38=route.operation_id, VAR_16=route.response_model_include, VAR_17=\n route.response_model_exclude, VAR_18=route.response_model_by_alias,\n VAR_19=route.response_model_exclude_unset, VAR_20=route.\n response_model_exclude_defaults, VAR_21=route.\n response_model_exclude_none, VAR_39=route.include_in_schema and self.\n include_in_schema and include_in_schema, VAR_14=use_response_class,\n VAR_28=route.name, VAR_49=type(route), VAR_40=current_callbacks)\n" ]
[ "def include_router(self, router: 'APIRouter', *, prefix: str='', tags:...\n", "if prefix:\n", "assert prefix.startswith('/'), \"A path prefix must start with '/'\"\n", "for r in router.routes:\n", "assert not prefix.endswith('/'\n ), \"A path prefix must not end with '/', as the routes will start with '/'\"\n", "path = getattr(r, 'path')\n", "if responses is None:\n", "name = getattr(r, 'name', 'unknown')\n", "responses = {}\n", "for route in router.routes:\n", "if path is not None and not path:\n", "if isinstance(route, APIRoute):\n", "for handler in router.on_startup:\n", "combined_responses = {**responses, **route.responses}\n", "if isinstance(route, routing.Route):\n", "self.add_event_handler('startup', handler)\n", "for handler in router.on_shutdown:\n", "use_response_class = get_value_or_default(route.response_class, router.\n default_response_class, default_response_class, self.default_response_class\n )\n", "methods = list(route.methods or [])\n", "if isinstance(route, APIWebSocketRoute):\n", "self.add_event_handler('shutdown', handler)\n", "current_tags = []\n", "self.add_route(prefix + route.path, route.endpoint, methods=methods,\n include_in_schema=route.include_in_schema, name=route.name)\n", "self.add_api_websocket_route(prefix + route.path, route.endpoint, name=\n route.name)\n", "if isinstance(route, routing.WebSocketRoute):\n", "if tags:\n", "self.add_websocket_route(prefix + route.path, route.endpoint, name=route.name)\n", "current_tags.extend(tags)\n", "if route.tags:\n", "current_tags.extend(route.tags)\n", "current_dependencies: List[params.Depends] = []\n", "if dependencies:\n", "current_dependencies.extend(dependencies)\n", "if route.dependencies:\n", "current_dependencies.extend(route.dependencies)\n", "current_callbacks = []\n", "if callbacks:\n", "current_callbacks.extend(callbacks)\n", "if route.callbacks:\n", "current_callbacks.extend(route.callbacks)\n", "self.add_api_route(prefix + route.path, route.endpoint, response_model=\n route.response_model, status_code=route.status_code, tags=current_tags,\n dependencies=current_dependencies, summary=route.summary, description=\n route.description, response_description=route.response_description,\n responses=combined_responses, deprecated=route.deprecated or deprecated or\n self.deprecated, methods=route.methods, operation_id=route.operation_id,\n response_model_include=route.response_model_include,\n response_model_exclude=route.response_model_exclude,\n response_model_by_alias=route.response_model_by_alias,\n response_model_exclude_unset=route.response_model_exclude_unset,\n response_model_exclude_defaults=route.response_model_exclude_defaults,\n response_model_exclude_none=route.response_model_exclude_none,\n include_in_schema=route.include_in_schema and self.include_in_schema and\n include_in_schema, response_class=use_response_class, name=route.name,\n route_class_override=type(route), callbacks=current_callbacks)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Condition", "Assert'", "For", "Assert'", "Assign'", "Condition", "Assign'", "Assign'", "For", "Condition", "Condition", "For", "Assign'", "Condition", "Expr'", "For", "Assign'", "Assign'", "Condition", "Expr'", "Assign'", "Expr'", "Expr'", "Condition", "Condition", "Expr'", "Expr'", "Condition", "Expr'", "AnnAssign'", "Condition", "Expr'", "Condition", "Expr'", "Assign'", "Condition", "Expr'", "Condition", "Expr'", "Expr'" ]
[ "def __init__(self, VAR_4):...\n", "self.response = VAR_4\n", "super(CLASS_1, self).__init__(force_text(VAR_4))\n" ]
[ "def __init__(self, response):...\n", "self.response = response\n", "super(ExceptionalResponse, self).__init__(force_text(response))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Expr'" ]
[ "def FUNC_31(VAR_37, VAR_38):...\n", "if VAR_37[1].get('_key', None) == 'plugin_corewizard_acl':\n", "return f'0:{to_unicode(VAR_37[0])}'\n", "if VAR_37[1].get('mandatory', False):\n", "return f'1:{to_unicode(VAR_37[0])}'\n", "return f'2:{to_unicode(VAR_37[0])}'\n" ]
[ "def wizard_key_extractor(d, k):...\n", "if d[1].get('_key', None) == 'plugin_corewizard_acl':\n", "return f'0:{to_unicode(d[0])}'\n", "if d[1].get('mandatory', False):\n", "return f'1:{to_unicode(d[0])}'\n", "return f'2:{to_unicode(d[0])}'\n" ]
[ 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Condition", "Return'", "Condition", "Return'", "Return'" ]
[ "@VAR_1.route('/ajax/mergebooks', methods=['POST'])...\n", "VAR_79 = VAR_32.get_json().get('Merge_books')\n", "VAR_81 = list()\n", "if VAR_79:\n", "VAR_100 = calibre_db.get_book(VAR_79[0])\n", "return ''\n", "VAR_79.pop(0)\n", "if VAR_100:\n", "for file in VAR_100.data:\n", "VAR_81.append(file.format)\n", "VAR_114 = helper.get_valid_filename(VAR_100.title\n ) + ' - ' + helper.get_valid_filename(VAR_100.authors[0].name)\n", "for VAR_15 in VAR_79:\n", "VAR_118 = calibre_db.get_book(VAR_15)\n", "if VAR_118:\n", "for element in VAR_118.data:\n", "if element.format not in VAR_81:\n", "FUNC_13(VAR_118.id, '', True)\n", "VAR_122 = os.path.normpath(os.path.join(config.config_calibre_dir, VAR_100.\n path, VAR_114 + '.' + element.format.lower()))\n", "return json.dumps({'success': True})\n", "VAR_123 = os.path.normpath(os.path.join(config.config_calibre_dir, VAR_118.\n path, element.name + '.' + element.format.lower()))\n", "copyfile(VAR_123, VAR_122)\n", "VAR_100.data.append(db.Data(VAR_100.id, element.format, element.\n uncompressed_size, VAR_114))\n" ]
[ "@editbook.route('/ajax/mergebooks', methods=['POST'])...\n", "vals = request.get_json().get('Merge_books')\n", "to_file = list()\n", "if vals:\n", "to_book = calibre_db.get_book(vals[0])\n", "return ''\n", "vals.pop(0)\n", "if to_book:\n", "for file in to_book.data:\n", "to_file.append(file.format)\n", "to_name = helper.get_valid_filename(to_book.title\n ) + ' - ' + helper.get_valid_filename(to_book.authors[0].name)\n", "for book_id in vals:\n", "from_book = calibre_db.get_book(book_id)\n", "if from_book:\n", "for element in from_book.data:\n", "if element.format not in to_file:\n", "delete_book(from_book.id, '', True)\n", "filepath_new = os.path.normpath(os.path.join(config.config_calibre_dir,\n to_book.path, to_name + '.' + element.format.lower()))\n", "return json.dumps({'success': True})\n", "filepath_old = os.path.normpath(os.path.join(config.config_calibre_dir,\n from_book.path, element.name + '.' + element.format.lower()))\n", "copyfile(filepath_old, filepath_new)\n", "to_book.data.append(db.Data(to_book.id, element.format, element.\n uncompressed_size, to_name))\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Return'", "Expr'", "Condition", "For", "Expr'", "Assign'", "For", "Assign'", "Condition", "For", "For", "Expr'", "Assign'", "Return'", "Assign'", "Expr'", "Expr'" ]
[ "def FUNC_15():...\n", "random.seed(VAR_24)\n", "return list(map(lambda x: x + VAR_5 * VAR_45, sorted(random.sample(list(\n range(VAR_5)), VAR_44))))\n" ]
[ "def generate_mix_indices():...\n", "random.seed(seed)\n", "return list(map(lambda x: x + ITEMS_PER_ITERATION * iteration_count, sorted\n (random.sample(list(range(ITEMS_PER_ITERATION)), mix_per_iteration))))\n" ]
[ 0, 0, 0 ]
[ "FunctionDef'", "Expr'", "Return'" ]
[ "@VAR_2.route('/ajax/editlistusers/<param>', methods=['POST'])...\n", "VAR_61 = request.form.to_dict(flat=False)\n", "VAR_47 = ub.session.query(ub.User)\n", "if not config.config_anonbrowse:\n", "VAR_47 = VAR_47.filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS)\n", "if 'pk' in VAR_61:\n", "VAR_54 = [VAR_47.filter(ub.User.id == VAR_61['pk'][0]).one_or_none()]\n", "if 'pk[]' in VAR_61:\n", "if 'field_index' in VAR_61:\n", "VAR_54 = VAR_47.filter(ub.User.id.in_(VAR_61['pk[]'])).all()\n", "return _('Malformed request'), 400\n", "VAR_61['field_index'] = VAR_61['field_index'][0]\n", "if 'value' in VAR_61:\n", "VAR_61['value'] = VAR_61['value'][0]\n", "if not 'value[]' in VAR_61:\n", "for VAR_12 in VAR_54:\n", "return _('Malformed request'), 400\n", "ub.session_commit()\n", "if VAR_4 in ['denied_tags', 'allowed_tags', 'allowed_column_value',\n", "VAR_0.debug_or_exception(ex)\n", "return ''\n", "if 'value[]' in VAR_61:\n", "VAR_61['value'] = VAR_61['value'].strip()\n", "return str(ex), 400\n", "setattr(VAR_12, VAR_4, FUNC_29(VAR_12, VAR_61['action'][0], VAR_4, VAR_61[\n 'value[]']))\n", "setattr(VAR_12, VAR_4, VAR_61['value'].strip())\n", "if VAR_4 == 'name':\n", "if VAR_12.name == 'Guest':\n", "if VAR_4 == 'email':\n", "VAR_12.name = check_username(VAR_61['value'])\n", "VAR_12.email = check_email(VAR_61['value'])\n", "if VAR_4 == 'kobo_only_shelves_sync':\n", "VAR_12.kobo_only_shelves_sync = VAR_119(VAR_61['value'] == 'true')\n", "if VAR_4 == 'kindle_mail':\n", "VAR_12.kindle_mail = valid_email(VAR_61['value']) if VAR_61['value'] else ''\n", "if VAR_4.endswith('role'):\n", "VAR_125 = VAR_119(VAR_61['field_index'])\n", "if VAR_4.startswith('sidebar'):\n", "if VAR_12.name == 'Guest' and VAR_125 in [constants.ROLE_ADMIN, constants.\n", "VAR_125 = VAR_119(VAR_61['field_index'])\n", "if VAR_4 == 'locale':\n", "if VAR_125 > 0 and VAR_125 <= constants.ROLE_VIEWER and (VAR_125 & VAR_125 -\n", "if VAR_12.name == 'Guest' and VAR_125 == constants.SIDEBAR_READ_AND_UNREAD:\n", "if VAR_12.name == 'Guest':\n", "if VAR_4 == 'default_language':\n", "if VAR_61['value'] == 'true':\n", "if VAR_125 > 0 and VAR_125 <= constants.SIDEBAR_LIST and (VAR_125 & VAR_125 -\n", "if VAR_61['value'] in _BABEL_TRANSLATIONS:\n", "VAR_23 = calibre_db.session.query(db.Languages).join(db.books_languages_link\n ).join(db.Books).filter(calibre_db.common_filters()).group_by(VAR_130(\n 'books_languages_link.lang_code')).all()\n", "return _('Parameter not found'), 400\n", "VAR_12.role |= VAR_125\n", "if VAR_61['value'] == 'false':\n", "if VAR_61['value'] == 'true':\n", "VAR_12.locale = VAR_61['value']\n", "VAR_131 = [lang.lang_code for lang in VAR_23] + ['all']\n", "if VAR_125 == constants.ROLE_ADMIN:\n", "VAR_12.sidebar_view |= VAR_125\n", "if VAR_61['value'] == 'false':\n", "if VAR_61['value'] in VAR_131:\n", "if not ub.session.query(ub.User).filter(ub.User.role.op('&')(constants.\n", "VAR_12.role &= ~VAR_125\n", "VAR_12.sidebar_view &= ~VAR_125\n", "VAR_12.default_language = VAR_61['value']\n", "return Response(json.dumps([{'type': 'danger', 'message': _(\n u\"No admin user remaining, can't remove admin role\", nick=user.name)}]),\n mimetype='application/json')\n" ]
[ "@admi.route('/ajax/editlistusers/<param>', methods=['POST'])...\n", "vals = request.form.to_dict(flat=False)\n", "all_user = ub.session.query(ub.User)\n", "if not config.config_anonbrowse:\n", "all_user = all_user.filter(ub.User.role.op('&')(constants.ROLE_ANONYMOUS) !=\n constants.ROLE_ANONYMOUS)\n", "if 'pk' in vals:\n", "users = [all_user.filter(ub.User.id == vals['pk'][0]).one_or_none()]\n", "if 'pk[]' in vals:\n", "if 'field_index' in vals:\n", "users = all_user.filter(ub.User.id.in_(vals['pk[]'])).all()\n", "return _('Malformed request'), 400\n", "vals['field_index'] = vals['field_index'][0]\n", "if 'value' in vals:\n", "vals['value'] = vals['value'][0]\n", "if not 'value[]' in vals:\n", "for user in users:\n", "return _('Malformed request'), 400\n", "ub.session_commit()\n", "if param in ['denied_tags', 'allowed_tags', 'allowed_column_value',\n", "log.debug_or_exception(ex)\n", "return ''\n", "if 'value[]' in vals:\n", "vals['value'] = vals['value'].strip()\n", "return str(ex), 400\n", "setattr(user, param, prepare_tags(user, vals['action'][0], param, vals[\n 'value[]']))\n", "setattr(user, param, vals['value'].strip())\n", "if param == 'name':\n", "if user.name == 'Guest':\n", "if param == 'email':\n", "user.name = check_username(vals['value'])\n", "user.email = check_email(vals['value'])\n", "if param == 'kobo_only_shelves_sync':\n", "user.kobo_only_shelves_sync = int(vals['value'] == 'true')\n", "if param == 'kindle_mail':\n", "user.kindle_mail = valid_email(vals['value']) if vals['value'] else ''\n", "if param.endswith('role'):\n", "value = int(vals['field_index'])\n", "if param.startswith('sidebar'):\n", "if user.name == 'Guest' and value in [constants.ROLE_ADMIN, constants.\n", "value = int(vals['field_index'])\n", "if param == 'locale':\n", "if value > 0 and value <= constants.ROLE_VIEWER and (value & value - 1 == 0 or\n", "if user.name == 'Guest' and value == constants.SIDEBAR_READ_AND_UNREAD:\n", "if user.name == 'Guest':\n", "if param == 'default_language':\n", "if vals['value'] == 'true':\n", "if value > 0 and value <= constants.SIDEBAR_LIST and (value & value - 1 == \n", "if vals['value'] in _BABEL_TRANSLATIONS:\n", "languages = calibre_db.session.query(db.Languages).join(db.books_languages_link\n ).join(db.Books).filter(calibre_db.common_filters()).group_by(text(\n 'books_languages_link.lang_code')).all()\n", "return _('Parameter not found'), 400\n", "user.role |= value\n", "if vals['value'] == 'false':\n", "if vals['value'] == 'true':\n", "user.locale = vals['value']\n", "lang_codes = [lang.lang_code for lang in languages] + ['all']\n", "if value == constants.ROLE_ADMIN:\n", "user.sidebar_view |= value\n", "if vals['value'] == 'false':\n", "if vals['value'] in lang_codes:\n", "if not ub.session.query(ub.User).filter(ub.User.role.op('&')(constants.\n", "user.role &= ~value\n", "user.sidebar_view &= ~value\n", "user.default_language = vals['value']\n", "return Response(json.dumps([{'type': 'danger', 'message': _(\n u\"No admin user remaining, can't remove admin role\", nick=user.name)}]),\n mimetype='application/json')\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Return'", "Assign'", "Condition", "Assign'", "Condition", "For", "Return'", "Expr'", "Condition", "Expr'", "Return'", "Condition", "Assign'", "Return'", "Expr'", "Expr'", "Condition", "Condition", "Condition", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Condition", "Assign'", "Condition", "Condition", "Condition", "Condition", "Condition", "Condition", "Condition", "Condition", "Assign'", "Return'", "AugAssign'", "Condition", "Condition", "Assign'", "Assign'", "Condition", "AugAssign'", "Condition", "Condition", "Condition", "AugAssign'", "AugAssign'", "Assign'", "Return'" ]
[ "def FUNC_10(self, VAR_0):...\n", "VAR_13 = VAR_0.patch('httpx.get')\n", "VAR_14 = VAR_0.patch('openapi_python_client.Path')\n", "VAR_15 = VAR_0.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "VAR_11 = VAR_2(VAR_6=mocker.MagicMock(), VAR_7=mocker.MagicMock())\n", "assert VAR_11 == GeneratorError(header='Provide URL or Path, not both.')\n", "VAR_13.assert_not_called()\n", "VAR_14.assert_not_called()\n", "VAR_15.assert_not_called()\n" ]
[ "def test__get_document_url_and_path(self, mocker):...\n", "get = mocker.patch('httpx.get')\n", "Path = mocker.patch('openapi_python_client.Path')\n", "loads = mocker.patch('yaml.safe_load')\n", "from openapi_python_client import _get_document\n", "result = _get_document(url=mocker.MagicMock(), path=mocker.MagicMock())\n", "assert result == GeneratorError(header='Provide URL or Path, not both.')\n", "get.assert_not_called()\n", "Path.assert_not_called()\n", "loads.assert_not_called()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Assign'", "Assign'", "Assign'", "ImportFrom'", "Assign'", "Assert'", "Expr'", "Expr'", "Expr'" ]
[ "@VAR_0.route('/simpleshelf/<int:shelf_id>')...\n", "return FUNC_15(2, VAR_3, 1, None)\n" ]
[ "@shelf.route('/simpleshelf/<int:shelf_id>')...\n", "return render_show_shelf(2, shelf_id, 1, None)\n" ]
[ 0, 0 ]
[ "Condition", "Return'" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "VAR_2 = FUNC_1('/groups/%s/roles', VAR_30)\n", "return self.client.get_json(VAR_5=destination, VAR_2=path, VAR_3={\n 'requester_user_id': requester_user_id}, VAR_15=True)\n" ]
[ "@log_function...\n", "\"\"\"docstring\"\"\"\n", "path = _create_v1_path('/groups/%s/roles', group_id)\n", "return self.client.get_json(destination=destination, path=path, args={\n 'requester_user_id': requester_user_id}, ignore_backoff=True)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Docstring", "Assign'", "Return'" ]
[ "@wraps(VAR_4)...\n", "if VAR_87.role_download():\n", "return VAR_4(*VAR_49, **kwargs)\n", "abort(403)\n" ]
[ "@wraps(f)...\n", "if current_user.role_download():\n", "return f(*args, **kwargs)\n", "abort(403)\n" ]
[ 0, 0, 0, 0 ]
[ "Condition", "Condition", "Return'", "Expr'" ]
[ "def FUNC_4(VAR_21: Text) ->Tuple[Optional[Text], Optional[Text]]:...\n", "\"\"\"docstring\"\"\"\n", "VAR_50 = os.path.join(VAR_21, DEFAULT_CORE_SUBDIRECTORY_NAME)\n", "VAR_51 = os.path.join(VAR_21, DEFAULT_NLU_SUBDIRECTORY_NAME)\n", "if not os.path.isdir(VAR_50):\n", "VAR_50 = None\n", "if not os.path.isdir(VAR_51):\n", "VAR_51 = None\n", "if not VAR_50 and not VAR_51:\n", "return VAR_50, VAR_51\n" ]
[ "def get_model_subdirectories(unpacked_model_path: Text) ->Tuple[Optional[...\n", "\"\"\"docstring\"\"\"\n", "core_path = os.path.join(unpacked_model_path, DEFAULT_CORE_SUBDIRECTORY_NAME)\n", "nlu_path = os.path.join(unpacked_model_path, DEFAULT_NLU_SUBDIRECTORY_NAME)\n", "if not os.path.isdir(core_path):\n", "core_path = None\n", "if not os.path.isdir(nlu_path):\n", "nlu_path = None\n", "if not core_path and not nlu_path:\n", "return core_path, nlu_path\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "FunctionDef'", "Docstring", "Assign'", "Assign'", "Condition", "Assign'", "Condition", "Assign'", "Condition", "Return'" ]
[ "async def FUNC_47(VAR_82):...\n", "for VAR_196 in VAR_82:\n", "return False\n", "await self.backfill(VAR_196, VAR_11, VAR_17=100, VAR_18=extremities)\n", "VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n", "return True\n", "if 400 <= VAR_170.code < 500:\n", "VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n", "if 400 <= VAR_170.code < 500:\n", "VAR_0.info('Failed to backfill from %s because %s', VAR_196, VAR_170)\n", "VAR_0.info(str(VAR_170))\n", "VAR_0.info('Failed to get backfill from %s because %s', VAR_196, VAR_170)\n", "VAR_0.info(VAR_170)\n", "VAR_0.exception('Failed to backfill from %s because %s', VAR_196, VAR_170)\n" ]
[ "async def try_backfill(domains):...\n", "for dom in domains:\n", "return False\n", "await self.backfill(dom, room_id, limit=100, extremities=extremities)\n", "logger.info('Failed to backfill from %s because %s', dom, e)\n", "return True\n", "if 400 <= e.code < 500:\n", "logger.info('Failed to backfill from %s because %s', dom, e)\n", "if 400 <= e.code < 500:\n", "logger.info('Failed to backfill from %s because %s', dom, e)\n", "logger.info(str(e))\n", "logger.info('Failed to get backfill from %s because %s', dom, e)\n", "logger.info(e)\n", "logger.exception('Failed to backfill from %s because %s', dom, e)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "AsyncFunctionDef'", "For", "Return'", "Expr'", "Expr'", "Return'", "Condition", "Expr'", "Condition", "Expr'", "Expr'", "Expr'", "Expr'", "Expr'" ]
[ "def FUNC_23(self):...\n", "return False\n" ]
[ "def role_admin(self):...\n", "return False\n" ]
[ 0, 0 ]
[ "FunctionDef'", "Return'" ]