problem_id
stringlengths 18
22
| source
stringclasses 1
value | task_type
stringclasses 1
value | in_source_id
stringlengths 13
58
| prompt
stringlengths 1.1k
25.4k
| golden_diff
stringlengths 145
5.13k
| verification_info
stringlengths 582
39.1k
| num_tokens
int64 271
4.1k
| num_tokens_diff
int64 47
1.02k
|
---|---|---|---|---|---|---|---|---|
gh_patches_debug_18040
|
rasdani/github-patches
|
git_diff
|
liqd__a4-meinberlin-1652
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
make it possible to detach plan from project
now that a connection of a project with a plan is no longer obligatory, initiators need to be able to undo a connection. Once I click on a plan I can no longer NOT connect to a plan, which should be possible.

--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `meinberlin/apps/plans/forms.py`
Content:
```
1 from django import forms
2 from django.conf import settings
3 from django.db.models import Q
4 from django.utils.translation import ugettext_lazy as _
5
6 from adhocracy4.dashboard.components.forms import ProjectDashboardForm
7 from adhocracy4.maps import widgets as maps_widgets
8 from adhocracy4.projects import models as project_models
9 from meinberlin.apps.contrib import widgets as contrib_widgets
10
11 from . import models
12
13
14 def get_theme_options():
15 return models.Plan.objects\
16 .filter(~Q(theme=''))\
17 .order_by('theme')\
18 .values_list('theme', flat=True)\
19 .distinct()
20
21
22 class PlanForm(forms.ModelForm):
23
24 class Meta:
25 model = models.Plan
26 fields = [
27 'title',
28 'description_image',
29 'contact',
30 'point',
31 'point_label',
32 'district',
33 'cost',
34 'description',
35 'theme',
36 'status',
37 'participation']
38 widgets = {
39 'point': maps_widgets.MapChoosePointWidget(
40 polygon=settings.BERLIN_POLYGON),
41 'theme': contrib_widgets.TextWithDatalistWidget(attrs={
42 'options': get_theme_options
43 })
44 }
45 error_messages = {
46 'point': {
47 'required': _('Please locate the plan on the map.')
48 }
49 }
50
51 def __init__(self, *args, **kwargs):
52 super().__init__(*args, **kwargs)
53 self.fields['district'].empty_label = _('City wide')
54
55
56 class CustomMultipleChoiceField(forms.ModelMultipleChoiceField):
57
58 widget = forms.RadioSelect
59
60 def clean(self, value):
61 if value is None:
62 return super().clean([])
63 return super().clean([value])
64
65
66 class ProjectPlansDashboardForm(ProjectDashboardForm):
67 plans = CustomMultipleChoiceField(queryset=None,
68 label=_('Plans'))
69
70 class Meta:
71 model = project_models.Project
72 fields = ['plans']
73 required = False
74
75 def save(self, commit=False):
76 plans = self.cleaned_data['plans']
77 self.instance.plans.set(plans)
78
79 def __init__(self, *args, **kwargs):
80 super().__init__(*args, **kwargs)
81 self.initial['plans'] = self.instance.plans.all()
82 self.fields['plans'
83 ].queryset = self.instance.organisation.plan_set.all()
84
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/meinberlin/apps/plans/forms.py b/meinberlin/apps/plans/forms.py
--- a/meinberlin/apps/plans/forms.py
+++ b/meinberlin/apps/plans/forms.py
@@ -55,10 +55,10 @@
class CustomMultipleChoiceField(forms.ModelMultipleChoiceField):
- widget = forms.RadioSelect
+ widget = forms.Select
def clean(self, value):
- if value is None:
+ if not value:
return super().clean([])
return super().clean([value])
@@ -79,5 +79,7 @@
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self.initial['plans'] = self.instance.plans.all()
- self.fields['plans'
- ].queryset = self.instance.organisation.plan_set.all()
+ self.fields['plans'].required = False
+ self.fields['plans'].empty_label = '----------'
+ self.fields['plans'].queryset = \
+ self.instance.organisation.plan_set.all()
|
{"golden_diff": "diff --git a/meinberlin/apps/plans/forms.py b/meinberlin/apps/plans/forms.py\n--- a/meinberlin/apps/plans/forms.py\n+++ b/meinberlin/apps/plans/forms.py\n@@ -55,10 +55,10 @@\n \n class CustomMultipleChoiceField(forms.ModelMultipleChoiceField):\n \n- widget = forms.RadioSelect\n+ widget = forms.Select\n \n def clean(self, value):\n- if value is None:\n+ if not value:\n return super().clean([])\n return super().clean([value])\n \n@@ -79,5 +79,7 @@\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.initial['plans'] = self.instance.plans.all()\n- self.fields['plans'\n- ].queryset = self.instance.organisation.plan_set.all()\n+ self.fields['plans'].required = False\n+ self.fields['plans'].empty_label = '----------'\n+ self.fields['plans'].queryset = \\\n+ self.instance.organisation.plan_set.all()\n", "issue": "make it possible to detach plan from project\nnow that a connection of a project with a plan is no longer obligatory, initiators need to be able to undo a connection. Once I click on a plan I can no longer NOT connect to a plan, which should be possible.\r\n\r\n\r\n\n", "before_files": [{"content": "from django import forms\nfrom django.conf import settings\nfrom django.db.models import Q\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom adhocracy4.dashboard.components.forms import ProjectDashboardForm\nfrom adhocracy4.maps import widgets as maps_widgets\nfrom adhocracy4.projects import models as project_models\nfrom meinberlin.apps.contrib import widgets as contrib_widgets\n\nfrom . import models\n\n\ndef get_theme_options():\n return models.Plan.objects\\\n .filter(~Q(theme=''))\\\n .order_by('theme')\\\n .values_list('theme', flat=True)\\\n .distinct()\n\n\nclass PlanForm(forms.ModelForm):\n\n class Meta:\n model = models.Plan\n fields = [\n 'title',\n 'description_image',\n 'contact',\n 'point',\n 'point_label',\n 'district',\n 'cost',\n 'description',\n 'theme',\n 'status',\n 'participation']\n widgets = {\n 'point': maps_widgets.MapChoosePointWidget(\n polygon=settings.BERLIN_POLYGON),\n 'theme': contrib_widgets.TextWithDatalistWidget(attrs={\n 'options': get_theme_options\n })\n }\n error_messages = {\n 'point': {\n 'required': _('Please locate the plan on the map.')\n }\n }\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['district'].empty_label = _('City wide')\n\n\nclass CustomMultipleChoiceField(forms.ModelMultipleChoiceField):\n\n widget = forms.RadioSelect\n\n def clean(self, value):\n if value is None:\n return super().clean([])\n return super().clean([value])\n\n\nclass ProjectPlansDashboardForm(ProjectDashboardForm):\n plans = CustomMultipleChoiceField(queryset=None,\n label=_('Plans'))\n\n class Meta:\n model = project_models.Project\n fields = ['plans']\n required = False\n\n def save(self, commit=False):\n plans = self.cleaned_data['plans']\n self.instance.plans.set(plans)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.initial['plans'] = self.instance.plans.all()\n self.fields['plans'\n ].queryset = self.instance.organisation.plan_set.all()\n", "path": "meinberlin/apps/plans/forms.py"}], "after_files": [{"content": "from django import forms\nfrom django.conf import settings\nfrom django.db.models import Q\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom adhocracy4.dashboard.components.forms import ProjectDashboardForm\nfrom adhocracy4.maps import widgets as maps_widgets\nfrom adhocracy4.projects import models as project_models\nfrom meinberlin.apps.contrib import widgets as contrib_widgets\n\nfrom . import models\n\n\ndef get_theme_options():\n return models.Plan.objects\\\n .filter(~Q(theme=''))\\\n .order_by('theme')\\\n .values_list('theme', flat=True)\\\n .distinct()\n\n\nclass PlanForm(forms.ModelForm):\n\n class Meta:\n model = models.Plan\n fields = [\n 'title',\n 'description_image',\n 'contact',\n 'point',\n 'point_label',\n 'district',\n 'cost',\n 'description',\n 'theme',\n 'status',\n 'participation']\n widgets = {\n 'point': maps_widgets.MapChoosePointWidget(\n polygon=settings.BERLIN_POLYGON),\n 'theme': contrib_widgets.TextWithDatalistWidget(attrs={\n 'options': get_theme_options\n })\n }\n error_messages = {\n 'point': {\n 'required': _('Please locate the plan on the map.')\n }\n }\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.fields['district'].empty_label = _('City wide')\n\n\nclass CustomMultipleChoiceField(forms.ModelMultipleChoiceField):\n\n widget = forms.Select\n\n def clean(self, value):\n if not value:\n return super().clean([])\n return super().clean([value])\n\n\nclass ProjectPlansDashboardForm(ProjectDashboardForm):\n plans = CustomMultipleChoiceField(queryset=None,\n label=_('Plans'))\n\n class Meta:\n model = project_models.Project\n fields = ['plans']\n required = False\n\n def save(self, commit=False):\n plans = self.cleaned_data['plans']\n self.instance.plans.set(plans)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.initial['plans'] = self.instance.plans.all()\n self.fields['plans'].required = False\n self.fields['plans'].empty_label = '----------'\n self.fields['plans'].queryset = \\\n self.instance.organisation.plan_set.all()\n", "path": "meinberlin/apps/plans/forms.py"}]}
| 1,038 | 240 |
gh_patches_debug_1022
|
rasdani/github-patches
|
git_diff
|
searxng__searxng-437
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Bug: microsoft academic engine
**Version of SearXNG, commit number if you are using on master branch and stipulate if you forked SearXNG**
Repository: https://github.com/tiekoetter/searxng
Branch: master
Version: 1.0.0-972-93548243
<!-- Check if these values are correct -->
**How did you install SearXNG?**
<!-- Did you install SearXNG using the official wiki or using searxng-docker
or manually by executing the searx/webapp.py file? -->
**What happened?**
<!-- A clear and concise description of what the bug is. -->
**How To Reproduce**
<!-- How can we reproduce this issue? (as minimally and as precisely as possible) -->
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Screenshots & Logs**
<!-- If applicable, add screenshots, logs to help explain your problem. -->
**Additional context**
<!-- Add any other context about the problem here. -->
**Technical report**
Error
* Error: httpx.TimeoutException
* Percentage: 50
* Parameters: `(None, None, None)`
* File name: `searx/search/processors/online.py:97`
* Function: `_send_http_request`
* Code: `response = req(params['url'], **request_args)`
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `searx/engines/microsoft_academic.py`
Content:
```
1 # SPDX-License-Identifier: AGPL-3.0-or-later
2 """
3 Microsoft Academic (Science)
4 """
5
6 from json import dumps, loads
7 from searx.utils import html_to_text
8
9 # about
10 about = {
11 "website": 'https://academic.microsoft.com',
12 "wikidata_id": 'Q28136779',
13 "official_api_documentation": 'http://ma-graph.org/',
14 "use_official_api": False,
15 "require_api_key": False,
16 "results": 'JSON',
17 }
18
19 categories = ['images']
20 paging = True
21 search_url = 'https://academic.microsoft.com/api/search'
22 _paper_url = 'https://academic.microsoft.com/paper/{id}/reference'
23
24
25 def request(query, params):
26 params['url'] = search_url
27 params['method'] = 'POST'
28 params['headers']['content-type'] = 'application/json; charset=utf-8'
29 params['data'] = dumps({
30 'query': query,
31 'queryExpression': '',
32 'filters': [],
33 'orderBy': 0,
34 'skip': (params['pageno'] - 1) * 10,
35 'sortAscending': True,
36 'take': 10,
37 'includeCitationContexts': False,
38 'profileId': '',
39 })
40
41 return params
42
43
44 def response(resp):
45 results = []
46 response_data = loads(resp.text)
47 if not response_data:
48 return results
49
50 for result in response_data['pr']:
51 if 'dn' not in result['paper']:
52 continue
53
54 title = result['paper']['dn']
55 content = _get_content(result['paper'])
56 url = _paper_url.format(id=result['paper']['id'])
57 results.append({
58 'url': url,
59 'title': html_to_text(title),
60 'content': html_to_text(content),
61 })
62
63 return results
64
65
66 def _get_content(result):
67 if 'd' in result:
68 content = result['d']
69 if len(content) > 300:
70 return content[:300] + '...'
71 return content
72
73 return ''
74
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/searx/engines/microsoft_academic.py b/searx/engines/microsoft_academic.py
--- a/searx/engines/microsoft_academic.py
+++ b/searx/engines/microsoft_academic.py
@@ -47,7 +47,7 @@
if not response_data:
return results
- for result in response_data['pr']:
+ for result in response_data.get('pr', {}):
if 'dn' not in result['paper']:
continue
|
{"golden_diff": "diff --git a/searx/engines/microsoft_academic.py b/searx/engines/microsoft_academic.py\n--- a/searx/engines/microsoft_academic.py\n+++ b/searx/engines/microsoft_academic.py\n@@ -47,7 +47,7 @@\n if not response_data:\n return results\n \n- for result in response_data['pr']:\n+ for result in response_data.get('pr', {}):\n if 'dn' not in result['paper']:\n continue\n", "issue": "Bug: microsoft academic engine\n**Version of SearXNG, commit number if you are using on master branch and stipulate if you forked SearXNG**\r\nRepository: https://github.com/tiekoetter/searxng\r\nBranch: master\r\nVersion: 1.0.0-972-93548243\r\n<!-- Check if these values are correct -->\r\n\r\n**How did you install SearXNG?**\r\n<!-- Did you install SearXNG using the official wiki or using searxng-docker\r\nor manually by executing the searx/webapp.py file? -->\r\n**What happened?**\r\n<!-- A clear and concise description of what the bug is. -->\r\n\r\n**How To Reproduce**\r\n<!-- How can we reproduce this issue? (as minimally and as precisely as possible) -->\r\n\r\n**Expected behavior**\r\n<!-- A clear and concise description of what you expected to happen. -->\r\n\r\n**Screenshots & Logs**\r\n<!-- If applicable, add screenshots, logs to help explain your problem. -->\r\n\r\n**Additional context**\r\n<!-- Add any other context about the problem here. -->\r\n\r\n**Technical report**\r\n\r\nError\r\n * Error: httpx.TimeoutException\r\n * Percentage: 50\r\n * Parameters: `(None, None, None)`\r\n * File name: `searx/search/processors/online.py:97`\r\n * Function: `_send_http_request`\r\n * Code: `response = req(params['url'], **request_args)`\r\n\r\n\n", "before_files": [{"content": "# SPDX-License-Identifier: AGPL-3.0-or-later\n\"\"\"\n Microsoft Academic (Science)\n\"\"\"\n\nfrom json import dumps, loads\nfrom searx.utils import html_to_text\n\n# about\nabout = {\n \"website\": 'https://academic.microsoft.com',\n \"wikidata_id\": 'Q28136779',\n \"official_api_documentation\": 'http://ma-graph.org/',\n \"use_official_api\": False,\n \"require_api_key\": False,\n \"results\": 'JSON',\n}\n\ncategories = ['images']\npaging = True\nsearch_url = 'https://academic.microsoft.com/api/search'\n_paper_url = 'https://academic.microsoft.com/paper/{id}/reference'\n\n\ndef request(query, params):\n params['url'] = search_url\n params['method'] = 'POST'\n params['headers']['content-type'] = 'application/json; charset=utf-8'\n params['data'] = dumps({\n 'query': query,\n 'queryExpression': '',\n 'filters': [],\n 'orderBy': 0,\n 'skip': (params['pageno'] - 1) * 10,\n 'sortAscending': True,\n 'take': 10,\n 'includeCitationContexts': False,\n 'profileId': '',\n })\n\n return params\n\n\ndef response(resp):\n results = []\n response_data = loads(resp.text)\n if not response_data:\n return results\n\n for result in response_data['pr']:\n if 'dn' not in result['paper']:\n continue\n\n title = result['paper']['dn']\n content = _get_content(result['paper'])\n url = _paper_url.format(id=result['paper']['id'])\n results.append({\n 'url': url,\n 'title': html_to_text(title),\n 'content': html_to_text(content),\n })\n\n return results\n\n\ndef _get_content(result):\n if 'd' in result:\n content = result['d']\n if len(content) > 300:\n return content[:300] + '...'\n return content\n\n return ''\n", "path": "searx/engines/microsoft_academic.py"}], "after_files": [{"content": "# SPDX-License-Identifier: AGPL-3.0-or-later\n\"\"\"\n Microsoft Academic (Science)\n\"\"\"\n\nfrom json import dumps, loads\nfrom searx.utils import html_to_text\n\n# about\nabout = {\n \"website\": 'https://academic.microsoft.com',\n \"wikidata_id\": 'Q28136779',\n \"official_api_documentation\": 'http://ma-graph.org/',\n \"use_official_api\": False,\n \"require_api_key\": False,\n \"results\": 'JSON',\n}\n\ncategories = ['images']\npaging = True\nsearch_url = 'https://academic.microsoft.com/api/search'\n_paper_url = 'https://academic.microsoft.com/paper/{id}/reference'\n\n\ndef request(query, params):\n params['url'] = search_url\n params['method'] = 'POST'\n params['headers']['content-type'] = 'application/json; charset=utf-8'\n params['data'] = dumps({\n 'query': query,\n 'queryExpression': '',\n 'filters': [],\n 'orderBy': 0,\n 'skip': (params['pageno'] - 1) * 10,\n 'sortAscending': True,\n 'take': 10,\n 'includeCitationContexts': False,\n 'profileId': '',\n })\n\n return params\n\n\ndef response(resp):\n results = []\n response_data = loads(resp.text)\n if not response_data:\n return results\n\n for result in response_data.get('pr', {}):\n if 'dn' not in result['paper']:\n continue\n\n title = result['paper']['dn']\n content = _get_content(result['paper'])\n url = _paper_url.format(id=result['paper']['id'])\n results.append({\n 'url': url,\n 'title': html_to_text(title),\n 'content': html_to_text(content),\n })\n\n return results\n\n\ndef _get_content(result):\n if 'd' in result:\n content = result['d']\n if len(content) > 300:\n return content[:300] + '...'\n return content\n\n return ''\n", "path": "searx/engines/microsoft_academic.py"}]}
| 1,175 | 115 |
gh_patches_debug_30987
|
rasdani/github-patches
|
git_diff
|
kubeflow__pipelines-5293
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
'kubeflow-pipelines-profile-controller' fails to deploy pods on profile creation when ResourceQuota is set in the profile.
In multi-user mode, it seems like `kubeflow-pipelines-profile-controller` deploy below pods on Kubeflow profile creation.
- `ml-pipeline-ui-artifact`
- `ml-pipeline-visualizationserver`
When `ResourceQuota` is set in the profile, `kubeflow-pipelines-profile-controller` fails to deploy `ml-pipeline-ui-artifact` and `ml-pipeline-visualizationserver` with below error.
```
Warning FailedCreate 17m replicaset-controller
Error creating: pods "ml-pipeline-ui-artifact-684c5db68-s74w8" is forbidden: failed quota: kf-resource-quota: must specify cpu,memory
```
- Related code: [https://github.com/kubeflow/manifests/blob/master/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py](https://github.com/kubeflow/manifests/blob/master/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py)
- Cause: The container resource limit & request is not set on the pod specs, so the pods can not be deployed in the namespace (which has `ResourceQuota` ).
Since Kubeflow profile supports setting `ResourceQuota`, `kubeflow-pipelines-profile-controller` should set container resource requests & limits in pod specs to avoid above errors.
I confirmed that with below patch, ml-pipeline pods are successfully deployed.
```python
diff --git a/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py b/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py
index 75c6e5db..a0e71fbf 100644
--- a/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py
+++ b/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py
@@ -104,6 +104,16 @@ class Controller(BaseHTTPRequestHandler):
"ports": [{
"containerPort": 8888
}],
+ "resources": {
+ "requests": {
+ "cpu": "50m",
+ "memory": "200Mi"
+ },
+ "limits": {
+ "cpu": "500m",
+ "memory": "2Gi"
+ },
+ }
}],
"serviceAccountName":
"default-editor",
@@ -204,7 +214,17 @@ class Controller(BaseHTTPRequestHandler):
"IfNotPresent",
"ports": [{
"containerPort": 3000
- }]
+ }],
+ "resources": {
+ "requests": {
+ "cpu": "50m",
+ "memory": "200Mi"
+ },
+ "limits": {
+ "cpu": "500m",
+ "memory": "2Gi"
+ },
+ }
}],
"serviceAccountName":
"default-editor"
```
Please take a look, thanks.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py`
Content:
```
1 # Copyright 2020 Google LLC
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 from http.server import BaseHTTPRequestHandler, HTTPServer
16 import json
17 import os
18 import base64
19
20 kfp_version = os.environ["KFP_VERSION"]
21 disable_istio_sidecar = os.environ.get("DISABLE_ISTIO_SIDECAR") == "true"
22 mlpipeline_minio_access_key = base64.b64encode(
23 bytes(os.environ.get("MINIO_ACCESS_KEY"), 'utf-8')).decode('utf-8')
24 mlpipeline_minio_secret_key = base64.b64encode(
25 bytes(os.environ.get("MINIO_SECRET_KEY"), 'utf-8')).decode('utf-8')
26
27
28 class Controller(BaseHTTPRequestHandler):
29 def sync(self, parent, children):
30 # HACK: Currently using serving.kubeflow.org/inferenceservice to identify
31 # kubeflow user namespaces.
32 # TODO: let Kubeflow profile controller add a pipeline specific label to
33 # user namespaces and use that label instead.
34 pipeline_enabled = parent.get("metadata", {}).get(
35 "labels", {}).get("serving.kubeflow.org/inferenceservice")
36
37 if not pipeline_enabled:
38 return {"status": {}, "children": []}
39
40 # Compute status based on observed state.
41 desired_status = {
42 "kubeflow-pipelines-ready": \
43 len(children["Secret.v1"]) == 1 and \
44 len(children["ConfigMap.v1"]) == 1 and \
45 len(children["Deployment.apps/v1"]) == 2 and \
46 len(children["Service.v1"]) == 2 and \
47 len(children["DestinationRule.networking.istio.io/v1alpha3"]) == 1 and \
48 len(children["AuthorizationPolicy.security.istio.io/v1beta1"]) == 1 and \
49 "True" or "False"
50 }
51
52 # Generate the desired child object(s).
53 # parent is a namespace
54 namespace = parent.get("metadata", {}).get("name")
55 desired_resources = [
56 {
57 "apiVersion": "v1",
58 "kind": "ConfigMap",
59 "metadata": {
60 "name": "metadata-grpc-configmap",
61 "namespace": namespace,
62 },
63 "data": {
64 "METADATA_GRPC_SERVICE_HOST":
65 "metadata-grpc-service.kubeflow",
66 "METADATA_GRPC_SERVICE_PORT": "8080",
67 },
68 },
69 # Visualization server related manifests below
70 {
71 "apiVersion": "apps/v1",
72 "kind": "Deployment",
73 "metadata": {
74 "labels": {
75 "app": "ml-pipeline-visualizationserver"
76 },
77 "name": "ml-pipeline-visualizationserver",
78 "namespace": namespace,
79 },
80 "spec": {
81 "selector": {
82 "matchLabels": {
83 "app": "ml-pipeline-visualizationserver"
84 },
85 },
86 "template": {
87 "metadata": {
88 "labels": {
89 "app": "ml-pipeline-visualizationserver"
90 },
91 "annotations": disable_istio_sidecar and {
92 "sidecar.istio.io/inject": "false"
93 } or {},
94 },
95 "spec": {
96 "containers": [{
97 "image":
98 "gcr.io/ml-pipeline/visualization-server:" +
99 kfp_version,
100 "imagePullPolicy":
101 "IfNotPresent",
102 "name":
103 "ml-pipeline-visualizationserver",
104 "ports": [{
105 "containerPort": 8888
106 }],
107 }],
108 "serviceAccountName":
109 "default-editor",
110 },
111 },
112 },
113 },
114 {
115 "apiVersion": "networking.istio.io/v1alpha3",
116 "kind": "DestinationRule",
117 "metadata": {
118 "name": "ml-pipeline-visualizationserver",
119 "namespace": namespace,
120 },
121 "spec": {
122 "host": "ml-pipeline-visualizationserver",
123 "trafficPolicy": {
124 "tls": {
125 "mode": "ISTIO_MUTUAL"
126 }
127 }
128 }
129 },
130 {
131 "apiVersion": "security.istio.io/v1beta1",
132 "kind": "AuthorizationPolicy",
133 "metadata": {
134 "name": "ml-pipeline-visualizationserver",
135 "namespace": namespace,
136 },
137 "spec": {
138 "selector": {
139 "matchLabels": {
140 "app": "ml-pipeline-visualizationserver"
141 }
142 },
143 "rules": [{
144 "from": [{
145 "source": {
146 "principals": ["cluster.local/ns/kubeflow/sa/ml-pipeline"]
147 }
148 }]
149 }]
150 }
151 },
152 {
153 "apiVersion": "v1",
154 "kind": "Service",
155 "metadata": {
156 "name": "ml-pipeline-visualizationserver",
157 "namespace": namespace,
158 },
159 "spec": {
160 "ports": [{
161 "name": "http",
162 "port": 8888,
163 "protocol": "TCP",
164 "targetPort": 8888,
165 }],
166 "selector": {
167 "app": "ml-pipeline-visualizationserver",
168 },
169 },
170 },
171 # Artifact fetcher related resources below.
172 {
173 "apiVersion": "apps/v1",
174 "kind": "Deployment",
175 "metadata": {
176 "labels": {
177 "app": "ml-pipeline-ui-artifact"
178 },
179 "name": "ml-pipeline-ui-artifact",
180 "namespace": namespace,
181 },
182 "spec": {
183 "selector": {
184 "matchLabels": {
185 "app": "ml-pipeline-ui-artifact"
186 }
187 },
188 "template": {
189 "metadata": {
190 "labels": {
191 "app": "ml-pipeline-ui-artifact"
192 },
193 "annotations": disable_istio_sidecar and {
194 "sidecar.istio.io/inject": "false"
195 } or {},
196 },
197 "spec": {
198 "containers": [{
199 "name":
200 "ml-pipeline-ui-artifact",
201 "image":
202 "gcr.io/ml-pipeline/frontend:" + kfp_version,
203 "imagePullPolicy":
204 "IfNotPresent",
205 "ports": [{
206 "containerPort": 3000
207 }]
208 }],
209 "serviceAccountName":
210 "default-editor"
211 }
212 }
213 }
214 },
215 {
216 "apiVersion": "v1",
217 "kind": "Service",
218 "metadata": {
219 "name": "ml-pipeline-ui-artifact",
220 "namespace": namespace,
221 "labels": {
222 "app": "ml-pipeline-ui-artifact"
223 }
224 },
225 "spec": {
226 "ports": [{
227 "name":
228 "http", # name is required to let istio understand request protocol
229 "port": 80,
230 "protocol": "TCP",
231 "targetPort": 3000
232 }],
233 "selector": {
234 "app": "ml-pipeline-ui-artifact"
235 }
236 }
237 },
238 ]
239 print('Received request:', parent)
240 print('Desired resources except secrets:', desired_resources)
241 # Moved after the print argument because this is sensitive data.
242 desired_resources.append({
243 "apiVersion": "v1",
244 "kind": "Secret",
245 "metadata": {
246 "name": "mlpipeline-minio-artifact",
247 "namespace": namespace,
248 },
249 "data": {
250 "accesskey": mlpipeline_minio_access_key,
251 "secretkey": mlpipeline_minio_secret_key,
252 },
253 })
254
255 return {"status": desired_status, "children": desired_resources}
256
257 def do_POST(self):
258 # Serve the sync() function as a JSON webhook.
259 observed = json.loads(
260 self.rfile.read(int(self.headers.get("content-length"))))
261 desired = self.sync(observed["parent"], observed["children"])
262
263 self.send_response(200)
264 self.send_header("Content-type", "application/json")
265 self.end_headers()
266 self.wfile.write(bytes(json.dumps(desired), 'utf-8'))
267
268
269 HTTPServer(("", 80), Controller).serve_forever()
270
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py b/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py
--- a/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py
+++ b/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py
@@ -104,6 +104,16 @@
"ports": [{
"containerPort": 8888
}],
+ "resources": {
+ "requests": {
+ "cpu": "50m",
+ "memory": "200Mi"
+ },
+ "limits": {
+ "cpu": "500m",
+ "memory": "1Gi"
+ },
+ }
}],
"serviceAccountName":
"default-editor",
@@ -204,7 +214,17 @@
"IfNotPresent",
"ports": [{
"containerPort": 3000
- }]
+ }],
+ "resources": {
+ "requests": {
+ "cpu": "10m",
+ "memory": "70Mi"
+ },
+ "limits": {
+ "cpu": "100m",
+ "memory": "500Mi"
+ },
+ }
}],
"serviceAccountName":
"default-editor"
|
{"golden_diff": "diff --git a/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py b/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py\n--- a/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py\n+++ b/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py\n@@ -104,6 +104,16 @@\n \"ports\": [{\n \"containerPort\": 8888\n }],\n+ \"resources\": {\n+ \"requests\": {\n+ \"cpu\": \"50m\",\n+ \"memory\": \"200Mi\"\n+ },\n+ \"limits\": {\n+ \"cpu\": \"500m\",\n+ \"memory\": \"1Gi\"\n+ },\n+ }\n }],\n \"serviceAccountName\":\n \"default-editor\",\n@@ -204,7 +214,17 @@\n \"IfNotPresent\",\n \"ports\": [{\n \"containerPort\": 3000\n- }]\n+ }],\n+ \"resources\": {\n+ \"requests\": {\n+ \"cpu\": \"10m\",\n+ \"memory\": \"70Mi\"\n+ },\n+ \"limits\": {\n+ \"cpu\": \"100m\",\n+ \"memory\": \"500Mi\"\n+ },\n+ }\n }],\n \"serviceAccountName\":\n \"default-editor\"\n", "issue": "'kubeflow-pipelines-profile-controller' fails to deploy pods on profile creation when ResourceQuota is set in the profile.\nIn multi-user mode, it seems like `kubeflow-pipelines-profile-controller` deploy below pods on Kubeflow profile creation.\r\n\r\n- `ml-pipeline-ui-artifact`\r\n- `ml-pipeline-visualizationserver`\r\n\r\nWhen `ResourceQuota` is set in the profile, `kubeflow-pipelines-profile-controller` fails to deploy `ml-pipeline-ui-artifact` and `ml-pipeline-visualizationserver` with below error.\r\n\r\n```\r\nWarning FailedCreate 17m replicaset-controller \r\nError creating: pods \"ml-pipeline-ui-artifact-684c5db68-s74w8\" is forbidden: failed quota: kf-resource-quota: must specify cpu,memory\r\n```\r\n\r\n- Related code: [https://github.com/kubeflow/manifests/blob/master/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py](https://github.com/kubeflow/manifests/blob/master/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py)\r\n- Cause: The container resource limit & request is not set on the pod specs, so the pods can not be deployed in the namespace (which has `ResourceQuota` ).\r\n\r\nSince Kubeflow profile supports setting `ResourceQuota`, `kubeflow-pipelines-profile-controller` should set container resource requests & limits in pod specs to avoid above errors. \r\n\r\nI confirmed that with below patch, ml-pipeline pods are successfully deployed.\r\n\r\n```python\r\ndiff --git a/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py b/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py\r\nindex 75c6e5db..a0e71fbf 100644\r\n--- a/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py\r\n+++ b/apps/pipeline/upstream/installs/multi-user/pipelines-profile-controller/sync.py\r\n@@ -104,6 +104,16 @@ class Controller(BaseHTTPRequestHandler):\r\n \"ports\": [{\r\n \"containerPort\": 8888\r\n }],\r\n+ \"resources\": {\r\n+ \"requests\": {\r\n+ \"cpu\": \"50m\",\r\n+ \"memory\": \"200Mi\"\r\n+ },\r\n+ \"limits\": {\r\n+ \"cpu\": \"500m\",\r\n+ \"memory\": \"2Gi\"\r\n+ },\r\n+ }\r\n }],\r\n \"serviceAccountName\":\r\n \"default-editor\",\r\n@@ -204,7 +214,17 @@ class Controller(BaseHTTPRequestHandler):\r\n \"IfNotPresent\",\r\n \"ports\": [{\r\n \"containerPort\": 3000\r\n- }]\r\n+ }],\r\n+ \"resources\": {\r\n+ \"requests\": {\r\n+ \"cpu\": \"50m\",\r\n+ \"memory\": \"200Mi\"\r\n+ },\r\n+ \"limits\": {\r\n+ \"cpu\": \"500m\",\r\n+ \"memory\": \"2Gi\"\r\n+ },\r\n+ }\r\n }],\r\n \"serviceAccountName\":\r\n \"default-editor\"\r\n```\r\n\r\nPlease take a look, thanks.\n", "before_files": [{"content": "# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\nimport json\nimport os\nimport base64\n\nkfp_version = os.environ[\"KFP_VERSION\"]\ndisable_istio_sidecar = os.environ.get(\"DISABLE_ISTIO_SIDECAR\") == \"true\"\nmlpipeline_minio_access_key = base64.b64encode(\n bytes(os.environ.get(\"MINIO_ACCESS_KEY\"), 'utf-8')).decode('utf-8')\nmlpipeline_minio_secret_key = base64.b64encode(\n bytes(os.environ.get(\"MINIO_SECRET_KEY\"), 'utf-8')).decode('utf-8')\n\n\nclass Controller(BaseHTTPRequestHandler):\n def sync(self, parent, children):\n # HACK: Currently using serving.kubeflow.org/inferenceservice to identify\n # kubeflow user namespaces.\n # TODO: let Kubeflow profile controller add a pipeline specific label to\n # user namespaces and use that label instead.\n pipeline_enabled = parent.get(\"metadata\", {}).get(\n \"labels\", {}).get(\"serving.kubeflow.org/inferenceservice\")\n\n if not pipeline_enabled:\n return {\"status\": {}, \"children\": []}\n\n # Compute status based on observed state.\n desired_status = {\n \"kubeflow-pipelines-ready\": \\\n len(children[\"Secret.v1\"]) == 1 and \\\n len(children[\"ConfigMap.v1\"]) == 1 and \\\n len(children[\"Deployment.apps/v1\"]) == 2 and \\\n len(children[\"Service.v1\"]) == 2 and \\\n len(children[\"DestinationRule.networking.istio.io/v1alpha3\"]) == 1 and \\\n len(children[\"AuthorizationPolicy.security.istio.io/v1beta1\"]) == 1 and \\\n \"True\" or \"False\"\n }\n\n # Generate the desired child object(s).\n # parent is a namespace\n namespace = parent.get(\"metadata\", {}).get(\"name\")\n desired_resources = [\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"ConfigMap\",\n \"metadata\": {\n \"name\": \"metadata-grpc-configmap\",\n \"namespace\": namespace,\n },\n \"data\": {\n \"METADATA_GRPC_SERVICE_HOST\":\n \"metadata-grpc-service.kubeflow\",\n \"METADATA_GRPC_SERVICE_PORT\": \"8080\",\n },\n },\n # Visualization server related manifests below\n {\n \"apiVersion\": \"apps/v1\",\n \"kind\": \"Deployment\",\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n },\n \"template\": {\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n \"annotations\": disable_istio_sidecar and {\n \"sidecar.istio.io/inject\": \"false\"\n } or {},\n },\n \"spec\": {\n \"containers\": [{\n \"image\":\n \"gcr.io/ml-pipeline/visualization-server:\" +\n kfp_version,\n \"imagePullPolicy\":\n \"IfNotPresent\",\n \"name\":\n \"ml-pipeline-visualizationserver\",\n \"ports\": [{\n \"containerPort\": 8888\n }],\n }],\n \"serviceAccountName\":\n \"default-editor\",\n },\n },\n },\n },\n {\n \"apiVersion\": \"networking.istio.io/v1alpha3\",\n \"kind\": \"DestinationRule\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"host\": \"ml-pipeline-visualizationserver\",\n \"trafficPolicy\": {\n \"tls\": {\n \"mode\": \"ISTIO_MUTUAL\"\n }\n }\n }\n },\n {\n \"apiVersion\": \"security.istio.io/v1beta1\",\n \"kind\": \"AuthorizationPolicy\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n }\n },\n \"rules\": [{\n \"from\": [{\n \"source\": {\n \"principals\": [\"cluster.local/ns/kubeflow/sa/ml-pipeline\"]\n }\n }]\n }]\n }\n },\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"Service\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"ports\": [{\n \"name\": \"http\",\n \"port\": 8888,\n \"protocol\": \"TCP\",\n \"targetPort\": 8888,\n }],\n \"selector\": {\n \"app\": \"ml-pipeline-visualizationserver\",\n },\n },\n },\n # Artifact fetcher related resources below.\n {\n \"apiVersion\": \"apps/v1\",\n \"kind\": \"Deployment\",\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n },\n \"name\": \"ml-pipeline-ui-artifact\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n },\n \"template\": {\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n },\n \"annotations\": disable_istio_sidecar and {\n \"sidecar.istio.io/inject\": \"false\"\n } or {},\n },\n \"spec\": {\n \"containers\": [{\n \"name\":\n \"ml-pipeline-ui-artifact\",\n \"image\":\n \"gcr.io/ml-pipeline/frontend:\" + kfp_version,\n \"imagePullPolicy\":\n \"IfNotPresent\",\n \"ports\": [{\n \"containerPort\": 3000\n }]\n }],\n \"serviceAccountName\":\n \"default-editor\"\n }\n }\n }\n },\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"Service\",\n \"metadata\": {\n \"name\": \"ml-pipeline-ui-artifact\",\n \"namespace\": namespace,\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n },\n \"spec\": {\n \"ports\": [{\n \"name\":\n \"http\", # name is required to let istio understand request protocol\n \"port\": 80,\n \"protocol\": \"TCP\",\n \"targetPort\": 3000\n }],\n \"selector\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n }\n },\n ]\n print('Received request:', parent)\n print('Desired resources except secrets:', desired_resources)\n # Moved after the print argument because this is sensitive data.\n desired_resources.append({\n \"apiVersion\": \"v1\",\n \"kind\": \"Secret\",\n \"metadata\": {\n \"name\": \"mlpipeline-minio-artifact\",\n \"namespace\": namespace,\n },\n \"data\": {\n \"accesskey\": mlpipeline_minio_access_key,\n \"secretkey\": mlpipeline_minio_secret_key,\n },\n })\n\n return {\"status\": desired_status, \"children\": desired_resources}\n\n def do_POST(self):\n # Serve the sync() function as a JSON webhook.\n observed = json.loads(\n self.rfile.read(int(self.headers.get(\"content-length\"))))\n desired = self.sync(observed[\"parent\"], observed[\"children\"])\n\n self.send_response(200)\n self.send_header(\"Content-type\", \"application/json\")\n self.end_headers()\n self.wfile.write(bytes(json.dumps(desired), 'utf-8'))\n\n\nHTTPServer((\"\", 80), Controller).serve_forever()\n", "path": "manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py"}], "after_files": [{"content": "# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\nimport json\nimport os\nimport base64\n\nkfp_version = os.environ[\"KFP_VERSION\"]\ndisable_istio_sidecar = os.environ.get(\"DISABLE_ISTIO_SIDECAR\") == \"true\"\nmlpipeline_minio_access_key = base64.b64encode(\n bytes(os.environ.get(\"MINIO_ACCESS_KEY\"), 'utf-8')).decode('utf-8')\nmlpipeline_minio_secret_key = base64.b64encode(\n bytes(os.environ.get(\"MINIO_SECRET_KEY\"), 'utf-8')).decode('utf-8')\n\n\nclass Controller(BaseHTTPRequestHandler):\n def sync(self, parent, children):\n # HACK: Currently using serving.kubeflow.org/inferenceservice to identify\n # kubeflow user namespaces.\n # TODO: let Kubeflow profile controller add a pipeline specific label to\n # user namespaces and use that label instead.\n pipeline_enabled = parent.get(\"metadata\", {}).get(\n \"labels\", {}).get(\"serving.kubeflow.org/inferenceservice\")\n\n if not pipeline_enabled:\n return {\"status\": {}, \"children\": []}\n\n # Compute status based on observed state.\n desired_status = {\n \"kubeflow-pipelines-ready\": \\\n len(children[\"Secret.v1\"]) == 1 and \\\n len(children[\"ConfigMap.v1\"]) == 1 and \\\n len(children[\"Deployment.apps/v1\"]) == 2 and \\\n len(children[\"Service.v1\"]) == 2 and \\\n len(children[\"DestinationRule.networking.istio.io/v1alpha3\"]) == 1 and \\\n len(children[\"AuthorizationPolicy.security.istio.io/v1beta1\"]) == 1 and \\\n \"True\" or \"False\"\n }\n\n # Generate the desired child object(s).\n # parent is a namespace\n namespace = parent.get(\"metadata\", {}).get(\"name\")\n desired_resources = [\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"ConfigMap\",\n \"metadata\": {\n \"name\": \"metadata-grpc-configmap\",\n \"namespace\": namespace,\n },\n \"data\": {\n \"METADATA_GRPC_SERVICE_HOST\":\n \"metadata-grpc-service.kubeflow\",\n \"METADATA_GRPC_SERVICE_PORT\": \"8080\",\n },\n },\n # Visualization server related manifests below\n {\n \"apiVersion\": \"apps/v1\",\n \"kind\": \"Deployment\",\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n },\n \"template\": {\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n },\n \"annotations\": disable_istio_sidecar and {\n \"sidecar.istio.io/inject\": \"false\"\n } or {},\n },\n \"spec\": {\n \"containers\": [{\n \"image\":\n \"gcr.io/ml-pipeline/visualization-server:\" +\n kfp_version,\n \"imagePullPolicy\":\n \"IfNotPresent\",\n \"name\":\n \"ml-pipeline-visualizationserver\",\n \"ports\": [{\n \"containerPort\": 8888\n }],\n \"resources\": {\n \"requests\": {\n \"cpu\": \"50m\",\n \"memory\": \"200Mi\"\n },\n \"limits\": {\n \"cpu\": \"500m\",\n \"memory\": \"1Gi\"\n },\n }\n }],\n \"serviceAccountName\":\n \"default-editor\",\n },\n },\n },\n },\n {\n \"apiVersion\": \"networking.istio.io/v1alpha3\",\n \"kind\": \"DestinationRule\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"host\": \"ml-pipeline-visualizationserver\",\n \"trafficPolicy\": {\n \"tls\": {\n \"mode\": \"ISTIO_MUTUAL\"\n }\n }\n }\n },\n {\n \"apiVersion\": \"security.istio.io/v1beta1\",\n \"kind\": \"AuthorizationPolicy\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-visualizationserver\"\n }\n },\n \"rules\": [{\n \"from\": [{\n \"source\": {\n \"principals\": [\"cluster.local/ns/kubeflow/sa/ml-pipeline\"]\n }\n }]\n }]\n }\n },\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"Service\",\n \"metadata\": {\n \"name\": \"ml-pipeline-visualizationserver\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"ports\": [{\n \"name\": \"http\",\n \"port\": 8888,\n \"protocol\": \"TCP\",\n \"targetPort\": 8888,\n }],\n \"selector\": {\n \"app\": \"ml-pipeline-visualizationserver\",\n },\n },\n },\n # Artifact fetcher related resources below.\n {\n \"apiVersion\": \"apps/v1\",\n \"kind\": \"Deployment\",\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n },\n \"name\": \"ml-pipeline-ui-artifact\",\n \"namespace\": namespace,\n },\n \"spec\": {\n \"selector\": {\n \"matchLabels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n },\n \"template\": {\n \"metadata\": {\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n },\n \"annotations\": disable_istio_sidecar and {\n \"sidecar.istio.io/inject\": \"false\"\n } or {},\n },\n \"spec\": {\n \"containers\": [{\n \"name\":\n \"ml-pipeline-ui-artifact\",\n \"image\":\n \"gcr.io/ml-pipeline/frontend:\" + kfp_version,\n \"imagePullPolicy\":\n \"IfNotPresent\",\n \"ports\": [{\n \"containerPort\": 3000\n }],\n \"resources\": {\n \"requests\": {\n \"cpu\": \"10m\",\n \"memory\": \"70Mi\"\n },\n \"limits\": {\n \"cpu\": \"100m\",\n \"memory\": \"500Mi\"\n },\n }\n }],\n \"serviceAccountName\":\n \"default-editor\"\n }\n }\n }\n },\n {\n \"apiVersion\": \"v1\",\n \"kind\": \"Service\",\n \"metadata\": {\n \"name\": \"ml-pipeline-ui-artifact\",\n \"namespace\": namespace,\n \"labels\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n },\n \"spec\": {\n \"ports\": [{\n \"name\":\n \"http\", # name is required to let istio understand request protocol\n \"port\": 80,\n \"protocol\": \"TCP\",\n \"targetPort\": 3000\n }],\n \"selector\": {\n \"app\": \"ml-pipeline-ui-artifact\"\n }\n }\n },\n ]\n print('Received request:', parent)\n print('Desired resources except secrets:', desired_resources)\n # Moved after the print argument because this is sensitive data.\n desired_resources.append({\n \"apiVersion\": \"v1\",\n \"kind\": \"Secret\",\n \"metadata\": {\n \"name\": \"mlpipeline-minio-artifact\",\n \"namespace\": namespace,\n },\n \"data\": {\n \"accesskey\": mlpipeline_minio_access_key,\n \"secretkey\": mlpipeline_minio_secret_key,\n },\n })\n\n return {\"status\": desired_status, \"children\": desired_resources}\n\n def do_POST(self):\n # Serve the sync() function as a JSON webhook.\n observed = json.loads(\n self.rfile.read(int(self.headers.get(\"content-length\"))))\n desired = self.sync(observed[\"parent\"], observed[\"children\"])\n\n self.send_response(200)\n self.send_header(\"Content-type\", \"application/json\")\n self.end_headers()\n self.wfile.write(bytes(json.dumps(desired), 'utf-8'))\n\n\nHTTPServer((\"\", 80), Controller).serve_forever()\n", "path": "manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py"}]}
| 3,575 | 331 |
gh_patches_debug_552
|
rasdani/github-patches
|
git_diff
|
pex-tool__pex-880
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Release 2.1.1
On the docket:
+ [x] PEX 2.1.0 regression: pex file won't build inside a running docker image as user #850
+ [x] Fully pin vendored requirements. #853
+ [x] Fix `tox -epackage` to create pex supporting 3.8. #843
+ [x] Pex erroneously warns about needing to use vendored `pkg_resources` for distributions with empty `namespace_packages.txt` metadata files. #840
+ [x] Interpreter discovery and pyenv don't interact well #782
+ [x] ensure_python_interpreter() bootstrapping broken on pypy shard #477
+ [x] Resolve error checking does not account for environment markers. #851
+ [x] Ensure Pex PEX contraints match pex wheel / sdist. #863
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pex/version.py`
Content:
```
1 # Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).
2 # Licensed under the Apache License, Version 2.0 (see LICENSE).
3
4 __version__ = '2.1.0'
5
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pex/version.py b/pex/version.py
--- a/pex/version.py
+++ b/pex/version.py
@@ -1,4 +1,4 @@
# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).
# Licensed under the Apache License, Version 2.0 (see LICENSE).
-__version__ = '2.1.0'
+__version__ = '2.1.1'
|
{"golden_diff": "diff --git a/pex/version.py b/pex/version.py\n--- a/pex/version.py\n+++ b/pex/version.py\n@@ -1,4 +1,4 @@\n # Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n # Licensed under the Apache License, Version 2.0 (see LICENSE).\n \n-__version__ = '2.1.0'\n+__version__ = '2.1.1'\n", "issue": "Release 2.1.1\nOn the docket:\r\n+ [x] PEX 2.1.0 regression: pex file won't build inside a running docker image as user #850\r\n+ [x] Fully pin vendored requirements. #853\r\n+ [x] Fix `tox -epackage` to create pex supporting 3.8. #843\r\n+ [x] Pex erroneously warns about needing to use vendored `pkg_resources` for distributions with empty `namespace_packages.txt` metadata files. #840\r\n+ [x] Interpreter discovery and pyenv don't interact well #782\r\n+ [x] ensure_python_interpreter() bootstrapping broken on pypy shard #477\r\n+ [x] Resolve error checking does not account for environment markers. #851\r\n+ [x] Ensure Pex PEX contraints match pex wheel / sdist. #863\r\n\r\n\n", "before_files": [{"content": "# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\n__version__ = '2.1.0'\n", "path": "pex/version.py"}], "after_files": [{"content": "# Copyright 2015 Pants project contributors (see CONTRIBUTORS.md).\n# Licensed under the Apache License, Version 2.0 (see LICENSE).\n\n__version__ = '2.1.1'\n", "path": "pex/version.py"}]}
| 515 | 94 |
gh_patches_debug_26729
|
rasdani/github-patches
|
git_diff
|
mathesar-foundation__mathesar-1398
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Allow the user to hide grouping separators for Number and Money
## Current behavior
- Look at this table of numbers. The right-most column is storing years as in integer, which I think will be common. But they are displayed with grouping separators. Ugh. That looks terrible!

- There is currently no way for the user to disable the display of grouping separators.
## Desired behavior
- The user can use display options to enable/disable grouping separators.
## Proposal
### Schema
- Use the `use_grouping` property as [spec'ed in this comment](https://github.com/centerofci/mathesar/discussions/1243#discussioncomment-2519195).
### UX
- Add a new form field labeled "Use Grouping Separators", with a dropdown input having options "Yes", "No", and "Auto". Display this new field before the already-existing "Format" field.
For integers, the default is "No".
For non-integers, and Money, the default is "Yes".
- Modify the "Format" form field to behave as follows:
- If we have an **integer with grouping separators**, then present the following options:
| option | value |
| --------- | ------- |
| Use browser locale | `null` |
| 1,234,567 | `'english'` |
| 1.234.567 | `'german'` |
| 1 234 567 | `'french'` |
| 12,34,567 | `'hindi'` |
| 1'234'567 | `'swiss'` |
- If we have an **integer without grouping separators**, then don't show the "format" option, because it's irrelevant. Keep value as `null`.
- If we have a **non-integer with grouping separators**, then present the following options:
| option | value |
| --------- | ------- |
| Use browser locale | `null` |
| 1,234,567.89 | `'english'` |
| 1.234.567,89 | `'german'` |
| 1 234 567,89 | `'french'` |
| 12,34,567.89 | `'hindi'` |
| 1'234'567.89 | `'swiss'` |
- If we have a **non-integer without grouping separators**, then present the following options:
| option | value |
| --------- | ------- |
| Use browser locale | `null` |
| 1234.56 | `'english'` |
| 1234,56 | `'german'` |
- When changing the "Use Grouping Separators" field from "No" to "Yes", we set a more specific value for the "Format" field, using the browser locale to resolve ambiguity. For example, if the user has configured `format` to `'german'` (because they just wanted a comma decimal separator), but their locale is `fr-FR`, then we change `'german'` to `'french'`, which also has a comma decimal separator but is more likely to match their desired grouping separator. We can use the browser locale to determine the format option by formatting a number and inspecting it. As a further edge case, if the user has configured `format` to `'german'`, but their locale is `en-US` (which is _not_ a more specific form of `'german'`), then we'd leave the value untouched.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mathesar/api/display_options.py`
Content:
```
1 import json
2 from mathesar.database.types import UIType
3 from lazydict import LazyDictionary
4
5
6 def _money_display_options_schema():
7 with open("currency_info.json", "r") as info_file:
8 currency_info = json.load(info_file)
9 currency_codes = list(currency_info.keys())
10 return {
11 "options": [
12 {"name": "currency_code", "type": "string", "enum": currency_codes},
13 {
14 "name": "currency_details",
15 "type": "object",
16 "items": [
17 {"name": "symbol", "type": "string"},
18 {"name": "symbol_location", "type": "number", "enum": [1, -1]},
19 {"name": "decimal_symbol", "type": "string", "enum": [",", "."]},
20 {"name": "digit_grouping", "type": "array"},
21 {"name": "digit_symbol", "type": "string", "enum": [",", ".", " "]}
22 ]
23 }]
24 }
25
26
27 DISPLAY_OPTIONS_BY_UI_TYPE = LazyDictionary(
28 {
29 UIType.BOOLEAN:
30 {
31 "options": [
32 {
33 "name": "input", "type": "string",
34 "enum": ['dropdown', 'checkbox']
35 },
36 {
37 'name': "custom_labels", "type": "object",
38 "items": [
39 {"name": "TRUE", "type": "string"},
40 {'name': "FALSE", "type": "string"}
41 ]
42 }
43 ]
44
45 },
46 UIType.NUMBER:
47 {
48 "options": [
49 {
50 "name": "show_as_percentage",
51 "type": "string",
52 "enum": ['dropdown', 'checkbox']
53 },
54 {
55 "name": "locale",
56 "type": "string"
57 }
58 ]
59 },
60 UIType.DATETIME:
61 {
62 "options": [{"name": "format", "type": "string"}]
63 },
64 UIType.TIME:
65 {
66 "options": [{"name": "format", "type": "string"}]
67 },
68 UIType.DATE:
69 {
70 "options": [{"name": "format", "type": "string"}]
71 },
72 UIType.DURATION:
73 {
74 "options": [
75 {"name": "min", "type": "string"},
76 {"name": "max", "type": "string"},
77 {"name": "show_units", "type": "boolean"},
78 ]
79 },
80 # NOTE: below callable will be evaluated lazily by LazyDictionary
81 UIType.MONEY: _money_display_options_schema,
82 }
83 )
84
```
Path: `mathesar/api/serializers/shared_serializers.py`
Content:
```
1 from django.core.exceptions import ImproperlyConfigured
2 from rest_framework import serializers
3
4 from mathesar.api.exceptions.mixins import MathesarErrorMessageMixin
5 from mathesar.database.types import UIType, get_ui_type_from_db_type
6
7
8 class ReadOnlyPolymorphicSerializerMappingMixin:
9 """
10 This serializer mixin is helpful in serializing polymorphic models,
11 by switching to correct serializer based on the mapping field value.
12 """
13 default_serializer = None
14
15 def __new__(cls, *args, **kwargs):
16 if cls.serializers_mapping is None:
17 raise ImproperlyConfigured(
18 '`{cls}` is missing a '
19 '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)
20 )
21 return super().__new__(cls, *args, **kwargs)
22
23 def _init_serializer(self, serializer_cls, *args, **kwargs):
24 if callable(serializer_cls):
25 serializer = serializer_cls(*args, **kwargs)
26 serializer.parent = self
27 else:
28 serializer = serializer_cls
29 return serializer
30
31 def __init__(self, *args, **kwargs):
32 super().__init__(*args, **kwargs)
33 self.serializers_cls_mapping = {}
34 serializers_mapping = self.serializers_mapping
35 self.serializers_mapping = {}
36 if self.default_serializer is not None:
37 self.default_serializer = self._init_serializer(self.default_serializer, *args, **kwargs)
38 for identifier, serializer_cls in serializers_mapping.items():
39 serializer = self._init_serializer(serializer_cls, *args, **kwargs)
40 self.serializers_mapping[identifier] = serializer
41 self.serializers_cls_mapping[identifier] = serializer_cls
42
43 def get_serializer_class(self, identifier):
44 if identifier in self.serializers_mapping:
45 return self.serializers_mapping.get(identifier)
46 else:
47 return self.default_serializer
48
49 def to_representation(self, instance):
50 serializer = self.get_serializer_class(self.get_mapping_field(instance))
51 if serializer is not None:
52 return serializer.to_representation(instance)
53 else:
54 return instance
55
56 def get_mapping_field(self, data):
57 mapping_field = getattr(self, "mapping_field", None)
58 if mapping_field is None:
59 raise Exception(
60 "Add a `mapping_field` to be used as a identifier"
61 "or override this method to return a identifier to identify a proper serializer"
62 )
63 return mapping_field
64
65
66 class ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):
67 def to_internal_value(self, data):
68 serializer = self.get_serializer_class(self.get_mapping_field(data))
69 if serializer is not None:
70 return serializer.to_internal_value(data=data)
71 else:
72 data = {}
73 return data
74
75 def validate(self, attrs):
76 serializer = self.serializers_mapping.get(self.get_mapping_field(attrs))
77 if serializer is not None:
78 return serializer.validate(attrs)
79 return {}
80
81
82 class MonkeyPatchPartial:
83 """
84 Work around bug #3847 in djangorestframework by monkey-patching the partial
85 attribute of the root serializer during the call to validate_empty_values.
86 https://github.com/encode/django-rest-framework/issues/3847
87 """
88
89 def __init__(self, root):
90 self._root = root
91
92 def __enter__(self):
93 self._old = getattr(self._root, 'partial')
94 setattr(self._root, 'partial', False)
95
96 def __exit__(self, *args):
97 setattr(self._root, 'partial', self._old)
98
99
100 class OverrideRootPartialMixin:
101 """
102 This mixin is used to convert a serializer into a partial serializer,
103 based on the serializer `partial` property rather than the parent's `partial` property.
104 Refer to the issue
105 https://github.com/encode/django-rest-framework/issues/3847
106 """
107
108 def run_validation(self, *args, **kwargs):
109 if not self.partial:
110 with MonkeyPatchPartial(self.root):
111 return super().run_validation(*args, **kwargs)
112 return super().run_validation(*args, **kwargs)
113
114
115 class MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):
116 def get_serializer_fields(self, data):
117 return self.serializers_mapping[self.get_mapping_field(data)].fields
118
119
120 class BaseDisplayOptionsSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):
121 show_fk_preview = serializers.BooleanField(default=True)
122
123
124 class CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):
125 TRUE = serializers.CharField()
126 FALSE = serializers.CharField()
127
128
129 # This is the key which will determine which display options serializer is used. Its value is
130 # supposed to be the column's DB type (a DatabaseType instance).
131 DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'
132
133
134 class BooleanDisplayOptionSerializer(BaseDisplayOptionsSerializer):
135 input = serializers.ChoiceField(choices=[("dropdown", "dropdown"), ("checkbox", "checkbox")])
136 custom_labels = CustomBooleanLabelSerializer(required=False)
137
138
139 class AbstractNumberDisplayOptionSerializer(BaseDisplayOptionsSerializer):
140 number_format = serializers.ChoiceField(
141 required=False,
142 allow_null=True,
143 choices=['english', 'german', 'french', 'hindi', 'swiss']
144 )
145
146
147 class NumberDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):
148 show_as_percentage = serializers.BooleanField(default=False)
149
150
151 class MoneyDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):
152 currency_symbol = serializers.CharField()
153 currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])
154
155
156 class TimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):
157 format = serializers.CharField(max_length=255)
158
159
160 class DateTimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):
161 time_format = serializers.CharField(max_length=255)
162 date_format = serializers.CharField(max_length=255)
163
164
165 class DurationDisplayOptionSerializer(BaseDisplayOptionsSerializer):
166 min = serializers.CharField(max_length=255)
167 max = serializers.CharField(max_length=255)
168 show_units = serializers.BooleanField()
169
170
171 class DisplayOptionsMappingSerializer(
172 OverrideRootPartialMixin,
173 MathesarPolymorphicErrorMixin,
174 ReadWritePolymorphicSerializerMappingMixin,
175 serializers.Serializer
176 ):
177 serializers_mapping = {
178 UIType.BOOLEAN: BooleanDisplayOptionSerializer,
179 UIType.NUMBER: NumberDisplayOptionSerializer,
180 UIType.DATETIME: DateTimeFormatDisplayOptionSerializer,
181 UIType.DATE: TimeFormatDisplayOptionSerializer,
182 UIType.TIME: TimeFormatDisplayOptionSerializer,
183 UIType.DURATION: DurationDisplayOptionSerializer,
184 UIType.MONEY: MoneyDisplayOptionSerializer,
185 }
186 default_serializer = BaseDisplayOptionsSerializer
187
188 def get_mapping_field(self, _):
189 return self._get_ui_type_of_column_being_serialized()
190
191 def _get_ui_type_of_column_being_serialized(self):
192 db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]
193 ui_type = get_ui_type_from_db_type(db_type)
194 return ui_type
195
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/mathesar/api/display_options.py b/mathesar/api/display_options.py
--- a/mathesar/api/display_options.py
+++ b/mathesar/api/display_options.py
@@ -51,6 +51,11 @@
"type": "string",
"enum": ['dropdown', 'checkbox']
},
+ {
+ "name": "use_grouping",
+ "type": "string",
+ "enum": ['true', 'false', 'auto']
+ },
{
"name": "locale",
"type": "string"
diff --git a/mathesar/api/serializers/shared_serializers.py b/mathesar/api/serializers/shared_serializers.py
--- a/mathesar/api/serializers/shared_serializers.py
+++ b/mathesar/api/serializers/shared_serializers.py
@@ -143,6 +143,18 @@
choices=['english', 'german', 'french', 'hindi', 'swiss']
)
+ use_grouping = serializers.ChoiceField(required=False, choices=['true', 'false', 'auto'], default='auto')
+ """
+ The choices here correspond to the options available for the `useGrouping`
+ property within the [Intl API][1]. True and False are encoded as strings
+ instead of booleans to maintain consistency with the Intl API and to keep
+ the type consistent. We did considering using an optional boolean but
+ decided a string would be better, especially if we want to support other
+ options eventually, like "min2".
+
+ [1]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Intl/NumberFormat/NumberFormat
+ """
+
class NumberDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):
show_as_percentage = serializers.BooleanField(default=False)
|
{"golden_diff": "diff --git a/mathesar/api/display_options.py b/mathesar/api/display_options.py\n--- a/mathesar/api/display_options.py\n+++ b/mathesar/api/display_options.py\n@@ -51,6 +51,11 @@\n \"type\": \"string\",\n \"enum\": ['dropdown', 'checkbox']\n },\n+ {\n+ \"name\": \"use_grouping\",\n+ \"type\": \"string\",\n+ \"enum\": ['true', 'false', 'auto']\n+ },\n {\n \"name\": \"locale\",\n \"type\": \"string\"\ndiff --git a/mathesar/api/serializers/shared_serializers.py b/mathesar/api/serializers/shared_serializers.py\n--- a/mathesar/api/serializers/shared_serializers.py\n+++ b/mathesar/api/serializers/shared_serializers.py\n@@ -143,6 +143,18 @@\n choices=['english', 'german', 'french', 'hindi', 'swiss']\n )\n \n+ use_grouping = serializers.ChoiceField(required=False, choices=['true', 'false', 'auto'], default='auto')\n+ \"\"\"\n+ The choices here correspond to the options available for the `useGrouping`\n+ property within the [Intl API][1]. True and False are encoded as strings\n+ instead of booleans to maintain consistency with the Intl API and to keep\n+ the type consistent. We did considering using an optional boolean but\n+ decided a string would be better, especially if we want to support other\n+ options eventually, like \"min2\".\n+\n+ [1]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Intl/NumberFormat/NumberFormat\n+ \"\"\"\n+\n \n class NumberDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):\n show_as_percentage = serializers.BooleanField(default=False)\n", "issue": "Allow the user to hide grouping separators for Number and Money\n## Current behavior\r\n\r\n- Look at this table of numbers. The right-most column is storing years as in integer, which I think will be common. But they are displayed with grouping separators. Ugh. That looks terrible!\r\n\r\n \r\n\r\n- There is currently no way for the user to disable the display of grouping separators.\r\n\r\n## Desired behavior\r\n\r\n- The user can use display options to enable/disable grouping separators.\r\n\r\n## Proposal\r\n\r\n### Schema\r\n\r\n- Use the `use_grouping` property as [spec'ed in this comment](https://github.com/centerofci/mathesar/discussions/1243#discussioncomment-2519195).\r\n\r\n### UX\r\n\r\n- Add a new form field labeled \"Use Grouping Separators\", with a dropdown input having options \"Yes\", \"No\", and \"Auto\". Display this new field before the already-existing \"Format\" field.\r\n\r\n For integers, the default is \"No\".\r\n\r\n For non-integers, and Money, the default is \"Yes\".\r\n\r\n- Modify the \"Format\" form field to behave as follows:\r\n\r\n - If we have an **integer with grouping separators**, then present the following options:\r\n\r\n | option | value |\r\n | --------- | ------- |\r\n | Use browser locale | `null` |\r\n | 1,234,567 | `'english'` |\r\n | 1.234.567 | `'german'` |\r\n | 1 234 567 | `'french'` |\r\n | 12,34,567 | `'hindi'` |\r\n | 1'234'567 | `'swiss'` |\r\n\r\n - If we have an **integer without grouping separators**, then don't show the \"format\" option, because it's irrelevant. Keep value as `null`.\r\n\r\n - If we have a **non-integer with grouping separators**, then present the following options:\r\n\r\n | option | value |\r\n | --------- | ------- |\r\n | Use browser locale | `null` |\r\n | 1,234,567.89 | `'english'` |\r\n | 1.234.567,89 | `'german'` |\r\n | 1 234 567,89 | `'french'` |\r\n | 12,34,567.89 | `'hindi'` |\r\n | 1'234'567.89 | `'swiss'` |\r\n\r\n - If we have a **non-integer without grouping separators**, then present the following options:\r\n\r\n | option | value |\r\n | --------- | ------- |\r\n | Use browser locale | `null` |\r\n | 1234.56 | `'english'` |\r\n | 1234,56 | `'german'` |\r\n\r\n- When changing the \"Use Grouping Separators\" field from \"No\" to \"Yes\", we set a more specific value for the \"Format\" field, using the browser locale to resolve ambiguity. For example, if the user has configured `format` to `'german'` (because they just wanted a comma decimal separator), but their locale is `fr-FR`, then we change `'german'` to `'french'`, which also has a comma decimal separator but is more likely to match their desired grouping separator. We can use the browser locale to determine the format option by formatting a number and inspecting it. As a further edge case, if the user has configured `format` to `'german'`, but their locale is `en-US` (which is _not_ a more specific form of `'german'`), then we'd leave the value untouched.\r\n\r\n\n", "before_files": [{"content": "import json\nfrom mathesar.database.types import UIType\nfrom lazydict import LazyDictionary\n\n\ndef _money_display_options_schema():\n with open(\"currency_info.json\", \"r\") as info_file:\n currency_info = json.load(info_file)\n currency_codes = list(currency_info.keys())\n return {\n \"options\": [\n {\"name\": \"currency_code\", \"type\": \"string\", \"enum\": currency_codes},\n {\n \"name\": \"currency_details\",\n \"type\": \"object\",\n \"items\": [\n {\"name\": \"symbol\", \"type\": \"string\"},\n {\"name\": \"symbol_location\", \"type\": \"number\", \"enum\": [1, -1]},\n {\"name\": \"decimal_symbol\", \"type\": \"string\", \"enum\": [\",\", \".\"]},\n {\"name\": \"digit_grouping\", \"type\": \"array\"},\n {\"name\": \"digit_symbol\", \"type\": \"string\", \"enum\": [\",\", \".\", \" \"]}\n ]\n }]\n }\n\n\nDISPLAY_OPTIONS_BY_UI_TYPE = LazyDictionary(\n {\n UIType.BOOLEAN:\n {\n \"options\": [\n {\n \"name\": \"input\", \"type\": \"string\",\n \"enum\": ['dropdown', 'checkbox']\n },\n {\n 'name': \"custom_labels\", \"type\": \"object\",\n \"items\": [\n {\"name\": \"TRUE\", \"type\": \"string\"},\n {'name': \"FALSE\", \"type\": \"string\"}\n ]\n }\n ]\n\n },\n UIType.NUMBER:\n {\n \"options\": [\n {\n \"name\": \"show_as_percentage\",\n \"type\": \"string\",\n \"enum\": ['dropdown', 'checkbox']\n },\n {\n \"name\": \"locale\",\n \"type\": \"string\"\n }\n ]\n },\n UIType.DATETIME:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.TIME:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.DATE:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.DURATION:\n {\n \"options\": [\n {\"name\": \"min\", \"type\": \"string\"},\n {\"name\": \"max\", \"type\": \"string\"},\n {\"name\": \"show_units\", \"type\": \"boolean\"},\n ]\n },\n # NOTE: below callable will be evaluated lazily by LazyDictionary\n UIType.MONEY: _money_display_options_schema,\n }\n)\n", "path": "mathesar/api/display_options.py"}, {"content": "from django.core.exceptions import ImproperlyConfigured\nfrom rest_framework import serializers\n\nfrom mathesar.api.exceptions.mixins import MathesarErrorMessageMixin\nfrom mathesar.database.types import UIType, get_ui_type_from_db_type\n\n\nclass ReadOnlyPolymorphicSerializerMappingMixin:\n \"\"\"\n This serializer mixin is helpful in serializing polymorphic models,\n by switching to correct serializer based on the mapping field value.\n \"\"\"\n default_serializer = None\n\n def __new__(cls, *args, **kwargs):\n if cls.serializers_mapping is None:\n raise ImproperlyConfigured(\n '`{cls}` is missing a '\n '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)\n )\n return super().__new__(cls, *args, **kwargs)\n\n def _init_serializer(self, serializer_cls, *args, **kwargs):\n if callable(serializer_cls):\n serializer = serializer_cls(*args, **kwargs)\n serializer.parent = self\n else:\n serializer = serializer_cls\n return serializer\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.serializers_cls_mapping = {}\n serializers_mapping = self.serializers_mapping\n self.serializers_mapping = {}\n if self.default_serializer is not None:\n self.default_serializer = self._init_serializer(self.default_serializer, *args, **kwargs)\n for identifier, serializer_cls in serializers_mapping.items():\n serializer = self._init_serializer(serializer_cls, *args, **kwargs)\n self.serializers_mapping[identifier] = serializer\n self.serializers_cls_mapping[identifier] = serializer_cls\n\n def get_serializer_class(self, identifier):\n if identifier in self.serializers_mapping:\n return self.serializers_mapping.get(identifier)\n else:\n return self.default_serializer\n\n def to_representation(self, instance):\n serializer = self.get_serializer_class(self.get_mapping_field(instance))\n if serializer is not None:\n return serializer.to_representation(instance)\n else:\n return instance\n\n def get_mapping_field(self, data):\n mapping_field = getattr(self, \"mapping_field\", None)\n if mapping_field is None:\n raise Exception(\n \"Add a `mapping_field` to be used as a identifier\"\n \"or override this method to return a identifier to identify a proper serializer\"\n )\n return mapping_field\n\n\nclass ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):\n def to_internal_value(self, data):\n serializer = self.get_serializer_class(self.get_mapping_field(data))\n if serializer is not None:\n return serializer.to_internal_value(data=data)\n else:\n data = {}\n return data\n\n def validate(self, attrs):\n serializer = self.serializers_mapping.get(self.get_mapping_field(attrs))\n if serializer is not None:\n return serializer.validate(attrs)\n return {}\n\n\nclass MonkeyPatchPartial:\n \"\"\"\n Work around bug #3847 in djangorestframework by monkey-patching the partial\n attribute of the root serializer during the call to validate_empty_values.\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def __init__(self, root):\n self._root = root\n\n def __enter__(self):\n self._old = getattr(self._root, 'partial')\n setattr(self._root, 'partial', False)\n\n def __exit__(self, *args):\n setattr(self._root, 'partial', self._old)\n\n\nclass OverrideRootPartialMixin:\n \"\"\"\n This mixin is used to convert a serializer into a partial serializer,\n based on the serializer `partial` property rather than the parent's `partial` property.\n Refer to the issue\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def run_validation(self, *args, **kwargs):\n if not self.partial:\n with MonkeyPatchPartial(self.root):\n return super().run_validation(*args, **kwargs)\n return super().run_validation(*args, **kwargs)\n\n\nclass MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):\n def get_serializer_fields(self, data):\n return self.serializers_mapping[self.get_mapping_field(data)].fields\n\n\nclass BaseDisplayOptionsSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n show_fk_preview = serializers.BooleanField(default=True)\n\n\nclass CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):\n TRUE = serializers.CharField()\n FALSE = serializers.CharField()\n\n\n# This is the key which will determine which display options serializer is used. Its value is\n# supposed to be the column's DB type (a DatabaseType instance).\nDISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'\n\n\nclass BooleanDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n input = serializers.ChoiceField(choices=[(\"dropdown\", \"dropdown\"), (\"checkbox\", \"checkbox\")])\n custom_labels = CustomBooleanLabelSerializer(required=False)\n\n\nclass AbstractNumberDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n number_format = serializers.ChoiceField(\n required=False,\n allow_null=True,\n choices=['english', 'german', 'french', 'hindi', 'swiss']\n )\n\n\nclass NumberDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):\n show_as_percentage = serializers.BooleanField(default=False)\n\n\nclass MoneyDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):\n currency_symbol = serializers.CharField()\n currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])\n\n\nclass TimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n format = serializers.CharField(max_length=255)\n\n\nclass DateTimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n time_format = serializers.CharField(max_length=255)\n date_format = serializers.CharField(max_length=255)\n\n\nclass DurationDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n min = serializers.CharField(max_length=255)\n max = serializers.CharField(max_length=255)\n show_units = serializers.BooleanField()\n\n\nclass DisplayOptionsMappingSerializer(\n OverrideRootPartialMixin,\n MathesarPolymorphicErrorMixin,\n ReadWritePolymorphicSerializerMappingMixin,\n serializers.Serializer\n):\n serializers_mapping = {\n UIType.BOOLEAN: BooleanDisplayOptionSerializer,\n UIType.NUMBER: NumberDisplayOptionSerializer,\n UIType.DATETIME: DateTimeFormatDisplayOptionSerializer,\n UIType.DATE: TimeFormatDisplayOptionSerializer,\n UIType.TIME: TimeFormatDisplayOptionSerializer,\n UIType.DURATION: DurationDisplayOptionSerializer,\n UIType.MONEY: MoneyDisplayOptionSerializer,\n }\n default_serializer = BaseDisplayOptionsSerializer\n\n def get_mapping_field(self, _):\n return self._get_ui_type_of_column_being_serialized()\n\n def _get_ui_type_of_column_being_serialized(self):\n db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]\n ui_type = get_ui_type_from_db_type(db_type)\n return ui_type\n", "path": "mathesar/api/serializers/shared_serializers.py"}], "after_files": [{"content": "import json\nfrom mathesar.database.types import UIType\nfrom lazydict import LazyDictionary\n\n\ndef _money_display_options_schema():\n with open(\"currency_info.json\", \"r\") as info_file:\n currency_info = json.load(info_file)\n currency_codes = list(currency_info.keys())\n return {\n \"options\": [\n {\"name\": \"currency_code\", \"type\": \"string\", \"enum\": currency_codes},\n {\n \"name\": \"currency_details\",\n \"type\": \"object\",\n \"items\": [\n {\"name\": \"symbol\", \"type\": \"string\"},\n {\"name\": \"symbol_location\", \"type\": \"number\", \"enum\": [1, -1]},\n {\"name\": \"decimal_symbol\", \"type\": \"string\", \"enum\": [\",\", \".\"]},\n {\"name\": \"digit_grouping\", \"type\": \"array\"},\n {\"name\": \"digit_symbol\", \"type\": \"string\", \"enum\": [\",\", \".\", \" \"]}\n ]\n }]\n }\n\n\nDISPLAY_OPTIONS_BY_UI_TYPE = LazyDictionary(\n {\n UIType.BOOLEAN:\n {\n \"options\": [\n {\n \"name\": \"input\", \"type\": \"string\",\n \"enum\": ['dropdown', 'checkbox']\n },\n {\n 'name': \"custom_labels\", \"type\": \"object\",\n \"items\": [\n {\"name\": \"TRUE\", \"type\": \"string\"},\n {'name': \"FALSE\", \"type\": \"string\"}\n ]\n }\n ]\n\n },\n UIType.NUMBER:\n {\n \"options\": [\n {\n \"name\": \"show_as_percentage\",\n \"type\": \"string\",\n \"enum\": ['dropdown', 'checkbox']\n },\n {\n \"name\": \"use_grouping\",\n \"type\": \"string\",\n \"enum\": ['true', 'false', 'auto']\n },\n {\n \"name\": \"locale\",\n \"type\": \"string\"\n }\n ]\n },\n UIType.DATETIME:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.TIME:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.DATE:\n {\n \"options\": [{\"name\": \"format\", \"type\": \"string\"}]\n },\n UIType.DURATION:\n {\n \"options\": [\n {\"name\": \"min\", \"type\": \"string\"},\n {\"name\": \"max\", \"type\": \"string\"},\n {\"name\": \"show_units\", \"type\": \"boolean\"},\n ]\n },\n # NOTE: below callable will be evaluated lazily by LazyDictionary\n UIType.MONEY: _money_display_options_schema,\n }\n)\n", "path": "mathesar/api/display_options.py"}, {"content": "from django.core.exceptions import ImproperlyConfigured\nfrom rest_framework import serializers\n\nfrom mathesar.api.exceptions.mixins import MathesarErrorMessageMixin\nfrom mathesar.database.types import UIType, get_ui_type_from_db_type\n\n\nclass ReadOnlyPolymorphicSerializerMappingMixin:\n \"\"\"\n This serializer mixin is helpful in serializing polymorphic models,\n by switching to correct serializer based on the mapping field value.\n \"\"\"\n default_serializer = None\n\n def __new__(cls, *args, **kwargs):\n if cls.serializers_mapping is None:\n raise ImproperlyConfigured(\n '`{cls}` is missing a '\n '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)\n )\n return super().__new__(cls, *args, **kwargs)\n\n def _init_serializer(self, serializer_cls, *args, **kwargs):\n if callable(serializer_cls):\n serializer = serializer_cls(*args, **kwargs)\n serializer.parent = self\n else:\n serializer = serializer_cls\n return serializer\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.serializers_cls_mapping = {}\n serializers_mapping = self.serializers_mapping\n self.serializers_mapping = {}\n if self.default_serializer is not None:\n self.default_serializer = self._init_serializer(self.default_serializer, *args, **kwargs)\n for identifier, serializer_cls in serializers_mapping.items():\n serializer = self._init_serializer(serializer_cls, *args, **kwargs)\n self.serializers_mapping[identifier] = serializer\n self.serializers_cls_mapping[identifier] = serializer_cls\n\n def get_serializer_class(self, identifier):\n if identifier in self.serializers_mapping:\n return self.serializers_mapping.get(identifier)\n else:\n return self.default_serializer\n\n def to_representation(self, instance):\n serializer = self.get_serializer_class(self.get_mapping_field(instance))\n if serializer is not None:\n return serializer.to_representation(instance)\n else:\n return instance\n\n def get_mapping_field(self, data):\n mapping_field = getattr(self, \"mapping_field\", None)\n if mapping_field is None:\n raise Exception(\n \"Add a `mapping_field` to be used as a identifier\"\n \"or override this method to return a identifier to identify a proper serializer\"\n )\n return mapping_field\n\n\nclass ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):\n def to_internal_value(self, data):\n serializer = self.get_serializer_class(self.get_mapping_field(data))\n if serializer is not None:\n return serializer.to_internal_value(data=data)\n else:\n data = {}\n return data\n\n def validate(self, attrs):\n serializer = self.serializers_mapping.get(self.get_mapping_field(attrs))\n if serializer is not None:\n return serializer.validate(attrs)\n return {}\n\n\nclass MonkeyPatchPartial:\n \"\"\"\n Work around bug #3847 in djangorestframework by monkey-patching the partial\n attribute of the root serializer during the call to validate_empty_values.\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def __init__(self, root):\n self._root = root\n\n def __enter__(self):\n self._old = getattr(self._root, 'partial')\n setattr(self._root, 'partial', False)\n\n def __exit__(self, *args):\n setattr(self._root, 'partial', self._old)\n\n\nclass OverrideRootPartialMixin:\n \"\"\"\n This mixin is used to convert a serializer into a partial serializer,\n based on the serializer `partial` property rather than the parent's `partial` property.\n Refer to the issue\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def run_validation(self, *args, **kwargs):\n if not self.partial:\n with MonkeyPatchPartial(self.root):\n return super().run_validation(*args, **kwargs)\n return super().run_validation(*args, **kwargs)\n\n\nclass MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):\n def get_serializer_fields(self, data):\n return self.serializers_mapping[self.get_mapping_field(data)].fields\n\n\nclass BaseDisplayOptionsSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n show_fk_preview = serializers.BooleanField(default=True)\n\n\nclass CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):\n TRUE = serializers.CharField()\n FALSE = serializers.CharField()\n\n\n# This is the key which will determine which display options serializer is used. Its value is\n# supposed to be the column's DB type (a DatabaseType instance).\nDISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'\n\n\nclass BooleanDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n input = serializers.ChoiceField(choices=[(\"dropdown\", \"dropdown\"), (\"checkbox\", \"checkbox\")])\n custom_labels = CustomBooleanLabelSerializer(required=False)\n\n\nclass AbstractNumberDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n number_format = serializers.ChoiceField(\n required=False,\n allow_null=True,\n choices=['english', 'german', 'french', 'hindi', 'swiss']\n )\n\n use_grouping = serializers.ChoiceField(required=False, choices=['true', 'false', 'auto'], default='auto')\n \"\"\"\n The choices here correspond to the options available for the `useGrouping`\n property within the [Intl API][1]. True and False are encoded as strings\n instead of booleans to maintain consistency with the Intl API and to keep\n the type consistent. We did considering using an optional boolean but\n decided a string would be better, especially if we want to support other\n options eventually, like \"min2\".\n\n [1]: https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Intl/NumberFormat/NumberFormat\n \"\"\"\n\n\nclass NumberDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):\n show_as_percentage = serializers.BooleanField(default=False)\n\n\nclass MoneyDisplayOptionSerializer(AbstractNumberDisplayOptionSerializer):\n currency_symbol = serializers.CharField()\n currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])\n\n\nclass TimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n format = serializers.CharField(max_length=255)\n\n\nclass DateTimeFormatDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n time_format = serializers.CharField(max_length=255)\n date_format = serializers.CharField(max_length=255)\n\n\nclass DurationDisplayOptionSerializer(BaseDisplayOptionsSerializer):\n min = serializers.CharField(max_length=255)\n max = serializers.CharField(max_length=255)\n show_units = serializers.BooleanField()\n\n\nclass DisplayOptionsMappingSerializer(\n OverrideRootPartialMixin,\n MathesarPolymorphicErrorMixin,\n ReadWritePolymorphicSerializerMappingMixin,\n serializers.Serializer\n):\n serializers_mapping = {\n UIType.BOOLEAN: BooleanDisplayOptionSerializer,\n UIType.NUMBER: NumberDisplayOptionSerializer,\n UIType.DATETIME: DateTimeFormatDisplayOptionSerializer,\n UIType.DATE: TimeFormatDisplayOptionSerializer,\n UIType.TIME: TimeFormatDisplayOptionSerializer,\n UIType.DURATION: DurationDisplayOptionSerializer,\n UIType.MONEY: MoneyDisplayOptionSerializer,\n }\n default_serializer = BaseDisplayOptionsSerializer\n\n def get_mapping_field(self, _):\n return self._get_ui_type_of_column_being_serialized()\n\n def _get_ui_type_of_column_being_serialized(self):\n db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]\n ui_type = get_ui_type_from_db_type(db_type)\n return ui_type\n", "path": "mathesar/api/serializers/shared_serializers.py"}]}
| 3,837 | 396 |
gh_patches_debug_29034
|
rasdani/github-patches
|
git_diff
|
bridgecrewio__checkov-4545
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[WARNI] Could not get the package version from pkg_resource
I execute a ADO pipeline that checks a terraform directory with subdirectories that have 4 modules locally.

Checkov version:
docker.io/bridgecrew/checkov:latest
Path of the Terraform maint.tf and directory that checks chekov: PoC\platform\main.tf
Execution:
LOG_LEVEL=DEBUG;docker run --volume $(pwd):/tf bridgecrew/checkov --directory /tf --framework terraform --output junitxml --soft-fail > $(pwd)/CheckovReport.xml
Result:
2023-02-22 12:23:07,252 [MainThread ] [WARNI] Could not get the package version from pkg_resources
I did a list of the files and I can see the checkovReport.xml but it doesn't have enough data to be published:
No Result Found to Publish '/home/vsts/work/1/s/PoC/platform/CheckovReport.xml'.
=====================List all the files of main folder============================
/home/vsts/work/1/s/PoC/platform
total 40
-rw-r--r-- 1 vsts docker 3728 Feb 22 12:22 main.tf
drwxr-xr-x 10 vsts docker 4096 Feb 22 12:22 azure-tfmodule
-rw-r--r-- 1 vsts docker 2563 Feb 22 12:22 README.md
-rw-r--r-- 1 vsts docker 962 Feb 22 12:22 .gitignore
drwxr-xr-x 3 vsts docker 4096 Feb 22 12:22 ..
-rw-r--r-- 1 vsts docker 1256 Feb 22 12:22 versions.tf
-rw-r--r-- 1 vsts docker 1477 Feb 22 12:22 variables.tf
-rw-r--r-- 1 vsts docker 1292 Feb 22 12:22 terraform.tfvars
drwxr-xr-x 3 vsts docker 4096 Feb 22 12:23 .
-rw-r--r-- 1 vsts docker 2191 Feb 22 12:23 CheckovReport.xml
This is an example of one of the modules with the WARNI, on the main.tf, where workingDirectory: "./PoC/platform"
module "azurerm_log_analytics_workspace" {
source = "./azure-tfmodule/azure-loganalytics-workspace"
law_name = module.naming_standard.standard.log-analytics-workspace
location = var.location
resource_group_name = var.resource_group_name
law_sku = var.law_sku
retention_in_days = var.law_retention_in_days
tags = var.tags
depends_on = [
module.naming_standard
]
}
If I try to check a terraform plan in json format not the main.tf I get the same warning.
If I execute same pipeline in the terraform directory without modules it works, the problem is when I try to execute the checkov and review a local modules or the terraform plan output that calls local modules. Is not on this example but I tested too.
This is my yml pipeline stage:
pool:
vmImage: ubuntu-latest
stages:
- stage: "runCheckov"
displayName: "Checkov - Scan Terraform files"
jobs:
- job: "runCheckov"
displayName: "Checkov > Pull, run and publish results of Checkov scan"
steps:
- bash: |
docker pull bridgecrew/checkov
workingDirectory: "./PoC/platform"
displayName: "Pull > bridgecrew/checkov"
- bash: |
LOG_LEVEL=DEBUG;docker run --volume $(pwd):/tf bridgecrew/checkov --directory /tf --framework terraform --output junitxml --soft-fail > $(pwd)/CheckovReport.xml
workingDirectory: "./PoC/platform"
displayName: "Run > checkov"
- script: |
echo "=====================List all the files of main folder============================"
cd ./PoC/platform
pwd
ls -ltra
- task: PublishTestResults@2
inputs:
testRunTitle: "Checkov Results"
failTaskOnFailedTests: true
testResultsFormat: "JUnit"
testResultsFiles: "CheckovReport.xml"
searchFolder: "./PoC/platform"
displayName: "Publish > Checkov scan results"
mergeTestResults: false
publishRunAttachments: true
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `checkov/common/util/stopit/utils.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 """
3 ============
4 stopit.utils
5 ============
6
7 Misc utilities and common resources
8 """
9
10 import functools
11 import logging
12 import sys
13
14 # Custom logger
15 LOG = logging.getLogger(name='stopit')
16
17 if sys.version_info < (2, 7):
18 class NullHandler(logging.Handler):
19 """Copied from Python 2.7 to avoid getting `No handlers could be found
20 for logger "xxx"` http://bugs.python.org/issue16539
21 """
22 def handle(self, record):
23 pass
24
25 def emit(self, record):
26 pass
27
28 def createLock(self):
29 self.lock = None # noqa
30 else:
31 from logging import NullHandler
32
33 LOG.addHandler(NullHandler())
34
35
36 class TimeoutException(Exception):
37 """Raised when the block under context management takes longer to complete
38 than the allowed maximum timeout value.
39 """
40 pass
41
42
43 class BaseTimeout(object):
44 """Context manager for limiting in the time the execution of a block
45
46 :param seconds: ``float`` or ``int`` duration enabled to run the context
47 manager block
48 :param swallow_exc: ``False`` if you want to manage the
49 ``TimeoutException`` (or any other) in an outer ``try ... except``
50 structure. ``True`` (default) if you just want to check the execution of
51 the block with the ``state`` attribute of the context manager.
52 """
53
54 def __init__(self, seconds, swallow_exc=True):
55
56 # Possible values for the ``state`` attribute, self explanative
57 self.EXECUTED, self.EXECUTING, self.TIMED_OUT, self.INTERRUPTED, self.CANCELED = range(5)
58
59 self.seconds = seconds
60 self.swallow_exc = swallow_exc
61 self.state = self.EXECUTED
62
63 def __bool__(self):
64 return self.state in (self.EXECUTED, self.EXECUTING, self.CANCELED)
65
66 def __repr__(self):
67 """Debug helper
68 """
69 return "<{0} in state: {1}>".format(self.__class__.__name__, self.state)
70
71 def __enter__(self):
72 self.state = self.EXECUTING
73 self.setup_interrupt()
74 return self
75
76 def __exit__(self, exc_type, exc_val, exc_tb):
77 if exc_type is TimeoutException:
78 if self.state != self.TIMED_OUT:
79 self.state = self.INTERRUPTED
80 self.suppress_interrupt()
81 LOG.warning("Code block execution exceeded {0} seconds timeout".format(self.seconds),
82 exc_info=(exc_type, exc_val, exc_tb))
83 return self.swallow_exc
84 else:
85 if exc_type is None:
86 self.state = self.EXECUTED
87 self.suppress_interrupt()
88 return False
89
90 def cancel(self):
91 """In case in the block you realize you don't need anymore
92 limitation"""
93 self.state = self.CANCELED
94 self.suppress_interrupt()
95
96 # Methods must be provided by subclasses
97 def suppress_interrupt(self):
98 """Removes/neutralizes the feature that interrupts the executed block
99 """
100 raise NotImplementedError
101
102 def setup_interrupt(self):
103 """Installs/initializes the feature that interrupts the executed block
104 """
105 raise NotImplementedError
106
107
108 class base_timeoutable(object):
109 """A base for function or method decorator that raises a ``TimeoutException`` to
110 decorated functions that should not last a certain amount of time.
111
112 Any decorated callable may receive a ``timeout`` optional parameter that
113 specifies the number of seconds allocated to the callable execution.
114
115 The decorated functions that exceed that timeout return ``None`` or the
116 value provided by the decorator.
117
118 :param default: The default value in case we timed out during the decorated
119 function execution. Default is None.
120
121 :param timeout_param: As adding dynamically a ``timeout`` named parameter
122 to the decorated callable may conflict with the callable signature, you
123 may choose another name to provide that parameter. Your decoration line
124 could look like ``@timeoutable(timeout_param='my_timeout')``
125
126 .. note::
127
128 This is a base class that must be subclassed. subclasses must override
129 thz ``to_ctx_mgr`` with a timeout context manager class which in turn
130 must subclasses of above ``BaseTimeout`` class.
131 """
132
133 def __init__(self, default=None, timeout_param='timeout'):
134 self.to_ctx_mgr = None
135 self.default, self.timeout_param = default, timeout_param
136
137 def __call__(self, func):
138 @functools.wraps(func)
139 def wrapper(*args, **kwargs):
140 timeout = kwargs.pop(self.timeout_param, None)
141 if timeout:
142 with self.to_ctx_mgr(timeout, swallow_exc=True):
143 result = self.default # noqa
144 # ``result`` may not be assigned below in case of timeout
145 result = func(*args, **kwargs)
146 return result
147 else:
148 return func(*args, **kwargs)
149 return wrapper
150
```
Path: `checkov/common/util/stopit/__init__.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 """
3 ======
4 stopit
5 ======
6
7 code from:
8 https://github.com/glenfant/stopit
9 package:
10 https://pypi.org/project/stopit
11
12 Public resources from ``stopit``
13 """
14
15 import pkg_resources
16
17 from .utils import LOG, TimeoutException
18 from .threadstop import ThreadingTimeout, async_raise, threading_timeoutable
19 from .signalstop import SignalTimeout, signal_timeoutable
20
21 # PEP 396 style version marker
22 try:
23 __version__ = pkg_resources.get_distribution(__name__).version
24 except Exception:
25 LOG.warning("Could not get the package version from pkg_resources")
26 __version__ = 'unknown'
27
28 __all__ = (
29 'ThreadingTimeout', 'async_raise', 'threading_timeoutable',
30 'SignalTimeout', 'signal_timeoutable', 'TimeoutException'
31 )
32
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/checkov/common/util/stopit/__init__.py b/checkov/common/util/stopit/__init__.py
--- a/checkov/common/util/stopit/__init__.py
+++ b/checkov/common/util/stopit/__init__.py
@@ -12,18 +12,10 @@
Public resources from ``stopit``
"""
-import pkg_resources
-
-from .utils import LOG, TimeoutException
+from .utils import TimeoutException
from .threadstop import ThreadingTimeout, async_raise, threading_timeoutable
from .signalstop import SignalTimeout, signal_timeoutable
-# PEP 396 style version marker
-try:
- __version__ = pkg_resources.get_distribution(__name__).version
-except Exception:
- LOG.warning("Could not get the package version from pkg_resources")
- __version__ = 'unknown'
__all__ = (
'ThreadingTimeout', 'async_raise', 'threading_timeoutable',
diff --git a/checkov/common/util/stopit/utils.py b/checkov/common/util/stopit/utils.py
--- a/checkov/common/util/stopit/utils.py
+++ b/checkov/common/util/stopit/utils.py
@@ -26,7 +26,7 @@
pass
def createLock(self):
- self.lock = None # noqa
+ self.lock = None
else:
from logging import NullHandler
@@ -140,7 +140,7 @@
timeout = kwargs.pop(self.timeout_param, None)
if timeout:
with self.to_ctx_mgr(timeout, swallow_exc=True):
- result = self.default # noqa
+ result = self.default
# ``result`` may not be assigned below in case of timeout
result = func(*args, **kwargs)
return result
|
{"golden_diff": "diff --git a/checkov/common/util/stopit/__init__.py b/checkov/common/util/stopit/__init__.py\n--- a/checkov/common/util/stopit/__init__.py\n+++ b/checkov/common/util/stopit/__init__.py\n@@ -12,18 +12,10 @@\n Public resources from ``stopit``\n \"\"\"\n \n-import pkg_resources\n-\n-from .utils import LOG, TimeoutException\n+from .utils import TimeoutException\n from .threadstop import ThreadingTimeout, async_raise, threading_timeoutable\n from .signalstop import SignalTimeout, signal_timeoutable\n \n-# PEP 396 style version marker\n-try:\n- __version__ = pkg_resources.get_distribution(__name__).version\n-except Exception:\n- LOG.warning(\"Could not get the package version from pkg_resources\")\n- __version__ = 'unknown'\n \n __all__ = (\n 'ThreadingTimeout', 'async_raise', 'threading_timeoutable',\ndiff --git a/checkov/common/util/stopit/utils.py b/checkov/common/util/stopit/utils.py\n--- a/checkov/common/util/stopit/utils.py\n+++ b/checkov/common/util/stopit/utils.py\n@@ -26,7 +26,7 @@\n pass\n \n def createLock(self):\n- self.lock = None # noqa\n+ self.lock = None\n else:\n from logging import NullHandler\n \n@@ -140,7 +140,7 @@\n timeout = kwargs.pop(self.timeout_param, None)\n if timeout:\n with self.to_ctx_mgr(timeout, swallow_exc=True):\n- result = self.default # noqa\n+ result = self.default\n # ``result`` may not be assigned below in case of timeout\n result = func(*args, **kwargs)\n return result\n", "issue": " [WARNI] Could not get the package version from pkg_resource\nI execute a ADO pipeline that checks a terraform directory with subdirectories that have 4 modules locally.\r\n\r\n\r\n\r\n\r\nCheckov version:\r\ndocker.io/bridgecrew/checkov:latest\r\n\r\nPath of the Terraform maint.tf and directory that checks chekov: PoC\\platform\\main.tf\r\n\r\n\r\nExecution:\r\nLOG_LEVEL=DEBUG;docker run --volume $(pwd):/tf bridgecrew/checkov --directory /tf --framework terraform --output junitxml --soft-fail > $(pwd)/CheckovReport.xml\r\nResult:\r\n2023-02-22 12:23:07,252 [MainThread ] [WARNI] Could not get the package version from pkg_resources\r\n\r\n\r\n\r\nI did a list of the files and I can see the checkovReport.xml but it doesn't have enough data to be published:\r\nNo Result Found to Publish '/home/vsts/work/1/s/PoC/platform/CheckovReport.xml'.\r\n\r\n\r\n=====================List all the files of main folder============================\r\n/home/vsts/work/1/s/PoC/platform\r\ntotal 40\r\n-rw-r--r-- 1 vsts docker 3728 Feb 22 12:22 main.tf\r\ndrwxr-xr-x 10 vsts docker 4096 Feb 22 12:22 azure-tfmodule\r\n-rw-r--r-- 1 vsts docker 2563 Feb 22 12:22 README.md\r\n-rw-r--r-- 1 vsts docker 962 Feb 22 12:22 .gitignore\r\ndrwxr-xr-x 3 vsts docker 4096 Feb 22 12:22 ..\r\n-rw-r--r-- 1 vsts docker 1256 Feb 22 12:22 versions.tf\r\n-rw-r--r-- 1 vsts docker 1477 Feb 22 12:22 variables.tf\r\n-rw-r--r-- 1 vsts docker 1292 Feb 22 12:22 terraform.tfvars\r\ndrwxr-xr-x 3 vsts docker 4096 Feb 22 12:23 .\r\n-rw-r--r-- 1 vsts docker 2191 Feb 22 12:23 CheckovReport.xml\r\n\r\n\r\nThis is an example of one of the modules with the WARNI, on the main.tf, where workingDirectory: \"./PoC/platform\"\r\nmodule \"azurerm_log_analytics_workspace\" {\r\n source = \"./azure-tfmodule/azure-loganalytics-workspace\"\r\n law_name = module.naming_standard.standard.log-analytics-workspace\r\n location = var.location\r\n resource_group_name = var.resource_group_name\r\n law_sku = var.law_sku\r\n retention_in_days = var.law_retention_in_days\r\n tags = var.tags \r\n\r\n depends_on = [\r\n module.naming_standard\r\n ]\r\n}\r\n\r\nIf I try to check a terraform plan in json format not the main.tf I get the same warning. \r\nIf I execute same pipeline in the terraform directory without modules it works, the problem is when I try to execute the checkov and review a local modules or the terraform plan output that calls local modules. Is not on this example but I tested too.\r\n\r\n\r\nThis is my yml pipeline stage:\r\npool:\r\n vmImage: ubuntu-latest\r\n\r\nstages:\r\n- stage: \"runCheckov\"\r\n displayName: \"Checkov - Scan Terraform files\"\r\n jobs:\r\n - job: \"runCheckov\"\r\n displayName: \"Checkov > Pull, run and publish results of Checkov scan\"\r\n steps:\r\n - bash: |\r\n docker pull bridgecrew/checkov\r\n workingDirectory: \"./PoC/platform\"\r\n displayName: \"Pull > bridgecrew/checkov\"\r\n - bash: |\r\n LOG_LEVEL=DEBUG;docker run --volume $(pwd):/tf bridgecrew/checkov --directory /tf --framework terraform --output junitxml --soft-fail > $(pwd)/CheckovReport.xml\r\n workingDirectory: \"./PoC/platform\"\r\n displayName: \"Run > checkov\"\r\n - script: |\r\n echo \"=====================List all the files of main folder============================\"\r\n cd ./PoC/platform\r\n pwd\r\n ls -ltra\r\n - task: PublishTestResults@2\r\n inputs:\r\n testRunTitle: \"Checkov Results\"\r\n failTaskOnFailedTests: true\r\n testResultsFormat: \"JUnit\"\r\n testResultsFiles: \"CheckovReport.xml\"\r\n searchFolder: \"./PoC/platform\"\r\n displayName: \"Publish > Checkov scan results\"\r\n mergeTestResults: false\r\n publishRunAttachments: true\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\n============\nstopit.utils\n============\n\nMisc utilities and common resources\n\"\"\"\n\nimport functools\nimport logging\nimport sys\n\n# Custom logger\nLOG = logging.getLogger(name='stopit')\n\nif sys.version_info < (2, 7):\n class NullHandler(logging.Handler):\n \"\"\"Copied from Python 2.7 to avoid getting `No handlers could be found\n for logger \"xxx\"` http://bugs.python.org/issue16539\n \"\"\"\n def handle(self, record):\n pass\n\n def emit(self, record):\n pass\n\n def createLock(self):\n self.lock = None # noqa\nelse:\n from logging import NullHandler\n\nLOG.addHandler(NullHandler())\n\n\nclass TimeoutException(Exception):\n \"\"\"Raised when the block under context management takes longer to complete\n than the allowed maximum timeout value.\n \"\"\"\n pass\n\n\nclass BaseTimeout(object):\n \"\"\"Context manager for limiting in the time the execution of a block\n\n :param seconds: ``float`` or ``int`` duration enabled to run the context\n manager block\n :param swallow_exc: ``False`` if you want to manage the\n ``TimeoutException`` (or any other) in an outer ``try ... except``\n structure. ``True`` (default) if you just want to check the execution of\n the block with the ``state`` attribute of the context manager.\n \"\"\"\n\n def __init__(self, seconds, swallow_exc=True):\n\n # Possible values for the ``state`` attribute, self explanative\n self.EXECUTED, self.EXECUTING, self.TIMED_OUT, self.INTERRUPTED, self.CANCELED = range(5)\n\n self.seconds = seconds\n self.swallow_exc = swallow_exc\n self.state = self.EXECUTED\n\n def __bool__(self):\n return self.state in (self.EXECUTED, self.EXECUTING, self.CANCELED)\n\n def __repr__(self):\n \"\"\"Debug helper\n \"\"\"\n return \"<{0} in state: {1}>\".format(self.__class__.__name__, self.state)\n\n def __enter__(self):\n self.state = self.EXECUTING\n self.setup_interrupt()\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n if exc_type is TimeoutException:\n if self.state != self.TIMED_OUT:\n self.state = self.INTERRUPTED\n self.suppress_interrupt()\n LOG.warning(\"Code block execution exceeded {0} seconds timeout\".format(self.seconds),\n exc_info=(exc_type, exc_val, exc_tb))\n return self.swallow_exc\n else:\n if exc_type is None:\n self.state = self.EXECUTED\n self.suppress_interrupt()\n return False\n\n def cancel(self):\n \"\"\"In case in the block you realize you don't need anymore\n limitation\"\"\"\n self.state = self.CANCELED\n self.suppress_interrupt()\n\n # Methods must be provided by subclasses\n def suppress_interrupt(self):\n \"\"\"Removes/neutralizes the feature that interrupts the executed block\n \"\"\"\n raise NotImplementedError\n\n def setup_interrupt(self):\n \"\"\"Installs/initializes the feature that interrupts the executed block\n \"\"\"\n raise NotImplementedError\n\n\nclass base_timeoutable(object):\n \"\"\"A base for function or method decorator that raises a ``TimeoutException`` to\n decorated functions that should not last a certain amount of time.\n\n Any decorated callable may receive a ``timeout`` optional parameter that\n specifies the number of seconds allocated to the callable execution.\n\n The decorated functions that exceed that timeout return ``None`` or the\n value provided by the decorator.\n\n :param default: The default value in case we timed out during the decorated\n function execution. Default is None.\n\n :param timeout_param: As adding dynamically a ``timeout`` named parameter\n to the decorated callable may conflict with the callable signature, you\n may choose another name to provide that parameter. Your decoration line\n could look like ``@timeoutable(timeout_param='my_timeout')``\n\n .. note::\n\n This is a base class that must be subclassed. subclasses must override\n thz ``to_ctx_mgr`` with a timeout context manager class which in turn\n must subclasses of above ``BaseTimeout`` class.\n \"\"\"\n\n def __init__(self, default=None, timeout_param='timeout'):\n self.to_ctx_mgr = None\n self.default, self.timeout_param = default, timeout_param\n\n def __call__(self, func):\n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n timeout = kwargs.pop(self.timeout_param, None)\n if timeout:\n with self.to_ctx_mgr(timeout, swallow_exc=True):\n result = self.default # noqa\n # ``result`` may not be assigned below in case of timeout\n result = func(*args, **kwargs)\n return result\n else:\n return func(*args, **kwargs)\n return wrapper\n", "path": "checkov/common/util/stopit/utils.py"}, {"content": "# -*- coding: utf-8 -*-\n\"\"\"\n======\nstopit\n======\n\ncode from:\nhttps://github.com/glenfant/stopit\npackage:\nhttps://pypi.org/project/stopit\n\nPublic resources from ``stopit``\n\"\"\"\n\nimport pkg_resources\n\nfrom .utils import LOG, TimeoutException\nfrom .threadstop import ThreadingTimeout, async_raise, threading_timeoutable\nfrom .signalstop import SignalTimeout, signal_timeoutable\n\n# PEP 396 style version marker\ntry:\n __version__ = pkg_resources.get_distribution(__name__).version\nexcept Exception:\n LOG.warning(\"Could not get the package version from pkg_resources\")\n __version__ = 'unknown'\n\n__all__ = (\n 'ThreadingTimeout', 'async_raise', 'threading_timeoutable',\n 'SignalTimeout', 'signal_timeoutable', 'TimeoutException'\n)\n", "path": "checkov/common/util/stopit/__init__.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\"\"\"\n============\nstopit.utils\n============\n\nMisc utilities and common resources\n\"\"\"\n\nimport functools\nimport logging\nimport sys\n\n# Custom logger\nLOG = logging.getLogger(name='stopit')\n\nif sys.version_info < (2, 7):\n class NullHandler(logging.Handler):\n \"\"\"Copied from Python 2.7 to avoid getting `No handlers could be found\n for logger \"xxx\"` http://bugs.python.org/issue16539\n \"\"\"\n def handle(self, record):\n pass\n\n def emit(self, record):\n pass\n\n def createLock(self):\n self.lock = None\nelse:\n from logging import NullHandler\n\nLOG.addHandler(NullHandler())\n\n\nclass TimeoutException(Exception):\n \"\"\"Raised when the block under context management takes longer to complete\n than the allowed maximum timeout value.\n \"\"\"\n pass\n\n\nclass BaseTimeout(object):\n \"\"\"Context manager for limiting in the time the execution of a block\n\n :param seconds: ``float`` or ``int`` duration enabled to run the context\n manager block\n :param swallow_exc: ``False`` if you want to manage the\n ``TimeoutException`` (or any other) in an outer ``try ... except``\n structure. ``True`` (default) if you just want to check the execution of\n the block with the ``state`` attribute of the context manager.\n \"\"\"\n\n def __init__(self, seconds, swallow_exc=True):\n\n # Possible values for the ``state`` attribute, self explanative\n self.EXECUTED, self.EXECUTING, self.TIMED_OUT, self.INTERRUPTED, self.CANCELED = range(5)\n\n self.seconds = seconds\n self.swallow_exc = swallow_exc\n self.state = self.EXECUTED\n\n def __bool__(self):\n return self.state in (self.EXECUTED, self.EXECUTING, self.CANCELED)\n\n def __repr__(self):\n \"\"\"Debug helper\n \"\"\"\n return \"<{0} in state: {1}>\".format(self.__class__.__name__, self.state)\n\n def __enter__(self):\n self.state = self.EXECUTING\n self.setup_interrupt()\n return self\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n if exc_type is TimeoutException:\n if self.state != self.TIMED_OUT:\n self.state = self.INTERRUPTED\n self.suppress_interrupt()\n LOG.warning(\"Code block execution exceeded {0} seconds timeout\".format(self.seconds),\n exc_info=(exc_type, exc_val, exc_tb))\n return self.swallow_exc\n else:\n if exc_type is None:\n self.state = self.EXECUTED\n self.suppress_interrupt()\n return False\n\n def cancel(self):\n \"\"\"In case in the block you realize you don't need anymore\n limitation\"\"\"\n self.state = self.CANCELED\n self.suppress_interrupt()\n\n # Methods must be provided by subclasses\n def suppress_interrupt(self):\n \"\"\"Removes/neutralizes the feature that interrupts the executed block\n \"\"\"\n raise NotImplementedError\n\n def setup_interrupt(self):\n \"\"\"Installs/initializes the feature that interrupts the executed block\n \"\"\"\n raise NotImplementedError\n\n\nclass base_timeoutable(object):\n \"\"\"A base for function or method decorator that raises a ``TimeoutException`` to\n decorated functions that should not last a certain amount of time.\n\n Any decorated callable may receive a ``timeout`` optional parameter that\n specifies the number of seconds allocated to the callable execution.\n\n The decorated functions that exceed that timeout return ``None`` or the\n value provided by the decorator.\n\n :param default: The default value in case we timed out during the decorated\n function execution. Default is None.\n\n :param timeout_param: As adding dynamically a ``timeout`` named parameter\n to the decorated callable may conflict with the callable signature, you\n may choose another name to provide that parameter. Your decoration line\n could look like ``@timeoutable(timeout_param='my_timeout')``\n\n .. note::\n\n This is a base class that must be subclassed. subclasses must override\n thz ``to_ctx_mgr`` with a timeout context manager class which in turn\n must subclasses of above ``BaseTimeout`` class.\n \"\"\"\n\n def __init__(self, default=None, timeout_param='timeout'):\n self.to_ctx_mgr = None\n self.default, self.timeout_param = default, timeout_param\n\n def __call__(self, func):\n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n timeout = kwargs.pop(self.timeout_param, None)\n if timeout:\n with self.to_ctx_mgr(timeout, swallow_exc=True):\n result = self.default\n # ``result`` may not be assigned below in case of timeout\n result = func(*args, **kwargs)\n return result\n else:\n return func(*args, **kwargs)\n return wrapper\n", "path": "checkov/common/util/stopit/utils.py"}, {"content": "# -*- coding: utf-8 -*-\n\"\"\"\n======\nstopit\n======\n\ncode from:\nhttps://github.com/glenfant/stopit\npackage:\nhttps://pypi.org/project/stopit\n\nPublic resources from ``stopit``\n\"\"\"\n\nfrom .utils import TimeoutException\nfrom .threadstop import ThreadingTimeout, async_raise, threading_timeoutable\nfrom .signalstop import SignalTimeout, signal_timeoutable\n\n\n__all__ = (\n 'ThreadingTimeout', 'async_raise', 'threading_timeoutable',\n 'SignalTimeout', 'signal_timeoutable', 'TimeoutException'\n)\n", "path": "checkov/common/util/stopit/__init__.py"}]}
| 3,069 | 391 |
gh_patches_debug_6017
|
rasdani/github-patches
|
git_diff
|
geopandas__geopandas-2289
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
TST: test_value_counts breaking against pandas latest
Recent CI builds on 38-dev.yaml are all failing due to geopandas/tests/test_pandas_methods.py::test_value_counts, which fails with the output:
```
______________________________ test_value_counts _______________________________
3536
[gw0] linux -- Python 3.8.12 /usr/share/miniconda3/envs/test/bin/python
3537
3538
def test_value_counts():
3539
# each object is considered unique
3540
s = GeoSeries([Point(0, 0), Point(1, 1), Point(0, 0)])
3541
res = s.value_counts()
3542
with compat.ignore_shapely2_warnings():
3543
exp = pd.Series([2, 1], index=[Point(0, 0), Point(1, 1)])
3544
> assert_series_equal(res, exp)
3545
3546
geopandas/tests/test_pandas_methods.py:454:
3547
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
3548
3549
left = Index([POINT (0 0), POINT (1 1)], dtype='geometry')
3550
right = Index([POINT (0 0), POINT (1 1)], dtype='object'), obj = 'Series.index'
3551
3552
def _check_types(left, right, obj="Index") -> None:
3553
if not exact:
3554
return
3555
3556
assert_class_equal(left, right, exact=exact, obj=obj)
3557
> assert_attr_equal("inferred_type", left, right, obj=obj)
3558
E AssertionError: Series.index are different
3559
E
3560
E Attribute "inferred_type" are different
3561
E [left]: unknown-array
3562
E [right]: mixed
```
It looks like this pandas change (https://github.com/pandas-dev/pandas/pull/43930) to allow storing extension arrays in the index (that is, not just casting everything to `object`) is responsible.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `geopandas/_compat.py`
Content:
```
1 import contextlib
2 from distutils.version import LooseVersion
3 import importlib
4 import os
5 import warnings
6
7 import numpy as np
8 import pandas as pd
9 import pyproj
10 import shapely
11 import shapely.geos
12
13
14 # -----------------------------------------------------------------------------
15 # pandas compat
16 # -----------------------------------------------------------------------------
17
18 PANDAS_GE_10 = str(pd.__version__) >= LooseVersion("1.0.0")
19 PANDAS_GE_11 = str(pd.__version__) >= LooseVersion("1.1.0")
20 PANDAS_GE_115 = str(pd.__version__) >= LooseVersion("1.1.5")
21 PANDAS_GE_12 = str(pd.__version__) >= LooseVersion("1.2.0")
22
23
24 # -----------------------------------------------------------------------------
25 # Shapely / PyGEOS compat
26 # -----------------------------------------------------------------------------
27
28
29 SHAPELY_GE_17 = str(shapely.__version__) >= LooseVersion("1.7.0")
30 SHAPELY_GE_18 = str(shapely.__version__) >= LooseVersion("1.8")
31 SHAPELY_GE_20 = str(shapely.__version__) >= LooseVersion("2.0")
32
33 GEOS_GE_390 = shapely.geos.geos_version >= (3, 9, 0)
34
35
36 HAS_PYGEOS = None
37 USE_PYGEOS = None
38 PYGEOS_SHAPELY_COMPAT = None
39
40 PYGEOS_GE_09 = None
41 PYGEOS_GE_010 = None
42
43 INSTALL_PYGEOS_ERROR = "To use PyGEOS within GeoPandas, you need to install PyGEOS: \
44 'conda install pygeos' or 'pip install pygeos'"
45
46 try:
47 import pygeos # noqa
48
49 # only automatically use pygeos if version is high enough
50 if str(pygeos.__version__) >= LooseVersion("0.8"):
51 HAS_PYGEOS = True
52 PYGEOS_GE_09 = str(pygeos.__version__) >= LooseVersion("0.9")
53 PYGEOS_GE_010 = str(pygeos.__version__) >= LooseVersion("0.10")
54 else:
55 warnings.warn(
56 "The installed version of PyGEOS is too old ({0} installed, 0.8 required),"
57 " and thus GeoPandas will not use PyGEOS.".format(pygeos.__version__),
58 UserWarning,
59 )
60 HAS_PYGEOS = False
61 except ImportError:
62 HAS_PYGEOS = False
63
64
65 def set_use_pygeos(val=None):
66 """
67 Set the global configuration on whether to use PyGEOS or not.
68
69 The default is use PyGEOS if it is installed. This can be overridden
70 with an environment variable USE_PYGEOS (this is only checked at
71 first import, cannot be changed during interactive session).
72
73 Alternatively, pass a value here to force a True/False value.
74 """
75 global USE_PYGEOS
76 global PYGEOS_SHAPELY_COMPAT
77
78 if val is not None:
79 USE_PYGEOS = bool(val)
80 else:
81 if USE_PYGEOS is None:
82
83 USE_PYGEOS = HAS_PYGEOS
84
85 env_use_pygeos = os.getenv("USE_PYGEOS", None)
86 if env_use_pygeos is not None:
87 USE_PYGEOS = bool(int(env_use_pygeos))
88
89 # validate the pygeos version
90 if USE_PYGEOS:
91 try:
92 import pygeos # noqa
93
94 # validate the pygeos version
95 if not str(pygeos.__version__) >= LooseVersion("0.8"):
96 raise ImportError(
97 "PyGEOS >= 0.8 is required, version {0} is installed".format(
98 pygeos.__version__
99 )
100 )
101
102 # Check whether Shapely and PyGEOS use the same GEOS version.
103 # Based on PyGEOS from_shapely implementation.
104
105 from shapely.geos import geos_version_string as shapely_geos_version
106 from pygeos import geos_capi_version_string
107
108 # shapely has something like: "3.6.2-CAPI-1.10.2 4d2925d6"
109 # pygeos has something like: "3.6.2-CAPI-1.10.2"
110 if not shapely_geos_version.startswith(geos_capi_version_string):
111 warnings.warn(
112 "The Shapely GEOS version ({}) is incompatible with the GEOS "
113 "version PyGEOS was compiled with ({}). Conversions between both "
114 "will be slow.".format(
115 shapely_geos_version, geos_capi_version_string
116 )
117 )
118 PYGEOS_SHAPELY_COMPAT = False
119 else:
120 PYGEOS_SHAPELY_COMPAT = True
121
122 except ImportError:
123 raise ImportError(INSTALL_PYGEOS_ERROR)
124
125
126 set_use_pygeos()
127
128
129 # compat related to deprecation warnings introduced in Shapely 1.8
130 # -> creating a numpy array from a list-like of Multi-part geometries,
131 # although doing the correct thing (not expanding in its parts), still raises
132 # the warning about iteration being deprecated
133 # This adds a context manager to explicitly ignore this warning
134
135
136 try:
137 from shapely.errors import ShapelyDeprecationWarning as shapely_warning
138 except ImportError:
139 shapely_warning = None
140
141
142 if shapely_warning is not None and not SHAPELY_GE_20:
143
144 @contextlib.contextmanager
145 def ignore_shapely2_warnings():
146 with warnings.catch_warnings():
147 warnings.filterwarnings(
148 "ignore", "Iteration|The array interface|__len__", shapely_warning
149 )
150 yield
151
152
153 elif (str(np.__version__) >= LooseVersion("1.21")) and not SHAPELY_GE_20:
154
155 @contextlib.contextmanager
156 def ignore_shapely2_warnings():
157 with warnings.catch_warnings():
158 # warning from numpy for existing Shapely releases (this is fixed
159 # with Shapely 1.8)
160 warnings.filterwarnings(
161 "ignore", "An exception was ignored while fetching", DeprecationWarning
162 )
163 yield
164
165
166 else:
167
168 @contextlib.contextmanager
169 def ignore_shapely2_warnings():
170 yield
171
172
173 def import_optional_dependency(name: str, extra: str = ""):
174 """
175 Import an optional dependency.
176
177 Adapted from pandas.compat._optional::import_optional_dependency
178
179 Raises a formatted ImportError if the module is not present.
180
181 Parameters
182 ----------
183 name : str
184 The module name.
185 extra : str
186 Additional text to include in the ImportError message.
187 Returns
188 -------
189 module
190 """
191 msg = """Missing optional dependency '{name}'. {extra} "
192 "Use pip or conda to install {name}.""".format(
193 name=name, extra=extra
194 )
195
196 if not isinstance(name, str):
197 raise ValueError(
198 "Invalid module name: '{name}'; must be a string".format(name=name)
199 )
200
201 try:
202 module = importlib.import_module(name)
203
204 except ImportError:
205 raise ImportError(msg) from None
206
207 return module
208
209
210 # -----------------------------------------------------------------------------
211 # RTree compat
212 # -----------------------------------------------------------------------------
213
214 HAS_RTREE = None
215 RTREE_GE_094 = False
216 try:
217 import rtree # noqa
218
219 HAS_RTREE = True
220 except ImportError:
221 HAS_RTREE = False
222
223 # -----------------------------------------------------------------------------
224 # pyproj compat
225 # -----------------------------------------------------------------------------
226
227 PYPROJ_LT_3 = LooseVersion(pyproj.__version__) < LooseVersion("3")
228 PYPROJ_GE_31 = LooseVersion(pyproj.__version__) >= LooseVersion("3.1")
229
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/geopandas/_compat.py b/geopandas/_compat.py
--- a/geopandas/_compat.py
+++ b/geopandas/_compat.py
@@ -19,6 +19,7 @@
PANDAS_GE_11 = str(pd.__version__) >= LooseVersion("1.1.0")
PANDAS_GE_115 = str(pd.__version__) >= LooseVersion("1.1.5")
PANDAS_GE_12 = str(pd.__version__) >= LooseVersion("1.2.0")
+PANDAS_GE_14 = str(pd.__version__) >= LooseVersion("1.4.0")
# -----------------------------------------------------------------------------
|
{"golden_diff": "diff --git a/geopandas/_compat.py b/geopandas/_compat.py\n--- a/geopandas/_compat.py\n+++ b/geopandas/_compat.py\n@@ -19,6 +19,7 @@\n PANDAS_GE_11 = str(pd.__version__) >= LooseVersion(\"1.1.0\")\n PANDAS_GE_115 = str(pd.__version__) >= LooseVersion(\"1.1.5\")\n PANDAS_GE_12 = str(pd.__version__) >= LooseVersion(\"1.2.0\")\n+PANDAS_GE_14 = str(pd.__version__) >= LooseVersion(\"1.4.0\")\n \n \n # -----------------------------------------------------------------------------\n", "issue": "TST: test_value_counts breaking against pandas latest\nRecent CI builds on 38-dev.yaml are all failing due to geopandas/tests/test_pandas_methods.py::test_value_counts, which fails with the output:\r\n```\r\n______________________________ test_value_counts _______________________________\r\n3536\r\n[gw0] linux -- Python 3.8.12 /usr/share/miniconda3/envs/test/bin/python\r\n3537\r\n\r\n3538\r\n def test_value_counts():\r\n3539\r\n # each object is considered unique\r\n3540\r\n s = GeoSeries([Point(0, 0), Point(1, 1), Point(0, 0)])\r\n3541\r\n res = s.value_counts()\r\n3542\r\n with compat.ignore_shapely2_warnings():\r\n3543\r\n exp = pd.Series([2, 1], index=[Point(0, 0), Point(1, 1)])\r\n3544\r\n> assert_series_equal(res, exp)\r\n3545\r\n\r\n3546\r\ngeopandas/tests/test_pandas_methods.py:454: \r\n3547\r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n3548\r\n\r\n3549\r\nleft = Index([POINT (0 0), POINT (1 1)], dtype='geometry')\r\n3550\r\nright = Index([POINT (0 0), POINT (1 1)], dtype='object'), obj = 'Series.index'\r\n3551\r\n\r\n3552\r\n def _check_types(left, right, obj=\"Index\") -> None:\r\n3553\r\n if not exact:\r\n3554\r\n return\r\n3555\r\n \r\n3556\r\n assert_class_equal(left, right, exact=exact, obj=obj)\r\n3557\r\n> assert_attr_equal(\"inferred_type\", left, right, obj=obj)\r\n3558\r\nE AssertionError: Series.index are different\r\n3559\r\nE \r\n3560\r\nE Attribute \"inferred_type\" are different\r\n3561\r\nE [left]: unknown-array\r\n3562\r\nE [right]: mixed\r\n```\r\n\r\nIt looks like this pandas change (https://github.com/pandas-dev/pandas/pull/43930) to allow storing extension arrays in the index (that is, not just casting everything to `object`) is responsible.\n", "before_files": [{"content": "import contextlib\nfrom distutils.version import LooseVersion\nimport importlib\nimport os\nimport warnings\n\nimport numpy as np\nimport pandas as pd\nimport pyproj\nimport shapely\nimport shapely.geos\n\n\n# -----------------------------------------------------------------------------\n# pandas compat\n# -----------------------------------------------------------------------------\n\nPANDAS_GE_10 = str(pd.__version__) >= LooseVersion(\"1.0.0\")\nPANDAS_GE_11 = str(pd.__version__) >= LooseVersion(\"1.1.0\")\nPANDAS_GE_115 = str(pd.__version__) >= LooseVersion(\"1.1.5\")\nPANDAS_GE_12 = str(pd.__version__) >= LooseVersion(\"1.2.0\")\n\n\n# -----------------------------------------------------------------------------\n# Shapely / PyGEOS compat\n# -----------------------------------------------------------------------------\n\n\nSHAPELY_GE_17 = str(shapely.__version__) >= LooseVersion(\"1.7.0\")\nSHAPELY_GE_18 = str(shapely.__version__) >= LooseVersion(\"1.8\")\nSHAPELY_GE_20 = str(shapely.__version__) >= LooseVersion(\"2.0\")\n\nGEOS_GE_390 = shapely.geos.geos_version >= (3, 9, 0)\n\n\nHAS_PYGEOS = None\nUSE_PYGEOS = None\nPYGEOS_SHAPELY_COMPAT = None\n\nPYGEOS_GE_09 = None\nPYGEOS_GE_010 = None\n\nINSTALL_PYGEOS_ERROR = \"To use PyGEOS within GeoPandas, you need to install PyGEOS: \\\n'conda install pygeos' or 'pip install pygeos'\"\n\ntry:\n import pygeos # noqa\n\n # only automatically use pygeos if version is high enough\n if str(pygeos.__version__) >= LooseVersion(\"0.8\"):\n HAS_PYGEOS = True\n PYGEOS_GE_09 = str(pygeos.__version__) >= LooseVersion(\"0.9\")\n PYGEOS_GE_010 = str(pygeos.__version__) >= LooseVersion(\"0.10\")\n else:\n warnings.warn(\n \"The installed version of PyGEOS is too old ({0} installed, 0.8 required),\"\n \" and thus GeoPandas will not use PyGEOS.\".format(pygeos.__version__),\n UserWarning,\n )\n HAS_PYGEOS = False\nexcept ImportError:\n HAS_PYGEOS = False\n\n\ndef set_use_pygeos(val=None):\n \"\"\"\n Set the global configuration on whether to use PyGEOS or not.\n\n The default is use PyGEOS if it is installed. This can be overridden\n with an environment variable USE_PYGEOS (this is only checked at\n first import, cannot be changed during interactive session).\n\n Alternatively, pass a value here to force a True/False value.\n \"\"\"\n global USE_PYGEOS\n global PYGEOS_SHAPELY_COMPAT\n\n if val is not None:\n USE_PYGEOS = bool(val)\n else:\n if USE_PYGEOS is None:\n\n USE_PYGEOS = HAS_PYGEOS\n\n env_use_pygeos = os.getenv(\"USE_PYGEOS\", None)\n if env_use_pygeos is not None:\n USE_PYGEOS = bool(int(env_use_pygeos))\n\n # validate the pygeos version\n if USE_PYGEOS:\n try:\n import pygeos # noqa\n\n # validate the pygeos version\n if not str(pygeos.__version__) >= LooseVersion(\"0.8\"):\n raise ImportError(\n \"PyGEOS >= 0.8 is required, version {0} is installed\".format(\n pygeos.__version__\n )\n )\n\n # Check whether Shapely and PyGEOS use the same GEOS version.\n # Based on PyGEOS from_shapely implementation.\n\n from shapely.geos import geos_version_string as shapely_geos_version\n from pygeos import geos_capi_version_string\n\n # shapely has something like: \"3.6.2-CAPI-1.10.2 4d2925d6\"\n # pygeos has something like: \"3.6.2-CAPI-1.10.2\"\n if not shapely_geos_version.startswith(geos_capi_version_string):\n warnings.warn(\n \"The Shapely GEOS version ({}) is incompatible with the GEOS \"\n \"version PyGEOS was compiled with ({}). Conversions between both \"\n \"will be slow.\".format(\n shapely_geos_version, geos_capi_version_string\n )\n )\n PYGEOS_SHAPELY_COMPAT = False\n else:\n PYGEOS_SHAPELY_COMPAT = True\n\n except ImportError:\n raise ImportError(INSTALL_PYGEOS_ERROR)\n\n\nset_use_pygeos()\n\n\n# compat related to deprecation warnings introduced in Shapely 1.8\n# -> creating a numpy array from a list-like of Multi-part geometries,\n# although doing the correct thing (not expanding in its parts), still raises\n# the warning about iteration being deprecated\n# This adds a context manager to explicitly ignore this warning\n\n\ntry:\n from shapely.errors import ShapelyDeprecationWarning as shapely_warning\nexcept ImportError:\n shapely_warning = None\n\n\nif shapely_warning is not None and not SHAPELY_GE_20:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n with warnings.catch_warnings():\n warnings.filterwarnings(\n \"ignore\", \"Iteration|The array interface|__len__\", shapely_warning\n )\n yield\n\n\nelif (str(np.__version__) >= LooseVersion(\"1.21\")) and not SHAPELY_GE_20:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n with warnings.catch_warnings():\n # warning from numpy for existing Shapely releases (this is fixed\n # with Shapely 1.8)\n warnings.filterwarnings(\n \"ignore\", \"An exception was ignored while fetching\", DeprecationWarning\n )\n yield\n\n\nelse:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n yield\n\n\ndef import_optional_dependency(name: str, extra: str = \"\"):\n \"\"\"\n Import an optional dependency.\n\n Adapted from pandas.compat._optional::import_optional_dependency\n\n Raises a formatted ImportError if the module is not present.\n\n Parameters\n ----------\n name : str\n The module name.\n extra : str\n Additional text to include in the ImportError message.\n Returns\n -------\n module\n \"\"\"\n msg = \"\"\"Missing optional dependency '{name}'. {extra} \"\n \"Use pip or conda to install {name}.\"\"\".format(\n name=name, extra=extra\n )\n\n if not isinstance(name, str):\n raise ValueError(\n \"Invalid module name: '{name}'; must be a string\".format(name=name)\n )\n\n try:\n module = importlib.import_module(name)\n\n except ImportError:\n raise ImportError(msg) from None\n\n return module\n\n\n# -----------------------------------------------------------------------------\n# RTree compat\n# -----------------------------------------------------------------------------\n\nHAS_RTREE = None\nRTREE_GE_094 = False\ntry:\n import rtree # noqa\n\n HAS_RTREE = True\nexcept ImportError:\n HAS_RTREE = False\n\n# -----------------------------------------------------------------------------\n# pyproj compat\n# -----------------------------------------------------------------------------\n\nPYPROJ_LT_3 = LooseVersion(pyproj.__version__) < LooseVersion(\"3\")\nPYPROJ_GE_31 = LooseVersion(pyproj.__version__) >= LooseVersion(\"3.1\")\n", "path": "geopandas/_compat.py"}], "after_files": [{"content": "import contextlib\nfrom distutils.version import LooseVersion\nimport importlib\nimport os\nimport warnings\n\nimport numpy as np\nimport pandas as pd\nimport pyproj\nimport shapely\nimport shapely.geos\n\n\n# -----------------------------------------------------------------------------\n# pandas compat\n# -----------------------------------------------------------------------------\n\nPANDAS_GE_10 = str(pd.__version__) >= LooseVersion(\"1.0.0\")\nPANDAS_GE_11 = str(pd.__version__) >= LooseVersion(\"1.1.0\")\nPANDAS_GE_115 = str(pd.__version__) >= LooseVersion(\"1.1.5\")\nPANDAS_GE_12 = str(pd.__version__) >= LooseVersion(\"1.2.0\")\nPANDAS_GE_14 = str(pd.__version__) >= LooseVersion(\"1.4.0\")\n\n\n# -----------------------------------------------------------------------------\n# Shapely / PyGEOS compat\n# -----------------------------------------------------------------------------\n\n\nSHAPELY_GE_17 = str(shapely.__version__) >= LooseVersion(\"1.7.0\")\nSHAPELY_GE_18 = str(shapely.__version__) >= LooseVersion(\"1.8\")\nSHAPELY_GE_20 = str(shapely.__version__) >= LooseVersion(\"2.0\")\n\nGEOS_GE_390 = shapely.geos.geos_version >= (3, 9, 0)\n\n\nHAS_PYGEOS = None\nUSE_PYGEOS = None\nPYGEOS_SHAPELY_COMPAT = None\n\nPYGEOS_GE_09 = None\nPYGEOS_GE_010 = None\n\nINSTALL_PYGEOS_ERROR = \"To use PyGEOS within GeoPandas, you need to install PyGEOS: \\\n'conda install pygeos' or 'pip install pygeos'\"\n\ntry:\n import pygeos # noqa\n\n # only automatically use pygeos if version is high enough\n if str(pygeos.__version__) >= LooseVersion(\"0.8\"):\n HAS_PYGEOS = True\n PYGEOS_GE_09 = str(pygeos.__version__) >= LooseVersion(\"0.9\")\n PYGEOS_GE_010 = str(pygeos.__version__) >= LooseVersion(\"0.10\")\n else:\n warnings.warn(\n \"The installed version of PyGEOS is too old ({0} installed, 0.8 required),\"\n \" and thus GeoPandas will not use PyGEOS.\".format(pygeos.__version__),\n UserWarning,\n )\n HAS_PYGEOS = False\nexcept ImportError:\n HAS_PYGEOS = False\n\n\ndef set_use_pygeos(val=None):\n \"\"\"\n Set the global configuration on whether to use PyGEOS or not.\n\n The default is use PyGEOS if it is installed. This can be overridden\n with an environment variable USE_PYGEOS (this is only checked at\n first import, cannot be changed during interactive session).\n\n Alternatively, pass a value here to force a True/False value.\n \"\"\"\n global USE_PYGEOS\n global PYGEOS_SHAPELY_COMPAT\n\n if val is not None:\n USE_PYGEOS = bool(val)\n else:\n if USE_PYGEOS is None:\n\n USE_PYGEOS = HAS_PYGEOS\n\n env_use_pygeos = os.getenv(\"USE_PYGEOS\", None)\n if env_use_pygeos is not None:\n USE_PYGEOS = bool(int(env_use_pygeos))\n\n # validate the pygeos version\n if USE_PYGEOS:\n try:\n import pygeos # noqa\n\n # validate the pygeos version\n if not str(pygeos.__version__) >= LooseVersion(\"0.8\"):\n raise ImportError(\n \"PyGEOS >= 0.8 is required, version {0} is installed\".format(\n pygeos.__version__\n )\n )\n\n # Check whether Shapely and PyGEOS use the same GEOS version.\n # Based on PyGEOS from_shapely implementation.\n\n from shapely.geos import geos_version_string as shapely_geos_version\n from pygeos import geos_capi_version_string\n\n # shapely has something like: \"3.6.2-CAPI-1.10.2 4d2925d6\"\n # pygeos has something like: \"3.6.2-CAPI-1.10.2\"\n if not shapely_geos_version.startswith(geos_capi_version_string):\n warnings.warn(\n \"The Shapely GEOS version ({}) is incompatible with the GEOS \"\n \"version PyGEOS was compiled with ({}). Conversions between both \"\n \"will be slow.\".format(\n shapely_geos_version, geos_capi_version_string\n )\n )\n PYGEOS_SHAPELY_COMPAT = False\n else:\n PYGEOS_SHAPELY_COMPAT = True\n\n except ImportError:\n raise ImportError(INSTALL_PYGEOS_ERROR)\n\n\nset_use_pygeos()\n\n\n# compat related to deprecation warnings introduced in Shapely 1.8\n# -> creating a numpy array from a list-like of Multi-part geometries,\n# although doing the correct thing (not expanding in its parts), still raises\n# the warning about iteration being deprecated\n# This adds a context manager to explicitly ignore this warning\n\n\ntry:\n from shapely.errors import ShapelyDeprecationWarning as shapely_warning\nexcept ImportError:\n shapely_warning = None\n\n\nif shapely_warning is not None and not SHAPELY_GE_20:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n with warnings.catch_warnings():\n warnings.filterwarnings(\n \"ignore\", \"Iteration|The array interface|__len__\", shapely_warning\n )\n yield\n\n\nelif (str(np.__version__) >= LooseVersion(\"1.21\")) and not SHAPELY_GE_20:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n with warnings.catch_warnings():\n # warning from numpy for existing Shapely releases (this is fixed\n # with Shapely 1.8)\n warnings.filterwarnings(\n \"ignore\", \"An exception was ignored while fetching\", DeprecationWarning\n )\n yield\n\n\nelse:\n\n @contextlib.contextmanager\n def ignore_shapely2_warnings():\n yield\n\n\ndef import_optional_dependency(name: str, extra: str = \"\"):\n \"\"\"\n Import an optional dependency.\n\n Adapted from pandas.compat._optional::import_optional_dependency\n\n Raises a formatted ImportError if the module is not present.\n\n Parameters\n ----------\n name : str\n The module name.\n extra : str\n Additional text to include in the ImportError message.\n Returns\n -------\n module\n \"\"\"\n msg = \"\"\"Missing optional dependency '{name}'. {extra} \"\n \"Use pip or conda to install {name}.\"\"\".format(\n name=name, extra=extra\n )\n\n if not isinstance(name, str):\n raise ValueError(\n \"Invalid module name: '{name}'; must be a string\".format(name=name)\n )\n\n try:\n module = importlib.import_module(name)\n\n except ImportError:\n raise ImportError(msg) from None\n\n return module\n\n\n# -----------------------------------------------------------------------------\n# RTree compat\n# -----------------------------------------------------------------------------\n\nHAS_RTREE = None\nRTREE_GE_094 = False\ntry:\n import rtree # noqa\n\n HAS_RTREE = True\nexcept ImportError:\n HAS_RTREE = False\n\n# -----------------------------------------------------------------------------\n# pyproj compat\n# -----------------------------------------------------------------------------\n\nPYPROJ_LT_3 = LooseVersion(pyproj.__version__) < LooseVersion(\"3\")\nPYPROJ_GE_31 = LooseVersion(pyproj.__version__) >= LooseVersion(\"3.1\")\n", "path": "geopandas/_compat.py"}]}
| 3,097 | 145 |
gh_patches_debug_11105
|
rasdani/github-patches
|
git_diff
|
pwndbg__pwndbg-130
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Error in the `search` command
The `search` command appears to have some issues while searching signed integer data
For example while debugging a x86 program:
```
pwndbg> search -4 0xf7eebf83
Traceback (most recent call last):
File "/home/bruce30262/pwndbg/pwndbg/commands/__init__.py", line 57, in __call__
return self.function(*args, **kwargs)
File "/home/bruce30262/pwndbg/pwndbg/commands/__init__.py", line 162, in _ArgparsedCommand
return function(**vars(args))
File "/home/bruce30262/pwndbg/pwndbg/commands/__init__.py", line 115, in _OnlyWhenRunning
return function(*a, **kw)
File "/home/bruce30262/pwndbg/pwndbg/commands/search.py", line 112, in search
value = struct.pack(fmt, value)
struct.error: argument out of range
```
This is because `pwndbg.commands.fix_int(value)` returns `-135348349` instead of an unsigned integer, and the packing format in [search.py](https://github.com/pwndbg/pwndbg/blob/b1892b27741a478cd5361061b0b4dda9ef46d02e/pwndbg/commands/search.py#L106) only support the unsigned data type, causing the error.
Commands like `search -4 -- -100` will also cause the same error.
Kind of curious why it only support the unsigned data type though. It might need some additional check for the appropriate packing format.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pwndbg/commands/search.py`
Content:
```
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3 from __future__ import absolute_import
4 from __future__ import division
5 from __future__ import print_function
6 from __future__ import unicode_literals
7
8 import argparse
9 import codecs
10 import os
11 import struct
12
13 import gdb
14
15 import pwndbg.color.memory as M
16 import pwndbg.commands
17 import pwndbg.config
18 import pwndbg.enhance
19 import pwndbg.search
20 import pwndbg.vmmap
21
22 saved = set()
23
24 def print_search_hit(address):
25 """Prints out a single search hit.
26
27 Arguments:
28 address(int): Address to print
29 """
30 if not address:
31 return
32
33 vmmap = pwndbg.vmmap.find(address)
34 if vmmap:
35 region = os.path.basename(vmmap.objfile)
36 else:
37 region = '[mapped]'
38
39 region = region.ljust(15)
40
41 region = M.get(address, region)
42 addr = M.get(address)
43 display = pwndbg.enhance.enhance(address)
44 print(region,addr,display)
45
46 auto_save = pwndbg.config.Parameter('auto-save-search', False,
47 'automatically pass --save to "search" command')
48
49 parser = argparse.ArgumentParser(description='''
50 Search memory for byte sequences, strings, pointers, and integer values
51 ''')
52 parser.add_argument('-t', '--type', choices=['byte','short','dword','qword','pointer','string','bytes'],
53 help='Size of search target', default='bytes', type=str)
54 parser.add_argument('-1', '--byte', dest='type', action='store_const', const='byte',
55 help='Search for a 1-byte integer')
56 parser.add_argument('-2', '--word', dest='type', action='store_const', const='word',
57 help='Search for a 2-byte integer')
58 parser.add_argument('-4', '--dword', dest='type', action='store_const', const='dword',
59 help='Search for a 4-byte integer')
60 parser.add_argument('-8', '--qword', dest='type', action='store_const', const='qword',
61 help='Search for an 8-byte integer')
62 parser.add_argument('-p', '--pointer', dest='type', action='store_const', const='pointer',
63 help='Search for a pointer-width integer')
64 parser.add_argument('-x', '--hex', action='store_true',
65 help='Target is a hex-encoded (for bytes/strings)')
66 parser.add_argument('-s', '--string', action='store_true',
67 help='Target is a raw string')
68 parser.add_argument('-e', '--executable', action='store_true',
69 help='Search executable segments only')
70 parser.add_argument('-w', '--writable', action='store_true',
71 help='Search writable segments only')
72 parser.add_argument('value', type=str,
73 help='Value to search for')
74 parser.add_argument('mapping', type=str, nargs='?', default=None,
75 help='Mapping to search [e.g. libc]')
76 parser.add_argument('--save', action='store_true', default=None,
77 help='Save results for --resume. Default comes from config %r' % auto_save.name)
78 parser.add_argument('--no-save', action='store_false', default=None, dest='save',
79 help='Invert --save')
80 parser.add_argument('-n', '--next', action='store_true',
81 help='Search only locations returned by previous search with --save')
82
83 @pwndbg.commands.ArgparsedCommand(parser)
84 @pwndbg.commands.OnlyWhenRunning
85 def search(type, hex, string, executable, writable, value, mapping, save, next):
86 # Adjust pointer sizes to the local architecture
87 if type == 'pointer':
88 type = {
89 4: 'dword',
90 8: 'qword'
91 }[pwndbg.arch.ptrsize]
92
93 if save is None:
94 save = bool(pwndbg.config.auto_save_search)
95
96 if hex:
97 value = codecs.decode(value, 'hex')
98
99 # Convert to an integer if needed, and pack to bytes
100 if type not in ('string', 'bytes'):
101 value = pwndbg.commands.fix_int(value)
102 fmt = {
103 'little': '<',
104 'big': '>'
105 }[pwndbg.arch.endian] + {
106 'byte': 'B',
107 'short': 'H',
108 'dword': 'L',
109 'qword': 'Q'
110 }[type]
111
112 value = struct.pack(fmt, value)
113
114 # Null-terminate strings
115 elif type == 'string':
116 value += b'\x00'
117
118 # Prep the saved set if necessary
119 global saved
120 if save:
121 saved = set()
122
123 # Perform the search
124 for address in pwndbg.search.search(value,
125 mapping=mapping,
126 executable=executable,
127 writable=writable):
128
129 if next and address not in saved:
130 continue
131
132 if save:
133 saved.add(address)
134
135 print_search_hit(address)
136
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pwndbg/commands/search.py b/pwndbg/commands/search.py
--- a/pwndbg/commands/search.py
+++ b/pwndbg/commands/search.py
@@ -12,6 +12,7 @@
import gdb
+import pwndbg.arch
import pwndbg.color.memory as M
import pwndbg.commands
import pwndbg.config
@@ -99,6 +100,7 @@
# Convert to an integer if needed, and pack to bytes
if type not in ('string', 'bytes'):
value = pwndbg.commands.fix_int(value)
+ value &= pwndbg.arch.ptrmask
fmt = {
'little': '<',
'big': '>'
|
{"golden_diff": "diff --git a/pwndbg/commands/search.py b/pwndbg/commands/search.py\n--- a/pwndbg/commands/search.py\n+++ b/pwndbg/commands/search.py\n@@ -12,6 +12,7 @@\n \n import gdb\n \n+import pwndbg.arch\n import pwndbg.color.memory as M\n import pwndbg.commands\n import pwndbg.config\n@@ -99,6 +100,7 @@\n # Convert to an integer if needed, and pack to bytes\n if type not in ('string', 'bytes'):\n value = pwndbg.commands.fix_int(value)\n+ value &= pwndbg.arch.ptrmask\n fmt = {\n 'little': '<',\n 'big': '>'\n", "issue": "Error in the `search` command\nThe `search` command appears to have some issues while searching signed integer data \nFor example while debugging a x86 program:\n\n```\npwndbg> search -4 0xf7eebf83\nTraceback (most recent call last):\n File \"/home/bruce30262/pwndbg/pwndbg/commands/__init__.py\", line 57, in __call__\n return self.function(*args, **kwargs)\n File \"/home/bruce30262/pwndbg/pwndbg/commands/__init__.py\", line 162, in _ArgparsedCommand\n return function(**vars(args))\n File \"/home/bruce30262/pwndbg/pwndbg/commands/__init__.py\", line 115, in _OnlyWhenRunning\n return function(*a, **kw)\n File \"/home/bruce30262/pwndbg/pwndbg/commands/search.py\", line 112, in search\n value = struct.pack(fmt, value)\nstruct.error: argument out of range\n```\n\nThis is because `pwndbg.commands.fix_int(value)` returns `-135348349` instead of an unsigned integer, and the packing format in [search.py](https://github.com/pwndbg/pwndbg/blob/b1892b27741a478cd5361061b0b4dda9ef46d02e/pwndbg/commands/search.py#L106) only support the unsigned data type, causing the error. \nCommands like `search -4 -- -100` will also cause the same error. \nKind of curious why it only support the unsigned data type though. It might need some additional check for the appropriate packing format.\n\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport argparse\nimport codecs\nimport os\nimport struct\n\nimport gdb\n\nimport pwndbg.color.memory as M\nimport pwndbg.commands\nimport pwndbg.config\nimport pwndbg.enhance\nimport pwndbg.search\nimport pwndbg.vmmap\n\nsaved = set()\n\ndef print_search_hit(address):\n \"\"\"Prints out a single search hit.\n\n Arguments:\n address(int): Address to print\n \"\"\"\n if not address:\n return\n\n vmmap = pwndbg.vmmap.find(address)\n if vmmap:\n region = os.path.basename(vmmap.objfile)\n else:\n region = '[mapped]'\n\n region = region.ljust(15)\n\n region = M.get(address, region)\n addr = M.get(address)\n display = pwndbg.enhance.enhance(address)\n print(region,addr,display)\n\nauto_save = pwndbg.config.Parameter('auto-save-search', False,\n 'automatically pass --save to \"search\" command')\n\nparser = argparse.ArgumentParser(description='''\nSearch memory for byte sequences, strings, pointers, and integer values\n''')\nparser.add_argument('-t', '--type', choices=['byte','short','dword','qword','pointer','string','bytes'],\n help='Size of search target', default='bytes', type=str)\nparser.add_argument('-1', '--byte', dest='type', action='store_const', const='byte',\n help='Search for a 1-byte integer')\nparser.add_argument('-2', '--word', dest='type', action='store_const', const='word',\n help='Search for a 2-byte integer')\nparser.add_argument('-4', '--dword', dest='type', action='store_const', const='dword',\n help='Search for a 4-byte integer')\nparser.add_argument('-8', '--qword', dest='type', action='store_const', const='qword',\n help='Search for an 8-byte integer')\nparser.add_argument('-p', '--pointer', dest='type', action='store_const', const='pointer',\n help='Search for a pointer-width integer')\nparser.add_argument('-x', '--hex', action='store_true',\n help='Target is a hex-encoded (for bytes/strings)')\nparser.add_argument('-s', '--string', action='store_true',\n help='Target is a raw string')\nparser.add_argument('-e', '--executable', action='store_true',\n help='Search executable segments only')\nparser.add_argument('-w', '--writable', action='store_true',\n help='Search writable segments only')\nparser.add_argument('value', type=str,\n help='Value to search for')\nparser.add_argument('mapping', type=str, nargs='?', default=None,\n help='Mapping to search [e.g. libc]')\nparser.add_argument('--save', action='store_true', default=None,\n help='Save results for --resume. Default comes from config %r' % auto_save.name)\nparser.add_argument('--no-save', action='store_false', default=None, dest='save',\n help='Invert --save')\nparser.add_argument('-n', '--next', action='store_true',\n help='Search only locations returned by previous search with --save')\n\[email protected](parser)\[email protected]\ndef search(type, hex, string, executable, writable, value, mapping, save, next):\n # Adjust pointer sizes to the local architecture\n if type == 'pointer':\n type = {\n 4: 'dword',\n 8: 'qword'\n }[pwndbg.arch.ptrsize]\n\n if save is None:\n save = bool(pwndbg.config.auto_save_search)\n\n if hex:\n value = codecs.decode(value, 'hex')\n\n # Convert to an integer if needed, and pack to bytes\n if type not in ('string', 'bytes'):\n value = pwndbg.commands.fix_int(value)\n fmt = {\n 'little': '<',\n 'big': '>'\n }[pwndbg.arch.endian] + {\n 'byte': 'B',\n 'short': 'H',\n 'dword': 'L',\n 'qword': 'Q'\n }[type]\n\n value = struct.pack(fmt, value)\n\n # Null-terminate strings\n elif type == 'string':\n value += b'\\x00'\n\n # Prep the saved set if necessary\n global saved\n if save:\n saved = set()\n\n # Perform the search\n for address in pwndbg.search.search(value,\n mapping=mapping,\n executable=executable,\n writable=writable):\n\n if next and address not in saved:\n continue\n\n if save:\n saved.add(address)\n\n print_search_hit(address)\n", "path": "pwndbg/commands/search.py"}], "after_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport argparse\nimport codecs\nimport os\nimport struct\n\nimport gdb\n\nimport pwndbg.arch\nimport pwndbg.color.memory as M\nimport pwndbg.commands\nimport pwndbg.config\nimport pwndbg.enhance\nimport pwndbg.search\nimport pwndbg.vmmap\n\nsaved = set()\n\ndef print_search_hit(address):\n \"\"\"Prints out a single search hit.\n\n Arguments:\n address(int): Address to print\n \"\"\"\n if not address:\n return\n\n vmmap = pwndbg.vmmap.find(address)\n if vmmap:\n region = os.path.basename(vmmap.objfile)\n else:\n region = '[mapped]'\n\n region = region.ljust(15)\n\n region = M.get(address, region)\n addr = M.get(address)\n display = pwndbg.enhance.enhance(address)\n print(region,addr,display)\n\nauto_save = pwndbg.config.Parameter('auto-save-search', False,\n 'automatically pass --save to \"search\" command')\n\nparser = argparse.ArgumentParser(description='''\nSearch memory for byte sequences, strings, pointers, and integer values\n''')\nparser.add_argument('-t', '--type', choices=['byte','short','dword','qword','pointer','string','bytes'],\n help='Size of search target', default='bytes', type=str)\nparser.add_argument('-1', '--byte', dest='type', action='store_const', const='byte',\n help='Search for a 1-byte integer')\nparser.add_argument('-2', '--word', dest='type', action='store_const', const='word',\n help='Search for a 2-byte integer')\nparser.add_argument('-4', '--dword', dest='type', action='store_const', const='dword',\n help='Search for a 4-byte integer')\nparser.add_argument('-8', '--qword', dest='type', action='store_const', const='qword',\n help='Search for an 8-byte integer')\nparser.add_argument('-p', '--pointer', dest='type', action='store_const', const='pointer',\n help='Search for a pointer-width integer')\nparser.add_argument('-x', '--hex', action='store_true',\n help='Target is a hex-encoded (for bytes/strings)')\nparser.add_argument('-s', '--string', action='store_true',\n help='Target is a raw string')\nparser.add_argument('-e', '--executable', action='store_true',\n help='Search executable segments only')\nparser.add_argument('-w', '--writable', action='store_true',\n help='Search writable segments only')\nparser.add_argument('value', type=str,\n help='Value to search for')\nparser.add_argument('mapping', type=str, nargs='?', default=None,\n help='Mapping to search [e.g. libc]')\nparser.add_argument('--save', action='store_true', default=None,\n help='Save results for --resume. Default comes from config %r' % auto_save.name)\nparser.add_argument('--no-save', action='store_false', default=None, dest='save',\n help='Invert --save')\nparser.add_argument('-n', '--next', action='store_true',\n help='Search only locations returned by previous search with --save')\n\[email protected](parser)\[email protected]\ndef search(type, hex, string, executable, writable, value, mapping, save, next):\n # Adjust pointer sizes to the local architecture\n if type == 'pointer':\n type = {\n 4: 'dword',\n 8: 'qword'\n }[pwndbg.arch.ptrsize]\n\n if save is None:\n save = bool(pwndbg.config.auto_save_search)\n\n if hex:\n value = codecs.decode(value, 'hex')\n\n # Convert to an integer if needed, and pack to bytes\n if type not in ('string', 'bytes'):\n value = pwndbg.commands.fix_int(value)\n value &= pwndbg.arch.ptrmask\n fmt = {\n 'little': '<',\n 'big': '>'\n }[pwndbg.arch.endian] + {\n 'byte': 'B',\n 'short': 'H',\n 'dword': 'L',\n 'qword': 'Q'\n }[type]\n\n value = struct.pack(fmt, value)\n\n # Null-terminate strings\n elif type == 'string':\n value += b'\\x00'\n\n # Prep the saved set if necessary\n global saved\n if save:\n saved = set()\n\n # Perform the search\n for address in pwndbg.search.search(value,\n mapping=mapping,\n executable=executable,\n writable=writable):\n\n if next and address not in saved:\n continue\n\n if save:\n saved.add(address)\n\n print_search_hit(address)\n", "path": "pwndbg/commands/search.py"}]}
| 2,037 | 160 |
gh_patches_debug_24
|
rasdani/github-patches
|
git_diff
|
hpcaitech__ColossalAI-2007
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG]: ModuleNotFoundError: No module named 'colossalai.nn.optimizer.zero_optimizer'
### 🐛 Describe the bug
I install colossalAI with the command `pip install colossalai==0.1.11rc3+torch1.12cu11.3 -f https://release.colossalai.org`
But I get an error when follow https://github.com/hpcaitech/ColossalAI/tree/main/examples/tutorial#-run-opt-finetuning-and-inference, I just run `bash ./run_clm_synthetic.sh` and get an error as follows:
```shell
╭─────────────────────────────── Traceback (most recent call last) ────────────────────────────────╮
│ /home/he.yan/ColossalAI/examples/tutorial/opt/opt/run_clm.py:46 in <module> │
│ │
│ 43 from colossalai.core import global_context as gpc │
│ 44 from colossalai.logging import disable_existing_loggers, get_dist_logger │
│ 45 from colossalai.nn.optimizer import HybridAdam │
│ ❱ 46 from colossalai.nn.optimizer.zero_optimizer import ZeroOptimizer │
│ 47 from colossalai.nn.parallel import ZeroDDP │
│ 48 from colossalai.tensor import ProcessGroup │
│ 49 from colossalai.utils import get_current_device, get_dataloader │
╰──────────────────────────────────────────────────────────────────────────────────────────────────╯
ModuleNotFoundError: No module named 'colossalai.nn.optimizer.zero_optimizer'
```
### Environment
Python 3.8.15
torch1.12cu11.3
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `colossalai/__init__.py`
Content:
```
1 from .initialize import (
2 get_default_parser,
3 initialize,
4 launch,
5 launch_from_openmpi,
6 launch_from_slurm,
7 launch_from_torch,
8 )
9
10 __version__ = '0.1.11rc2'
11
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/colossalai/__init__.py b/colossalai/__init__.py
--- a/colossalai/__init__.py
+++ b/colossalai/__init__.py
@@ -7,4 +7,4 @@
launch_from_torch,
)
-__version__ = '0.1.11rc2'
+__version__ = '0.1.11rc4'
|
{"golden_diff": "diff --git a/colossalai/__init__.py b/colossalai/__init__.py\n--- a/colossalai/__init__.py\n+++ b/colossalai/__init__.py\n@@ -7,4 +7,4 @@\n launch_from_torch,\n )\n \n-__version__ = '0.1.11rc2'\n+__version__ = '0.1.11rc4'\n", "issue": "[BUG]: ModuleNotFoundError: No module named 'colossalai.nn.optimizer.zero_optimizer'\n### \ud83d\udc1b Describe the bug\r\n\r\nI install colossalAI with the command `pip install colossalai==0.1.11rc3+torch1.12cu11.3 -f https://release.colossalai.org`\r\nBut I get an error when follow https://github.com/hpcaitech/ColossalAI/tree/main/examples/tutorial#-run-opt-finetuning-and-inference, I just run `bash ./run_clm_synthetic.sh` and get an error as follows:\r\n\r\n```shell\r\n\u256d\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500 Traceback (most recent call last) \u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u256e\r\n\u2502 /home/he.yan/ColossalAI/examples/tutorial/opt/opt/run_clm.py:46 in <module> \u2502\r\n\u2502 \u2502\r\n\u2502 43 from colossalai.core import global_context as gpc \u2502\r\n\u2502 44 from colossalai.logging import disable_existing_loggers, get_dist_logger \u2502\r\n\u2502 45 from colossalai.nn.optimizer import HybridAdam \u2502\r\n\u2502 \u2771 46 from colossalai.nn.optimizer.zero_optimizer import ZeroOptimizer \u2502\r\n\u2502 47 from colossalai.nn.parallel import ZeroDDP \u2502\r\n\u2502 48 from colossalai.tensor import ProcessGroup \u2502\r\n\u2502 49 from colossalai.utils import get_current_device, get_dataloader \u2502\r\n\u2570\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u256f\r\nModuleNotFoundError: No module named 'colossalai.nn.optimizer.zero_optimizer'\r\n```\r\n\r\n### Environment\r\n\r\nPython 3.8.15\r\ntorch1.12cu11.3\n", "before_files": [{"content": "from .initialize import (\n get_default_parser,\n initialize,\n launch,\n launch_from_openmpi,\n launch_from_slurm,\n launch_from_torch,\n)\n\n__version__ = '0.1.11rc2'\n", "path": "colossalai/__init__.py"}], "after_files": [{"content": "from .initialize import (\n get_default_parser,\n initialize,\n launch,\n launch_from_openmpi,\n launch_from_slurm,\n launch_from_torch,\n)\n\n__version__ = '0.1.11rc4'\n", "path": "colossalai/__init__.py"}]}
| 690 | 92 |
gh_patches_debug_2151
|
rasdani/github-patches
|
git_diff
|
WeblateOrg__weblate-4665
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
migrations fail for database name containing "-"
**Describe the bug**
Applying memory.0007_use_trigram...Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/django/db/backends/utils.py", line 84, in _execute
return self.cursor.execute(sql, params)
psycopg2.errors.SyntaxError: syntax error at or near "-"
LINE 1: ALTER DATABASE weblate-staging SET pg_trgm.similarity_thresh...
^
**To Reproduce**
Set the database name to "weblate-staging"
I worked around this by changing of
ALTER DATABASE {} SET
to
ALTER DATABASE \"{}\" SET
in 0007_use_trigram.py and 0008_adjust_similarity.py.
weblate-4.1.1
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `weblate/memory/migrations/0008_adjust_similarity.py`
Content:
```
1 # Generated by Django 3.0.5 on 2020-05-12 11:44
2
3 from django.db import migrations
4
5
6 def update_index(apps, schema_editor):
7 if schema_editor.connection.vendor != "postgresql":
8 return
9 # This ensures that extensions are loaded into the session. Without that
10 # the next ALTER database fails unless we're running as superuser (which
11 # is allowed to set non existing parameters, so missing extension doesn't
12 # matter)
13 # See https://www.postgresql.org/message-id/6376.1533675236%40sss.pgh.pa.us
14 schema_editor.execute("SELECT show_limit()")
15
16 schema_editor.execute(
17 "ALTER ROLE {} SET pg_trgm.similarity_threshold = 0.5".format(
18 schema_editor.connection.settings_dict["USER"]
19 )
20 )
21
22
23 class Migration(migrations.Migration):
24
25 dependencies = [
26 ("memory", "0007_use_trigram"),
27 ]
28
29 operations = [
30 migrations.RunPython(
31 update_index, migrations.RunPython.noop, elidable=False, atomic=False
32 )
33 ]
34
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/weblate/memory/migrations/0008_adjust_similarity.py b/weblate/memory/migrations/0008_adjust_similarity.py
--- a/weblate/memory/migrations/0008_adjust_similarity.py
+++ b/weblate/memory/migrations/0008_adjust_similarity.py
@@ -15,7 +15,7 @@
schema_editor.execute(
"ALTER ROLE {} SET pg_trgm.similarity_threshold = 0.5".format(
- schema_editor.connection.settings_dict["USER"]
+ schema_editor.quote_name(schema_editor.connection.settings_dict["USER"])
)
)
|
{"golden_diff": "diff --git a/weblate/memory/migrations/0008_adjust_similarity.py b/weblate/memory/migrations/0008_adjust_similarity.py\n--- a/weblate/memory/migrations/0008_adjust_similarity.py\n+++ b/weblate/memory/migrations/0008_adjust_similarity.py\n@@ -15,7 +15,7 @@\n \n schema_editor.execute(\n \"ALTER ROLE {} SET pg_trgm.similarity_threshold = 0.5\".format(\n- schema_editor.connection.settings_dict[\"USER\"]\n+ schema_editor.quote_name(schema_editor.connection.settings_dict[\"USER\"])\n )\n )\n", "issue": "migrations fail for database name containing \"-\"\n**Describe the bug**\r\n Applying memory.0007_use_trigram...Traceback (most recent call last):\r\n File \"/usr/lib/python3.6/site-packages/django/db/backends/utils.py\", line 84, in _execute\r\n return self.cursor.execute(sql, params)\r\npsycopg2.errors.SyntaxError: syntax error at or near \"-\"\r\nLINE 1: ALTER DATABASE weblate-staging SET pg_trgm.similarity_thresh...\r\n ^\r\n**To Reproduce**\r\nSet the database name to \"weblate-staging\"\r\n\r\nI worked around this by changing of\r\nALTER DATABASE {} SET\r\nto\r\nALTER DATABASE \\\"{}\\\" SET\r\nin 0007_use_trigram.py and 0008_adjust_similarity.py.\r\n\r\nweblate-4.1.1\n", "before_files": [{"content": "# Generated by Django 3.0.5 on 2020-05-12 11:44\n\nfrom django.db import migrations\n\n\ndef update_index(apps, schema_editor):\n if schema_editor.connection.vendor != \"postgresql\":\n return\n # This ensures that extensions are loaded into the session. Without that\n # the next ALTER database fails unless we're running as superuser (which\n # is allowed to set non existing parameters, so missing extension doesn't\n # matter)\n # See https://www.postgresql.org/message-id/6376.1533675236%40sss.pgh.pa.us\n schema_editor.execute(\"SELECT show_limit()\")\n\n schema_editor.execute(\n \"ALTER ROLE {} SET pg_trgm.similarity_threshold = 0.5\".format(\n schema_editor.connection.settings_dict[\"USER\"]\n )\n )\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\"memory\", \"0007_use_trigram\"),\n ]\n\n operations = [\n migrations.RunPython(\n update_index, migrations.RunPython.noop, elidable=False, atomic=False\n )\n ]\n", "path": "weblate/memory/migrations/0008_adjust_similarity.py"}], "after_files": [{"content": "# Generated by Django 3.0.5 on 2020-05-12 11:44\n\nfrom django.db import migrations\n\n\ndef update_index(apps, schema_editor):\n if schema_editor.connection.vendor != \"postgresql\":\n return\n # This ensures that extensions are loaded into the session. Without that\n # the next ALTER database fails unless we're running as superuser (which\n # is allowed to set non existing parameters, so missing extension doesn't\n # matter)\n # See https://www.postgresql.org/message-id/6376.1533675236%40sss.pgh.pa.us\n schema_editor.execute(\"SELECT show_limit()\")\n\n schema_editor.execute(\n \"ALTER ROLE {} SET pg_trgm.similarity_threshold = 0.5\".format(\n schema_editor.quote_name(schema_editor.connection.settings_dict[\"USER\"])\n )\n )\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\"memory\", \"0007_use_trigram\"),\n ]\n\n operations = [\n migrations.RunPython(\n update_index, migrations.RunPython.noop, elidable=False, atomic=False\n )\n ]\n", "path": "weblate/memory/migrations/0008_adjust_similarity.py"}]}
| 758 | 134 |
gh_patches_debug_2298
|
rasdani/github-patches
|
git_diff
|
iterative__dvc-1436
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
repro: --all-pipelines doesn't do anything.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `dvc/command/repro.py`
Content:
```
1 import os
2
3 from dvc.command.base import CmdBase
4 from dvc.command.status import CmdDataStatus
5 from dvc.exceptions import DvcException
6
7
8 class CmdRepro(CmdBase):
9 def run(self):
10 recursive = not self.args.single_item
11 saved_dir = os.path.realpath(os.curdir)
12 if self.args.cwd:
13 os.chdir(self.args.cwd)
14
15 ret = 0
16 for target in self.args.targets:
17 try:
18 stages = self.project.reproduce(
19 target,
20 recursive=recursive,
21 force=self.args.force,
22 dry=self.args.dry,
23 interactive=self.args.interactive,
24 pipeline=self.args.pipeline,
25 all_pipelines=self.args.all_pipelines,
26 ignore_build_cache=self.args.ignore_build_cache)
27
28 if len(stages) == 0:
29 self.project.logger.info(CmdDataStatus.UP_TO_DATE_MSG)
30
31 if self.args.metrics:
32 self.project.metrics_show()
33 except DvcException as ex:
34 msg = 'Failed to reproduce \'{}\''.format(target)
35 self.project.logger.error(msg, ex)
36 ret = 1
37 break
38
39 os.chdir(saved_dir)
40 return ret
41
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/dvc/command/repro.py b/dvc/command/repro.py
--- a/dvc/command/repro.py
+++ b/dvc/command/repro.py
@@ -12,6 +12,10 @@
if self.args.cwd:
os.chdir(self.args.cwd)
+ # Dirty hack so the for loop below can at least enter once
+ if self.args.all_pipelines:
+ self.args.targets = [None]
+
ret = 0
for target in self.args.targets:
try:
|
{"golden_diff": "diff --git a/dvc/command/repro.py b/dvc/command/repro.py\n--- a/dvc/command/repro.py\n+++ b/dvc/command/repro.py\n@@ -12,6 +12,10 @@\n if self.args.cwd:\n os.chdir(self.args.cwd)\n \n+ # Dirty hack so the for loop below can at least enter once\n+ if self.args.all_pipelines:\n+ self.args.targets = [None]\n+\n ret = 0\n for target in self.args.targets:\n try:\n", "issue": "repro: --all-pipelines doesn't do anything.\n\n", "before_files": [{"content": "import os\n\nfrom dvc.command.base import CmdBase\nfrom dvc.command.status import CmdDataStatus\nfrom dvc.exceptions import DvcException\n\n\nclass CmdRepro(CmdBase):\n def run(self):\n recursive = not self.args.single_item\n saved_dir = os.path.realpath(os.curdir)\n if self.args.cwd:\n os.chdir(self.args.cwd)\n\n ret = 0\n for target in self.args.targets:\n try:\n stages = self.project.reproduce(\n target,\n recursive=recursive,\n force=self.args.force,\n dry=self.args.dry,\n interactive=self.args.interactive,\n pipeline=self.args.pipeline,\n all_pipelines=self.args.all_pipelines,\n ignore_build_cache=self.args.ignore_build_cache)\n\n if len(stages) == 0:\n self.project.logger.info(CmdDataStatus.UP_TO_DATE_MSG)\n\n if self.args.metrics:\n self.project.metrics_show()\n except DvcException as ex:\n msg = 'Failed to reproduce \\'{}\\''.format(target)\n self.project.logger.error(msg, ex)\n ret = 1\n break\n\n os.chdir(saved_dir)\n return ret\n", "path": "dvc/command/repro.py"}], "after_files": [{"content": "import os\n\nfrom dvc.command.base import CmdBase\nfrom dvc.command.status import CmdDataStatus\nfrom dvc.exceptions import DvcException\n\n\nclass CmdRepro(CmdBase):\n def run(self):\n recursive = not self.args.single_item\n saved_dir = os.path.realpath(os.curdir)\n if self.args.cwd:\n os.chdir(self.args.cwd)\n\n # Dirty hack so the for loop below can at least enter once\n if self.args.all_pipelines:\n self.args.targets = [None]\n\n ret = 0\n for target in self.args.targets:\n try:\n stages = self.project.reproduce(\n target,\n recursive=recursive,\n force=self.args.force,\n dry=self.args.dry,\n interactive=self.args.interactive,\n pipeline=self.args.pipeline,\n all_pipelines=self.args.all_pipelines,\n ignore_build_cache=self.args.ignore_build_cache)\n\n if len(stages) == 0:\n self.project.logger.info(CmdDataStatus.UP_TO_DATE_MSG)\n\n if self.args.metrics:\n self.project.metrics_show()\n except DvcException as ex:\n msg = 'Failed to reproduce \\'{}\\''.format(target)\n self.project.logger.error(msg, ex)\n ret = 1\n break\n\n os.chdir(saved_dir)\n return ret\n", "path": "dvc/command/repro.py"}]}
| 594 | 115 |
gh_patches_debug_6720
|
rasdani/github-patches
|
git_diff
|
lutris__lutris-2315
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
wine sandboxing error in 0.5.3
Hello,
it seems you resolved issue #2114 but now, when I launch any of my wine games, I obtain this error message:
`[Errno 21] est un dossier: '/home/legluondunet/Games/test/drive_c/users/legluondunet/Downloads'`
* "est un dossier" = "is a folder"
then an other error message:
`en error prevented the game to launch`
I can reproduce this issue with all my installed wine game, **even with new installed game**.
Workaround: disable sandboxing...
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `lutris/util/wine/prefix.py`
Content:
```
1 """Wine prefix management"""
2 import os
3 from lutris.util.wine.registry import WineRegistry
4 from lutris.util.log import logger
5 from lutris.util import joypad, system, i18n
6 from lutris.util.display import DISPLAY_MANAGER
7
8 DESKTOP_KEYS = ["Desktop", "My Music", "My Pictures", "My Videos", "Personal"]
9
10
11 class WinePrefixManager:
12 """Class to allow modification of Wine prefixes without the use of Wine"""
13
14 hkcu_prefix = "HKEY_CURRENT_USER"
15
16 def __init__(self, path):
17 if not path:
18 logger.warning("No path specified for Wine prefix")
19 self.path = path
20
21 def setup_defaults(self):
22 """Sets the defaults for newly created prefixes"""
23 self.override_dll("winemenubuilder.exe", "")
24 self.override_dll("steamwebhelper.exe", "")
25 try:
26 self.desktop_integration()
27 except OSError as ex:
28 logger.error("Failed to setup desktop integration, the prefix may not be valid.")
29 logger.exception(ex)
30
31 def get_registry_path(self, key):
32 """Matches registry keys to a registry file
33
34 Currently, only HKEY_CURRENT_USER keys are supported.
35 """
36 if key.startswith(self.hkcu_prefix):
37 return os.path.join(self.path, "user.reg")
38 raise ValueError("Unsupported key '{}'".format(key))
39
40 def get_key_path(self, key):
41 if key.startswith(self.hkcu_prefix):
42 return key[len(self.hkcu_prefix) + 1:]
43 raise ValueError(
44 "The key {} is currently not supported by WinePrefixManager".format(key)
45 )
46
47 def get_registry_key(self,key,subkey):
48 registry = WineRegistry(self.get_registry_path(key))
49 return registry.query(self.get_key_path(key),subkey)
50
51 def set_registry_key(self, key, subkey, value):
52 registry = WineRegistry(self.get_registry_path(key))
53 registry.set_value(self.get_key_path(key), subkey, value)
54 registry.save()
55
56 def clear_registry_key(self, key):
57 registry = WineRegistry(self.get_registry_path(key))
58 registry.clear_key(self.get_key_path(key))
59 registry.save()
60
61 def clear_registry_subkeys(self, key, subkeys):
62 registry = WineRegistry(self.get_registry_path(key))
63 registry.clear_subkeys(self.get_key_path(key), subkeys)
64 registry.save()
65
66 def override_dll(self, dll, mode):
67 key = self.hkcu_prefix + "/Software/Wine/DllOverrides"
68 if mode.startswith("dis"):
69 mode = ""
70 if mode not in ("builtin", "native", "builtin,native", "native,builtin", ""):
71 logger.error("DLL override '%s' mode is not valid", mode)
72 return
73 self.set_registry_key(key, dll, mode)
74
75 def desktop_integration(self, desktop_dir=None):
76 """Overwrite desktop integration"""
77 DESKTOP_FOLDERS = []
78
79 user = os.getenv("USER")
80 user_dir = os.path.join(self.path, "drive_c/users/", user)
81
82 for key in DESKTOP_KEYS:
83 folder = self.get_registry_key(self.hkcu_prefix+"/Software/Microsoft/Windows/CurrentVersion/Explorer/Shell Folders",key)
84 DESKTOP_FOLDERS.append(folder[folder.rfind("\\")+1:])
85
86 if not desktop_dir:
87 desktop_dir = user_dir
88 else:
89 desktop_dir = os.path.expanduser(desktop_dir)
90
91 if system.path_exists(user_dir):
92 # Replace desktop integration symlinks
93 for item in DESKTOP_FOLDERS:
94 path = os.path.join(user_dir, item)
95 old_path = path + ".winecfg"
96
97 if os.path.islink(path):
98 os.unlink(path)
99 elif os.path.isdir(path):
100 try:
101 os.rmdir(path)
102 # We can't delete nonempty dir, so we rename as wine do.
103 except OSError:
104 os.rename(path, old_path)
105
106 if desktop_dir != user_dir:
107 src_path = os.path.join(desktop_dir, item)
108 os.makedirs(src_path, exist_ok=True)
109 os.symlink(src_path, path)
110 else:
111 # We use first the renamed dir, otherwise we make it.
112 if os.path.isdir(old_path):
113 os.rename(old_path, path)
114 else:
115 os.makedirs(path, exist_ok=True)
116
117 # Security: Remove other symlinks.
118 for item in os.listdir(user_dir):
119 if item not in DESKTOP_FOLDERS and os.path.islink(item):
120 path = os.path.join(user_dir, item)
121 os.unlink(path)
122 os.makedirs(path)
123
124 def set_crash_dialogs(self, enabled):
125 """Enable or diable Wine crash dialogs"""
126 self.set_registry_key(
127 self.hkcu_prefix + "/Software/Wine/WineDbg",
128 "ShowCrashDialog",
129 1 if enabled else 0,
130 )
131
132 def set_virtual_desktop(self, enabled):
133 """Enable or disable wine virtual desktop.
134 The Lutris virtual desktop is refered to as 'WineDesktop', in Wine the
135 virtual desktop name is 'default'.
136 """
137 path = self.hkcu_prefix + "/Software/Wine/Explorer"
138 if enabled:
139 self.set_registry_key(path, "Desktop", "WineDesktop")
140 default_resolution = "x".join(DISPLAY_MANAGER.get_current_resolution())
141 logger.debug("Enabling wine virtual desktop with default resolution of %s",
142 default_resolution)
143 self.set_registry_key(
144 self.hkcu_prefix + "/Software/Wine/Explorer/Desktops",
145 "WineDesktop",
146 default_resolution
147 )
148 else:
149 self.clear_registry_key(path)
150
151 def set_desktop_size(self, desktop_size):
152 """Sets the desktop size if one is given but do not reset the key if
153 one isn't.
154 """
155 path = self.hkcu_prefix + "/Software/Wine/Explorer/Desktops"
156 if desktop_size:
157 self.set_registry_key(path, "WineDesktop", desktop_size)
158
159 def use_xvid_mode(self, enabled):
160 """Set this to "Y" to allow wine switch the resolution using XVidMode extension."""
161 self.set_registry_key(
162 self.hkcu_prefix + "/Software/Wine/X11 Driver",
163 "UseXVidMode",
164 "Y" if enabled else "N",
165 )
166
167 def configure_joypads(self):
168 joypads = joypad.get_joypads()
169 key = self.hkcu_prefix + "/Software/Wine/DirectInput/Joysticks"
170 self.clear_registry_key(key)
171 for device, joypad_name in joypads:
172 if "event" in device:
173 disabled_joypad = "{} (js)".format(joypad_name)
174 else:
175 disabled_joypad = "{} (event)".format(joypad_name)
176 self.set_registry_key(key, disabled_joypad, "disabled")
177
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/lutris/util/wine/prefix.py b/lutris/util/wine/prefix.py
--- a/lutris/util/wine/prefix.py
+++ b/lutris/util/wine/prefix.py
@@ -116,8 +116,8 @@
# Security: Remove other symlinks.
for item in os.listdir(user_dir):
- if item not in DESKTOP_FOLDERS and os.path.islink(item):
- path = os.path.join(user_dir, item)
+ path = os.path.join(user_dir, item)
+ if item not in DESKTOP_FOLDERS and os.path.islink(path):
os.unlink(path)
os.makedirs(path)
|
{"golden_diff": "diff --git a/lutris/util/wine/prefix.py b/lutris/util/wine/prefix.py\n--- a/lutris/util/wine/prefix.py\n+++ b/lutris/util/wine/prefix.py\n@@ -116,8 +116,8 @@\n \n # Security: Remove other symlinks.\n for item in os.listdir(user_dir):\n- if item not in DESKTOP_FOLDERS and os.path.islink(item):\n- path = os.path.join(user_dir, item)\n+ path = os.path.join(user_dir, item)\n+ if item not in DESKTOP_FOLDERS and os.path.islink(path):\n os.unlink(path)\n os.makedirs(path)\n", "issue": "wine sandboxing error in 0.5.3\nHello,\r\nit seems you resolved issue #2114 but now, when I launch any of my wine games, I obtain this error message:\r\n`[Errno 21] est un dossier: '/home/legluondunet/Games/test/drive_c/users/legluondunet/Downloads'`\r\n\r\n* \"est un dossier\" = \"is a folder\"\r\n\r\nthen an other error message:\r\n`en error prevented the game to launch`\r\n\r\nI can reproduce this issue with all my installed wine game, **even with new installed game**.\r\nWorkaround: disable sandboxing...\n", "before_files": [{"content": "\"\"\"Wine prefix management\"\"\"\nimport os\nfrom lutris.util.wine.registry import WineRegistry\nfrom lutris.util.log import logger\nfrom lutris.util import joypad, system, i18n\nfrom lutris.util.display import DISPLAY_MANAGER\n\nDESKTOP_KEYS = [\"Desktop\", \"My Music\", \"My Pictures\", \"My Videos\", \"Personal\"]\n\n\nclass WinePrefixManager:\n \"\"\"Class to allow modification of Wine prefixes without the use of Wine\"\"\"\n\n hkcu_prefix = \"HKEY_CURRENT_USER\"\n\n def __init__(self, path):\n if not path:\n logger.warning(\"No path specified for Wine prefix\")\n self.path = path\n\n def setup_defaults(self):\n \"\"\"Sets the defaults for newly created prefixes\"\"\"\n self.override_dll(\"winemenubuilder.exe\", \"\")\n self.override_dll(\"steamwebhelper.exe\", \"\")\n try:\n self.desktop_integration()\n except OSError as ex:\n logger.error(\"Failed to setup desktop integration, the prefix may not be valid.\")\n logger.exception(ex)\n\n def get_registry_path(self, key):\n \"\"\"Matches registry keys to a registry file\n\n Currently, only HKEY_CURRENT_USER keys are supported.\n \"\"\"\n if key.startswith(self.hkcu_prefix):\n return os.path.join(self.path, \"user.reg\")\n raise ValueError(\"Unsupported key '{}'\".format(key))\n\n def get_key_path(self, key):\n if key.startswith(self.hkcu_prefix):\n return key[len(self.hkcu_prefix) + 1:]\n raise ValueError(\n \"The key {} is currently not supported by WinePrefixManager\".format(key)\n )\n\n def get_registry_key(self,key,subkey):\n registry = WineRegistry(self.get_registry_path(key))\n return registry.query(self.get_key_path(key),subkey)\n\n def set_registry_key(self, key, subkey, value):\n registry = WineRegistry(self.get_registry_path(key))\n registry.set_value(self.get_key_path(key), subkey, value)\n registry.save()\n\n def clear_registry_key(self, key):\n registry = WineRegistry(self.get_registry_path(key))\n registry.clear_key(self.get_key_path(key))\n registry.save()\n\n def clear_registry_subkeys(self, key, subkeys):\n registry = WineRegistry(self.get_registry_path(key))\n registry.clear_subkeys(self.get_key_path(key), subkeys)\n registry.save()\n\n def override_dll(self, dll, mode):\n key = self.hkcu_prefix + \"/Software/Wine/DllOverrides\"\n if mode.startswith(\"dis\"):\n mode = \"\"\n if mode not in (\"builtin\", \"native\", \"builtin,native\", \"native,builtin\", \"\"):\n logger.error(\"DLL override '%s' mode is not valid\", mode)\n return\n self.set_registry_key(key, dll, mode)\n\n def desktop_integration(self, desktop_dir=None):\n \"\"\"Overwrite desktop integration\"\"\"\n DESKTOP_FOLDERS = []\n\n user = os.getenv(\"USER\")\n user_dir = os.path.join(self.path, \"drive_c/users/\", user)\n\n for key in DESKTOP_KEYS:\n folder = self.get_registry_key(self.hkcu_prefix+\"/Software/Microsoft/Windows/CurrentVersion/Explorer/Shell Folders\",key)\n DESKTOP_FOLDERS.append(folder[folder.rfind(\"\\\\\")+1:]) \n\n if not desktop_dir:\n desktop_dir = user_dir\n else:\n desktop_dir = os.path.expanduser(desktop_dir)\n\n if system.path_exists(user_dir):\n # Replace desktop integration symlinks\n for item in DESKTOP_FOLDERS:\n path = os.path.join(user_dir, item)\n old_path = path + \".winecfg\"\n\n if os.path.islink(path):\n os.unlink(path)\n elif os.path.isdir(path):\n try:\n os.rmdir(path)\n # We can't delete nonempty dir, so we rename as wine do.\n except OSError:\n os.rename(path, old_path)\n\n if desktop_dir != user_dir:\n src_path = os.path.join(desktop_dir, item)\n os.makedirs(src_path, exist_ok=True)\n os.symlink(src_path, path)\n else:\n # We use first the renamed dir, otherwise we make it.\n if os.path.isdir(old_path):\n os.rename(old_path, path)\n else:\n os.makedirs(path, exist_ok=True)\n\n # Security: Remove other symlinks.\n for item in os.listdir(user_dir):\n if item not in DESKTOP_FOLDERS and os.path.islink(item):\n path = os.path.join(user_dir, item)\n os.unlink(path)\n os.makedirs(path)\n\n def set_crash_dialogs(self, enabled):\n \"\"\"Enable or diable Wine crash dialogs\"\"\"\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/WineDbg\",\n \"ShowCrashDialog\",\n 1 if enabled else 0,\n )\n\n def set_virtual_desktop(self, enabled):\n \"\"\"Enable or disable wine virtual desktop.\n The Lutris virtual desktop is refered to as 'WineDesktop', in Wine the\n virtual desktop name is 'default'.\n \"\"\"\n path = self.hkcu_prefix + \"/Software/Wine/Explorer\"\n if enabled:\n self.set_registry_key(path, \"Desktop\", \"WineDesktop\")\n default_resolution = \"x\".join(DISPLAY_MANAGER.get_current_resolution())\n logger.debug(\"Enabling wine virtual desktop with default resolution of %s\",\n default_resolution)\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/Explorer/Desktops\",\n \"WineDesktop\",\n default_resolution\n )\n else:\n self.clear_registry_key(path)\n\n def set_desktop_size(self, desktop_size):\n \"\"\"Sets the desktop size if one is given but do not reset the key if\n one isn't.\n \"\"\"\n path = self.hkcu_prefix + \"/Software/Wine/Explorer/Desktops\"\n if desktop_size:\n self.set_registry_key(path, \"WineDesktop\", desktop_size)\n\n def use_xvid_mode(self, enabled):\n \"\"\"Set this to \"Y\" to allow wine switch the resolution using XVidMode extension.\"\"\"\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/X11 Driver\",\n \"UseXVidMode\",\n \"Y\" if enabled else \"N\",\n )\n\n def configure_joypads(self):\n joypads = joypad.get_joypads()\n key = self.hkcu_prefix + \"/Software/Wine/DirectInput/Joysticks\"\n self.clear_registry_key(key)\n for device, joypad_name in joypads:\n if \"event\" in device:\n disabled_joypad = \"{} (js)\".format(joypad_name)\n else:\n disabled_joypad = \"{} (event)\".format(joypad_name)\n self.set_registry_key(key, disabled_joypad, \"disabled\")\n", "path": "lutris/util/wine/prefix.py"}], "after_files": [{"content": "\"\"\"Wine prefix management\"\"\"\nimport os\nfrom lutris.util.wine.registry import WineRegistry\nfrom lutris.util.log import logger\nfrom lutris.util import joypad, system, i18n\nfrom lutris.util.display import DISPLAY_MANAGER\n\nDESKTOP_KEYS = [\"Desktop\", \"My Music\", \"My Pictures\", \"My Videos\", \"Personal\"]\n\n\nclass WinePrefixManager:\n \"\"\"Class to allow modification of Wine prefixes without the use of Wine\"\"\"\n\n hkcu_prefix = \"HKEY_CURRENT_USER\"\n\n def __init__(self, path):\n if not path:\n logger.warning(\"No path specified for Wine prefix\")\n self.path = path\n\n def setup_defaults(self):\n \"\"\"Sets the defaults for newly created prefixes\"\"\"\n self.override_dll(\"winemenubuilder.exe\", \"\")\n self.override_dll(\"steamwebhelper.exe\", \"\")\n try:\n self.desktop_integration()\n except OSError as ex:\n logger.error(\"Failed to setup desktop integration, the prefix may not be valid.\")\n logger.exception(ex)\n\n def get_registry_path(self, key):\n \"\"\"Matches registry keys to a registry file\n\n Currently, only HKEY_CURRENT_USER keys are supported.\n \"\"\"\n if key.startswith(self.hkcu_prefix):\n return os.path.join(self.path, \"user.reg\")\n raise ValueError(\"Unsupported key '{}'\".format(key))\n\n def get_key_path(self, key):\n if key.startswith(self.hkcu_prefix):\n return key[len(self.hkcu_prefix) + 1:]\n raise ValueError(\n \"The key {} is currently not supported by WinePrefixManager\".format(key)\n )\n\n def get_registry_key(self,key,subkey):\n registry = WineRegistry(self.get_registry_path(key))\n return registry.query(self.get_key_path(key),subkey)\n\n def set_registry_key(self, key, subkey, value):\n registry = WineRegistry(self.get_registry_path(key))\n registry.set_value(self.get_key_path(key), subkey, value)\n registry.save()\n\n def clear_registry_key(self, key):\n registry = WineRegistry(self.get_registry_path(key))\n registry.clear_key(self.get_key_path(key))\n registry.save()\n\n def clear_registry_subkeys(self, key, subkeys):\n registry = WineRegistry(self.get_registry_path(key))\n registry.clear_subkeys(self.get_key_path(key), subkeys)\n registry.save()\n\n def override_dll(self, dll, mode):\n key = self.hkcu_prefix + \"/Software/Wine/DllOverrides\"\n if mode.startswith(\"dis\"):\n mode = \"\"\n if mode not in (\"builtin\", \"native\", \"builtin,native\", \"native,builtin\", \"\"):\n logger.error(\"DLL override '%s' mode is not valid\", mode)\n return\n self.set_registry_key(key, dll, mode)\n\n def desktop_integration(self, desktop_dir=None):\n \"\"\"Overwrite desktop integration\"\"\"\n DESKTOP_FOLDERS = []\n\n user = os.getenv(\"USER\")\n user_dir = os.path.join(self.path, \"drive_c/users/\", user)\n\n for key in DESKTOP_KEYS:\n folder = self.get_registry_key(self.hkcu_prefix+\"/Software/Microsoft/Windows/CurrentVersion/Explorer/Shell Folders\",key)\n DESKTOP_FOLDERS.append(folder[folder.rfind(\"\\\\\")+1:]) \n\n if not desktop_dir:\n desktop_dir = user_dir\n else:\n desktop_dir = os.path.expanduser(desktop_dir)\n\n if system.path_exists(user_dir):\n # Replace desktop integration symlinks\n for item in DESKTOP_FOLDERS:\n path = os.path.join(user_dir, item)\n old_path = path + \".winecfg\"\n\n if os.path.islink(path):\n os.unlink(path)\n elif os.path.isdir(path):\n try:\n os.rmdir(path)\n # We can't delete nonempty dir, so we rename as wine do.\n except OSError:\n os.rename(path, old_path)\n\n if desktop_dir != user_dir:\n src_path = os.path.join(desktop_dir, item)\n os.makedirs(src_path, exist_ok=True)\n os.symlink(src_path, path)\n else:\n # We use first the renamed dir, otherwise we make it.\n if os.path.isdir(old_path):\n os.rename(old_path, path)\n else:\n os.makedirs(path, exist_ok=True)\n\n # Security: Remove other symlinks.\n for item in os.listdir(user_dir):\n path = os.path.join(user_dir, item)\n if item not in DESKTOP_FOLDERS and os.path.islink(path):\n os.unlink(path)\n os.makedirs(path)\n\n def set_crash_dialogs(self, enabled):\n \"\"\"Enable or diable Wine crash dialogs\"\"\"\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/WineDbg\",\n \"ShowCrashDialog\",\n 1 if enabled else 0,\n )\n\n def set_virtual_desktop(self, enabled):\n \"\"\"Enable or disable wine virtual desktop.\n The Lutris virtual desktop is refered to as 'WineDesktop', in Wine the\n virtual desktop name is 'default'.\n \"\"\"\n path = self.hkcu_prefix + \"/Software/Wine/Explorer\"\n if enabled:\n self.set_registry_key(path, \"Desktop\", \"WineDesktop\")\n default_resolution = \"x\".join(DISPLAY_MANAGER.get_current_resolution())\n logger.debug(\"Enabling wine virtual desktop with default resolution of %s\",\n default_resolution)\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/Explorer/Desktops\",\n \"WineDesktop\",\n default_resolution\n )\n else:\n self.clear_registry_key(path)\n\n def set_desktop_size(self, desktop_size):\n \"\"\"Sets the desktop size if one is given but do not reset the key if\n one isn't.\n \"\"\"\n path = self.hkcu_prefix + \"/Software/Wine/Explorer/Desktops\"\n if desktop_size:\n self.set_registry_key(path, \"WineDesktop\", desktop_size)\n\n def use_xvid_mode(self, enabled):\n \"\"\"Set this to \"Y\" to allow wine switch the resolution using XVidMode extension.\"\"\"\n self.set_registry_key(\n self.hkcu_prefix + \"/Software/Wine/X11 Driver\",\n \"UseXVidMode\",\n \"Y\" if enabled else \"N\",\n )\n\n def configure_joypads(self):\n joypads = joypad.get_joypads()\n key = self.hkcu_prefix + \"/Software/Wine/DirectInput/Joysticks\"\n self.clear_registry_key(key)\n for device, joypad_name in joypads:\n if \"event\" in device:\n disabled_joypad = \"{} (js)\".format(joypad_name)\n else:\n disabled_joypad = \"{} (event)\".format(joypad_name)\n self.set_registry_key(key, disabled_joypad, \"disabled\")\n", "path": "lutris/util/wine/prefix.py"}]}
| 2,304 | 154 |
gh_patches_debug_36027
|
rasdani/github-patches
|
git_diff
|
learningequality__kolibri-5540
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
$computedClass warning might be too sensitive
### Observed behavior
We added a warning:
> Used the $computedClass method for a class definition without any pseudo selectors ... Please use a v-bind:style directive instead.
However I think there might be some cases where it's necessary to use this even without pseudo-selectors:
* I've been having a bit of trouble with CSS specificity rules, which are different depending on whether the style is defined using `:style` vs `:class`. It has been a bit challenging to get the two to have the right precedence related to each other. (Side-note - I think source maps would have helped here)
* There might be situations where a class is necessary in non-pseudo-selector cases - for example when defining the classes in vue transitions.
### Expected behavior
we should investigate these cases and decide whether the warning is appropriate, or if there was another way of implementing
### User-facing consequences
N/A
### Errors and logs

### Context
0.12
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `kolibri/core/hooks.py`
Content:
```
1 """
2 Kolibri Core hooks
3 ------------------
4
5 WIP! Many applications are supposed to live inside the core namespace to make
6 it explicit that they are part of the core.
7
8 Do we put all their hooks in one module or should each app have its own hooks
9 module?
10
11 Anyways, for now to get hooks started, we have some defined here...
12 """
13 from __future__ import absolute_import
14 from __future__ import print_function
15 from __future__ import unicode_literals
16
17 import logging
18
19 from kolibri.plugins.hooks import KolibriHook
20 from kolibri.plugins.utils import plugin_url
21
22 logger = logging.getLogger(__name__)
23
24
25 class NavigationHook(KolibriHook):
26
27 # : A string label for the menu item
28 label = "Untitled"
29
30 # : A string or lazy proxy for the url
31 url = "/"
32
33 # Set this to True so that any time this is mixed in with a
34 # frontend asset hook, the resulting frontend code will be rendered inline.
35 inline = True
36
37 def get_menu(self):
38 menu = {}
39 for hook in self.registered_hooks:
40 menu[hook.label] = self.url
41 return menu
42
43 class Meta:
44
45 abstract = True
46
47
48 class RoleBasedRedirectHook(KolibriHook):
49 # User role to redirect for
50 role = None
51
52 # URL to redirect to
53 url = None
54
55 # Special flag to only redirect on first login
56 # Default to False
57 first_login = False
58
59 def plugin_url(self, plugin_class, url_name):
60 return plugin_url(plugin_class, url_name)
61
62 class Meta:
63
64 abstract = True
65
```
Path: `kolibri/core/templatetags/kolibri_tags.py`
Content:
```
1 """
2 Kolibri template tags
3 =====================
4 """
5 from __future__ import absolute_import
6 from __future__ import print_function
7 from __future__ import unicode_literals
8
9 import copy
10 import json
11 import logging
12 import re
13
14 from django import template
15 from django.conf import settings
16 from django.contrib.staticfiles.templatetags.staticfiles import static
17 from django.core.urlresolvers import get_resolver
18 from django.core.urlresolvers import get_script_prefix
19 from django.core.urlresolvers import resolve
20 from django.core.urlresolvers import reverse
21 from django.template.loader import render_to_string
22 from django.utils.html import mark_safe
23 from django.utils.translation import get_language
24 from django.utils.translation import get_language_bidi
25 from django.utils.translation import get_language_info
26 from django_js_reverse.core import prepare_url_list
27 from django_js_reverse.rjsmin import jsmin
28 from rest_framework.renderers import JSONRenderer
29 from six import iteritems
30
31 import kolibri
32 from kolibri.core.device.models import ContentCacheKey
33 from kolibri.core.hooks import NavigationHook
34 from kolibri.core.webpack.utils import webpack_asset_render
35 from kolibri.utils import conf
36 from kolibri.utils import i18n
37
38 register = template.Library()
39
40 logger = logging.getLogger(__name__)
41
42
43 @register.simple_tag()
44 def kolibri_content_cache_key():
45 js = """
46 <script>
47 var contentCacheKey = '{cache_key}';
48 </script>
49 """.format(
50 cache_key=ContentCacheKey.get_cache_key()
51 )
52 return mark_safe(js)
53
54
55 @register.simple_tag(takes_context=True)
56 def kolibri_language_globals(context):
57
58 template = """
59 <script>
60 var languageCode = '{lang_code}';
61 var languageDir = '{lang_dir}';
62 var languages = JSON.parse('{languages}');
63 var fullCSSFileModern = '{full_css_file_modern}?v={version}';
64 var fullCSSFileBasic = '{full_css_file_basic}?v={version}';
65 </script>
66 <link type="text/css" href="{common_css_file}?v={version}" rel="stylesheet"/>
67 <link type="text/css" href="{subset_css_file}?v={version}" rel="stylesheet"/>
68 """
69
70 language_code = get_language()
71 lang_dir = "rtl" if get_language_bidi() else "ltr"
72
73 languages = {}
74 for code, language_name in settings.LANGUAGES:
75 lang_info = next(
76 (
77 lang
78 for lang in i18n.KOLIBRI_SUPPORTED_LANGUAGES
79 if lang["intl_code"] == code
80 ),
81 None,
82 )
83 languages[code] = {
84 # Format to match the schema of the content Language model
85 "id": code,
86 "lang_name": language_name,
87 "english_name": lang_info["english_name"]
88 if lang_info
89 else get_language_info(code)["name"],
90 "lang_direction": get_language_info(code)["bidi"],
91 }
92
93 common_file = static("assets/fonts/noto-common.css")
94 subset_file = static("assets/fonts/noto-subset.{}.css".format(language_code))
95 full_file = "assets/fonts/noto-full.{}.{}.css"
96 full_file_modern = static(full_file.format(language_code, "modern"))
97 full_file_basic = static(full_file.format(language_code, "basic"))
98
99 return mark_safe(
100 template.format(
101 lang_code=language_code,
102 lang_dir=lang_dir,
103 languages=json.dumps(languages),
104 common_css_file=common_file,
105 subset_css_file=subset_file,
106 full_css_file_modern=full_file_modern,
107 full_css_file_basic=full_file_basic,
108 # Temporary cache busting strategy.
109 # Would be better to use ManifestStaticFilesStorage
110 version=kolibri.__version__,
111 )
112 )
113
114
115 @register.simple_tag()
116 def kolibri_navigation_actions():
117 """
118 A tag to include an initial JS-object to bootstrap nav action data into the app.
119 :return: An html string
120 """
121 return webpack_asset_render(NavigationHook)
122
123
124 @register.simple_tag(takes_context=True)
125 def kolibri_set_urls(context):
126 # Modified from:
127 # https://github.com/ierror/django-js-reverse/blob/master/django_js_reverse/core.py#L101
128 js_global_object_name = "window"
129 js_var_name = "kolibriUrls"
130 script_prefix = get_script_prefix()
131
132 if "request" in context:
133 default_urlresolver = get_resolver(getattr(context["request"], "urlconf", None))
134 else:
135 default_urlresolver = get_resolver(None)
136
137 js = render_to_string(
138 "django_js_reverse/urls_js.tpl",
139 {
140 "urls": sorted(list(prepare_url_list(default_urlresolver))),
141 "url_prefix": script_prefix,
142 "js_var_name": js_var_name,
143 "js_global_object_name": js_global_object_name,
144 },
145 )
146
147 js = jsmin(js)
148
149 js = (
150 """<script type="text/javascript">"""
151 + js
152 + """
153 {global_object}.staticUrl = '{static_url}';
154 </script>
155 """.format(
156 global_object=js_global_object_name, static_url=settings.STATIC_URL
157 )
158 )
159 return mark_safe(js)
160
161
162 @register.simple_tag(takes_context=True)
163 def kolibri_bootstrap_model(context, base_name, api_resource, **kwargs):
164 response, kwargs = _kolibri_bootstrap_helper(
165 context, base_name, api_resource, "detail", **kwargs
166 )
167 html = (
168 "<script type='text/javascript'>"
169 "var model = {0}.resources.{1}.createModel(JSON.parse({2}));"
170 "model.synced = true;"
171 "</script>".format(
172 conf.KOLIBRI_CORE_JS_NAME,
173 api_resource,
174 json.dumps(JSONRenderer().render(response.data).decode("utf-8")),
175 )
176 )
177 return mark_safe(html)
178
179
180 @register.simple_tag(takes_context=True)
181 def kolibri_bootstrap_collection(context, base_name, api_resource, **kwargs):
182 response, kwargs = _kolibri_bootstrap_helper(
183 context, base_name, api_resource, "list", **kwargs
184 )
185 html = (
186 "<script type='text/javascript'>"
187 "var collection = {0}.resources.{1}.createCollection({2}, JSON.parse({3}));"
188 "collection.synced = true;"
189 "</script>".format(
190 conf.KOLIBRI_CORE_JS_NAME,
191 api_resource,
192 json.dumps(kwargs),
193 json.dumps(JSONRenderer().render(response.data).decode("utf-8")),
194 )
195 )
196 return mark_safe(html)
197
198
199 def _replace_dict_values(check, replace, dict):
200 for (key, value) in iteritems(dict):
201 if dict[key] is check:
202 dict[key] = replace
203
204
205 def _kolibri_bootstrap_helper(context, base_name, api_resource, route, **kwargs):
206 reversal = dict()
207 kwargs_check = "kwargs_"
208 # remove prepended string and matching items from kwargs
209 for key in list(kwargs.keys()):
210 if kwargs_check in key:
211 item = kwargs.pop(key)
212 key = re.sub(kwargs_check, "", key)
213 reversal[key] = item
214 view, view_args, view_kwargs = resolve(
215 reverse("kolibri:core:{0}-{1}".format(base_name, route), kwargs=reversal)
216 )
217 # switch out None temporarily because invalid filtering and caching can occur
218 _replace_dict_values(None, str(""), kwargs)
219 request = copy.copy(context["request"])
220 request.GET = request.GET.copy()
221 for key in kwargs:
222 request.GET[key] = kwargs[key]
223 response = view(request, **view_kwargs)
224 _replace_dict_values(str(""), None, kwargs)
225 return response, kwargs
226
227
228 @register.simple_tag()
229 def kolibri_sentry_error_reporting():
230
231 if not conf.OPTIONS["Debug"]["SENTRY_FRONTEND_DSN"]:
232 return ""
233
234 template = """
235 <script>
236 var sentryDSN = '{dsn}';
237 var sentryEnv = '{env}';
238 </script>
239 """
240 return mark_safe(
241 template.format(
242 dsn=conf.OPTIONS["Debug"]["SENTRY_FRONTEND_DSN"],
243 env=conf.OPTIONS["Debug"]["SENTRY_ENVIRONMENT"],
244 )
245 )
246
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/kolibri/core/hooks.py b/kolibri/core/hooks.py
--- a/kolibri/core/hooks.py
+++ b/kolibri/core/hooks.py
@@ -15,7 +15,9 @@
from __future__ import unicode_literals
import logging
+import warnings
+from kolibri.plugins.hooks import abstract_method
from kolibri.plugins.hooks import KolibriHook
from kolibri.plugins.utils import plugin_url
@@ -62,3 +64,47 @@
class Meta:
abstract = True
+
+
+class MultipleThemesWarning(UserWarning):
+ pass
+
+
+class ThemeHook(KolibriHook):
+ """
+ A hook to allow custom theming of Kolibri
+ """
+
+ class Meta:
+
+ abstract = True
+
+ @property
+ @abstract_method
+ def theme(self):
+ default = {
+ # Whether to show the Kolibri log
+ # Boolean
+ "showKolibriLogo": True,
+ # URL for custom logo
+ "customLogoURL": None,
+ # URL for custom login background image
+ "splashBackgroundURL": None,
+ # Color Palette specification
+ "paletteColors": {},
+ # Brand Color specification
+ "brandColors": {},
+ # Mapping from colors to particular usage
+ "tokenMapping": {},
+ }
+ theme = {}
+ once = False
+ for hook in self.registered_hooks:
+ if once:
+ warnings.warn("Multiple themes defined by plugins, ignoring all themes")
+ return default
+ for key in default:
+ theme[key] = getattr(hook, key, theme[key])
+ once = True
+
+ return theme or default
diff --git a/kolibri/core/templatetags/kolibri_tags.py b/kolibri/core/templatetags/kolibri_tags.py
--- a/kolibri/core/templatetags/kolibri_tags.py
+++ b/kolibri/core/templatetags/kolibri_tags.py
@@ -31,6 +31,7 @@
import kolibri
from kolibri.core.device.models import ContentCacheKey
from kolibri.core.hooks import NavigationHook
+from kolibri.core.hooks import ThemeHook
from kolibri.core.webpack.utils import webpack_asset_render
from kolibri.utils import conf
from kolibri.utils import i18n
@@ -121,6 +122,20 @@
return webpack_asset_render(NavigationHook)
[email protected]_tag()
+def kolibri_theme():
+ """
+ A tag to include a theme configuration object to add custom theming to Kolibri.
+ :return: An html string
+ """
+ template = """
+ <script>
+ var customTheme = JSON.parse('{theme}');
+ </script>
+ """
+ return mark_safe(template.format(theme=json.dumps(ThemeHook().theme)))
+
+
@register.simple_tag(takes_context=True)
def kolibri_set_urls(context):
# Modified from:
|
{"golden_diff": "diff --git a/kolibri/core/hooks.py b/kolibri/core/hooks.py\n--- a/kolibri/core/hooks.py\n+++ b/kolibri/core/hooks.py\n@@ -15,7 +15,9 @@\n from __future__ import unicode_literals\n \n import logging\n+import warnings\n \n+from kolibri.plugins.hooks import abstract_method\n from kolibri.plugins.hooks import KolibriHook\n from kolibri.plugins.utils import plugin_url\n \n@@ -62,3 +64,47 @@\n class Meta:\n \n abstract = True\n+\n+\n+class MultipleThemesWarning(UserWarning):\n+ pass\n+\n+\n+class ThemeHook(KolibriHook):\n+ \"\"\"\n+ A hook to allow custom theming of Kolibri\n+ \"\"\"\n+\n+ class Meta:\n+\n+ abstract = True\n+\n+ @property\n+ @abstract_method\n+ def theme(self):\n+ default = {\n+ # Whether to show the Kolibri log\n+ # Boolean\n+ \"showKolibriLogo\": True,\n+ # URL for custom logo\n+ \"customLogoURL\": None,\n+ # URL for custom login background image\n+ \"splashBackgroundURL\": None,\n+ # Color Palette specification\n+ \"paletteColors\": {},\n+ # Brand Color specification\n+ \"brandColors\": {},\n+ # Mapping from colors to particular usage\n+ \"tokenMapping\": {},\n+ }\n+ theme = {}\n+ once = False\n+ for hook in self.registered_hooks:\n+ if once:\n+ warnings.warn(\"Multiple themes defined by plugins, ignoring all themes\")\n+ return default\n+ for key in default:\n+ theme[key] = getattr(hook, key, theme[key])\n+ once = True\n+\n+ return theme or default\ndiff --git a/kolibri/core/templatetags/kolibri_tags.py b/kolibri/core/templatetags/kolibri_tags.py\n--- a/kolibri/core/templatetags/kolibri_tags.py\n+++ b/kolibri/core/templatetags/kolibri_tags.py\n@@ -31,6 +31,7 @@\n import kolibri\n from kolibri.core.device.models import ContentCacheKey\n from kolibri.core.hooks import NavigationHook\n+from kolibri.core.hooks import ThemeHook\n from kolibri.core.webpack.utils import webpack_asset_render\n from kolibri.utils import conf\n from kolibri.utils import i18n\n@@ -121,6 +122,20 @@\n return webpack_asset_render(NavigationHook)\n \n \[email protected]_tag()\n+def kolibri_theme():\n+ \"\"\"\n+ A tag to include a theme configuration object to add custom theming to Kolibri.\n+ :return: An html string\n+ \"\"\"\n+ template = \"\"\"\n+ <script>\n+ var customTheme = JSON.parse('{theme}');\n+ </script>\n+ \"\"\"\n+ return mark_safe(template.format(theme=json.dumps(ThemeHook().theme)))\n+\n+\n @register.simple_tag(takes_context=True)\n def kolibri_set_urls(context):\n # Modified from:\n", "issue": "$computedClass warning might be too sensitive\n### Observed behavior\r\n\r\nWe added a warning:\r\n\r\n> Used the $computedClass method for a class definition without any pseudo selectors ... Please use a v-bind:style directive instead.\r\n\r\nHowever I think there might be some cases where it's necessary to use this even without pseudo-selectors:\r\n\r\n* I've been having a bit of trouble with CSS specificity rules, which are different depending on whether the style is defined using `:style` vs `:class`. It has been a bit challenging to get the two to have the right precedence related to each other. (Side-note - I think source maps would have helped here)\r\n* There might be situations where a class is necessary in non-pseudo-selector cases - for example when defining the classes in vue transitions.\r\n\r\n\r\n\r\n### Expected behavior\r\n\r\nwe should investigate these cases and decide whether the warning is appropriate, or if there was another way of implementing\r\n\r\n### User-facing consequences\r\n\r\nN/A\r\n\r\n### Errors and logs\r\n\r\n\r\n\r\n\r\n### Context\r\n\r\n0.12\n", "before_files": [{"content": "\"\"\"\nKolibri Core hooks\n------------------\n\nWIP! Many applications are supposed to live inside the core namespace to make\nit explicit that they are part of the core.\n\nDo we put all their hooks in one module or should each app have its own hooks\nmodule?\n\nAnyways, for now to get hooks started, we have some defined here...\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport logging\n\nfrom kolibri.plugins.hooks import KolibriHook\nfrom kolibri.plugins.utils import plugin_url\n\nlogger = logging.getLogger(__name__)\n\n\nclass NavigationHook(KolibriHook):\n\n # : A string label for the menu item\n label = \"Untitled\"\n\n # : A string or lazy proxy for the url\n url = \"/\"\n\n # Set this to True so that any time this is mixed in with a\n # frontend asset hook, the resulting frontend code will be rendered inline.\n inline = True\n\n def get_menu(self):\n menu = {}\n for hook in self.registered_hooks:\n menu[hook.label] = self.url\n return menu\n\n class Meta:\n\n abstract = True\n\n\nclass RoleBasedRedirectHook(KolibriHook):\n # User role to redirect for\n role = None\n\n # URL to redirect to\n url = None\n\n # Special flag to only redirect on first login\n # Default to False\n first_login = False\n\n def plugin_url(self, plugin_class, url_name):\n return plugin_url(plugin_class, url_name)\n\n class Meta:\n\n abstract = True\n", "path": "kolibri/core/hooks.py"}, {"content": "\"\"\"\nKolibri template tags\n=====================\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport copy\nimport json\nimport logging\nimport re\n\nfrom django import template\nfrom django.conf import settings\nfrom django.contrib.staticfiles.templatetags.staticfiles import static\nfrom django.core.urlresolvers import get_resolver\nfrom django.core.urlresolvers import get_script_prefix\nfrom django.core.urlresolvers import resolve\nfrom django.core.urlresolvers import reverse\nfrom django.template.loader import render_to_string\nfrom django.utils.html import mark_safe\nfrom django.utils.translation import get_language\nfrom django.utils.translation import get_language_bidi\nfrom django.utils.translation import get_language_info\nfrom django_js_reverse.core import prepare_url_list\nfrom django_js_reverse.rjsmin import jsmin\nfrom rest_framework.renderers import JSONRenderer\nfrom six import iteritems\n\nimport kolibri\nfrom kolibri.core.device.models import ContentCacheKey\nfrom kolibri.core.hooks import NavigationHook\nfrom kolibri.core.webpack.utils import webpack_asset_render\nfrom kolibri.utils import conf\nfrom kolibri.utils import i18n\n\nregister = template.Library()\n\nlogger = logging.getLogger(__name__)\n\n\[email protected]_tag()\ndef kolibri_content_cache_key():\n js = \"\"\"\n <script>\n var contentCacheKey = '{cache_key}';\n </script>\n \"\"\".format(\n cache_key=ContentCacheKey.get_cache_key()\n )\n return mark_safe(js)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_language_globals(context):\n\n template = \"\"\"\n <script>\n var languageCode = '{lang_code}';\n var languageDir = '{lang_dir}';\n var languages = JSON.parse('{languages}');\n var fullCSSFileModern = '{full_css_file_modern}?v={version}';\n var fullCSSFileBasic = '{full_css_file_basic}?v={version}';\n </script>\n <link type=\"text/css\" href=\"{common_css_file}?v={version}\" rel=\"stylesheet\"/>\n <link type=\"text/css\" href=\"{subset_css_file}?v={version}\" rel=\"stylesheet\"/>\n \"\"\"\n\n language_code = get_language()\n lang_dir = \"rtl\" if get_language_bidi() else \"ltr\"\n\n languages = {}\n for code, language_name in settings.LANGUAGES:\n lang_info = next(\n (\n lang\n for lang in i18n.KOLIBRI_SUPPORTED_LANGUAGES\n if lang[\"intl_code\"] == code\n ),\n None,\n )\n languages[code] = {\n # Format to match the schema of the content Language model\n \"id\": code,\n \"lang_name\": language_name,\n \"english_name\": lang_info[\"english_name\"]\n if lang_info\n else get_language_info(code)[\"name\"],\n \"lang_direction\": get_language_info(code)[\"bidi\"],\n }\n\n common_file = static(\"assets/fonts/noto-common.css\")\n subset_file = static(\"assets/fonts/noto-subset.{}.css\".format(language_code))\n full_file = \"assets/fonts/noto-full.{}.{}.css\"\n full_file_modern = static(full_file.format(language_code, \"modern\"))\n full_file_basic = static(full_file.format(language_code, \"basic\"))\n\n return mark_safe(\n template.format(\n lang_code=language_code,\n lang_dir=lang_dir,\n languages=json.dumps(languages),\n common_css_file=common_file,\n subset_css_file=subset_file,\n full_css_file_modern=full_file_modern,\n full_css_file_basic=full_file_basic,\n # Temporary cache busting strategy.\n # Would be better to use ManifestStaticFilesStorage\n version=kolibri.__version__,\n )\n )\n\n\[email protected]_tag()\ndef kolibri_navigation_actions():\n \"\"\"\n A tag to include an initial JS-object to bootstrap nav action data into the app.\n :return: An html string\n \"\"\"\n return webpack_asset_render(NavigationHook)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_set_urls(context):\n # Modified from:\n # https://github.com/ierror/django-js-reverse/blob/master/django_js_reverse/core.py#L101\n js_global_object_name = \"window\"\n js_var_name = \"kolibriUrls\"\n script_prefix = get_script_prefix()\n\n if \"request\" in context:\n default_urlresolver = get_resolver(getattr(context[\"request\"], \"urlconf\", None))\n else:\n default_urlresolver = get_resolver(None)\n\n js = render_to_string(\n \"django_js_reverse/urls_js.tpl\",\n {\n \"urls\": sorted(list(prepare_url_list(default_urlresolver))),\n \"url_prefix\": script_prefix,\n \"js_var_name\": js_var_name,\n \"js_global_object_name\": js_global_object_name,\n },\n )\n\n js = jsmin(js)\n\n js = (\n \"\"\"<script type=\"text/javascript\">\"\"\"\n + js\n + \"\"\"\n {global_object}.staticUrl = '{static_url}';\n </script>\n \"\"\".format(\n global_object=js_global_object_name, static_url=settings.STATIC_URL\n )\n )\n return mark_safe(js)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_bootstrap_model(context, base_name, api_resource, **kwargs):\n response, kwargs = _kolibri_bootstrap_helper(\n context, base_name, api_resource, \"detail\", **kwargs\n )\n html = (\n \"<script type='text/javascript'>\"\n \"var model = {0}.resources.{1}.createModel(JSON.parse({2}));\"\n \"model.synced = true;\"\n \"</script>\".format(\n conf.KOLIBRI_CORE_JS_NAME,\n api_resource,\n json.dumps(JSONRenderer().render(response.data).decode(\"utf-8\")),\n )\n )\n return mark_safe(html)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_bootstrap_collection(context, base_name, api_resource, **kwargs):\n response, kwargs = _kolibri_bootstrap_helper(\n context, base_name, api_resource, \"list\", **kwargs\n )\n html = (\n \"<script type='text/javascript'>\"\n \"var collection = {0}.resources.{1}.createCollection({2}, JSON.parse({3}));\"\n \"collection.synced = true;\"\n \"</script>\".format(\n conf.KOLIBRI_CORE_JS_NAME,\n api_resource,\n json.dumps(kwargs),\n json.dumps(JSONRenderer().render(response.data).decode(\"utf-8\")),\n )\n )\n return mark_safe(html)\n\n\ndef _replace_dict_values(check, replace, dict):\n for (key, value) in iteritems(dict):\n if dict[key] is check:\n dict[key] = replace\n\n\ndef _kolibri_bootstrap_helper(context, base_name, api_resource, route, **kwargs):\n reversal = dict()\n kwargs_check = \"kwargs_\"\n # remove prepended string and matching items from kwargs\n for key in list(kwargs.keys()):\n if kwargs_check in key:\n item = kwargs.pop(key)\n key = re.sub(kwargs_check, \"\", key)\n reversal[key] = item\n view, view_args, view_kwargs = resolve(\n reverse(\"kolibri:core:{0}-{1}\".format(base_name, route), kwargs=reversal)\n )\n # switch out None temporarily because invalid filtering and caching can occur\n _replace_dict_values(None, str(\"\"), kwargs)\n request = copy.copy(context[\"request\"])\n request.GET = request.GET.copy()\n for key in kwargs:\n request.GET[key] = kwargs[key]\n response = view(request, **view_kwargs)\n _replace_dict_values(str(\"\"), None, kwargs)\n return response, kwargs\n\n\[email protected]_tag()\ndef kolibri_sentry_error_reporting():\n\n if not conf.OPTIONS[\"Debug\"][\"SENTRY_FRONTEND_DSN\"]:\n return \"\"\n\n template = \"\"\"\n <script>\n var sentryDSN = '{dsn}';\n var sentryEnv = '{env}';\n </script>\n \"\"\"\n return mark_safe(\n template.format(\n dsn=conf.OPTIONS[\"Debug\"][\"SENTRY_FRONTEND_DSN\"],\n env=conf.OPTIONS[\"Debug\"][\"SENTRY_ENVIRONMENT\"],\n )\n )\n", "path": "kolibri/core/templatetags/kolibri_tags.py"}], "after_files": [{"content": "\"\"\"\nKolibri Core hooks\n------------------\n\nWIP! Many applications are supposed to live inside the core namespace to make\nit explicit that they are part of the core.\n\nDo we put all their hooks in one module or should each app have its own hooks\nmodule?\n\nAnyways, for now to get hooks started, we have some defined here...\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport logging\nimport warnings\n\nfrom kolibri.plugins.hooks import abstract_method\nfrom kolibri.plugins.hooks import KolibriHook\nfrom kolibri.plugins.utils import plugin_url\n\nlogger = logging.getLogger(__name__)\n\n\nclass NavigationHook(KolibriHook):\n\n # : A string label for the menu item\n label = \"Untitled\"\n\n # : A string or lazy proxy for the url\n url = \"/\"\n\n # Set this to True so that any time this is mixed in with a\n # frontend asset hook, the resulting frontend code will be rendered inline.\n inline = True\n\n def get_menu(self):\n menu = {}\n for hook in self.registered_hooks:\n menu[hook.label] = self.url\n return menu\n\n class Meta:\n\n abstract = True\n\n\nclass RoleBasedRedirectHook(KolibriHook):\n # User role to redirect for\n role = None\n\n # URL to redirect to\n url = None\n\n # Special flag to only redirect on first login\n # Default to False\n first_login = False\n\n def plugin_url(self, plugin_class, url_name):\n return plugin_url(plugin_class, url_name)\n\n class Meta:\n\n abstract = True\n\n\nclass MultipleThemesWarning(UserWarning):\n pass\n\n\nclass ThemeHook(KolibriHook):\n \"\"\"\n A hook to allow custom theming of Kolibri\n \"\"\"\n\n class Meta:\n\n abstract = True\n\n @property\n @abstract_method\n def theme(self):\n default = {\n # Whether to show the Kolibri log\n # Boolean\n \"showKolibriLogo\": True,\n # URL for custom logo\n \"customLogoURL\": None,\n # URL for custom login background image\n \"splashBackgroundURL\": None,\n # Color Palette specification\n \"paletteColors\": {},\n # Brand Color specification\n \"brandColors\": {},\n # Mapping from colors to particular usage\n \"tokenMapping\": {},\n }\n theme = {}\n once = False\n for hook in self.registered_hooks:\n if once:\n warnings.warn(\"Multiple themes defined by plugins, ignoring all themes\")\n return default\n for key in default:\n theme[key] = getattr(hook, key, theme[key])\n once = True\n\n return theme or default\n", "path": "kolibri/core/hooks.py"}, {"content": "\"\"\"\nKolibri template tags\n=====================\n\"\"\"\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport copy\nimport json\nimport logging\nimport re\n\nfrom django import template\nfrom django.conf import settings\nfrom django.contrib.staticfiles.templatetags.staticfiles import static\nfrom django.core.urlresolvers import get_resolver\nfrom django.core.urlresolvers import get_script_prefix\nfrom django.core.urlresolvers import resolve\nfrom django.core.urlresolvers import reverse\nfrom django.template.loader import render_to_string\nfrom django.utils.html import mark_safe\nfrom django.utils.translation import get_language\nfrom django.utils.translation import get_language_bidi\nfrom django.utils.translation import get_language_info\nfrom django_js_reverse.core import prepare_url_list\nfrom django_js_reverse.rjsmin import jsmin\nfrom rest_framework.renderers import JSONRenderer\nfrom six import iteritems\n\nimport kolibri\nfrom kolibri.core.device.models import ContentCacheKey\nfrom kolibri.core.hooks import NavigationHook\nfrom kolibri.core.hooks import ThemeHook\nfrom kolibri.core.webpack.utils import webpack_asset_render\nfrom kolibri.utils import conf\nfrom kolibri.utils import i18n\n\nregister = template.Library()\n\nlogger = logging.getLogger(__name__)\n\n\[email protected]_tag()\ndef kolibri_content_cache_key():\n js = \"\"\"\n <script>\n var contentCacheKey = '{cache_key}';\n </script>\n \"\"\".format(\n cache_key=ContentCacheKey.get_cache_key()\n )\n return mark_safe(js)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_language_globals(context):\n\n template = \"\"\"\n <script>\n var languageCode = '{lang_code}';\n var languageDir = '{lang_dir}';\n var languages = JSON.parse('{languages}');\n var fullCSSFileModern = '{full_css_file_modern}?v={version}';\n var fullCSSFileBasic = '{full_css_file_basic}?v={version}';\n </script>\n <link type=\"text/css\" href=\"{common_css_file}?v={version}\" rel=\"stylesheet\"/>\n <link type=\"text/css\" href=\"{subset_css_file}?v={version}\" rel=\"stylesheet\"/>\n \"\"\"\n\n language_code = get_language()\n lang_dir = \"rtl\" if get_language_bidi() else \"ltr\"\n\n languages = {}\n for code, language_name in settings.LANGUAGES:\n lang_info = next(\n (\n lang\n for lang in i18n.KOLIBRI_SUPPORTED_LANGUAGES\n if lang[\"intl_code\"] == code\n ),\n None,\n )\n languages[code] = {\n # Format to match the schema of the content Language model\n \"id\": code,\n \"lang_name\": language_name,\n \"english_name\": lang_info[\"english_name\"]\n if lang_info\n else get_language_info(code)[\"name\"],\n \"lang_direction\": get_language_info(code)[\"bidi\"],\n }\n\n common_file = static(\"assets/fonts/noto-common.css\")\n subset_file = static(\"assets/fonts/noto-subset.{}.css\".format(language_code))\n full_file = \"assets/fonts/noto-full.{}.{}.css\"\n full_file_modern = static(full_file.format(language_code, \"modern\"))\n full_file_basic = static(full_file.format(language_code, \"basic\"))\n\n return mark_safe(\n template.format(\n lang_code=language_code,\n lang_dir=lang_dir,\n languages=json.dumps(languages),\n common_css_file=common_file,\n subset_css_file=subset_file,\n full_css_file_modern=full_file_modern,\n full_css_file_basic=full_file_basic,\n # Temporary cache busting strategy.\n # Would be better to use ManifestStaticFilesStorage\n version=kolibri.__version__,\n )\n )\n\n\[email protected]_tag()\ndef kolibri_navigation_actions():\n \"\"\"\n A tag to include an initial JS-object to bootstrap nav action data into the app.\n :return: An html string\n \"\"\"\n return webpack_asset_render(NavigationHook)\n\n\[email protected]_tag()\ndef kolibri_theme():\n \"\"\"\n A tag to include a theme configuration object to add custom theming to Kolibri.\n :return: An html string\n \"\"\"\n template = \"\"\"\n <script>\n var customTheme = JSON.parse('{theme}');\n </script>\n \"\"\"\n return mark_safe(template.format(theme=json.dumps(ThemeHook().theme)))\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_set_urls(context):\n # Modified from:\n # https://github.com/ierror/django-js-reverse/blob/master/django_js_reverse/core.py#L101\n js_global_object_name = \"window\"\n js_var_name = \"kolibriUrls\"\n script_prefix = get_script_prefix()\n\n if \"request\" in context:\n default_urlresolver = get_resolver(getattr(context[\"request\"], \"urlconf\", None))\n else:\n default_urlresolver = get_resolver(None)\n\n js = render_to_string(\n \"django_js_reverse/urls_js.tpl\",\n {\n \"urls\": sorted(list(prepare_url_list(default_urlresolver))),\n \"url_prefix\": script_prefix,\n \"js_var_name\": js_var_name,\n \"js_global_object_name\": js_global_object_name,\n },\n )\n\n js = jsmin(js)\n\n js = (\n \"\"\"<script type=\"text/javascript\">\"\"\"\n + js\n + \"\"\"\n {global_object}.staticUrl = '{static_url}';\n </script>\n \"\"\".format(\n global_object=js_global_object_name, static_url=settings.STATIC_URL\n )\n )\n return mark_safe(js)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_bootstrap_model(context, base_name, api_resource, **kwargs):\n response, kwargs = _kolibri_bootstrap_helper(\n context, base_name, api_resource, \"detail\", **kwargs\n )\n html = (\n \"<script type='text/javascript'>\"\n \"var model = {0}.resources.{1}.createModel(JSON.parse({2}));\"\n \"model.synced = true;\"\n \"</script>\".format(\n conf.KOLIBRI_CORE_JS_NAME,\n api_resource,\n json.dumps(JSONRenderer().render(response.data).decode(\"utf-8\")),\n )\n )\n return mark_safe(html)\n\n\[email protected]_tag(takes_context=True)\ndef kolibri_bootstrap_collection(context, base_name, api_resource, **kwargs):\n response, kwargs = _kolibri_bootstrap_helper(\n context, base_name, api_resource, \"list\", **kwargs\n )\n html = (\n \"<script type='text/javascript'>\"\n \"var collection = {0}.resources.{1}.createCollection({2}, JSON.parse({3}));\"\n \"collection.synced = true;\"\n \"</script>\".format(\n conf.KOLIBRI_CORE_JS_NAME,\n api_resource,\n json.dumps(kwargs),\n json.dumps(JSONRenderer().render(response.data).decode(\"utf-8\")),\n )\n )\n return mark_safe(html)\n\n\ndef _replace_dict_values(check, replace, dict):\n for (key, value) in iteritems(dict):\n if dict[key] is check:\n dict[key] = replace\n\n\ndef _kolibri_bootstrap_helper(context, base_name, api_resource, route, **kwargs):\n reversal = dict()\n kwargs_check = \"kwargs_\"\n # remove prepended string and matching items from kwargs\n for key in list(kwargs.keys()):\n if kwargs_check in key:\n item = kwargs.pop(key)\n key = re.sub(kwargs_check, \"\", key)\n reversal[key] = item\n view, view_args, view_kwargs = resolve(\n reverse(\"kolibri:core:{0}-{1}\".format(base_name, route), kwargs=reversal)\n )\n # switch out None temporarily because invalid filtering and caching can occur\n _replace_dict_values(None, str(\"\"), kwargs)\n request = copy.copy(context[\"request\"])\n request.GET = request.GET.copy()\n for key in kwargs:\n request.GET[key] = kwargs[key]\n response = view(request, **view_kwargs)\n _replace_dict_values(str(\"\"), None, kwargs)\n return response, kwargs\n\n\[email protected]_tag()\ndef kolibri_sentry_error_reporting():\n\n if not conf.OPTIONS[\"Debug\"][\"SENTRY_FRONTEND_DSN\"]:\n return \"\"\n\n template = \"\"\"\n <script>\n var sentryDSN = '{dsn}';\n var sentryEnv = '{env}';\n </script>\n \"\"\"\n return mark_safe(\n template.format(\n dsn=conf.OPTIONS[\"Debug\"][\"SENTRY_FRONTEND_DSN\"],\n env=conf.OPTIONS[\"Debug\"][\"SENTRY_ENVIRONMENT\"],\n )\n )\n", "path": "kolibri/core/templatetags/kolibri_tags.py"}]}
| 3,432 | 674 |
gh_patches_debug_3798
|
rasdani/github-patches
|
git_diff
|
Parsl__parsl-618
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Error building docs with new env
```
Running Sphinx v1.8.1
loading pickled environment... failed: build environment version not current
Theme error:
sphinx_rtd_theme is no longer a hard dependency since version 1.4.0. Please install it manually.(pip install sphinx_rtd_theme)
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 from setuptools import setup, find_packages
2
3 with open('parsl/version.py') as f:
4 exec(f.read())
5
6 with open('requirements.txt') as f:
7 install_requires = f.readlines()
8
9 setup(
10 name='parsl',
11 version=VERSION,
12 description='Simple data dependent workflows in Python',
13 long_description='Simple parallel workflows system for Python',
14 url='https://github.com/Parsl/parsl',
15 author='The Parsl Team',
16 author_email='[email protected]',
17 license='Apache 2.0',
18 download_url='https://github.com/Parsl/parsl/archive/{}.tar.gz'.format(VERSION),
19 package_data={'': ['LICENSE']},
20 packages=find_packages(),
21 install_requires=install_requires,
22 scripts = ['parsl/executors/high_throughput/process_worker_pool.py',
23 'parsl/executors/extreme_scale/mpi_worker_pool.py'],
24 extras_require = {
25 'db_logging' : ['CMRESHandler', 'psutil', 'sqlalchemy'],
26 'aws' : ['boto3'],
27 'jetstream' : ['python-novaclient'],
28 'extreme_scale' : ['mpi4py'],
29 'docs' : ['nbsphinx'],
30 'google_cloud' : ['google-auth', 'google-api-python-client']
31 },
32 classifiers = [
33 # Maturity
34 'Development Status :: 3 - Alpha',
35 # Intended audience
36 'Intended Audience :: Developers',
37 # Licence, must match with licence above
38 'License :: OSI Approved :: Apache Software License',
39 # Python versions supported
40 'Programming Language :: Python :: 3.5',
41 'Programming Language :: Python :: 3.6',
42 ],
43 keywords=['Workflows', 'Scientific computing'],
44 )
45
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -26,7 +26,7 @@
'aws' : ['boto3'],
'jetstream' : ['python-novaclient'],
'extreme_scale' : ['mpi4py'],
- 'docs' : ['nbsphinx'],
+ 'docs' : ['nbsphinx', 'sphinx_rtd_theme'],
'google_cloud' : ['google-auth', 'google-api-python-client']
},
classifiers = [
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -26,7 +26,7 @@\n 'aws' : ['boto3'],\n 'jetstream' : ['python-novaclient'],\n 'extreme_scale' : ['mpi4py'],\n- 'docs' : ['nbsphinx'],\n+ 'docs' : ['nbsphinx', 'sphinx_rtd_theme'],\n 'google_cloud' : ['google-auth', 'google-api-python-client']\n },\n classifiers = [\n", "issue": "Error building docs with new env\n\r\n```\r\nRunning Sphinx v1.8.1\r\nloading pickled environment... failed: build environment version not current\r\n\r\nTheme error:\r\nsphinx_rtd_theme is no longer a hard dependency since version 1.4.0. Please install it manually.(pip install sphinx_rtd_theme)\r\n```\n", "before_files": [{"content": "from setuptools import setup, find_packages\n\nwith open('parsl/version.py') as f:\n exec(f.read())\n\nwith open('requirements.txt') as f:\n install_requires = f.readlines()\n\nsetup(\n name='parsl',\n version=VERSION,\n description='Simple data dependent workflows in Python',\n long_description='Simple parallel workflows system for Python',\n url='https://github.com/Parsl/parsl',\n author='The Parsl Team',\n author_email='[email protected]',\n license='Apache 2.0',\n download_url='https://github.com/Parsl/parsl/archive/{}.tar.gz'.format(VERSION),\n package_data={'': ['LICENSE']},\n packages=find_packages(),\n install_requires=install_requires,\n scripts = ['parsl/executors/high_throughput/process_worker_pool.py',\n 'parsl/executors/extreme_scale/mpi_worker_pool.py'],\n extras_require = {\n 'db_logging' : ['CMRESHandler', 'psutil', 'sqlalchemy'],\n 'aws' : ['boto3'],\n 'jetstream' : ['python-novaclient'],\n 'extreme_scale' : ['mpi4py'],\n 'docs' : ['nbsphinx'],\n 'google_cloud' : ['google-auth', 'google-api-python-client']\n },\n classifiers = [\n # Maturity\n 'Development Status :: 3 - Alpha',\n # Intended audience\n 'Intended Audience :: Developers',\n # Licence, must match with licence above\n 'License :: OSI Approved :: Apache Software License',\n # Python versions supported\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n ],\n keywords=['Workflows', 'Scientific computing'],\n)\n", "path": "setup.py"}], "after_files": [{"content": "from setuptools import setup, find_packages\n\nwith open('parsl/version.py') as f:\n exec(f.read())\n\nwith open('requirements.txt') as f:\n install_requires = f.readlines()\n\nsetup(\n name='parsl',\n version=VERSION,\n description='Simple data dependent workflows in Python',\n long_description='Simple parallel workflows system for Python',\n url='https://github.com/Parsl/parsl',\n author='The Parsl Team',\n author_email='[email protected]',\n license='Apache 2.0',\n download_url='https://github.com/Parsl/parsl/archive/{}.tar.gz'.format(VERSION),\n package_data={'': ['LICENSE']},\n packages=find_packages(),\n install_requires=install_requires,\n scripts = ['parsl/executors/high_throughput/process_worker_pool.py',\n 'parsl/executors/extreme_scale/mpi_worker_pool.py'],\n extras_require = {\n 'db_logging' : ['CMRESHandler', 'psutil', 'sqlalchemy'],\n 'aws' : ['boto3'],\n 'jetstream' : ['python-novaclient'],\n 'extreme_scale' : ['mpi4py'],\n 'docs' : ['nbsphinx', 'sphinx_rtd_theme'],\n 'google_cloud' : ['google-auth', 'google-api-python-client']\n },\n classifiers = [\n # Maturity\n 'Development Status :: 3 - Alpha',\n # Intended audience\n 'Intended Audience :: Developers',\n # Licence, must match with licence above\n 'License :: OSI Approved :: Apache Software License',\n # Python versions supported\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n ],\n keywords=['Workflows', 'Scientific computing'],\n)\n", "path": "setup.py"}]}
| 786 | 117 |
gh_patches_debug_31952
|
rasdani/github-patches
|
git_diff
|
svthalia__concrexit-3174
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Re-add rudamentary version of organisers to the event list API endpoint
<!--
Please add the appropriate label for what change should be made:
docs: changes to the documentation)
refactor: refactoring production code, eg. renaming a variable or rewriting a function
test: adding missing tests, refactoring tests; no production code change
chore: updating poetry etc; no production code change
-->
### Describe the change
Add a id-name based list of event organisers to the event list API endpoint.
### Motivation
Because it would make me and my Reaxit PR very happy.
### Current implementation
It got removed :(
### Suggested implementation
Just do it Kappa
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `website/activemembers/api/v2/serializers/member_group.py`
Content:
```
1 from rest_framework import serializers
2
3 from activemembers.api.v2.serializers.member_group_membership import (
4 MemberGroupMembershipSerializer,
5 )
6 from activemembers.models import MemberGroup
7 from thaliawebsite.api.v2.serializers import ThumbnailSerializer
8 from thaliawebsite.api.v2.serializers.cleaned_model_serializer import (
9 CleanedModelSerializer,
10 )
11
12
13 class MemberGroupSerializer(CleanedModelSerializer):
14 """API serializer for member groups."""
15
16 def __init__(self, *args, **kwargs):
17 super().__init__(*args, **kwargs)
18
19 if "get_memberships" not in self.context and "members" in self.fields:
20 self.fields.pop("members")
21
22 class Meta:
23 """Meta class for the serializer."""
24
25 model = MemberGroup
26 fields = (
27 "pk",
28 "name",
29 "type",
30 "description",
31 "since",
32 "until",
33 "contact_address",
34 "photo",
35 "members",
36 )
37
38 members = serializers.SerializerMethodField("_members")
39 type = serializers.SerializerMethodField("_type")
40 photo = ThumbnailSerializer(placeholder="activemembers/images/placeholder.png")
41
42 def _members(self, instance):
43 memberships = self.context["get_memberships"](instance).prefetch_related(
44 "member__membergroupmembership_set"
45 )
46 return MemberGroupMembershipSerializer(
47 many=True, context=self.context
48 ).to_representation(memberships)
49
50 def _type(self, instance):
51 if hasattr(instance, "board"):
52 return "board"
53 if hasattr(instance, "committee"):
54 return "committee"
55 if hasattr(instance, "society"):
56 return "society"
57 return None
58
59
60 class MemberGroupListSerializer(MemberGroupSerializer):
61 class Meta:
62 """Meta class for the serializer."""
63
64 model = MemberGroup
65 fields = (
66 "pk",
67 "name",
68 "type",
69 "description",
70 "since",
71 "until",
72 "contact_address",
73 "photo",
74 )
75
```
Path: `website/events/api/v2/serializers/event.py`
Content:
```
1 from rest_framework import serializers
2 from rest_framework.reverse import reverse
3
4 from activemembers.api.v2.serializers.member_group import MemberGroupSerializer
5 from documents.api.v2.serializers.document import DocumentSerializer
6 from events import services
7 from events.api.v2.serializers.event_registration import EventRegistrationSerializer
8 from events.models import Event
9 from payments.api.v2.serializers.payment_amount import PaymentAmountSerializer
10 from thaliawebsite.api.v2.serializers import CleanedHTMLSerializer
11 from thaliawebsite.api.v2.serializers.cleaned_model_serializer import (
12 CleanedModelSerializer,
13 )
14 from utils.snippets import create_google_maps_url
15
16
17 class EventSerializer(CleanedModelSerializer):
18 """Serializer for events."""
19
20 class Meta:
21 model = Event
22 fields = (
23 "pk",
24 "slug",
25 "url",
26 "title",
27 "description",
28 "caption",
29 "start",
30 "end",
31 "category",
32 "registration_start",
33 "registration_end",
34 "cancel_deadline",
35 "optional_registrations",
36 "location",
37 "price",
38 "fine",
39 "num_participants",
40 "max_participants",
41 "no_registration_message",
42 "registration_status",
43 "cancel_too_late_message",
44 "has_fields",
45 "food_event",
46 "maps_url",
47 "user_permissions",
48 "user_registration",
49 "organisers",
50 "documents",
51 )
52
53 description = CleanedHTMLSerializer()
54 organisers = MemberGroupSerializer(many=True)
55 user_registration = serializers.SerializerMethodField("_user_registration")
56 num_participants = serializers.SerializerMethodField("_num_participants")
57 maps_url = serializers.SerializerMethodField("_maps_url")
58 registration_status = serializers.SerializerMethodField("_registration_status")
59 price = PaymentAmountSerializer()
60 fine = PaymentAmountSerializer()
61 documents = DocumentSerializer(many=True)
62 user_permissions = serializers.SerializerMethodField("_user_permissions")
63 url = serializers.SerializerMethodField("_url")
64
65 def _user_registration(self, instance: Event):
66 if self.context["request"].member and len(instance.member_registration) > 0:
67 registration = instance.member_registration[-1]
68 return EventRegistrationSerializer(
69 registration,
70 context=self.context,
71 fields=(
72 "pk",
73 "present",
74 "queue_position",
75 "is_cancelled",
76 "is_late_cancellation",
77 "date",
78 "payment",
79 ),
80 ).data
81 return None
82
83 def _registration_status(self, instance: Event):
84 if self.context["request"].member and len(instance.member_registration) > 0:
85 registration = instance.member_registration[-1]
86 else:
87 registration = None
88 status = services.registration_status(
89 instance, registration, self.context["request"].member
90 )
91 cancel_status = services.cancel_status(instance, registration)
92
93 status_str = services.registration_status_string(status, instance, registration)
94 cancel_str = services.cancel_info_string(instance, cancel_status, status)
95 if services.show_cancel_status(status) and cancel_str != "":
96 return f"{status_str} {cancel_str}"
97 return f"{status_str}"
98
99 def _num_participants(self, instance: Event):
100 if instance.max_participants:
101 return min(instance.participant_count, instance.max_participants)
102 return instance.participant_count
103
104 def _user_permissions(self, instance):
105 member = self.context["request"].member
106 return services.event_permissions(member, instance, registration_prefetch=True)
107
108 def _url(self, instance: Event):
109 if instance.slug is None:
110 return reverse(
111 "events:event",
112 kwargs={"pk": instance.pk},
113 request=self.context["request"],
114 )
115 return reverse(
116 "events:event",
117 kwargs={"slug": instance.slug},
118 request=self.context["request"],
119 )
120
121 def _maps_url(self, instance):
122 return create_google_maps_url(instance.map_location, zoom=13, size="450x250")
123
124
125 class EventListSerializer(EventSerializer):
126 class Meta:
127 model = Event
128 fields = (
129 "pk",
130 "slug",
131 "url",
132 "title",
133 "description",
134 "caption",
135 "start",
136 "end",
137 "category",
138 "registration_start",
139 "registration_end",
140 "cancel_deadline",
141 "optional_registrations",
142 "location",
143 "price",
144 "fine",
145 "num_participants",
146 "max_participants",
147 "no_registration_message",
148 "registration_status",
149 "cancel_too_late_message",
150 "has_fields",
151 "food_event",
152 "maps_url",
153 "user_permissions",
154 "user_registration",
155 "documents",
156 )
157
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/website/activemembers/api/v2/serializers/member_group.py b/website/activemembers/api/v2/serializers/member_group.py
--- a/website/activemembers/api/v2/serializers/member_group.py
+++ b/website/activemembers/api/v2/serializers/member_group.py
@@ -72,3 +72,16 @@
"contact_address",
"photo",
)
+
+
+class MemberGroupShortSerializer(MemberGroupListSerializer):
+ class Meta:
+ model = MemberGroup
+ fields = (
+ "pk",
+ "name",
+ "type",
+ "since",
+ "until",
+ "contact_address",
+ )
diff --git a/website/events/api/v2/serializers/event.py b/website/events/api/v2/serializers/event.py
--- a/website/events/api/v2/serializers/event.py
+++ b/website/events/api/v2/serializers/event.py
@@ -1,7 +1,10 @@
from rest_framework import serializers
from rest_framework.reverse import reverse
-from activemembers.api.v2.serializers.member_group import MemberGroupSerializer
+from activemembers.api.v2.serializers.member_group import (
+ MemberGroupSerializer,
+ MemberGroupShortSerializer,
+)
from documents.api.v2.serializers.document import DocumentSerializer
from events import services
from events.api.v2.serializers.event_registration import EventRegistrationSerializer
@@ -123,34 +126,4 @@
class EventListSerializer(EventSerializer):
- class Meta:
- model = Event
- fields = (
- "pk",
- "slug",
- "url",
- "title",
- "description",
- "caption",
- "start",
- "end",
- "category",
- "registration_start",
- "registration_end",
- "cancel_deadline",
- "optional_registrations",
- "location",
- "price",
- "fine",
- "num_participants",
- "max_participants",
- "no_registration_message",
- "registration_status",
- "cancel_too_late_message",
- "has_fields",
- "food_event",
- "maps_url",
- "user_permissions",
- "user_registration",
- "documents",
- )
+ organisers = MemberGroupShortSerializer(many=True)
|
{"golden_diff": "diff --git a/website/activemembers/api/v2/serializers/member_group.py b/website/activemembers/api/v2/serializers/member_group.py\n--- a/website/activemembers/api/v2/serializers/member_group.py\n+++ b/website/activemembers/api/v2/serializers/member_group.py\n@@ -72,3 +72,16 @@\n \"contact_address\",\n \"photo\",\n )\n+\n+\n+class MemberGroupShortSerializer(MemberGroupListSerializer):\n+ class Meta:\n+ model = MemberGroup\n+ fields = (\n+ \"pk\",\n+ \"name\",\n+ \"type\",\n+ \"since\",\n+ \"until\",\n+ \"contact_address\",\n+ )\ndiff --git a/website/events/api/v2/serializers/event.py b/website/events/api/v2/serializers/event.py\n--- a/website/events/api/v2/serializers/event.py\n+++ b/website/events/api/v2/serializers/event.py\n@@ -1,7 +1,10 @@\n from rest_framework import serializers\n from rest_framework.reverse import reverse\n \n-from activemembers.api.v2.serializers.member_group import MemberGroupSerializer\n+from activemembers.api.v2.serializers.member_group import (\n+ MemberGroupSerializer,\n+ MemberGroupShortSerializer,\n+)\n from documents.api.v2.serializers.document import DocumentSerializer\n from events import services\n from events.api.v2.serializers.event_registration import EventRegistrationSerializer\n@@ -123,34 +126,4 @@\n \n \n class EventListSerializer(EventSerializer):\n- class Meta:\n- model = Event\n- fields = (\n- \"pk\",\n- \"slug\",\n- \"url\",\n- \"title\",\n- \"description\",\n- \"caption\",\n- \"start\",\n- \"end\",\n- \"category\",\n- \"registration_start\",\n- \"registration_end\",\n- \"cancel_deadline\",\n- \"optional_registrations\",\n- \"location\",\n- \"price\",\n- \"fine\",\n- \"num_participants\",\n- \"max_participants\",\n- \"no_registration_message\",\n- \"registration_status\",\n- \"cancel_too_late_message\",\n- \"has_fields\",\n- \"food_event\",\n- \"maps_url\",\n- \"user_permissions\",\n- \"user_registration\",\n- \"documents\",\n- )\n+ organisers = MemberGroupShortSerializer(many=True)\n", "issue": "Re-add rudamentary version of organisers to the event list API endpoint\n<!--\r\n\r\nPlease add the appropriate label for what change should be made:\r\ndocs: changes to the documentation)\r\nrefactor: refactoring production code, eg. renaming a variable or rewriting a function\r\ntest: adding missing tests, refactoring tests; no production code change\r\nchore: updating poetry etc; no production code change\r\n\r\n-->\r\n\r\n### Describe the change\r\nAdd a id-name based list of event organisers to the event list API endpoint.\r\n\r\n### Motivation\r\nBecause it would make me and my Reaxit PR very happy.\r\n\r\n### Current implementation\r\nIt got removed :(\r\n\r\n### Suggested implementation\r\nJust do it Kappa\r\n\n", "before_files": [{"content": "from rest_framework import serializers\n\nfrom activemembers.api.v2.serializers.member_group_membership import (\n MemberGroupMembershipSerializer,\n)\nfrom activemembers.models import MemberGroup\nfrom thaliawebsite.api.v2.serializers import ThumbnailSerializer\nfrom thaliawebsite.api.v2.serializers.cleaned_model_serializer import (\n CleanedModelSerializer,\n)\n\n\nclass MemberGroupSerializer(CleanedModelSerializer):\n \"\"\"API serializer for member groups.\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n\n if \"get_memberships\" not in self.context and \"members\" in self.fields:\n self.fields.pop(\"members\")\n\n class Meta:\n \"\"\"Meta class for the serializer.\"\"\"\n\n model = MemberGroup\n fields = (\n \"pk\",\n \"name\",\n \"type\",\n \"description\",\n \"since\",\n \"until\",\n \"contact_address\",\n \"photo\",\n \"members\",\n )\n\n members = serializers.SerializerMethodField(\"_members\")\n type = serializers.SerializerMethodField(\"_type\")\n photo = ThumbnailSerializer(placeholder=\"activemembers/images/placeholder.png\")\n\n def _members(self, instance):\n memberships = self.context[\"get_memberships\"](instance).prefetch_related(\n \"member__membergroupmembership_set\"\n )\n return MemberGroupMembershipSerializer(\n many=True, context=self.context\n ).to_representation(memberships)\n\n def _type(self, instance):\n if hasattr(instance, \"board\"):\n return \"board\"\n if hasattr(instance, \"committee\"):\n return \"committee\"\n if hasattr(instance, \"society\"):\n return \"society\"\n return None\n\n\nclass MemberGroupListSerializer(MemberGroupSerializer):\n class Meta:\n \"\"\"Meta class for the serializer.\"\"\"\n\n model = MemberGroup\n fields = (\n \"pk\",\n \"name\",\n \"type\",\n \"description\",\n \"since\",\n \"until\",\n \"contact_address\",\n \"photo\",\n )\n", "path": "website/activemembers/api/v2/serializers/member_group.py"}, {"content": "from rest_framework import serializers\nfrom rest_framework.reverse import reverse\n\nfrom activemembers.api.v2.serializers.member_group import MemberGroupSerializer\nfrom documents.api.v2.serializers.document import DocumentSerializer\nfrom events import services\nfrom events.api.v2.serializers.event_registration import EventRegistrationSerializer\nfrom events.models import Event\nfrom payments.api.v2.serializers.payment_amount import PaymentAmountSerializer\nfrom thaliawebsite.api.v2.serializers import CleanedHTMLSerializer\nfrom thaliawebsite.api.v2.serializers.cleaned_model_serializer import (\n CleanedModelSerializer,\n)\nfrom utils.snippets import create_google_maps_url\n\n\nclass EventSerializer(CleanedModelSerializer):\n \"\"\"Serializer for events.\"\"\"\n\n class Meta:\n model = Event\n fields = (\n \"pk\",\n \"slug\",\n \"url\",\n \"title\",\n \"description\",\n \"caption\",\n \"start\",\n \"end\",\n \"category\",\n \"registration_start\",\n \"registration_end\",\n \"cancel_deadline\",\n \"optional_registrations\",\n \"location\",\n \"price\",\n \"fine\",\n \"num_participants\",\n \"max_participants\",\n \"no_registration_message\",\n \"registration_status\",\n \"cancel_too_late_message\",\n \"has_fields\",\n \"food_event\",\n \"maps_url\",\n \"user_permissions\",\n \"user_registration\",\n \"organisers\",\n \"documents\",\n )\n\n description = CleanedHTMLSerializer()\n organisers = MemberGroupSerializer(many=True)\n user_registration = serializers.SerializerMethodField(\"_user_registration\")\n num_participants = serializers.SerializerMethodField(\"_num_participants\")\n maps_url = serializers.SerializerMethodField(\"_maps_url\")\n registration_status = serializers.SerializerMethodField(\"_registration_status\")\n price = PaymentAmountSerializer()\n fine = PaymentAmountSerializer()\n documents = DocumentSerializer(many=True)\n user_permissions = serializers.SerializerMethodField(\"_user_permissions\")\n url = serializers.SerializerMethodField(\"_url\")\n\n def _user_registration(self, instance: Event):\n if self.context[\"request\"].member and len(instance.member_registration) > 0:\n registration = instance.member_registration[-1]\n return EventRegistrationSerializer(\n registration,\n context=self.context,\n fields=(\n \"pk\",\n \"present\",\n \"queue_position\",\n \"is_cancelled\",\n \"is_late_cancellation\",\n \"date\",\n \"payment\",\n ),\n ).data\n return None\n\n def _registration_status(self, instance: Event):\n if self.context[\"request\"].member and len(instance.member_registration) > 0:\n registration = instance.member_registration[-1]\n else:\n registration = None\n status = services.registration_status(\n instance, registration, self.context[\"request\"].member\n )\n cancel_status = services.cancel_status(instance, registration)\n\n status_str = services.registration_status_string(status, instance, registration)\n cancel_str = services.cancel_info_string(instance, cancel_status, status)\n if services.show_cancel_status(status) and cancel_str != \"\":\n return f\"{status_str} {cancel_str}\"\n return f\"{status_str}\"\n\n def _num_participants(self, instance: Event):\n if instance.max_participants:\n return min(instance.participant_count, instance.max_participants)\n return instance.participant_count\n\n def _user_permissions(self, instance):\n member = self.context[\"request\"].member\n return services.event_permissions(member, instance, registration_prefetch=True)\n\n def _url(self, instance: Event):\n if instance.slug is None:\n return reverse(\n \"events:event\",\n kwargs={\"pk\": instance.pk},\n request=self.context[\"request\"],\n )\n return reverse(\n \"events:event\",\n kwargs={\"slug\": instance.slug},\n request=self.context[\"request\"],\n )\n\n def _maps_url(self, instance):\n return create_google_maps_url(instance.map_location, zoom=13, size=\"450x250\")\n\n\nclass EventListSerializer(EventSerializer):\n class Meta:\n model = Event\n fields = (\n \"pk\",\n \"slug\",\n \"url\",\n \"title\",\n \"description\",\n \"caption\",\n \"start\",\n \"end\",\n \"category\",\n \"registration_start\",\n \"registration_end\",\n \"cancel_deadline\",\n \"optional_registrations\",\n \"location\",\n \"price\",\n \"fine\",\n \"num_participants\",\n \"max_participants\",\n \"no_registration_message\",\n \"registration_status\",\n \"cancel_too_late_message\",\n \"has_fields\",\n \"food_event\",\n \"maps_url\",\n \"user_permissions\",\n \"user_registration\",\n \"documents\",\n )\n", "path": "website/events/api/v2/serializers/event.py"}], "after_files": [{"content": "from rest_framework import serializers\n\nfrom activemembers.api.v2.serializers.member_group_membership import (\n MemberGroupMembershipSerializer,\n)\nfrom activemembers.models import MemberGroup\nfrom thaliawebsite.api.v2.serializers import ThumbnailSerializer\nfrom thaliawebsite.api.v2.serializers.cleaned_model_serializer import (\n CleanedModelSerializer,\n)\n\n\nclass MemberGroupSerializer(CleanedModelSerializer):\n \"\"\"API serializer for member groups.\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n\n if \"get_memberships\" not in self.context and \"members\" in self.fields:\n self.fields.pop(\"members\")\n\n class Meta:\n \"\"\"Meta class for the serializer.\"\"\"\n\n model = MemberGroup\n fields = (\n \"pk\",\n \"name\",\n \"type\",\n \"description\",\n \"since\",\n \"until\",\n \"contact_address\",\n \"photo\",\n \"members\",\n )\n\n members = serializers.SerializerMethodField(\"_members\")\n type = serializers.SerializerMethodField(\"_type\")\n photo = ThumbnailSerializer(placeholder=\"activemembers/images/placeholder.png\")\n\n def _members(self, instance):\n memberships = self.context[\"get_memberships\"](instance).prefetch_related(\n \"member__membergroupmembership_set\"\n )\n return MemberGroupMembershipSerializer(\n many=True, context=self.context\n ).to_representation(memberships)\n\n def _type(self, instance):\n if hasattr(instance, \"board\"):\n return \"board\"\n if hasattr(instance, \"committee\"):\n return \"committee\"\n if hasattr(instance, \"society\"):\n return \"society\"\n return None\n\n\nclass MemberGroupListSerializer(MemberGroupSerializer):\n class Meta:\n \"\"\"Meta class for the serializer.\"\"\"\n\n model = MemberGroup\n fields = (\n \"pk\",\n \"name\",\n \"type\",\n \"description\",\n \"since\",\n \"until\",\n \"contact_address\",\n \"photo\",\n )\n\n\nclass MemberGroupShortSerializer(MemberGroupListSerializer):\n class Meta:\n model = MemberGroup\n fields = (\n \"pk\",\n \"name\",\n \"type\",\n \"since\",\n \"until\",\n \"contact_address\",\n )\n", "path": "website/activemembers/api/v2/serializers/member_group.py"}, {"content": "from rest_framework import serializers\nfrom rest_framework.reverse import reverse\n\nfrom activemembers.api.v2.serializers.member_group import (\n MemberGroupSerializer,\n MemberGroupShortSerializer,\n)\nfrom documents.api.v2.serializers.document import DocumentSerializer\nfrom events import services\nfrom events.api.v2.serializers.event_registration import EventRegistrationSerializer\nfrom events.models import Event\nfrom payments.api.v2.serializers.payment_amount import PaymentAmountSerializer\nfrom thaliawebsite.api.v2.serializers import CleanedHTMLSerializer\nfrom thaliawebsite.api.v2.serializers.cleaned_model_serializer import (\n CleanedModelSerializer,\n)\nfrom utils.snippets import create_google_maps_url\n\n\nclass EventSerializer(CleanedModelSerializer):\n \"\"\"Serializer for events.\"\"\"\n\n class Meta:\n model = Event\n fields = (\n \"pk\",\n \"slug\",\n \"url\",\n \"title\",\n \"description\",\n \"caption\",\n \"start\",\n \"end\",\n \"category\",\n \"registration_start\",\n \"registration_end\",\n \"cancel_deadline\",\n \"optional_registrations\",\n \"location\",\n \"price\",\n \"fine\",\n \"num_participants\",\n \"max_participants\",\n \"no_registration_message\",\n \"registration_status\",\n \"cancel_too_late_message\",\n \"has_fields\",\n \"food_event\",\n \"maps_url\",\n \"user_permissions\",\n \"user_registration\",\n \"organisers\",\n \"documents\",\n )\n\n description = CleanedHTMLSerializer()\n organisers = MemberGroupSerializer(many=True)\n user_registration = serializers.SerializerMethodField(\"_user_registration\")\n num_participants = serializers.SerializerMethodField(\"_num_participants\")\n maps_url = serializers.SerializerMethodField(\"_maps_url\")\n registration_status = serializers.SerializerMethodField(\"_registration_status\")\n price = PaymentAmountSerializer()\n fine = PaymentAmountSerializer()\n documents = DocumentSerializer(many=True)\n user_permissions = serializers.SerializerMethodField(\"_user_permissions\")\n url = serializers.SerializerMethodField(\"_url\")\n\n def _user_registration(self, instance: Event):\n if self.context[\"request\"].member and len(instance.member_registration) > 0:\n registration = instance.member_registration[-1]\n return EventRegistrationSerializer(\n registration,\n context=self.context,\n fields=(\n \"pk\",\n \"present\",\n \"queue_position\",\n \"is_cancelled\",\n \"is_late_cancellation\",\n \"date\",\n \"payment\",\n ),\n ).data\n return None\n\n def _registration_status(self, instance: Event):\n if self.context[\"request\"].member and len(instance.member_registration) > 0:\n registration = instance.member_registration[-1]\n else:\n registration = None\n status = services.registration_status(\n instance, registration, self.context[\"request\"].member\n )\n cancel_status = services.cancel_status(instance, registration)\n\n status_str = services.registration_status_string(status, instance, registration)\n cancel_str = services.cancel_info_string(instance, cancel_status, status)\n if services.show_cancel_status(status) and cancel_str != \"\":\n return f\"{status_str} {cancel_str}\"\n return f\"{status_str}\"\n\n def _num_participants(self, instance: Event):\n if instance.max_participants:\n return min(instance.participant_count, instance.max_participants)\n return instance.participant_count\n\n def _user_permissions(self, instance):\n member = self.context[\"request\"].member\n return services.event_permissions(member, instance, registration_prefetch=True)\n\n def _url(self, instance: Event):\n if instance.slug is None:\n return reverse(\n \"events:event\",\n kwargs={\"pk\": instance.pk},\n request=self.context[\"request\"],\n )\n return reverse(\n \"events:event\",\n kwargs={\"slug\": instance.slug},\n request=self.context[\"request\"],\n )\n\n def _maps_url(self, instance):\n return create_google_maps_url(instance.map_location, zoom=13, size=\"450x250\")\n\n\nclass EventListSerializer(EventSerializer):\n organisers = MemberGroupShortSerializer(many=True)\n", "path": "website/events/api/v2/serializers/event.py"}]}
| 2,351 | 529 |
gh_patches_debug_28357
|
rasdani/github-patches
|
git_diff
|
pyca__cryptography-4200
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Some OpenSSL bignum arithmetic operations are not in constant time
### What's wrong:
1. Some arithmetic operations exposed in the [OpenSSL bignum binding](https://github.com/pyca/cryptography/blob/master/src/_cffi_src/openssl/bignum.py) are not in constant time.
2. These functions are specifically:
1. `BN_div`
2. `BN_mod_inverse`
3. `BN_mod_exp`
### How to resolve:
1. The OpenSSL solution to this is to call `BN_set_flags` on the secret BIGNUM and set the `BN_FLG_CONSTTIME` flag. [The OpenSSL docs reference this here.](https://github.com/openssl/openssl/blob/master/include/openssl/bn.h#L61)
2. Expose the needed operations:
1. `BN_MONT_CTX` operations.
2. `BN_mod_exp_mont`, and `BN_mod_exp_mont_consttime`
3. `BN_set_flags`, `BN_get_flags`, and the `BN_FLG_CONSTTIME` flag.
### Notes:
1. Cryptography.io isn't affected by these timing issues. I have clarified with @alex and @reaperhulk.
2. If you are a downstream user utilizing these operations on private/secret values, then _you are affected_.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/_cffi_src/openssl/bignum.py`
Content:
```
1 # This file is dual licensed under the terms of the Apache License, Version
2 # 2.0, and the BSD License. See the LICENSE file in the root of this repository
3 # for complete details.
4
5 from __future__ import absolute_import, division, print_function
6
7 INCLUDES = """
8 #include <openssl/bn.h>
9 """
10
11 TYPES = """
12 typedef ... BN_CTX;
13 typedef ... BIGNUM;
14 typedef int... BN_ULONG;
15 """
16
17 FUNCTIONS = """
18 BIGNUM *BN_new(void);
19 void BN_free(BIGNUM *);
20 void BN_clear_free(BIGNUM *);
21
22 int BN_rand(BIGNUM *, int, int, int);
23 int BN_rand_range(BIGNUM *, BIGNUM *);
24
25 BN_CTX *BN_CTX_new(void);
26 void BN_CTX_free(BN_CTX *);
27
28 void BN_CTX_start(BN_CTX *);
29 BIGNUM *BN_CTX_get(BN_CTX *);
30 void BN_CTX_end(BN_CTX *);
31
32 BIGNUM *BN_copy(BIGNUM *, const BIGNUM *);
33 BIGNUM *BN_dup(const BIGNUM *);
34
35 int BN_set_word(BIGNUM *, BN_ULONG);
36 BN_ULONG BN_get_word(const BIGNUM *);
37
38 const BIGNUM *BN_value_one(void);
39
40 char *BN_bn2hex(const BIGNUM *);
41 int BN_hex2bn(BIGNUM **, const char *);
42 int BN_dec2bn(BIGNUM **, const char *);
43
44 int BN_bn2bin(const BIGNUM *, unsigned char *);
45 BIGNUM *BN_bin2bn(const unsigned char *, int, BIGNUM *);
46
47 int BN_num_bits(const BIGNUM *);
48
49 int BN_cmp(const BIGNUM *, const BIGNUM *);
50 int BN_add(BIGNUM *, const BIGNUM *, const BIGNUM *);
51 int BN_sub(BIGNUM *, const BIGNUM *, const BIGNUM *);
52 int BN_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
53 int BN_sqr(BIGNUM *, const BIGNUM *, BN_CTX *);
54 int BN_div(BIGNUM *, BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
55 int BN_nnmod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
56 int BN_mod_add(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
57 BN_CTX *);
58 int BN_mod_sub(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
59 BN_CTX *);
60 int BN_mod_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
61 BN_CTX *);
62 int BN_mod_sqr(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
63 int BN_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
64 int BN_mod_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
65 BN_CTX *);
66 int BN_gcd(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
67 BIGNUM *BN_mod_inverse(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
68
69 int BN_set_bit(BIGNUM *, int);
70 int BN_clear_bit(BIGNUM *, int);
71
72 int BN_is_bit_set(const BIGNUM *, int);
73
74 int BN_mask_bits(BIGNUM *, int);
75
76 int BN_num_bytes(const BIGNUM *);
77
78 int BN_zero(BIGNUM *);
79 int BN_one(BIGNUM *);
80 int BN_mod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
81
82 int BN_lshift(BIGNUM *, const BIGNUM *, int);
83 int BN_lshift1(BIGNUM *, BIGNUM *);
84
85 int BN_rshift(BIGNUM *, BIGNUM *, int);
86 int BN_rshift1(BIGNUM *, BIGNUM *);
87 """
88
89 CUSTOMIZATIONS = """
90 """
91
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/_cffi_src/openssl/bignum.py b/src/_cffi_src/openssl/bignum.py
--- a/src/_cffi_src/openssl/bignum.py
+++ b/src/_cffi_src/openssl/bignum.py
@@ -10,11 +10,17 @@
TYPES = """
typedef ... BN_CTX;
+typedef ... BN_MONT_CTX;
typedef ... BIGNUM;
typedef int... BN_ULONG;
"""
FUNCTIONS = """
+#define BN_FLG_CONSTTIME ...
+
+void BN_set_flags(BIGNUM *, int);
+int BN_get_flags(const BIGNUM *, int);
+
BIGNUM *BN_new(void);
void BN_free(BIGNUM *);
void BN_clear_free(BIGNUM *);
@@ -29,6 +35,10 @@
BIGNUM *BN_CTX_get(BN_CTX *);
void BN_CTX_end(BN_CTX *);
+BN_MONT_CTX *BN_MONT_CTX_new(void);
+int BN_MONT_CTX_set(BN_MONT_CTX *, BIGNUM *, BN_CTX *);
+void BN_MONT_CTX_free(BN_MONT_CTX *);
+
BIGNUM *BN_copy(BIGNUM *, const BIGNUM *);
BIGNUM *BN_dup(const BIGNUM *);
@@ -63,6 +73,10 @@
int BN_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
int BN_mod_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
BN_CTX *);
+int BN_mod_exp_mont(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,
+ BN_CTX *, BN_MONT_CTX *);
+int BN_mod_exp_mont_consttime(BIGNUM *, const BIGNUM *, const BIGNUM *,
+ const BIGNUM *, BN_CTX *, BN_MONT_CTX *);
int BN_gcd(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
BIGNUM *BN_mod_inverse(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);
|
{"golden_diff": "diff --git a/src/_cffi_src/openssl/bignum.py b/src/_cffi_src/openssl/bignum.py\n--- a/src/_cffi_src/openssl/bignum.py\n+++ b/src/_cffi_src/openssl/bignum.py\n@@ -10,11 +10,17 @@\n \n TYPES = \"\"\"\n typedef ... BN_CTX;\n+typedef ... BN_MONT_CTX;\n typedef ... BIGNUM;\n typedef int... BN_ULONG;\n \"\"\"\n \n FUNCTIONS = \"\"\"\n+#define BN_FLG_CONSTTIME ...\n+\n+void BN_set_flags(BIGNUM *, int);\n+int BN_get_flags(const BIGNUM *, int);\n+\n BIGNUM *BN_new(void);\n void BN_free(BIGNUM *);\n void BN_clear_free(BIGNUM *);\n@@ -29,6 +35,10 @@\n BIGNUM *BN_CTX_get(BN_CTX *);\n void BN_CTX_end(BN_CTX *);\n \n+BN_MONT_CTX *BN_MONT_CTX_new(void);\n+int BN_MONT_CTX_set(BN_MONT_CTX *, BIGNUM *, BN_CTX *);\n+void BN_MONT_CTX_free(BN_MONT_CTX *);\n+\n BIGNUM *BN_copy(BIGNUM *, const BIGNUM *);\n BIGNUM *BN_dup(const BIGNUM *);\n \n@@ -63,6 +73,10 @@\n int BN_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n int BN_mod_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\n+int BN_mod_exp_mont(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n+ BN_CTX *, BN_MONT_CTX *);\n+int BN_mod_exp_mont_consttime(BIGNUM *, const BIGNUM *, const BIGNUM *,\n+ const BIGNUM *, BN_CTX *, BN_MONT_CTX *);\n int BN_gcd(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n BIGNUM *BN_mod_inverse(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n", "issue": "Some OpenSSL bignum arithmetic operations are not in constant time\n### What's wrong:\r\n1. Some arithmetic operations exposed in the [OpenSSL bignum binding](https://github.com/pyca/cryptography/blob/master/src/_cffi_src/openssl/bignum.py) are not in constant time.\r\n2. These functions are specifically:\r\n 1. `BN_div`\r\n 2. `BN_mod_inverse`\r\n 3. `BN_mod_exp`\r\n\r\n### How to resolve:\r\n1. The OpenSSL solution to this is to call `BN_set_flags` on the secret BIGNUM and set the `BN_FLG_CONSTTIME` flag. [The OpenSSL docs reference this here.](https://github.com/openssl/openssl/blob/master/include/openssl/bn.h#L61)\r\n2. Expose the needed operations:\r\n 1. `BN_MONT_CTX` operations.\r\n 2. `BN_mod_exp_mont`, and `BN_mod_exp_mont_consttime` \r\n 3. `BN_set_flags`, `BN_get_flags`, and the `BN_FLG_CONSTTIME` flag.\r\n\r\n### Notes:\r\n1. Cryptography.io isn't affected by these timing issues. I have clarified with @alex and @reaperhulk.\r\n2. If you are a downstream user utilizing these operations on private/secret values, then _you are affected_.\n", "before_files": [{"content": "# This file is dual licensed under the terms of the Apache License, Version\n# 2.0, and the BSD License. See the LICENSE file in the root of this repository\n# for complete details.\n\nfrom __future__ import absolute_import, division, print_function\n\nINCLUDES = \"\"\"\n#include <openssl/bn.h>\n\"\"\"\n\nTYPES = \"\"\"\ntypedef ... BN_CTX;\ntypedef ... BIGNUM;\ntypedef int... BN_ULONG;\n\"\"\"\n\nFUNCTIONS = \"\"\"\nBIGNUM *BN_new(void);\nvoid BN_free(BIGNUM *);\nvoid BN_clear_free(BIGNUM *);\n\nint BN_rand(BIGNUM *, int, int, int);\nint BN_rand_range(BIGNUM *, BIGNUM *);\n\nBN_CTX *BN_CTX_new(void);\nvoid BN_CTX_free(BN_CTX *);\n\nvoid BN_CTX_start(BN_CTX *);\nBIGNUM *BN_CTX_get(BN_CTX *);\nvoid BN_CTX_end(BN_CTX *);\n\nBIGNUM *BN_copy(BIGNUM *, const BIGNUM *);\nBIGNUM *BN_dup(const BIGNUM *);\n\nint BN_set_word(BIGNUM *, BN_ULONG);\nBN_ULONG BN_get_word(const BIGNUM *);\n\nconst BIGNUM *BN_value_one(void);\n\nchar *BN_bn2hex(const BIGNUM *);\nint BN_hex2bn(BIGNUM **, const char *);\nint BN_dec2bn(BIGNUM **, const char *);\n\nint BN_bn2bin(const BIGNUM *, unsigned char *);\nBIGNUM *BN_bin2bn(const unsigned char *, int, BIGNUM *);\n\nint BN_num_bits(const BIGNUM *);\n\nint BN_cmp(const BIGNUM *, const BIGNUM *);\nint BN_add(BIGNUM *, const BIGNUM *, const BIGNUM *);\nint BN_sub(BIGNUM *, const BIGNUM *, const BIGNUM *);\nint BN_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_sqr(BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_div(BIGNUM *, BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_nnmod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_mod_add(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_sub(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_sqr(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_mod_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_gcd(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nBIGNUM *BN_mod_inverse(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n\nint BN_set_bit(BIGNUM *, int);\nint BN_clear_bit(BIGNUM *, int);\n\nint BN_is_bit_set(const BIGNUM *, int);\n\nint BN_mask_bits(BIGNUM *, int);\n\nint BN_num_bytes(const BIGNUM *);\n\nint BN_zero(BIGNUM *);\nint BN_one(BIGNUM *);\nint BN_mod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n\nint BN_lshift(BIGNUM *, const BIGNUM *, int);\nint BN_lshift1(BIGNUM *, BIGNUM *);\n\nint BN_rshift(BIGNUM *, BIGNUM *, int);\nint BN_rshift1(BIGNUM *, BIGNUM *);\n\"\"\"\n\nCUSTOMIZATIONS = \"\"\"\n\"\"\"\n", "path": "src/_cffi_src/openssl/bignum.py"}], "after_files": [{"content": "# This file is dual licensed under the terms of the Apache License, Version\n# 2.0, and the BSD License. See the LICENSE file in the root of this repository\n# for complete details.\n\nfrom __future__ import absolute_import, division, print_function\n\nINCLUDES = \"\"\"\n#include <openssl/bn.h>\n\"\"\"\n\nTYPES = \"\"\"\ntypedef ... BN_CTX;\ntypedef ... BN_MONT_CTX;\ntypedef ... BIGNUM;\ntypedef int... BN_ULONG;\n\"\"\"\n\nFUNCTIONS = \"\"\"\n#define BN_FLG_CONSTTIME ...\n\nvoid BN_set_flags(BIGNUM *, int);\nint BN_get_flags(const BIGNUM *, int);\n\nBIGNUM *BN_new(void);\nvoid BN_free(BIGNUM *);\nvoid BN_clear_free(BIGNUM *);\n\nint BN_rand(BIGNUM *, int, int, int);\nint BN_rand_range(BIGNUM *, BIGNUM *);\n\nBN_CTX *BN_CTX_new(void);\nvoid BN_CTX_free(BN_CTX *);\n\nvoid BN_CTX_start(BN_CTX *);\nBIGNUM *BN_CTX_get(BN_CTX *);\nvoid BN_CTX_end(BN_CTX *);\n\nBN_MONT_CTX *BN_MONT_CTX_new(void);\nint BN_MONT_CTX_set(BN_MONT_CTX *, BIGNUM *, BN_CTX *);\nvoid BN_MONT_CTX_free(BN_MONT_CTX *);\n\nBIGNUM *BN_copy(BIGNUM *, const BIGNUM *);\nBIGNUM *BN_dup(const BIGNUM *);\n\nint BN_set_word(BIGNUM *, BN_ULONG);\nBN_ULONG BN_get_word(const BIGNUM *);\n\nconst BIGNUM *BN_value_one(void);\n\nchar *BN_bn2hex(const BIGNUM *);\nint BN_hex2bn(BIGNUM **, const char *);\nint BN_dec2bn(BIGNUM **, const char *);\n\nint BN_bn2bin(const BIGNUM *, unsigned char *);\nBIGNUM *BN_bin2bn(const unsigned char *, int, BIGNUM *);\n\nint BN_num_bits(const BIGNUM *);\n\nint BN_cmp(const BIGNUM *, const BIGNUM *);\nint BN_add(BIGNUM *, const BIGNUM *, const BIGNUM *);\nint BN_sub(BIGNUM *, const BIGNUM *, const BIGNUM *);\nint BN_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_sqr(BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_div(BIGNUM *, BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_nnmod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_mod_add(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_sub(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_mul(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_sqr(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nint BN_mod_exp(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *);\nint BN_mod_exp_mont(BIGNUM *, const BIGNUM *, const BIGNUM *, const BIGNUM *,\n BN_CTX *, BN_MONT_CTX *);\nint BN_mod_exp_mont_consttime(BIGNUM *, const BIGNUM *, const BIGNUM *,\n const BIGNUM *, BN_CTX *, BN_MONT_CTX *);\nint BN_gcd(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\nBIGNUM *BN_mod_inverse(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n\nint BN_set_bit(BIGNUM *, int);\nint BN_clear_bit(BIGNUM *, int);\n\nint BN_is_bit_set(const BIGNUM *, int);\n\nint BN_mask_bits(BIGNUM *, int);\n\nint BN_num_bytes(const BIGNUM *);\n\nint BN_zero(BIGNUM *);\nint BN_one(BIGNUM *);\nint BN_mod(BIGNUM *, const BIGNUM *, const BIGNUM *, BN_CTX *);\n\nint BN_lshift(BIGNUM *, const BIGNUM *, int);\nint BN_lshift1(BIGNUM *, BIGNUM *);\n\nint BN_rshift(BIGNUM *, BIGNUM *, int);\nint BN_rshift1(BIGNUM *, BIGNUM *);\n\"\"\"\n\nCUSTOMIZATIONS = \"\"\"\n\"\"\"\n", "path": "src/_cffi_src/openssl/bignum.py"}]}
| 1,531 | 443 |
gh_patches_debug_14895
|
rasdani/github-patches
|
git_diff
|
arviz-devs__arviz-619
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
densityplot warning tight_layout
plot density returns the following warning:
_UserWarning: This figure was using constrained_layout==True, but that is incompatible with subplots_adjust and or tight_layout: setting constrained_layout==False.
warnings.warn("This figure was using constrained_layout==True, ")_
The solution is to do not use tight_layout()
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `arviz/plots/densityplot.py`
Content:
```
1 """KDE and histogram plots for multiple variables."""
2 import numpy as np
3
4 from ..data import convert_to_dataset
5 from ..stats import hpd
6 from .kdeplot import _fast_kde
7 from .plot_utils import (
8 _scale_fig_size,
9 make_label,
10 xarray_var_iter,
11 default_grid,
12 _create_axes_grid,
13 )
14 from ..utils import _var_names
15
16
17 # pylint:disable-msg=too-many-function-args
18 def plot_density(
19 data,
20 group="posterior",
21 data_labels=None,
22 var_names=None,
23 credible_interval=0.94,
24 point_estimate="mean",
25 colors="cycle",
26 outline=True,
27 hpd_markers="",
28 shade=0.0,
29 bw=4.5,
30 figsize=None,
31 textsize=None,
32 ):
33 """Generate KDE plots for continuous variables and histograms for discrete ones.
34
35 Plots are truncated at their 100*(1-alpha)% credible intervals. Plots are grouped per variable
36 and colors assigned to models.
37
38 Parameters
39 ----------
40 data : Union[Object, Iterator[Object]]
41 Any object that can be converted to an az.InferenceData object, or an Iterator returning
42 a sequence of such objects.
43 Refer to documentation of az.convert_to_dataset for details about such objects.
44 group: Optional[str]
45 Specifies which InferenceData group should be plotted. Defaults to 'posterior'.
46 Alternative values include 'prior' and any other strings used as dataset keys in the
47 InferenceData.
48 data_labels : Optional[List[str]]
49 List with names for the datasets passed as "data." Useful when plotting more than one
50 dataset. Must be the same shape as the data parameter. Defaults to None.
51 var_names: Optional[List[str]]
52 List of variables to plot. If multiple datasets are supplied and var_names is not None,
53 will print the same set of variables for each dataset. Defaults to None, which results in
54 all the variables being plotted.
55 credible_interval : float
56 Credible intervals. Should be in the interval (0, 1]. Defaults to 0.94.
57 point_estimate : Optional[str]
58 Plot point estimate per variable. Values should be 'mean', 'median' or None.
59 Defaults to 'mean'.
60 colors : Optional[Union[List[str],str]]
61 List with valid matplotlib colors, one color per model. Alternative a string can be passed.
62 If the string is `cycle`, it will automatically choose a color per model from matplolib's
63 cycle. If a single color is passed, e.g. 'k', 'C2' or 'red' this color will be used for all
64 models. Defaults to `cycle`.
65 outline : bool
66 Use a line to draw KDEs and histograms. Default to True
67 hpd_markers : str
68 A valid `matplotlib.markers` like 'v', used to indicate the limits of the hpd interval.
69 Defaults to empty string (no marker).
70 shade : Optional[float]
71 Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1
72 (opaque). Defaults to 0.
73 bw : Optional[float]
74 Bandwidth scaling factor for the KDE. Should be larger than 0. The higher this number the
75 smoother the KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule
76 of thumb (the default rule used by SciPy).
77 figsize : Optional[Tuple[int, int]]
78 Figure size. If None it will be defined automatically.
79 textsize: Optional[float]
80 Text size scaling factor for labels, titles and lines. If None it will be autoscaled based
81 on figsize.
82
83 Returns
84 -------
85 ax : Matplotlib axes
86
87
88 Examples
89 --------
90 Plot default density plot
91
92 .. plot::
93 :context: close-figs
94
95 >>> import arviz as az
96 >>> centered = az.load_arviz_data('centered_eight')
97 >>> non_centered = az.load_arviz_data('non_centered_eight')
98 >>> az.plot_density([centered, non_centered])
99
100 Plot subset variables by specifying variable name exactly
101
102 .. plot::
103 :context: close-figs
104
105 >>> az.plot_density([centered, non_centered], var_names=["mu"])
106
107 Plot a specific `az.InferenceData` group
108
109 .. plot::
110 :context: close-figs
111
112 >>> az.plot_density([centered, non_centered], var_names=["mu"], group="prior")
113
114 Specify credible interval
115
116 .. plot::
117 :context: close-figs
118
119 >>> az.plot_density([centered, non_centered], var_names=["mu"], credible_interval=.5)
120
121 Shade plots and/or remove outlines
122
123 .. plot::
124 :context: close-figs
125
126 >>> az.plot_density([centered, non_centered], var_names=["mu"], outline=False, shade=.8)
127
128 Specify binwidth for kernel density estimation
129
130 .. plot::
131 :context: close-figs
132
133 >>> az.plot_density([centered, non_centered], var_names=["mu"], bw=.9)
134 """
135 if not isinstance(data, (list, tuple)):
136 datasets = [convert_to_dataset(data, group=group)]
137 else:
138 datasets = [convert_to_dataset(datum, group=group) for datum in data]
139
140 var_names = _var_names(var_names, datasets)
141
142 if point_estimate not in ("mean", "median", None):
143 raise ValueError(
144 "Point estimate should be 'mean'," "median' or None, not {}".format(point_estimate)
145 )
146
147 n_data = len(datasets)
148
149 if data_labels is None:
150 if n_data > 1:
151 data_labels = ["{}".format(idx) for idx in range(n_data)]
152 else:
153 data_labels = [""]
154 elif len(data_labels) != n_data:
155 raise ValueError(
156 "The number of names for the models ({}) "
157 "does not match the number of models ({})".format(len(data_labels), n_data)
158 )
159
160 if colors == "cycle":
161 colors = ["C{}".format(idx % 10) for idx in range(n_data)]
162 elif isinstance(colors, str):
163 colors = [colors for _ in range(n_data)]
164
165 if not 1 >= credible_interval > 0:
166 raise ValueError("The value of credible_interval should be in the interval (0, 1]")
167
168 to_plot = [list(xarray_var_iter(data, var_names, combined=True)) for data in datasets]
169 all_labels = []
170 length_plotters = []
171 for plotters in to_plot:
172 length_plotters.append(len(plotters))
173 for var_name, selection, _ in plotters:
174 label = make_label(var_name, selection)
175 if label not in all_labels:
176 all_labels.append(label)
177 length_plotters = max(length_plotters)
178 rows, cols = default_grid(length_plotters, max_cols=3)
179
180 (figsize, _, titlesize, xt_labelsize, linewidth, markersize) = _scale_fig_size(
181 figsize, textsize, rows, cols
182 )
183
184 fig, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)
185
186 axis_map = {label: ax_ for label, ax_ in zip(all_labels, ax.flatten())}
187 for m_idx, plotters in enumerate(to_plot):
188 for var_name, selection, values in plotters:
189 label = make_label(var_name, selection)
190 _d_helper(
191 values.flatten(),
192 label,
193 colors[m_idx],
194 bw,
195 titlesize,
196 xt_labelsize,
197 linewidth,
198 markersize,
199 credible_interval,
200 point_estimate,
201 hpd_markers,
202 outline,
203 shade,
204 axis_map[label],
205 )
206
207 if n_data > 1:
208 for m_idx, label in enumerate(data_labels):
209 ax[0].plot([], label=label, c=colors[m_idx], markersize=markersize)
210 ax[0].legend(fontsize=xt_labelsize)
211
212 fig.tight_layout()
213
214 return ax
215
216
217 def _d_helper(
218 vec,
219 vname,
220 color,
221 bw,
222 titlesize,
223 xt_labelsize,
224 linewidth,
225 markersize,
226 credible_interval,
227 point_estimate,
228 hpd_markers,
229 outline,
230 shade,
231 ax,
232 ):
233 """Plot an individual dimension.
234
235 Parameters
236 ----------
237 vec : array
238 1D array from trace
239 vname : str
240 variable name
241 color : str
242 matplotlib color
243 bw : float
244 Bandwidth scaling factor. Should be larger than 0. The higher this number the smoother the
245 KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule of thumb
246 (the default used rule by SciPy).
247 titlesize : float
248 font size for title
249 xt_labelsize : float
250 fontsize for xticks
251 linewidth : float
252 Thickness of lines
253 markersize : float
254 Size of markers
255 credible_interval : float
256 Credible intervals. Defaults to 0.94
257 point_estimate : str or None
258 'mean' or 'median'
259 shade : float
260 Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1
261 (opaque). Defaults to 0.
262 ax : matplotlib axes
263 """
264 if vec.dtype.kind == "f":
265 if credible_interval != 1:
266 hpd_ = hpd(vec, credible_interval)
267 new_vec = vec[(vec >= hpd_[0]) & (vec <= hpd_[1])]
268 else:
269 new_vec = vec
270
271 density, xmin, xmax = _fast_kde(new_vec, bw=bw)
272 density *= credible_interval
273 x = np.linspace(xmin, xmax, len(density))
274 ymin = density[0]
275 ymax = density[-1]
276
277 if outline:
278 ax.plot(x, density, color=color, lw=linewidth)
279 ax.plot([xmin, xmin], [-ymin / 100, ymin], color=color, ls="-", lw=linewidth)
280 ax.plot([xmax, xmax], [-ymax / 100, ymax], color=color, ls="-", lw=linewidth)
281
282 if shade:
283 ax.fill_between(x, density, color=color, alpha=shade)
284
285 else:
286 xmin, xmax = hpd(vec, credible_interval)
287 bins = range(xmin, xmax + 2)
288 if outline:
289 ax.hist(vec, bins=bins, color=color, histtype="step", align="left")
290 if shade:
291 ax.hist(vec, bins=bins, color=color, alpha=shade)
292
293 if hpd_markers:
294 ax.plot(xmin, 0, hpd_markers, color=color, markeredgecolor="k", markersize=markersize)
295 ax.plot(xmax, 0, hpd_markers, color=color, markeredgecolor="k", markersize=markersize)
296
297 if point_estimate is not None:
298 if point_estimate == "mean":
299 est = np.mean(vec)
300 elif point_estimate == "median":
301 est = np.median(vec)
302 ax.plot(est, 0, "o", color=color, markeredgecolor="k", markersize=markersize)
303
304 ax.set_yticks([])
305 ax.set_title(vname, fontsize=titlesize, wrap=True)
306 for pos in ["left", "right", "top"]:
307 ax.spines[pos].set_visible(False)
308 ax.tick_params(labelsize=xt_labelsize)
309
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/arviz/plots/densityplot.py b/arviz/plots/densityplot.py
--- a/arviz/plots/densityplot.py
+++ b/arviz/plots/densityplot.py
@@ -181,7 +181,7 @@
figsize, textsize, rows, cols
)
- fig, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)
+ _, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)
axis_map = {label: ax_ for label, ax_ in zip(all_labels, ax.flatten())}
for m_idx, plotters in enumerate(to_plot):
@@ -209,8 +209,6 @@
ax[0].plot([], label=label, c=colors[m_idx], markersize=markersize)
ax[0].legend(fontsize=xt_labelsize)
- fig.tight_layout()
-
return ax
|
{"golden_diff": "diff --git a/arviz/plots/densityplot.py b/arviz/plots/densityplot.py\n--- a/arviz/plots/densityplot.py\n+++ b/arviz/plots/densityplot.py\n@@ -181,7 +181,7 @@\n figsize, textsize, rows, cols\n )\n \n- fig, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)\n+ _, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)\n \n axis_map = {label: ax_ for label, ax_ in zip(all_labels, ax.flatten())}\n for m_idx, plotters in enumerate(to_plot):\n@@ -209,8 +209,6 @@\n ax[0].plot([], label=label, c=colors[m_idx], markersize=markersize)\n ax[0].legend(fontsize=xt_labelsize)\n \n- fig.tight_layout()\n-\n return ax\n", "issue": "densityplot warning tight_layout\nplot density returns the following warning:\r\n\r\n_UserWarning: This figure was using constrained_layout==True, but that is incompatible with subplots_adjust and or tight_layout: setting constrained_layout==False. \r\n warnings.warn(\"This figure was using constrained_layout==True, \")_\r\n\r\nThe solution is to do not use tight_layout()\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"KDE and histogram plots for multiple variables.\"\"\"\nimport numpy as np\n\nfrom ..data import convert_to_dataset\nfrom ..stats import hpd\nfrom .kdeplot import _fast_kde\nfrom .plot_utils import (\n _scale_fig_size,\n make_label,\n xarray_var_iter,\n default_grid,\n _create_axes_grid,\n)\nfrom ..utils import _var_names\n\n\n# pylint:disable-msg=too-many-function-args\ndef plot_density(\n data,\n group=\"posterior\",\n data_labels=None,\n var_names=None,\n credible_interval=0.94,\n point_estimate=\"mean\",\n colors=\"cycle\",\n outline=True,\n hpd_markers=\"\",\n shade=0.0,\n bw=4.5,\n figsize=None,\n textsize=None,\n):\n \"\"\"Generate KDE plots for continuous variables and histograms for discrete ones.\n\n Plots are truncated at their 100*(1-alpha)% credible intervals. Plots are grouped per variable\n and colors assigned to models.\n\n Parameters\n ----------\n data : Union[Object, Iterator[Object]]\n Any object that can be converted to an az.InferenceData object, or an Iterator returning\n a sequence of such objects.\n Refer to documentation of az.convert_to_dataset for details about such objects.\n group: Optional[str]\n Specifies which InferenceData group should be plotted. Defaults to 'posterior'.\n Alternative values include 'prior' and any other strings used as dataset keys in the\n InferenceData.\n data_labels : Optional[List[str]]\n List with names for the datasets passed as \"data.\" Useful when plotting more than one\n dataset. Must be the same shape as the data parameter. Defaults to None.\n var_names: Optional[List[str]]\n List of variables to plot. If multiple datasets are supplied and var_names is not None,\n will print the same set of variables for each dataset. Defaults to None, which results in\n all the variables being plotted.\n credible_interval : float\n Credible intervals. Should be in the interval (0, 1]. Defaults to 0.94.\n point_estimate : Optional[str]\n Plot point estimate per variable. Values should be 'mean', 'median' or None.\n Defaults to 'mean'.\n colors : Optional[Union[List[str],str]]\n List with valid matplotlib colors, one color per model. Alternative a string can be passed.\n If the string is `cycle`, it will automatically choose a color per model from matplolib's\n cycle. If a single color is passed, e.g. 'k', 'C2' or 'red' this color will be used for all\n models. Defaults to `cycle`.\n outline : bool\n Use a line to draw KDEs and histograms. Default to True\n hpd_markers : str\n A valid `matplotlib.markers` like 'v', used to indicate the limits of the hpd interval.\n Defaults to empty string (no marker).\n shade : Optional[float]\n Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1\n (opaque). Defaults to 0.\n bw : Optional[float]\n Bandwidth scaling factor for the KDE. Should be larger than 0. The higher this number the\n smoother the KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule\n of thumb (the default rule used by SciPy).\n figsize : Optional[Tuple[int, int]]\n Figure size. If None it will be defined automatically.\n textsize: Optional[float]\n Text size scaling factor for labels, titles and lines. If None it will be autoscaled based\n on figsize.\n\n Returns\n -------\n ax : Matplotlib axes\n\n\n Examples\n --------\n Plot default density plot\n\n .. plot::\n :context: close-figs\n\n >>> import arviz as az\n >>> centered = az.load_arviz_data('centered_eight')\n >>> non_centered = az.load_arviz_data('non_centered_eight')\n >>> az.plot_density([centered, non_centered])\n\n Plot subset variables by specifying variable name exactly\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"])\n\n Plot a specific `az.InferenceData` group\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], group=\"prior\")\n\n Specify credible interval\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], credible_interval=.5)\n\n Shade plots and/or remove outlines\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], outline=False, shade=.8)\n\n Specify binwidth for kernel density estimation\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], bw=.9)\n \"\"\"\n if not isinstance(data, (list, tuple)):\n datasets = [convert_to_dataset(data, group=group)]\n else:\n datasets = [convert_to_dataset(datum, group=group) for datum in data]\n\n var_names = _var_names(var_names, datasets)\n\n if point_estimate not in (\"mean\", \"median\", None):\n raise ValueError(\n \"Point estimate should be 'mean',\" \"median' or None, not {}\".format(point_estimate)\n )\n\n n_data = len(datasets)\n\n if data_labels is None:\n if n_data > 1:\n data_labels = [\"{}\".format(idx) for idx in range(n_data)]\n else:\n data_labels = [\"\"]\n elif len(data_labels) != n_data:\n raise ValueError(\n \"The number of names for the models ({}) \"\n \"does not match the number of models ({})\".format(len(data_labels), n_data)\n )\n\n if colors == \"cycle\":\n colors = [\"C{}\".format(idx % 10) for idx in range(n_data)]\n elif isinstance(colors, str):\n colors = [colors for _ in range(n_data)]\n\n if not 1 >= credible_interval > 0:\n raise ValueError(\"The value of credible_interval should be in the interval (0, 1]\")\n\n to_plot = [list(xarray_var_iter(data, var_names, combined=True)) for data in datasets]\n all_labels = []\n length_plotters = []\n for plotters in to_plot:\n length_plotters.append(len(plotters))\n for var_name, selection, _ in plotters:\n label = make_label(var_name, selection)\n if label not in all_labels:\n all_labels.append(label)\n length_plotters = max(length_plotters)\n rows, cols = default_grid(length_plotters, max_cols=3)\n\n (figsize, _, titlesize, xt_labelsize, linewidth, markersize) = _scale_fig_size(\n figsize, textsize, rows, cols\n )\n\n fig, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)\n\n axis_map = {label: ax_ for label, ax_ in zip(all_labels, ax.flatten())}\n for m_idx, plotters in enumerate(to_plot):\n for var_name, selection, values in plotters:\n label = make_label(var_name, selection)\n _d_helper(\n values.flatten(),\n label,\n colors[m_idx],\n bw,\n titlesize,\n xt_labelsize,\n linewidth,\n markersize,\n credible_interval,\n point_estimate,\n hpd_markers,\n outline,\n shade,\n axis_map[label],\n )\n\n if n_data > 1:\n for m_idx, label in enumerate(data_labels):\n ax[0].plot([], label=label, c=colors[m_idx], markersize=markersize)\n ax[0].legend(fontsize=xt_labelsize)\n\n fig.tight_layout()\n\n return ax\n\n\ndef _d_helper(\n vec,\n vname,\n color,\n bw,\n titlesize,\n xt_labelsize,\n linewidth,\n markersize,\n credible_interval,\n point_estimate,\n hpd_markers,\n outline,\n shade,\n ax,\n):\n \"\"\"Plot an individual dimension.\n\n Parameters\n ----------\n vec : array\n 1D array from trace\n vname : str\n variable name\n color : str\n matplotlib color\n bw : float\n Bandwidth scaling factor. Should be larger than 0. The higher this number the smoother the\n KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule of thumb\n (the default used rule by SciPy).\n titlesize : float\n font size for title\n xt_labelsize : float\n fontsize for xticks\n linewidth : float\n Thickness of lines\n markersize : float\n Size of markers\n credible_interval : float\n Credible intervals. Defaults to 0.94\n point_estimate : str or None\n 'mean' or 'median'\n shade : float\n Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1\n (opaque). Defaults to 0.\n ax : matplotlib axes\n \"\"\"\n if vec.dtype.kind == \"f\":\n if credible_interval != 1:\n hpd_ = hpd(vec, credible_interval)\n new_vec = vec[(vec >= hpd_[0]) & (vec <= hpd_[1])]\n else:\n new_vec = vec\n\n density, xmin, xmax = _fast_kde(new_vec, bw=bw)\n density *= credible_interval\n x = np.linspace(xmin, xmax, len(density))\n ymin = density[0]\n ymax = density[-1]\n\n if outline:\n ax.plot(x, density, color=color, lw=linewidth)\n ax.plot([xmin, xmin], [-ymin / 100, ymin], color=color, ls=\"-\", lw=linewidth)\n ax.plot([xmax, xmax], [-ymax / 100, ymax], color=color, ls=\"-\", lw=linewidth)\n\n if shade:\n ax.fill_between(x, density, color=color, alpha=shade)\n\n else:\n xmin, xmax = hpd(vec, credible_interval)\n bins = range(xmin, xmax + 2)\n if outline:\n ax.hist(vec, bins=bins, color=color, histtype=\"step\", align=\"left\")\n if shade:\n ax.hist(vec, bins=bins, color=color, alpha=shade)\n\n if hpd_markers:\n ax.plot(xmin, 0, hpd_markers, color=color, markeredgecolor=\"k\", markersize=markersize)\n ax.plot(xmax, 0, hpd_markers, color=color, markeredgecolor=\"k\", markersize=markersize)\n\n if point_estimate is not None:\n if point_estimate == \"mean\":\n est = np.mean(vec)\n elif point_estimate == \"median\":\n est = np.median(vec)\n ax.plot(est, 0, \"o\", color=color, markeredgecolor=\"k\", markersize=markersize)\n\n ax.set_yticks([])\n ax.set_title(vname, fontsize=titlesize, wrap=True)\n for pos in [\"left\", \"right\", \"top\"]:\n ax.spines[pos].set_visible(False)\n ax.tick_params(labelsize=xt_labelsize)\n", "path": "arviz/plots/densityplot.py"}], "after_files": [{"content": "\"\"\"KDE and histogram plots for multiple variables.\"\"\"\nimport numpy as np\n\nfrom ..data import convert_to_dataset\nfrom ..stats import hpd\nfrom .kdeplot import _fast_kde\nfrom .plot_utils import (\n _scale_fig_size,\n make_label,\n xarray_var_iter,\n default_grid,\n _create_axes_grid,\n)\nfrom ..utils import _var_names\n\n\n# pylint:disable-msg=too-many-function-args\ndef plot_density(\n data,\n group=\"posterior\",\n data_labels=None,\n var_names=None,\n credible_interval=0.94,\n point_estimate=\"mean\",\n colors=\"cycle\",\n outline=True,\n hpd_markers=\"\",\n shade=0.0,\n bw=4.5,\n figsize=None,\n textsize=None,\n):\n \"\"\"Generate KDE plots for continuous variables and histograms for discrete ones.\n\n Plots are truncated at their 100*(1-alpha)% credible intervals. Plots are grouped per variable\n and colors assigned to models.\n\n Parameters\n ----------\n data : Union[Object, Iterator[Object]]\n Any object that can be converted to an az.InferenceData object, or an Iterator returning\n a sequence of such objects.\n Refer to documentation of az.convert_to_dataset for details about such objects.\n group: Optional[str]\n Specifies which InferenceData group should be plotted. Defaults to 'posterior'.\n Alternative values include 'prior' and any other strings used as dataset keys in the\n InferenceData.\n data_labels : Optional[List[str]]\n List with names for the datasets passed as \"data.\" Useful when plotting more than one\n dataset. Must be the same shape as the data parameter. Defaults to None.\n var_names: Optional[List[str]]\n List of variables to plot. If multiple datasets are supplied and var_names is not None,\n will print the same set of variables for each dataset. Defaults to None, which results in\n all the variables being plotted.\n credible_interval : float\n Credible intervals. Should be in the interval (0, 1]. Defaults to 0.94.\n point_estimate : Optional[str]\n Plot point estimate per variable. Values should be 'mean', 'median' or None.\n Defaults to 'mean'.\n colors : Optional[Union[List[str],str]]\n List with valid matplotlib colors, one color per model. Alternative a string can be passed.\n If the string is `cycle`, it will automatically choose a color per model from matplolib's\n cycle. If a single color is passed, e.g. 'k', 'C2' or 'red' this color will be used for all\n models. Defaults to `cycle`.\n outline : bool\n Use a line to draw KDEs and histograms. Default to True\n hpd_markers : str\n A valid `matplotlib.markers` like 'v', used to indicate the limits of the hpd interval.\n Defaults to empty string (no marker).\n shade : Optional[float]\n Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1\n (opaque). Defaults to 0.\n bw : Optional[float]\n Bandwidth scaling factor for the KDE. Should be larger than 0. The higher this number the\n smoother the KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule\n of thumb (the default rule used by SciPy).\n figsize : Optional[Tuple[int, int]]\n Figure size. If None it will be defined automatically.\n textsize: Optional[float]\n Text size scaling factor for labels, titles and lines. If None it will be autoscaled based\n on figsize.\n\n Returns\n -------\n ax : Matplotlib axes\n\n\n Examples\n --------\n Plot default density plot\n\n .. plot::\n :context: close-figs\n\n >>> import arviz as az\n >>> centered = az.load_arviz_data('centered_eight')\n >>> non_centered = az.load_arviz_data('non_centered_eight')\n >>> az.plot_density([centered, non_centered])\n\n Plot subset variables by specifying variable name exactly\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"])\n\n Plot a specific `az.InferenceData` group\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], group=\"prior\")\n\n Specify credible interval\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], credible_interval=.5)\n\n Shade plots and/or remove outlines\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], outline=False, shade=.8)\n\n Specify binwidth for kernel density estimation\n\n .. plot::\n :context: close-figs\n\n >>> az.plot_density([centered, non_centered], var_names=[\"mu\"], bw=.9)\n \"\"\"\n if not isinstance(data, (list, tuple)):\n datasets = [convert_to_dataset(data, group=group)]\n else:\n datasets = [convert_to_dataset(datum, group=group) for datum in data]\n\n var_names = _var_names(var_names, datasets)\n\n if point_estimate not in (\"mean\", \"median\", None):\n raise ValueError(\n \"Point estimate should be 'mean',\" \"median' or None, not {}\".format(point_estimate)\n )\n\n n_data = len(datasets)\n\n if data_labels is None:\n if n_data > 1:\n data_labels = [\"{}\".format(idx) for idx in range(n_data)]\n else:\n data_labels = [\"\"]\n elif len(data_labels) != n_data:\n raise ValueError(\n \"The number of names for the models ({}) \"\n \"does not match the number of models ({})\".format(len(data_labels), n_data)\n )\n\n if colors == \"cycle\":\n colors = [\"C{}\".format(idx % 10) for idx in range(n_data)]\n elif isinstance(colors, str):\n colors = [colors for _ in range(n_data)]\n\n if not 1 >= credible_interval > 0:\n raise ValueError(\"The value of credible_interval should be in the interval (0, 1]\")\n\n to_plot = [list(xarray_var_iter(data, var_names, combined=True)) for data in datasets]\n all_labels = []\n length_plotters = []\n for plotters in to_plot:\n length_plotters.append(len(plotters))\n for var_name, selection, _ in plotters:\n label = make_label(var_name, selection)\n if label not in all_labels:\n all_labels.append(label)\n length_plotters = max(length_plotters)\n rows, cols = default_grid(length_plotters, max_cols=3)\n\n (figsize, _, titlesize, xt_labelsize, linewidth, markersize) = _scale_fig_size(\n figsize, textsize, rows, cols\n )\n\n _, ax = _create_axes_grid(length_plotters, rows, cols, figsize=figsize, squeeze=False)\n\n axis_map = {label: ax_ for label, ax_ in zip(all_labels, ax.flatten())}\n for m_idx, plotters in enumerate(to_plot):\n for var_name, selection, values in plotters:\n label = make_label(var_name, selection)\n _d_helper(\n values.flatten(),\n label,\n colors[m_idx],\n bw,\n titlesize,\n xt_labelsize,\n linewidth,\n markersize,\n credible_interval,\n point_estimate,\n hpd_markers,\n outline,\n shade,\n axis_map[label],\n )\n\n if n_data > 1:\n for m_idx, label in enumerate(data_labels):\n ax[0].plot([], label=label, c=colors[m_idx], markersize=markersize)\n ax[0].legend(fontsize=xt_labelsize)\n\n return ax\n\n\ndef _d_helper(\n vec,\n vname,\n color,\n bw,\n titlesize,\n xt_labelsize,\n linewidth,\n markersize,\n credible_interval,\n point_estimate,\n hpd_markers,\n outline,\n shade,\n ax,\n):\n \"\"\"Plot an individual dimension.\n\n Parameters\n ----------\n vec : array\n 1D array from trace\n vname : str\n variable name\n color : str\n matplotlib color\n bw : float\n Bandwidth scaling factor. Should be larger than 0. The higher this number the smoother the\n KDE will be. Defaults to 4.5 which is essentially the same as the Scott's rule of thumb\n (the default used rule by SciPy).\n titlesize : float\n font size for title\n xt_labelsize : float\n fontsize for xticks\n linewidth : float\n Thickness of lines\n markersize : float\n Size of markers\n credible_interval : float\n Credible intervals. Defaults to 0.94\n point_estimate : str or None\n 'mean' or 'median'\n shade : float\n Alpha blending value for the shaded area under the curve, between 0 (no shade) and 1\n (opaque). Defaults to 0.\n ax : matplotlib axes\n \"\"\"\n if vec.dtype.kind == \"f\":\n if credible_interval != 1:\n hpd_ = hpd(vec, credible_interval)\n new_vec = vec[(vec >= hpd_[0]) & (vec <= hpd_[1])]\n else:\n new_vec = vec\n\n density, xmin, xmax = _fast_kde(new_vec, bw=bw)\n density *= credible_interval\n x = np.linspace(xmin, xmax, len(density))\n ymin = density[0]\n ymax = density[-1]\n\n if outline:\n ax.plot(x, density, color=color, lw=linewidth)\n ax.plot([xmin, xmin], [-ymin / 100, ymin], color=color, ls=\"-\", lw=linewidth)\n ax.plot([xmax, xmax], [-ymax / 100, ymax], color=color, ls=\"-\", lw=linewidth)\n\n if shade:\n ax.fill_between(x, density, color=color, alpha=shade)\n\n else:\n xmin, xmax = hpd(vec, credible_interval)\n bins = range(xmin, xmax + 2)\n if outline:\n ax.hist(vec, bins=bins, color=color, histtype=\"step\", align=\"left\")\n if shade:\n ax.hist(vec, bins=bins, color=color, alpha=shade)\n\n if hpd_markers:\n ax.plot(xmin, 0, hpd_markers, color=color, markeredgecolor=\"k\", markersize=markersize)\n ax.plot(xmax, 0, hpd_markers, color=color, markeredgecolor=\"k\", markersize=markersize)\n\n if point_estimate is not None:\n if point_estimate == \"mean\":\n est = np.mean(vec)\n elif point_estimate == \"median\":\n est = np.median(vec)\n ax.plot(est, 0, \"o\", color=color, markeredgecolor=\"k\", markersize=markersize)\n\n ax.set_yticks([])\n ax.set_title(vname, fontsize=titlesize, wrap=True)\n for pos in [\"left\", \"right\", \"top\"]:\n ax.spines[pos].set_visible(False)\n ax.tick_params(labelsize=xt_labelsize)\n", "path": "arviz/plots/densityplot.py"}]}
| 3,693 | 222 |
gh_patches_debug_27355
|
rasdani/github-patches
|
git_diff
|
mathesar-foundation__mathesar-3121
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add a Docker image containing the Mathesar server along with a Postgres server
## Problem
The current docker image requires the user to already have a Postgres server running on their machine. This is a hassle for someone looking to try out Matheasar quickly.
## Proposed solution
Provide a docker image containing a Postgres server along with the Mathesar server
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `config/settings/common_settings.py`
Content:
```
1 """
2 Base settings to build other settings files upon.
3
4 Generated by 'django-admin startproject' using Django 3.1.7.
5
6 For more information on this file, see
7 https://docs.djangoproject.com/en/3.1/topics/settings/
8
9 For the full list of settings and their values, see
10 https://docs.djangoproject.com/en/3.1/ref/settings/
11 """
12
13 import os
14 from pathlib import Path
15
16 from decouple import Csv, config as decouple_config
17 from dj_database_url import parse as db_url
18
19
20 # We use a 'tuple' with pipes as delimiters as decople naively splits the global
21 # variables on commas when casting to Csv()
22 def pipe_delim(pipe_string):
23 # Remove opening and closing brackets
24 pipe_string = pipe_string[1:-1]
25 # Split on pipe delim
26 return pipe_string.split("|")
27
28
29 # Build paths inside the project like this: BASE_DIR / 'subdir'.
30 BASE_DIR = Path(__file__).resolve().parent.parent.parent
31
32 # Application definition
33
34 INSTALLED_APPS = [
35 "django.contrib.admin",
36 "django.contrib.auth",
37 "django.contrib.contenttypes",
38 "django.contrib.sessions",
39 "django.contrib.messages",
40 "whitenoise.runserver_nostatic",
41 "django.contrib.staticfiles",
42 "rest_framework",
43 "django_filters",
44 "django_property_filter",
45 "drf_spectacular",
46 "mathesar",
47 ]
48
49 MIDDLEWARE = [
50 "django.middleware.security.SecurityMiddleware",
51 "whitenoise.middleware.WhiteNoiseMiddleware",
52 "django.contrib.sessions.middleware.SessionMiddleware",
53 "django.middleware.common.CommonMiddleware",
54 "django.middleware.csrf.CsrfViewMiddleware",
55 "django.contrib.auth.middleware.AuthenticationMiddleware",
56 "django.contrib.messages.middleware.MessageMiddleware",
57 "django.middleware.clickjacking.XFrameOptionsMiddleware",
58 "mathesar.middleware.CursorClosedHandlerMiddleware",
59 "mathesar.middleware.PasswordChangeNeededMiddleware",
60 'django_userforeignkey.middleware.UserForeignKeyMiddleware',
61 'django_request_cache.middleware.RequestCacheMiddleware',
62 ]
63
64 ROOT_URLCONF = "config.urls"
65
66 TEMPLATES = [
67 {
68 "BACKEND": "django.template.backends.django.DjangoTemplates",
69 "DIRS": [],
70 "APP_DIRS": True,
71 "OPTIONS": {
72 "context_processors": [
73 "config.context_processors.frontend_settings",
74 "django.template.context_processors.debug",
75 "django.template.context_processors.request",
76 "django.contrib.auth.context_processors.auth",
77 "django.contrib.messages.context_processors.messages",
78 "mathesar.template_context_processors.base_template_extensions.script_extension_templates"
79 ],
80 },
81 },
82 ]
83
84 WSGI_APPLICATION = "config.wsgi.application"
85
86 # Database
87 # https://docs.djangoproject.com/en/3.1/ref/settings/#databases
88
89 # TODO: Add to documentation that database keys should not be than 128 characters.
90
91 # MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'
92 # See pipe_delim above for why we use pipes as delimiters
93 DATABASES = {
94 db_key: db_url(url_string)
95 for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))
96 }
97
98 DATABASES[decouple_config('DJANGO_DATABASE_KEY', default="default")] = decouple_config('DJANGO_DATABASE_URL', cast=db_url, default='sqlite:///db.sqlite3')
99
100 for db_key, db_dict in DATABASES.items():
101 # Engine should be '.postgresql' or '.postgresql_psycopg2' for all db(s),
102 # however for the internal 'default' db 'sqlite3' can be used.
103 if not db_dict['ENGINE'].startswith('django.db.backends.postgresql') and db_key != 'default':
104 raise ValueError(
105 f"{db_key} is not a PostgreSQL database. "
106 f"{db_dict['ENGINE']} found for {db_key}'s engine."
107 )
108
109 # pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'
110 # and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']
111 TEST = decouple_config('TEST', default=False, cast=bool)
112 if TEST:
113 for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):
114 DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}
115
116
117 # SECURITY WARNING: keep the secret key used in production secret!
118 SECRET_KEY = decouple_config('SECRET_KEY')
119
120 # SECURITY WARNING: don't run with debug turned on in production!
121 DEBUG = decouple_config('DEBUG', default=False, cast=bool)
122
123 ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())
124
125 # Password validation
126 # https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators
127
128 AUTH_PASSWORD_VALIDATORS = [
129 {
130 "NAME": "django.contrib.auth.password_validation.UserAttributeSimilarityValidator",
131 },
132 {
133 "NAME": "django.contrib.auth.password_validation.MinimumLengthValidator",
134 },
135 {
136 "NAME": "django.contrib.auth.password_validation.CommonPasswordValidator",
137 },
138 {
139 "NAME": "django.contrib.auth.password_validation.NumericPasswordValidator",
140 },
141 ]
142
143 # Internationalization
144 # https://docs.djangoproject.com/en/3.1/topics/i18n/
145
146 LANGUAGE_CODE = "en-us"
147
148 TIME_ZONE = "UTC"
149
150 USE_I18N = True
151
152 USE_L10N = True
153
154 USE_TZ = True
155
156 # Static files (CSS, JavaScript, Images)
157 # https://docs.djangoproject.com/en/3.1/howto/static-files/
158 # https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/
159
160 STATIC_URL = "/static/"
161
162 # When running with DEBUG=False, the webserver needs to serve files from this location
163 # python manage.py collectstatic has to be run to collect all static files into this location
164 # The files need to served in brotli or gzip compressed format
165 STATIC_ROOT = os.path.join(BASE_DIR, 'static/')
166
167 # Media files (uploaded by the user)
168 DEFAULT_MEDIA_ROOT = os.path.join(BASE_DIR, '.media/')
169 MEDIA_ROOT = decouple_config('MEDIA_ROOT', default=DEFAULT_MEDIA_ROOT)
170
171 MEDIA_URL = "/media/"
172
173 # Update Authentication classes, removed BasicAuthentication
174 # Defaults: https://www.django-rest-framework.org/api-guide/settings/
175 REST_FRAMEWORK = {
176 'DEFAULT_AUTHENTICATION_CLASSES': [
177 'rest_framework.authentication.TokenAuthentication',
178 'rest_framework.authentication.SessionAuthentication'
179 ],
180 'DEFAULT_PERMISSION_CLASSES': [
181 'rest_framework.permissions.IsAuthenticated',
182 ],
183 'DEFAULT_FILTER_BACKENDS': (
184 'django_filters.rest_framework.DjangoFilterBackend',
185 'rest_framework.filters.OrderingFilter',
186 ),
187 'TEST_REQUEST_DEFAULT_FORMAT': 'json',
188 'EXCEPTION_HANDLER':
189 'mathesar.exception_handlers.mathesar_exception_handler',
190 'DEFAULT_SCHEMA_CLASS': 'drf_spectacular.openapi.AutoSchema'
191 }
192 SPECTACULAR_SETTINGS = {
193 'TITLE': 'Mathesar API',
194 'DESCRIPTION': '',
195 'VERSION': '1.0.0',
196 'SERVE_INCLUDE_SCHEMA': False,
197 'PREPROCESSING_HOOKS': ['config.settings.openapi.custom_preprocessing_hook'],
198 'POSTPROCESSING_HOOKS': [
199 'config.settings.openapi.remove_url_prefix_hook',
200 ],
201 # OTHER SETTINGS
202 }
203 FRIENDLY_ERRORS = {
204 'FIELD_ERRORS': {
205 # By default drf-friendly-errors does contain error codes for ListSerializer type
206 'ListSerializer': {
207 'required': 2007,
208 'null': 2027,
209 'invalid_choice': 2083,
210 'not_a_list': 2123,
211 'empty': 2093
212 },
213 'PermittedPkRelatedField': {
214 'required': 2007,
215 'null': 2027,
216 'does_not_exist': 2151,
217 'incorrect_type': 2161
218 },
219 'PermittedSlugRelatedField': {
220 'required': 2007, 'invalid': 2002, 'null': 2027,
221 'does_not_exist': 2151, 'incorrect_type': 2161
222 },
223 },
224 'EXCEPTION_DICT': {
225 'Http404': 4005
226 }
227 }
228 # Mathesar settings
229 MATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')
230 MATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')
231 MATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')
232 MATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'
233 MATHESAR_UI_SOURCE_LOCATION = os.path.join(BASE_DIR, 'mathesar_ui/')
234 MATHESAR_CAPTURE_UNHANDLED_EXCEPTION = decouple_config('CAPTURE_UNHANDLED_EXCEPTION', default=False)
235 MATHESAR_STATIC_NON_CODE_FILES_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/non-code/')
236
237 # UI source files have to be served by Django in order for static assets to be included during dev mode
238 # https://vitejs.dev/guide/assets.html
239 # https://vitejs.dev/guide/backend-integration.html
240 STATICFILES_DIRS = [MATHESAR_UI_SOURCE_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION] if MATHESAR_MODE == 'DEVELOPMENT' else [MATHESAR_UI_BUILD_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION]
241 STATICFILES_STORAGE = "whitenoise.storage.CompressedManifestStaticFilesStorage"
242
243 # Accounts
244 AUTH_USER_MODEL = 'mathesar.User'
245 LOGIN_URL = '/auth/login/'
246 LOGIN_REDIRECT_URL = '/'
247 LOGOUT_REDIRECT_URL = LOGIN_URL
248 DRF_ACCESS_POLICY = {
249 'reusable_conditions': ['mathesar.api.permission_conditions']
250 }
251 # List of Template names that contains additional script tags to be added to the base template
252 BASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES = []
253
```
Path: `db/install.py`
Content:
```
1 from psycopg.errors import InsufficientPrivilege
2 from sqlalchemy import text
3 from sqlalchemy.exc import OperationalError, ProgrammingError
4
5 from db import engine
6 from db.sql import install as sql_install
7 from db.types import install as types_install
8
9
10 def install_mathesar(
11 database_name, username, password, hostname, port, skip_confirm
12 ):
13 """Create database and install Mathesar on it."""
14 user_db_engine = engine.create_future_engine(
15 username, password, hostname, database_name, port,
16 connect_args={"connect_timeout": 10}
17 )
18 try:
19 user_db_engine.connect()
20 print(f"Installing Mathesar on preexisting PostgreSQL database {database_name} at host {hostname}...")
21 sql_install.install(user_db_engine)
22 types_install.install_mathesar_on_database(user_db_engine)
23 user_db_engine.dispose()
24 except OperationalError:
25 database_created = _create_database(
26 database_name=database_name,
27 hostname=hostname,
28 username=username,
29 password=password,
30 port=port,
31 skip_confirm=skip_confirm
32 )
33 if database_created:
34 print(f"Installing Mathesar on PostgreSQL database {database_name} at host {hostname}...")
35 sql_install.install(user_db_engine)
36 types_install.install_mathesar_on_database(user_db_engine)
37 user_db_engine.dispose()
38 else:
39 print(f"Skipping installing on DB with key {database_name}.")
40
41
42 def _create_database(database_name, hostname, username, password, port, skip_confirm=True):
43 if skip_confirm is True:
44 create_database = "y"
45 else:
46 create_database = input(
47 f"Create a new Database called {database_name}? (y/n) > "
48 )
49 if create_database.lower() in ["y", "yes"]:
50 # We need to connect to an existing database inorder to create a new Database.
51 # So we use the default Database `postgres` that comes with postgres.
52 # TODO Throw correct error when the default postgres database does not exists(which is very rare but still possible)
53 root_database = "postgres"
54 root_db_engine = engine.create_future_engine(
55 username, password, hostname, root_database, port,
56 connect_args={"connect_timeout": 10}
57 )
58 try:
59 with root_db_engine.connect() as conn:
60 conn.execution_options(isolation_level="AUTOCOMMIT")
61 conn.execute(text(f'CREATE DATABASE "{database_name}"'))
62 root_db_engine.dispose()
63 print(f"Created DB is {database_name}.")
64 return True
65 except ProgrammingError as e:
66 if isinstance(e.orig, InsufficientPrivilege):
67 print(f"Database {database_name} could not be created due to Insufficient Privilege")
68 return False
69 except Exception:
70 print(f"Database {database_name} could not be created!")
71 return False
72 else:
73 print(f"Database {database_name} not created!")
74 return False
75
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/config/settings/common_settings.py b/config/settings/common_settings.py
--- a/config/settings/common_settings.py
+++ b/config/settings/common_settings.py
@@ -115,12 +115,12 @@
# SECURITY WARNING: keep the secret key used in production secret!
-SECRET_KEY = decouple_config('SECRET_KEY')
+SECRET_KEY = decouple_config('SECRET_KEY', default="2gr6ud88x=(p855_5nbj_+7^gw-iz&n7ldqv%94mjaecl+b9=4")
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = decouple_config('DEBUG', default=False, cast=bool)
-ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())
+ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv(), default=".localhost, 127.0.0.1, [::1]")
# Password validation
# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators
diff --git a/db/install.py b/db/install.py
--- a/db/install.py
+++ b/db/install.py
@@ -48,7 +48,7 @@
)
if create_database.lower() in ["y", "yes"]:
# We need to connect to an existing database inorder to create a new Database.
- # So we use the default Database `postgres` that comes with postgres.
+ # So we use the default database `postgres` that comes with postgres.
# TODO Throw correct error when the default postgres database does not exists(which is very rare but still possible)
root_database = "postgres"
root_db_engine = engine.create_future_engine(
|
{"golden_diff": "diff --git a/config/settings/common_settings.py b/config/settings/common_settings.py\n--- a/config/settings/common_settings.py\n+++ b/config/settings/common_settings.py\n@@ -115,12 +115,12 @@\n \n \n # SECURITY WARNING: keep the secret key used in production secret!\n-SECRET_KEY = decouple_config('SECRET_KEY')\n+SECRET_KEY = decouple_config('SECRET_KEY', default=\"2gr6ud88x=(p855_5nbj_+7^gw-iz&n7ldqv%94mjaecl+b9=4\")\n \n # SECURITY WARNING: don't run with debug turned on in production!\n DEBUG = decouple_config('DEBUG', default=False, cast=bool)\n \n-ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())\n+ALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv(), default=\".localhost, 127.0.0.1, [::1]\")\n \n # Password validation\n # https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators\ndiff --git a/db/install.py b/db/install.py\n--- a/db/install.py\n+++ b/db/install.py\n@@ -48,7 +48,7 @@\n )\n if create_database.lower() in [\"y\", \"yes\"]:\n # We need to connect to an existing database inorder to create a new Database.\n- # So we use the default Database `postgres` that comes with postgres.\n+ # So we use the default database `postgres` that comes with postgres.\n # TODO Throw correct error when the default postgres database does not exists(which is very rare but still possible)\n root_database = \"postgres\"\n root_db_engine = engine.create_future_engine(\n", "issue": "Add a Docker image containing the Mathesar server along with a Postgres server\n## Problem\r\nThe current docker image requires the user to already have a Postgres server running on their machine. This is a hassle for someone looking to try out Matheasar quickly.\r\n\r\n## Proposed solution\r\nProvide a docker image containing a Postgres server along with the Mathesar server\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nBase settings to build other settings files upon.\n\nGenerated by 'django-admin startproject' using Django 3.1.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.1/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.1/ref/settings/\n\"\"\"\n\nimport os\nfrom pathlib import Path\n\nfrom decouple import Csv, config as decouple_config\nfrom dj_database_url import parse as db_url\n\n\n# We use a 'tuple' with pipes as delimiters as decople naively splits the global\n# variables on commas when casting to Csv()\ndef pipe_delim(pipe_string):\n # Remove opening and closing brackets\n pipe_string = pipe_string[1:-1]\n # Split on pipe delim\n return pipe_string.split(\"|\")\n\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent.parent\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"whitenoise.runserver_nostatic\",\n \"django.contrib.staticfiles\",\n \"rest_framework\",\n \"django_filters\",\n \"django_property_filter\",\n \"drf_spectacular\",\n \"mathesar\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"whitenoise.middleware.WhiteNoiseMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n \"mathesar.middleware.CursorClosedHandlerMiddleware\",\n \"mathesar.middleware.PasswordChangeNeededMiddleware\",\n 'django_userforeignkey.middleware.UserForeignKeyMiddleware',\n 'django_request_cache.middleware.RequestCacheMiddleware',\n]\n\nROOT_URLCONF = \"config.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"config.context_processors.frontend_settings\",\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n \"mathesar.template_context_processors.base_template_extensions.script_extension_templates\"\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"config.wsgi.application\"\n\n# Database\n# https://docs.djangoproject.com/en/3.1/ref/settings/#databases\n\n# TODO: Add to documentation that database keys should not be than 128 characters.\n\n# MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'\n# See pipe_delim above for why we use pipes as delimiters\nDATABASES = {\n db_key: db_url(url_string)\n for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))\n}\n\nDATABASES[decouple_config('DJANGO_DATABASE_KEY', default=\"default\")] = decouple_config('DJANGO_DATABASE_URL', cast=db_url, default='sqlite:///db.sqlite3')\n\nfor db_key, db_dict in DATABASES.items():\n # Engine should be '.postgresql' or '.postgresql_psycopg2' for all db(s),\n # however for the internal 'default' db 'sqlite3' can be used.\n if not db_dict['ENGINE'].startswith('django.db.backends.postgresql') and db_key != 'default':\n raise ValueError(\n f\"{db_key} is not a PostgreSQL database. \"\n f\"{db_dict['ENGINE']} found for {db_key}'s engine.\"\n )\n\n# pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'\n# and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']\nTEST = decouple_config('TEST', default=False, cast=bool)\nif TEST:\n for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):\n DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}\n\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = decouple_config('SECRET_KEY')\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = decouple_config('DEBUG', default=False, cast=bool)\n\nALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv())\n\n# Password validation\n# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.1/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.1/howto/static-files/\n# https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/\n\nSTATIC_URL = \"/static/\"\n\n# When running with DEBUG=False, the webserver needs to serve files from this location\n# python manage.py collectstatic has to be run to collect all static files into this location\n# The files need to served in brotli or gzip compressed format\nSTATIC_ROOT = os.path.join(BASE_DIR, 'static/')\n\n# Media files (uploaded by the user)\nDEFAULT_MEDIA_ROOT = os.path.join(BASE_DIR, '.media/')\nMEDIA_ROOT = decouple_config('MEDIA_ROOT', default=DEFAULT_MEDIA_ROOT)\n\nMEDIA_URL = \"/media/\"\n\n# Update Authentication classes, removed BasicAuthentication\n# Defaults: https://www.django-rest-framework.org/api-guide/settings/\nREST_FRAMEWORK = {\n 'DEFAULT_AUTHENTICATION_CLASSES': [\n 'rest_framework.authentication.TokenAuthentication',\n 'rest_framework.authentication.SessionAuthentication'\n ],\n 'DEFAULT_PERMISSION_CLASSES': [\n 'rest_framework.permissions.IsAuthenticated',\n ],\n 'DEFAULT_FILTER_BACKENDS': (\n 'django_filters.rest_framework.DjangoFilterBackend',\n 'rest_framework.filters.OrderingFilter',\n ),\n 'TEST_REQUEST_DEFAULT_FORMAT': 'json',\n 'EXCEPTION_HANDLER':\n 'mathesar.exception_handlers.mathesar_exception_handler',\n 'DEFAULT_SCHEMA_CLASS': 'drf_spectacular.openapi.AutoSchema'\n}\nSPECTACULAR_SETTINGS = {\n 'TITLE': 'Mathesar API',\n 'DESCRIPTION': '',\n 'VERSION': '1.0.0',\n 'SERVE_INCLUDE_SCHEMA': False,\n 'PREPROCESSING_HOOKS': ['config.settings.openapi.custom_preprocessing_hook'],\n 'POSTPROCESSING_HOOKS': [\n 'config.settings.openapi.remove_url_prefix_hook',\n ],\n # OTHER SETTINGS\n}\nFRIENDLY_ERRORS = {\n 'FIELD_ERRORS': {\n # By default drf-friendly-errors does contain error codes for ListSerializer type\n 'ListSerializer': {\n 'required': 2007,\n 'null': 2027,\n 'invalid_choice': 2083,\n 'not_a_list': 2123,\n 'empty': 2093\n },\n 'PermittedPkRelatedField': {\n 'required': 2007,\n 'null': 2027,\n 'does_not_exist': 2151,\n 'incorrect_type': 2161\n },\n 'PermittedSlugRelatedField': {\n 'required': 2007, 'invalid': 2002, 'null': 2027,\n 'does_not_exist': 2151, 'incorrect_type': 2161\n },\n },\n 'EXCEPTION_DICT': {\n 'Http404': 4005\n }\n}\n# Mathesar settings\nMATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')\nMATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')\nMATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')\nMATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'\nMATHESAR_UI_SOURCE_LOCATION = os.path.join(BASE_DIR, 'mathesar_ui/')\nMATHESAR_CAPTURE_UNHANDLED_EXCEPTION = decouple_config('CAPTURE_UNHANDLED_EXCEPTION', default=False)\nMATHESAR_STATIC_NON_CODE_FILES_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/non-code/')\n\n# UI source files have to be served by Django in order for static assets to be included during dev mode\n# https://vitejs.dev/guide/assets.html\n# https://vitejs.dev/guide/backend-integration.html\nSTATICFILES_DIRS = [MATHESAR_UI_SOURCE_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION] if MATHESAR_MODE == 'DEVELOPMENT' else [MATHESAR_UI_BUILD_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION]\nSTATICFILES_STORAGE = \"whitenoise.storage.CompressedManifestStaticFilesStorage\"\n\n# Accounts\nAUTH_USER_MODEL = 'mathesar.User'\nLOGIN_URL = '/auth/login/'\nLOGIN_REDIRECT_URL = '/'\nLOGOUT_REDIRECT_URL = LOGIN_URL\nDRF_ACCESS_POLICY = {\n 'reusable_conditions': ['mathesar.api.permission_conditions']\n}\n# List of Template names that contains additional script tags to be added to the base template\nBASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES = []\n", "path": "config/settings/common_settings.py"}, {"content": "from psycopg.errors import InsufficientPrivilege\nfrom sqlalchemy import text\nfrom sqlalchemy.exc import OperationalError, ProgrammingError\n\nfrom db import engine\nfrom db.sql import install as sql_install\nfrom db.types import install as types_install\n\n\ndef install_mathesar(\n database_name, username, password, hostname, port, skip_confirm\n):\n \"\"\"Create database and install Mathesar on it.\"\"\"\n user_db_engine = engine.create_future_engine(\n username, password, hostname, database_name, port,\n connect_args={\"connect_timeout\": 10}\n )\n try:\n user_db_engine.connect()\n print(f\"Installing Mathesar on preexisting PostgreSQL database {database_name} at host {hostname}...\")\n sql_install.install(user_db_engine)\n types_install.install_mathesar_on_database(user_db_engine)\n user_db_engine.dispose()\n except OperationalError:\n database_created = _create_database(\n database_name=database_name,\n hostname=hostname,\n username=username,\n password=password,\n port=port,\n skip_confirm=skip_confirm\n )\n if database_created:\n print(f\"Installing Mathesar on PostgreSQL database {database_name} at host {hostname}...\")\n sql_install.install(user_db_engine)\n types_install.install_mathesar_on_database(user_db_engine)\n user_db_engine.dispose()\n else:\n print(f\"Skipping installing on DB with key {database_name}.\")\n\n\ndef _create_database(database_name, hostname, username, password, port, skip_confirm=True):\n if skip_confirm is True:\n create_database = \"y\"\n else:\n create_database = input(\n f\"Create a new Database called {database_name}? (y/n) > \"\n )\n if create_database.lower() in [\"y\", \"yes\"]:\n # We need to connect to an existing database inorder to create a new Database.\n # So we use the default Database `postgres` that comes with postgres.\n # TODO Throw correct error when the default postgres database does not exists(which is very rare but still possible)\n root_database = \"postgres\"\n root_db_engine = engine.create_future_engine(\n username, password, hostname, root_database, port,\n connect_args={\"connect_timeout\": 10}\n )\n try:\n with root_db_engine.connect() as conn:\n conn.execution_options(isolation_level=\"AUTOCOMMIT\")\n conn.execute(text(f'CREATE DATABASE \"{database_name}\"'))\n root_db_engine.dispose()\n print(f\"Created DB is {database_name}.\")\n return True\n except ProgrammingError as e:\n if isinstance(e.orig, InsufficientPrivilege):\n print(f\"Database {database_name} could not be created due to Insufficient Privilege\")\n return False\n except Exception:\n print(f\"Database {database_name} could not be created!\")\n return False\n else:\n print(f\"Database {database_name} not created!\")\n return False\n", "path": "db/install.py"}], "after_files": [{"content": "\"\"\"\nBase settings to build other settings files upon.\n\nGenerated by 'django-admin startproject' using Django 3.1.7.\n\nFor more information on this file, see\nhttps://docs.djangoproject.com/en/3.1/topics/settings/\n\nFor the full list of settings and their values, see\nhttps://docs.djangoproject.com/en/3.1/ref/settings/\n\"\"\"\n\nimport os\nfrom pathlib import Path\n\nfrom decouple import Csv, config as decouple_config\nfrom dj_database_url import parse as db_url\n\n\n# We use a 'tuple' with pipes as delimiters as decople naively splits the global\n# variables on commas when casting to Csv()\ndef pipe_delim(pipe_string):\n # Remove opening and closing brackets\n pipe_string = pipe_string[1:-1]\n # Split on pipe delim\n return pipe_string.split(\"|\")\n\n\n# Build paths inside the project like this: BASE_DIR / 'subdir'.\nBASE_DIR = Path(__file__).resolve().parent.parent.parent\n\n# Application definition\n\nINSTALLED_APPS = [\n \"django.contrib.admin\",\n \"django.contrib.auth\",\n \"django.contrib.contenttypes\",\n \"django.contrib.sessions\",\n \"django.contrib.messages\",\n \"whitenoise.runserver_nostatic\",\n \"django.contrib.staticfiles\",\n \"rest_framework\",\n \"django_filters\",\n \"django_property_filter\",\n \"drf_spectacular\",\n \"mathesar\",\n]\n\nMIDDLEWARE = [\n \"django.middleware.security.SecurityMiddleware\",\n \"whitenoise.middleware.WhiteNoiseMiddleware\",\n \"django.contrib.sessions.middleware.SessionMiddleware\",\n \"django.middleware.common.CommonMiddleware\",\n \"django.middleware.csrf.CsrfViewMiddleware\",\n \"django.contrib.auth.middleware.AuthenticationMiddleware\",\n \"django.contrib.messages.middleware.MessageMiddleware\",\n \"django.middleware.clickjacking.XFrameOptionsMiddleware\",\n \"mathesar.middleware.CursorClosedHandlerMiddleware\",\n \"mathesar.middleware.PasswordChangeNeededMiddleware\",\n 'django_userforeignkey.middleware.UserForeignKeyMiddleware',\n 'django_request_cache.middleware.RequestCacheMiddleware',\n]\n\nROOT_URLCONF = \"config.urls\"\n\nTEMPLATES = [\n {\n \"BACKEND\": \"django.template.backends.django.DjangoTemplates\",\n \"DIRS\": [],\n \"APP_DIRS\": True,\n \"OPTIONS\": {\n \"context_processors\": [\n \"config.context_processors.frontend_settings\",\n \"django.template.context_processors.debug\",\n \"django.template.context_processors.request\",\n \"django.contrib.auth.context_processors.auth\",\n \"django.contrib.messages.context_processors.messages\",\n \"mathesar.template_context_processors.base_template_extensions.script_extension_templates\"\n ],\n },\n },\n]\n\nWSGI_APPLICATION = \"config.wsgi.application\"\n\n# Database\n# https://docs.djangoproject.com/en/3.1/ref/settings/#databases\n\n# TODO: Add to documentation that database keys should not be than 128 characters.\n\n# MATHESAR_DATABASES should be of the form '({db_name}|{db_url}), ({db_name}|{db_url})'\n# See pipe_delim above for why we use pipes as delimiters\nDATABASES = {\n db_key: db_url(url_string)\n for db_key, url_string in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim))\n}\n\nDATABASES[decouple_config('DJANGO_DATABASE_KEY', default=\"default\")] = decouple_config('DJANGO_DATABASE_URL', cast=db_url, default='sqlite:///db.sqlite3')\n\nfor db_key, db_dict in DATABASES.items():\n # Engine should be '.postgresql' or '.postgresql_psycopg2' for all db(s),\n # however for the internal 'default' db 'sqlite3' can be used.\n if not db_dict['ENGINE'].startswith('django.db.backends.postgresql') and db_key != 'default':\n raise ValueError(\n f\"{db_key} is not a PostgreSQL database. \"\n f\"{db_dict['ENGINE']} found for {db_key}'s engine.\"\n )\n\n# pytest-django will create a new database named 'test_{DATABASES[table_db]['NAME']}'\n# and use it for our API tests if we don't specify DATABASES[table_db]['TEST']['NAME']\nTEST = decouple_config('TEST', default=False, cast=bool)\nif TEST:\n for db_key, _ in decouple_config('MATHESAR_DATABASES', cast=Csv(pipe_delim)):\n DATABASES[db_key]['TEST'] = {'NAME': DATABASES[db_key]['NAME']}\n\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = decouple_config('SECRET_KEY', default=\"2gr6ud88x=(p855_5nbj_+7^gw-iz&n7ldqv%94mjaecl+b9=4\")\n\n# SECURITY WARNING: don't run with debug turned on in production!\nDEBUG = decouple_config('DEBUG', default=False, cast=bool)\n\nALLOWED_HOSTS = decouple_config('ALLOWED_HOSTS', cast=Csv(), default=\".localhost, 127.0.0.1, [::1]\")\n\n# Password validation\n# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators\n\nAUTH_PASSWORD_VALIDATORS = [\n {\n \"NAME\": \"django.contrib.auth.password_validation.UserAttributeSimilarityValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.MinimumLengthValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.CommonPasswordValidator\",\n },\n {\n \"NAME\": \"django.contrib.auth.password_validation.NumericPasswordValidator\",\n },\n]\n\n# Internationalization\n# https://docs.djangoproject.com/en/3.1/topics/i18n/\n\nLANGUAGE_CODE = \"en-us\"\n\nTIME_ZONE = \"UTC\"\n\nUSE_I18N = True\n\nUSE_L10N = True\n\nUSE_TZ = True\n\n# Static files (CSS, JavaScript, Images)\n# https://docs.djangoproject.com/en/3.1/howto/static-files/\n# https://docs.djangoproject.com/en/3.1/ref/contrib/staticfiles/\n\nSTATIC_URL = \"/static/\"\n\n# When running with DEBUG=False, the webserver needs to serve files from this location\n# python manage.py collectstatic has to be run to collect all static files into this location\n# The files need to served in brotli or gzip compressed format\nSTATIC_ROOT = os.path.join(BASE_DIR, 'static/')\n\n# Media files (uploaded by the user)\nDEFAULT_MEDIA_ROOT = os.path.join(BASE_DIR, '.media/')\nMEDIA_ROOT = decouple_config('MEDIA_ROOT', default=DEFAULT_MEDIA_ROOT)\n\nMEDIA_URL = \"/media/\"\n\n# Update Authentication classes, removed BasicAuthentication\n# Defaults: https://www.django-rest-framework.org/api-guide/settings/\nREST_FRAMEWORK = {\n 'DEFAULT_AUTHENTICATION_CLASSES': [\n 'rest_framework.authentication.TokenAuthentication',\n 'rest_framework.authentication.SessionAuthentication'\n ],\n 'DEFAULT_PERMISSION_CLASSES': [\n 'rest_framework.permissions.IsAuthenticated',\n ],\n 'DEFAULT_FILTER_BACKENDS': (\n 'django_filters.rest_framework.DjangoFilterBackend',\n 'rest_framework.filters.OrderingFilter',\n ),\n 'TEST_REQUEST_DEFAULT_FORMAT': 'json',\n 'EXCEPTION_HANDLER':\n 'mathesar.exception_handlers.mathesar_exception_handler',\n 'DEFAULT_SCHEMA_CLASS': 'drf_spectacular.openapi.AutoSchema'\n}\nSPECTACULAR_SETTINGS = {\n 'TITLE': 'Mathesar API',\n 'DESCRIPTION': '',\n 'VERSION': '1.0.0',\n 'SERVE_INCLUDE_SCHEMA': False,\n 'PREPROCESSING_HOOKS': ['config.settings.openapi.custom_preprocessing_hook'],\n 'POSTPROCESSING_HOOKS': [\n 'config.settings.openapi.remove_url_prefix_hook',\n ],\n # OTHER SETTINGS\n}\nFRIENDLY_ERRORS = {\n 'FIELD_ERRORS': {\n # By default drf-friendly-errors does contain error codes for ListSerializer type\n 'ListSerializer': {\n 'required': 2007,\n 'null': 2027,\n 'invalid_choice': 2083,\n 'not_a_list': 2123,\n 'empty': 2093\n },\n 'PermittedPkRelatedField': {\n 'required': 2007,\n 'null': 2027,\n 'does_not_exist': 2151,\n 'incorrect_type': 2161\n },\n 'PermittedSlugRelatedField': {\n 'required': 2007, 'invalid': 2002, 'null': 2027,\n 'does_not_exist': 2151, 'incorrect_type': 2161\n },\n },\n 'EXCEPTION_DICT': {\n 'Http404': 4005\n }\n}\n# Mathesar settings\nMATHESAR_MODE = decouple_config('MODE', default='PRODUCTION')\nMATHESAR_UI_BUILD_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/mathesar/')\nMATHESAR_MANIFEST_LOCATION = os.path.join(MATHESAR_UI_BUILD_LOCATION, 'manifest.json')\nMATHESAR_CLIENT_DEV_URL = 'http://localhost:3000'\nMATHESAR_UI_SOURCE_LOCATION = os.path.join(BASE_DIR, 'mathesar_ui/')\nMATHESAR_CAPTURE_UNHANDLED_EXCEPTION = decouple_config('CAPTURE_UNHANDLED_EXCEPTION', default=False)\nMATHESAR_STATIC_NON_CODE_FILES_LOCATION = os.path.join(BASE_DIR, 'mathesar/static/non-code/')\n\n# UI source files have to be served by Django in order for static assets to be included during dev mode\n# https://vitejs.dev/guide/assets.html\n# https://vitejs.dev/guide/backend-integration.html\nSTATICFILES_DIRS = [MATHESAR_UI_SOURCE_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION] if MATHESAR_MODE == 'DEVELOPMENT' else [MATHESAR_UI_BUILD_LOCATION, MATHESAR_STATIC_NON_CODE_FILES_LOCATION]\nSTATICFILES_STORAGE = \"whitenoise.storage.CompressedManifestStaticFilesStorage\"\n\n# Accounts\nAUTH_USER_MODEL = 'mathesar.User'\nLOGIN_URL = '/auth/login/'\nLOGIN_REDIRECT_URL = '/'\nLOGOUT_REDIRECT_URL = LOGIN_URL\nDRF_ACCESS_POLICY = {\n 'reusable_conditions': ['mathesar.api.permission_conditions']\n}\n# List of Template names that contains additional script tags to be added to the base template\nBASE_TEMPLATE_ADDITIONAL_SCRIPT_TEMPLATES = []\n", "path": "config/settings/common_settings.py"}, {"content": "from psycopg.errors import InsufficientPrivilege\nfrom sqlalchemy import text\nfrom sqlalchemy.exc import OperationalError, ProgrammingError\n\nfrom db import engine\nfrom db.sql import install as sql_install\nfrom db.types import install as types_install\n\n\ndef install_mathesar(\n database_name, username, password, hostname, port, skip_confirm\n):\n \"\"\"Create database and install Mathesar on it.\"\"\"\n user_db_engine = engine.create_future_engine(\n username, password, hostname, database_name, port,\n connect_args={\"connect_timeout\": 10}\n )\n try:\n user_db_engine.connect()\n print(f\"Installing Mathesar on preexisting PostgreSQL database {database_name} at host {hostname}...\")\n sql_install.install(user_db_engine)\n types_install.install_mathesar_on_database(user_db_engine)\n user_db_engine.dispose()\n except OperationalError:\n database_created = _create_database(\n database_name=database_name,\n hostname=hostname,\n username=username,\n password=password,\n port=port,\n skip_confirm=skip_confirm\n )\n if database_created:\n print(f\"Installing Mathesar on PostgreSQL database {database_name} at host {hostname}...\")\n sql_install.install(user_db_engine)\n types_install.install_mathesar_on_database(user_db_engine)\n user_db_engine.dispose()\n else:\n print(f\"Skipping installing on DB with key {database_name}.\")\n\n\ndef _create_database(database_name, hostname, username, password, port, skip_confirm=True):\n if skip_confirm is True:\n create_database = \"y\"\n else:\n create_database = input(\n f\"Create a new Database called {database_name}? (y/n) > \"\n )\n if create_database.lower() in [\"y\", \"yes\"]:\n # We need to connect to an existing database inorder to create a new Database.\n # So we use the default database `postgres` that comes with postgres.\n # TODO Throw correct error when the default postgres database does not exists(which is very rare but still possible)\n root_database = \"postgres\"\n root_db_engine = engine.create_future_engine(\n username, password, hostname, root_database, port,\n connect_args={\"connect_timeout\": 10}\n )\n try:\n with root_db_engine.connect() as conn:\n conn.execution_options(isolation_level=\"AUTOCOMMIT\")\n conn.execute(text(f'CREATE DATABASE \"{database_name}\"'))\n root_db_engine.dispose()\n print(f\"Created DB is {database_name}.\")\n return True\n except ProgrammingError as e:\n if isinstance(e.orig, InsufficientPrivilege):\n print(f\"Database {database_name} could not be created due to Insufficient Privilege\")\n return False\n except Exception:\n print(f\"Database {database_name} could not be created!\")\n return False\n else:\n print(f\"Database {database_name} not created!\")\n return False\n", "path": "db/install.py"}]}
| 3,908 | 378 |
gh_patches_debug_19759
|
rasdani/github-patches
|
git_diff
|
sql-machine-learning__elasticdl-761
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Provide default gpu resource name and validation
Currently users can only pass "gpu" as part of the resource name. However, k8s requires it to be either `"nvidia.com/gpu"` or `"amd.com/gpu"` if AMD plugin is enabled. There are other different vendors as well but a pattern to use for validation would be `"<vendor>.com/gpu"`.
We should consider adding `"nvidia.com/gpu"` as the default and validate for user provided gpu resource name based on the pattern `"<vendor>.com/gpu"`.
Reference: https://kubernetes.io/docs/tasks/manage-gpus/scheduling-gpus/
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `elasticdl/python/elasticdl/common/k8s_utils.py`
Content:
```
1 import re
2
3
4 _ALLOWED_RESOURCE_TYPES = ["memory", "disk", "ephemeral-storage", "cpu", "gpu"]
5
6
7 def _is_numeric(n):
8 try:
9 float(n)
10 except ValueError:
11 return False
12 return True
13
14
15 def _valid_gpu_spec(gpu_str):
16 if not gpu_str.isnumeric():
17 raise ValueError("invalid gpu request spec: " + gpu_str)
18 return gpu_str
19
20
21 def _valid_cpu_spec(cpu_str):
22 regexp = re.compile("([1-9]{1})([0-9]*)m$")
23 if not regexp.match(cpu_str) and not _is_numeric(cpu_str):
24 raise ValueError("invalid cpu request spec: " + cpu_str)
25 return cpu_str
26
27
28 def _valid_mem_spec(mem_str):
29 regexp = re.compile("([1-9]{1})([0-9]*)(E|P|T|G|M|K|Ei|Pi|Ti|Gi|Mi|Ki)$")
30 if not regexp.match(mem_str):
31 raise ValueError("invalid memory request spec: " + mem_str)
32 return mem_str
33
34
35 def parse_resource(resource_str):
36 """Parse combined k8s resource string into a dict.
37
38 Args:
39 resource_str: The string representation for k8s resource,
40 e.g. "cpu=250m,memory=32Mi,disk=64Mi,gpu=1,ephemeral-storage=32Mi".
41
42 Return:
43 A Python dictionary parsed from the given resource string.
44 """
45 kvs = resource_str.split(",")
46 resource_names = []
47 parsed_res_dict = {}
48 for kv in kvs:
49 k, v = kv.split("=")
50 if k not in resource_names:
51 resource_names.append(k)
52 else:
53 raise ValueError(
54 "The resource string contains duplicate resource names: %s" % k
55 )
56 if k in ["memory", "disk", "ephemeral-storage"]:
57 _valid_mem_spec(v)
58 elif k == "cpu":
59 _valid_cpu_spec(v)
60 elif k == "gpu":
61 _valid_gpu_spec(v)
62 else:
63 raise ValueError(
64 "%s is not in the allowed list of resource types: %s"
65 % (k, _ALLOWED_RESOURCE_TYPES)
66 )
67 parsed_res_dict[k] = v
68 return parsed_res_dict
69
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/elasticdl/python/elasticdl/common/k8s_utils.py b/elasticdl/python/elasticdl/common/k8s_utils.py
--- a/elasticdl/python/elasticdl/common/k8s_utils.py
+++ b/elasticdl/python/elasticdl/common/k8s_utils.py
@@ -2,6 +2,10 @@
_ALLOWED_RESOURCE_TYPES = ["memory", "disk", "ephemeral-storage", "cpu", "gpu"]
+# Any domain name is (syntactically) valid if it's a dot-separated list of
+# identifiers, each no longer than 63 characters, and made up of letters,
+# digits and dashes (no underscores).
+_GPU_VENDOR_REGEX_STR = r"^[a-zA-Z\d-]{,63}(\.[a-zA-Z\d-]{,63})*/gpu$"
def _is_numeric(n):
@@ -57,7 +61,14 @@
_valid_mem_spec(v)
elif k == "cpu":
_valid_cpu_spec(v)
- elif k == "gpu":
+ elif "gpu" in k:
+ if k == "gpu":
+ k = "nvidia.com/gpu"
+ elif not re.compile(_GPU_VENDOR_REGEX_STR).match(k):
+ raise ValueError(
+ "gpu resource name does not have a valid vendor name: %s"
+ % k
+ )
_valid_gpu_spec(v)
else:
raise ValueError(
|
{"golden_diff": "diff --git a/elasticdl/python/elasticdl/common/k8s_utils.py b/elasticdl/python/elasticdl/common/k8s_utils.py\n--- a/elasticdl/python/elasticdl/common/k8s_utils.py\n+++ b/elasticdl/python/elasticdl/common/k8s_utils.py\n@@ -2,6 +2,10 @@\n \n \n _ALLOWED_RESOURCE_TYPES = [\"memory\", \"disk\", \"ephemeral-storage\", \"cpu\", \"gpu\"]\n+# Any domain name is (syntactically) valid if it's a dot-separated list of\n+# identifiers, each no longer than 63 characters, and made up of letters,\n+# digits and dashes (no underscores).\n+_GPU_VENDOR_REGEX_STR = r\"^[a-zA-Z\\d-]{,63}(\\.[a-zA-Z\\d-]{,63})*/gpu$\"\n \n \n def _is_numeric(n):\n@@ -57,7 +61,14 @@\n _valid_mem_spec(v)\n elif k == \"cpu\":\n _valid_cpu_spec(v)\n- elif k == \"gpu\":\n+ elif \"gpu\" in k:\n+ if k == \"gpu\":\n+ k = \"nvidia.com/gpu\"\n+ elif not re.compile(_GPU_VENDOR_REGEX_STR).match(k):\n+ raise ValueError(\n+ \"gpu resource name does not have a valid vendor name: %s\"\n+ % k\n+ )\n _valid_gpu_spec(v)\n else:\n raise ValueError(\n", "issue": "Provide default gpu resource name and validation\nCurrently users can only pass \"gpu\" as part of the resource name. However, k8s requires it to be either `\"nvidia.com/gpu\"` or `\"amd.com/gpu\"` if AMD plugin is enabled. There are other different vendors as well but a pattern to use for validation would be `\"<vendor>.com/gpu\"`.\r\n\r\nWe should consider adding `\"nvidia.com/gpu\"` as the default and validate for user provided gpu resource name based on the pattern `\"<vendor>.com/gpu\"`.\r\n\r\nReference: https://kubernetes.io/docs/tasks/manage-gpus/scheduling-gpus/\n", "before_files": [{"content": "import re\n\n\n_ALLOWED_RESOURCE_TYPES = [\"memory\", \"disk\", \"ephemeral-storage\", \"cpu\", \"gpu\"]\n\n\ndef _is_numeric(n):\n try:\n float(n)\n except ValueError:\n return False\n return True\n\n\ndef _valid_gpu_spec(gpu_str):\n if not gpu_str.isnumeric():\n raise ValueError(\"invalid gpu request spec: \" + gpu_str)\n return gpu_str\n\n\ndef _valid_cpu_spec(cpu_str):\n regexp = re.compile(\"([1-9]{1})([0-9]*)m$\")\n if not regexp.match(cpu_str) and not _is_numeric(cpu_str):\n raise ValueError(\"invalid cpu request spec: \" + cpu_str)\n return cpu_str\n\n\ndef _valid_mem_spec(mem_str):\n regexp = re.compile(\"([1-9]{1})([0-9]*)(E|P|T|G|M|K|Ei|Pi|Ti|Gi|Mi|Ki)$\")\n if not regexp.match(mem_str):\n raise ValueError(\"invalid memory request spec: \" + mem_str)\n return mem_str\n\n\ndef parse_resource(resource_str):\n \"\"\"Parse combined k8s resource string into a dict.\n\n Args:\n resource_str: The string representation for k8s resource,\n e.g. \"cpu=250m,memory=32Mi,disk=64Mi,gpu=1,ephemeral-storage=32Mi\".\n\n Return:\n A Python dictionary parsed from the given resource string.\n \"\"\"\n kvs = resource_str.split(\",\")\n resource_names = []\n parsed_res_dict = {}\n for kv in kvs:\n k, v = kv.split(\"=\")\n if k not in resource_names:\n resource_names.append(k)\n else:\n raise ValueError(\n \"The resource string contains duplicate resource names: %s\" % k\n )\n if k in [\"memory\", \"disk\", \"ephemeral-storage\"]:\n _valid_mem_spec(v)\n elif k == \"cpu\":\n _valid_cpu_spec(v)\n elif k == \"gpu\":\n _valid_gpu_spec(v)\n else:\n raise ValueError(\n \"%s is not in the allowed list of resource types: %s\"\n % (k, _ALLOWED_RESOURCE_TYPES)\n )\n parsed_res_dict[k] = v\n return parsed_res_dict\n", "path": "elasticdl/python/elasticdl/common/k8s_utils.py"}], "after_files": [{"content": "import re\n\n\n_ALLOWED_RESOURCE_TYPES = [\"memory\", \"disk\", \"ephemeral-storage\", \"cpu\", \"gpu\"]\n# Any domain name is (syntactically) valid if it's a dot-separated list of\n# identifiers, each no longer than 63 characters, and made up of letters,\n# digits and dashes (no underscores).\n_GPU_VENDOR_REGEX_STR = r\"^[a-zA-Z\\d-]{,63}(\\.[a-zA-Z\\d-]{,63})*/gpu$\"\n\n\ndef _is_numeric(n):\n try:\n float(n)\n except ValueError:\n return False\n return True\n\n\ndef _valid_gpu_spec(gpu_str):\n if not gpu_str.isnumeric():\n raise ValueError(\"invalid gpu request spec: \" + gpu_str)\n return gpu_str\n\n\ndef _valid_cpu_spec(cpu_str):\n regexp = re.compile(\"([1-9]{1})([0-9]*)m$\")\n if not regexp.match(cpu_str) and not _is_numeric(cpu_str):\n raise ValueError(\"invalid cpu request spec: \" + cpu_str)\n return cpu_str\n\n\ndef _valid_mem_spec(mem_str):\n regexp = re.compile(\"([1-9]{1})([0-9]*)(E|P|T|G|M|K|Ei|Pi|Ti|Gi|Mi|Ki)$\")\n if not regexp.match(mem_str):\n raise ValueError(\"invalid memory request spec: \" + mem_str)\n return mem_str\n\n\ndef parse_resource(resource_str):\n \"\"\"Parse combined k8s resource string into a dict.\n\n Args:\n resource_str: The string representation for k8s resource,\n e.g. \"cpu=250m,memory=32Mi,disk=64Mi,gpu=1,ephemeral-storage=32Mi\".\n\n Return:\n A Python dictionary parsed from the given resource string.\n \"\"\"\n kvs = resource_str.split(\",\")\n resource_names = []\n parsed_res_dict = {}\n for kv in kvs:\n k, v = kv.split(\"=\")\n if k not in resource_names:\n resource_names.append(k)\n else:\n raise ValueError(\n \"The resource string contains duplicate resource names: %s\" % k\n )\n if k in [\"memory\", \"disk\", \"ephemeral-storage\"]:\n _valid_mem_spec(v)\n elif k == \"cpu\":\n _valid_cpu_spec(v)\n elif \"gpu\" in k:\n if k == \"gpu\":\n k = \"nvidia.com/gpu\"\n elif not re.compile(_GPU_VENDOR_REGEX_STR).match(k):\n raise ValueError(\n \"gpu resource name does not have a valid vendor name: %s\"\n % k\n )\n _valid_gpu_spec(v)\n else:\n raise ValueError(\n \"%s is not in the allowed list of resource types: %s\"\n % (k, _ALLOWED_RESOURCE_TYPES)\n )\n parsed_res_dict[k] = v\n return parsed_res_dict\n", "path": "elasticdl/python/elasticdl/common/k8s_utils.py"}]}
| 1,033 | 318 |
gh_patches_debug_778
|
rasdani/github-patches
|
git_diff
|
microsoft__DeepSpeed-2611
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] pydantic DeepSpeedConfigModel has no validator for <class:object>
**Describe the bug**
During ```from deepspeed.inference.config import DtypeEnum```, got error
```
File "pydantic/main.py", line 299, in pydantic.main.ModelMetaclass.__new__
File "pydantic/fields.py", line 411, in pydantic.fields.ModelField.infer
File "pydantic/fields.py", line 342, in pydantic.fields.ModelField.__init__
File "pydantic/fields.py", line 456, in pydantic.fields.ModelField.prepare
File "pydantic/fields.py", line 670, in pydantic.fields.ModelField.populate_validators
File "pydantic/validators.py", line 715, in find_validators
RuntimeError: no validator found for <class 'object'>, see `arbitrary_types_allowed` in Config
```
**To Reproduce**
Steps to reproduce the behavior:
1. Simple inference script to reproduce ```from deepspeed.inference.config import DtypeEnum```
2. pydantic 1.8.2, deepspeed 0.8.0+384f17b
**Expected behavior**
successful import with no error
**ds_report output**
cannot produce due to the same import error
**System info (please complete the following information):**
- OS: Red Hat Enterprise Linux Server 7.9 (Maipo)
- GPU count and types: one machine with 8 A100s, three machines with 8 A100s each
- Hugging Face Transformers 4.19.2, no accelerate
- Python version 3.8.13
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `deepspeed/runtime/config_utils.py`
Content:
```
1 """
2 Copyright (c) Microsoft Corporation
3 Licensed under the MIT license.
4 """
5 """
6 Collection of DeepSpeed configuration utilities
7 """
8 import json
9 import collections
10 import collections.abc
11 from functools import reduce
12 from pydantic import BaseModel
13 from deepspeed.utils import logger
14
15
16 class DeepSpeedConfigModel(BaseModel):
17 """
18 This class should be used as a base for all DeepSpeed configs. It extends
19 pydantic.BaseModel to allow for deprecated fields. To enable this feature,
20 add deprecated=True to pydantic.Field:
21
22 my_dep_field: int = Field(0, deprecated=True)
23
24 Deprecated Field kwargs:
25 - deprecated: [True|False], default False
26 Enables / Disables deprecated fields
27 - new_param: str, default ""
28 Name of the field replacing the deprecated field
29 - set_new_param: [True|False], default True
30 If new_param is provided, enables setting the value of that param with
31 deprecated field value
32 - new_param_fn: callable, default (lambda x: x)
33 If new_param is provided and set_new_param is True, this function will
34 modify the value of the deprecated field before placing that value in
35 the new_param field
36
37 Example:
38 my_new_field is replacing a deprecated my_old_field. The expected type
39 for my_new_field is int while the expected type for my_old_field is
40 str. We want to maintain backward compatibility with our configs, so we
41 define the fields with:
42
43 class MyExampleConfig(DeepSpeedConfigModel):
44 my_new_field: int = 0
45 my_old_field: str = Field('0',
46 deprecated=True,
47 new_param='my_new_field',
48 new_param_fn=(lambda x: int(x)))
49 """
50 def __init__(self, strict=False, **data):
51 if (
52 not strict
53 ): # This is temporary until we refactor all DS configs, allows HF to load models
54 data = {k: v for k, v in data.items() if v != "auto"}
55 super().__init__(**data)
56 self._deprecated_fields_check(self)
57
58 def _process_deprecated_field(self, pydantic_config, field):
59 # Get information about the deprecated field
60 fields_set = pydantic_config.__fields_set__
61 dep_param = field.name
62 kwargs = field.field_info.extra
63 new_param_fn = kwargs.get("new_param_fn", lambda x: x)
64 param_value = new_param_fn(getattr(pydantic_config, dep_param))
65 new_param = kwargs.get("new_param", "")
66 if dep_param in fields_set:
67 logger.warning(f"Config parameter {dep_param} is deprecated" +
68 (f" use {new_param} instead" if new_param else ""))
69 # Check if there is a new param and if it should be set with a value
70 if new_param and kwargs.get("set_new_param", True):
71 # Remove the deprecate field if there is a replacing field
72 try:
73 delattr(pydantic_config, dep_param)
74 except Exception as e:
75 logger.error(f"Tried removing deprecated '{dep_param}' from config")
76 raise e
77
78 # Set new param value
79 new_param_nested = new_param.split(".")
80 if len(new_param_nested) > 1:
81 # If the new param exists in a subconfig, we need to get
82 # the fields set for that subconfig
83 pydantic_config = reduce(getattr,
84 new_param_nested[:-1],
85 pydantic_config)
86 fields_set = pydantic_config.__fields_set__
87 new_param_name = new_param_nested[-1]
88 assert (
89 new_param_name not in fields_set
90 ), f"Cannot provide deprecated parameter '{dep_param}' and replacing parameter '{new_param}' together"
91 # A custom function for converting the old param value to new param value can be provided
92 try:
93 setattr(pydantic_config, new_param_name, param_value)
94 except Exception as e:
95 logger.error(
96 f"Tried setting value for '{new_param}' with value from deprecated '{dep_param}'"
97 )
98 raise e
99
100 def _deprecated_fields_check(self, pydantic_config):
101 fields = pydantic_config.__fields__
102 for field in fields.values():
103 if field.field_info.extra.get("deprecated", False):
104 self._process_deprecated_field(pydantic_config, field)
105
106 class Config:
107 validate_all = True
108 validate_assignment = True
109 use_enum_values = True
110 allow_population_by_field_name = True
111 extra = "forbid"
112
113
114 class pp_int(int):
115 """
116 A wrapper for integers that will return a custom string or comma-formatted
117 string of the integer. For example, print(pp_int(1e5)) will return
118 "10,000". This is useful mainly for auto-generated documentation purposes.
119 """
120 def __new__(cls, val, custom_print_str=None):
121 inst = super().__new__(cls, val)
122 inst.custom_print_str = custom_print_str
123 return inst
124
125 def __repr__(self):
126 if self.custom_print_str:
127 return self.custom_print_str
128 return f"{self.real:,}"
129
130
131 # adapted from https://stackoverflow.com/a/50701137/9201239
132 class ScientificNotationEncoder(json.JSONEncoder):
133 """
134 This class overrides ``json.dumps`` default formatter.
135
136 This version keeps everything as normal except formats numbers bigger than 1e3 using scientific notation.
137
138 Just pass ``cls=ScientificNotationEncoder`` to ``json.dumps`` to activate it
139
140 """
141 def iterencode(self, o, _one_shot=False, level=0):
142 indent = self.indent if self.indent is not None else 4
143 prefix_close = " " * level * indent
144 level += 1
145 prefix = " " * level * indent
146 if isinstance(o, bool):
147 return "true" if o else "false"
148 elif isinstance(o, float) or isinstance(o, int):
149 if o > 1e3:
150 return f"{o:e}"
151 else:
152 return f"{o}"
153 elif isinstance(o, collections.abc.Mapping):
154 x = [
155 f'\n{prefix}"{k}": {self.iterencode(v, level=level)}' for k,
156 v in o.items()
157 ]
158 return "{" + ", ".join(x) + f"\n{prefix_close}" + "}"
159 elif isinstance(o, collections.abc.Sequence) and not isinstance(o, str):
160 return f"[{ f', '.join(map(self.iterencode, o)) }]"
161 return "\n, ".join(super().iterencode(o, _one_shot))
162
163
164 class DeepSpeedConfigObject(object):
165 """
166 For json serialization
167 """
168 def repr(self):
169 return self.__dict__
170
171 def __repr__(self):
172 return json.dumps(
173 self.__dict__,
174 sort_keys=True,
175 indent=4,
176 cls=ScientificNotationEncoder,
177 )
178
179
180 def get_scalar_param(param_dict, param_name, param_default_value):
181 return param_dict.get(param_name, param_default_value)
182
183
184 def get_list_param(param_dict, param_name, param_default_value):
185 return param_dict.get(param_name, param_default_value)
186
187
188 def get_dict_param(param_dict, param_name, param_default_value):
189 return param_dict.get(param_name, param_default_value)
190
191
192 def dict_raise_error_on_duplicate_keys(ordered_pairs):
193 """Reject duplicate keys."""
194 d = dict((k, v) for k, v in ordered_pairs)
195 if len(d) != len(ordered_pairs):
196 counter = collections.Counter([pair[0] for pair in ordered_pairs])
197 keys = [key for key, value in counter.items() if value > 1]
198 raise ValueError("Duplicate keys in DeepSpeed config: {}".format(keys))
199 return d
200
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/deepspeed/runtime/config_utils.py b/deepspeed/runtime/config_utils.py
--- a/deepspeed/runtime/config_utils.py
+++ b/deepspeed/runtime/config_utils.py
@@ -109,6 +109,7 @@
use_enum_values = True
allow_population_by_field_name = True
extra = "forbid"
+ arbitrary_types_allowed = True
class pp_int(int):
|
{"golden_diff": "diff --git a/deepspeed/runtime/config_utils.py b/deepspeed/runtime/config_utils.py\n--- a/deepspeed/runtime/config_utils.py\n+++ b/deepspeed/runtime/config_utils.py\n@@ -109,6 +109,7 @@\n use_enum_values = True\n allow_population_by_field_name = True\n extra = \"forbid\"\n+ arbitrary_types_allowed = True\n \n \n class pp_int(int):\n", "issue": "[BUG] pydantic DeepSpeedConfigModel has no validator for <class:object>\n**Describe the bug**\r\nDuring ```from deepspeed.inference.config import DtypeEnum```, got error \r\n```\r\n File \"pydantic/main.py\", line 299, in pydantic.main.ModelMetaclass.__new__\r\n File \"pydantic/fields.py\", line 411, in pydantic.fields.ModelField.infer\r\n File \"pydantic/fields.py\", line 342, in pydantic.fields.ModelField.__init__\r\n File \"pydantic/fields.py\", line 456, in pydantic.fields.ModelField.prepare\r\n File \"pydantic/fields.py\", line 670, in pydantic.fields.ModelField.populate_validators\r\n File \"pydantic/validators.py\", line 715, in find_validators\r\nRuntimeError: no validator found for <class 'object'>, see `arbitrary_types_allowed` in Config\r\n```\r\n\r\n**To Reproduce**\r\nSteps to reproduce the behavior:\r\n1. Simple inference script to reproduce ```from deepspeed.inference.config import DtypeEnum```\r\n2. pydantic 1.8.2, deepspeed 0.8.0+384f17b\r\n\r\n\r\n**Expected behavior**\r\nsuccessful import with no error\r\n\r\n**ds_report output**\r\ncannot produce due to the same import error\r\n\r\n**System info (please complete the following information):**\r\n - OS: Red Hat Enterprise Linux Server 7.9 (Maipo)\r\n - GPU count and types: one machine with 8 A100s, three machines with 8 A100s each\r\n - Hugging Face Transformers 4.19.2, no accelerate\r\n - Python version 3.8.13\r\n\n", "before_files": [{"content": "\"\"\"\nCopyright (c) Microsoft Corporation\nLicensed under the MIT license.\n\"\"\"\n\"\"\"\nCollection of DeepSpeed configuration utilities\n\"\"\"\nimport json\nimport collections\nimport collections.abc\nfrom functools import reduce\nfrom pydantic import BaseModel\nfrom deepspeed.utils import logger\n\n\nclass DeepSpeedConfigModel(BaseModel):\n \"\"\"\n This class should be used as a base for all DeepSpeed configs. It extends\n pydantic.BaseModel to allow for deprecated fields. To enable this feature,\n add deprecated=True to pydantic.Field:\n\n my_dep_field: int = Field(0, deprecated=True)\n\n Deprecated Field kwargs:\n - deprecated: [True|False], default False\n Enables / Disables deprecated fields\n - new_param: str, default \"\"\n Name of the field replacing the deprecated field\n - set_new_param: [True|False], default True\n If new_param is provided, enables setting the value of that param with\n deprecated field value\n - new_param_fn: callable, default (lambda x: x)\n If new_param is provided and set_new_param is True, this function will\n modify the value of the deprecated field before placing that value in\n the new_param field\n\n Example:\n my_new_field is replacing a deprecated my_old_field. The expected type\n for my_new_field is int while the expected type for my_old_field is\n str. We want to maintain backward compatibility with our configs, so we\n define the fields with:\n\n class MyExampleConfig(DeepSpeedConfigModel):\n my_new_field: int = 0\n my_old_field: str = Field('0',\n deprecated=True,\n new_param='my_new_field',\n new_param_fn=(lambda x: int(x)))\n \"\"\"\n def __init__(self, strict=False, **data):\n if (\n not strict\n ): # This is temporary until we refactor all DS configs, allows HF to load models\n data = {k: v for k, v in data.items() if v != \"auto\"}\n super().__init__(**data)\n self._deprecated_fields_check(self)\n\n def _process_deprecated_field(self, pydantic_config, field):\n # Get information about the deprecated field\n fields_set = pydantic_config.__fields_set__\n dep_param = field.name\n kwargs = field.field_info.extra\n new_param_fn = kwargs.get(\"new_param_fn\", lambda x: x)\n param_value = new_param_fn(getattr(pydantic_config, dep_param))\n new_param = kwargs.get(\"new_param\", \"\")\n if dep_param in fields_set:\n logger.warning(f\"Config parameter {dep_param} is deprecated\" +\n (f\" use {new_param} instead\" if new_param else \"\"))\n # Check if there is a new param and if it should be set with a value\n if new_param and kwargs.get(\"set_new_param\", True):\n # Remove the deprecate field if there is a replacing field\n try:\n delattr(pydantic_config, dep_param)\n except Exception as e:\n logger.error(f\"Tried removing deprecated '{dep_param}' from config\")\n raise e\n\n # Set new param value\n new_param_nested = new_param.split(\".\")\n if len(new_param_nested) > 1:\n # If the new param exists in a subconfig, we need to get\n # the fields set for that subconfig\n pydantic_config = reduce(getattr,\n new_param_nested[:-1],\n pydantic_config)\n fields_set = pydantic_config.__fields_set__\n new_param_name = new_param_nested[-1]\n assert (\n new_param_name not in fields_set\n ), f\"Cannot provide deprecated parameter '{dep_param}' and replacing parameter '{new_param}' together\"\n # A custom function for converting the old param value to new param value can be provided\n try:\n setattr(pydantic_config, new_param_name, param_value)\n except Exception as e:\n logger.error(\n f\"Tried setting value for '{new_param}' with value from deprecated '{dep_param}'\"\n )\n raise e\n\n def _deprecated_fields_check(self, pydantic_config):\n fields = pydantic_config.__fields__\n for field in fields.values():\n if field.field_info.extra.get(\"deprecated\", False):\n self._process_deprecated_field(pydantic_config, field)\n\n class Config:\n validate_all = True\n validate_assignment = True\n use_enum_values = True\n allow_population_by_field_name = True\n extra = \"forbid\"\n\n\nclass pp_int(int):\n \"\"\"\n A wrapper for integers that will return a custom string or comma-formatted\n string of the integer. For example, print(pp_int(1e5)) will return\n \"10,000\". This is useful mainly for auto-generated documentation purposes.\n \"\"\"\n def __new__(cls, val, custom_print_str=None):\n inst = super().__new__(cls, val)\n inst.custom_print_str = custom_print_str\n return inst\n\n def __repr__(self):\n if self.custom_print_str:\n return self.custom_print_str\n return f\"{self.real:,}\"\n\n\n# adapted from https://stackoverflow.com/a/50701137/9201239\nclass ScientificNotationEncoder(json.JSONEncoder):\n \"\"\"\n This class overrides ``json.dumps`` default formatter.\n\n This version keeps everything as normal except formats numbers bigger than 1e3 using scientific notation.\n\n Just pass ``cls=ScientificNotationEncoder`` to ``json.dumps`` to activate it\n\n \"\"\"\n def iterencode(self, o, _one_shot=False, level=0):\n indent = self.indent if self.indent is not None else 4\n prefix_close = \" \" * level * indent\n level += 1\n prefix = \" \" * level * indent\n if isinstance(o, bool):\n return \"true\" if o else \"false\"\n elif isinstance(o, float) or isinstance(o, int):\n if o > 1e3:\n return f\"{o:e}\"\n else:\n return f\"{o}\"\n elif isinstance(o, collections.abc.Mapping):\n x = [\n f'\\n{prefix}\"{k}\": {self.iterencode(v, level=level)}' for k,\n v in o.items()\n ]\n return \"{\" + \", \".join(x) + f\"\\n{prefix_close}\" + \"}\"\n elif isinstance(o, collections.abc.Sequence) and not isinstance(o, str):\n return f\"[{ f', '.join(map(self.iterencode, o)) }]\"\n return \"\\n, \".join(super().iterencode(o, _one_shot))\n\n\nclass DeepSpeedConfigObject(object):\n \"\"\"\n For json serialization\n \"\"\"\n def repr(self):\n return self.__dict__\n\n def __repr__(self):\n return json.dumps(\n self.__dict__,\n sort_keys=True,\n indent=4,\n cls=ScientificNotationEncoder,\n )\n\n\ndef get_scalar_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef get_list_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef get_dict_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef dict_raise_error_on_duplicate_keys(ordered_pairs):\n \"\"\"Reject duplicate keys.\"\"\"\n d = dict((k, v) for k, v in ordered_pairs)\n if len(d) != len(ordered_pairs):\n counter = collections.Counter([pair[0] for pair in ordered_pairs])\n keys = [key for key, value in counter.items() if value > 1]\n raise ValueError(\"Duplicate keys in DeepSpeed config: {}\".format(keys))\n return d\n", "path": "deepspeed/runtime/config_utils.py"}], "after_files": [{"content": "\"\"\"\nCopyright (c) Microsoft Corporation\nLicensed under the MIT license.\n\"\"\"\n\"\"\"\nCollection of DeepSpeed configuration utilities\n\"\"\"\nimport json\nimport collections\nimport collections.abc\nfrom functools import reduce\nfrom pydantic import BaseModel\nfrom deepspeed.utils import logger\n\n\nclass DeepSpeedConfigModel(BaseModel):\n \"\"\"\n This class should be used as a base for all DeepSpeed configs. It extends\n pydantic.BaseModel to allow for deprecated fields. To enable this feature,\n add deprecated=True to pydantic.Field:\n\n my_dep_field: int = Field(0, deprecated=True)\n\n Deprecated Field kwargs:\n - deprecated: [True|False], default False\n Enables / Disables deprecated fields\n - new_param: str, default \"\"\n Name of the field replacing the deprecated field\n - set_new_param: [True|False], default True\n If new_param is provided, enables setting the value of that param with\n deprecated field value\n - new_param_fn: callable, default (lambda x: x)\n If new_param is provided and set_new_param is True, this function will\n modify the value of the deprecated field before placing that value in\n the new_param field\n\n Example:\n my_new_field is replacing a deprecated my_old_field. The expected type\n for my_new_field is int while the expected type for my_old_field is\n str. We want to maintain backward compatibility with our configs, so we\n define the fields with:\n\n class MyExampleConfig(DeepSpeedConfigModel):\n my_new_field: int = 0\n my_old_field: str = Field('0',\n deprecated=True,\n new_param='my_new_field',\n new_param_fn=(lambda x: int(x)))\n \"\"\"\n def __init__(self, strict=False, **data):\n if (\n not strict\n ): # This is temporary until we refactor all DS configs, allows HF to load models\n data = {k: v for k, v in data.items() if v != \"auto\"}\n super().__init__(**data)\n self._deprecated_fields_check(self)\n\n def _process_deprecated_field(self, pydantic_config, field):\n # Get information about the deprecated field\n fields_set = pydantic_config.__fields_set__\n dep_param = field.name\n kwargs = field.field_info.extra\n new_param_fn = kwargs.get(\"new_param_fn\", lambda x: x)\n param_value = new_param_fn(getattr(pydantic_config, dep_param))\n new_param = kwargs.get(\"new_param\", \"\")\n if dep_param in fields_set:\n logger.warning(f\"Config parameter {dep_param} is deprecated\" +\n (f\" use {new_param} instead\" if new_param else \"\"))\n # Check if there is a new param and if it should be set with a value\n if new_param and kwargs.get(\"set_new_param\", True):\n # Remove the deprecate field if there is a replacing field\n try:\n delattr(pydantic_config, dep_param)\n except Exception as e:\n logger.error(f\"Tried removing deprecated '{dep_param}' from config\")\n raise e\n\n # Set new param value\n new_param_nested = new_param.split(\".\")\n if len(new_param_nested) > 1:\n # If the new param exists in a subconfig, we need to get\n # the fields set for that subconfig\n pydantic_config = reduce(getattr,\n new_param_nested[:-1],\n pydantic_config)\n fields_set = pydantic_config.__fields_set__\n new_param_name = new_param_nested[-1]\n assert (\n new_param_name not in fields_set\n ), f\"Cannot provide deprecated parameter '{dep_param}' and replacing parameter '{new_param}' together\"\n # A custom function for converting the old param value to new param value can be provided\n try:\n setattr(pydantic_config, new_param_name, param_value)\n except Exception as e:\n logger.error(\n f\"Tried setting value for '{new_param}' with value from deprecated '{dep_param}'\"\n )\n raise e\n\n def _deprecated_fields_check(self, pydantic_config):\n fields = pydantic_config.__fields__\n for field in fields.values():\n if field.field_info.extra.get(\"deprecated\", False):\n self._process_deprecated_field(pydantic_config, field)\n\n class Config:\n validate_all = True\n validate_assignment = True\n use_enum_values = True\n allow_population_by_field_name = True\n extra = \"forbid\"\n arbitrary_types_allowed = True\n\n\nclass pp_int(int):\n \"\"\"\n A wrapper for integers that will return a custom string or comma-formatted\n string of the integer. For example, print(pp_int(1e5)) will return\n \"10,000\". This is useful mainly for auto-generated documentation purposes.\n \"\"\"\n def __new__(cls, val, custom_print_str=None):\n inst = super().__new__(cls, val)\n inst.custom_print_str = custom_print_str\n return inst\n\n def __repr__(self):\n if self.custom_print_str:\n return self.custom_print_str\n return f\"{self.real:,}\"\n\n\n# adapted from https://stackoverflow.com/a/50701137/9201239\nclass ScientificNotationEncoder(json.JSONEncoder):\n \"\"\"\n This class overrides ``json.dumps`` default formatter.\n\n This version keeps everything as normal except formats numbers bigger than 1e3 using scientific notation.\n\n Just pass ``cls=ScientificNotationEncoder`` to ``json.dumps`` to activate it\n\n \"\"\"\n def iterencode(self, o, _one_shot=False, level=0):\n indent = self.indent if self.indent is not None else 4\n prefix_close = \" \" * level * indent\n level += 1\n prefix = \" \" * level * indent\n if isinstance(o, bool):\n return \"true\" if o else \"false\"\n elif isinstance(o, float) or isinstance(o, int):\n if o > 1e3:\n return f\"{o:e}\"\n else:\n return f\"{o}\"\n elif isinstance(o, collections.abc.Mapping):\n x = [\n f'\\n{prefix}\"{k}\": {self.iterencode(v, level=level)}' for k,\n v in o.items()\n ]\n return \"{\" + \", \".join(x) + f\"\\n{prefix_close}\" + \"}\"\n elif isinstance(o, collections.abc.Sequence) and not isinstance(o, str):\n return f\"[{ f', '.join(map(self.iterencode, o)) }]\"\n return \"\\n, \".join(super().iterencode(o, _one_shot))\n\n\nclass DeepSpeedConfigObject(object):\n \"\"\"\n For json serialization\n \"\"\"\n def repr(self):\n return self.__dict__\n\n def __repr__(self):\n return json.dumps(\n self.__dict__,\n sort_keys=True,\n indent=4,\n cls=ScientificNotationEncoder,\n )\n\n\ndef get_scalar_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef get_list_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef get_dict_param(param_dict, param_name, param_default_value):\n return param_dict.get(param_name, param_default_value)\n\n\ndef dict_raise_error_on_duplicate_keys(ordered_pairs):\n \"\"\"Reject duplicate keys.\"\"\"\n d = dict((k, v) for k, v in ordered_pairs)\n if len(d) != len(ordered_pairs):\n counter = collections.Counter([pair[0] for pair in ordered_pairs])\n keys = [key for key, value in counter.items() if value > 1]\n raise ValueError(\"Duplicate keys in DeepSpeed config: {}\".format(keys))\n return d\n", "path": "deepspeed/runtime/config_utils.py"}]}
| 2,835 | 91 |
gh_patches_debug_43770
|
rasdani/github-patches
|
git_diff
|
beetbox__beets-1384
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
duplicates: Do not report duplicates when only `mb_albumid` is duplicated
```
$ beet duplicates -F
/home/simon/Música/2 Unlimited/Get Ready/03 - The Magic Friend.mp3
/home/simon/Música/2 Unlimited/Get Ready/04 - Contrast.mp3
/home/simon/Música/2 Unlimited/Get Ready/05 - Desire.mp3
/home/simon/Música/2 Unlimited/Get Ready/06 - Pacific Walk.mp3
/home/simon/Música/2 Unlimited/Get Ready/07 - Workaholic.mp3
/home/simon/Música/2 Unlimited/Get Ready/08 - Rougher Than The Average.mp3
/home/simon/Música/2 Unlimited/Get Ready/09 - Delight.mp3
/home/simon/Música/2 Unlimited/Get Ready/10 - Eternally Yours.mp3
/home/simon/Música/2 Unlimited/Get Ready/11 - Twilight Zone (Rave Version Edit).mp3
/home/simon/Música/2 Unlimited/Get Ready/12 - Get Ready For This (Vocal Version).mp3
/home/simon/Música/2 Unlimited/Get Ready/13 - The Magic Friend (Instrumental).mp3
/home/simon/Música/2 Unlimited/Get Ready/14 - Workaholic (Instrumental).mp3
```
```
$ ll Música/2\ Unlimited/Get\ Ready/
total 83M
-rw-rw-r-- 1 simon simon 7,5M oct 10 16:00 03 - The Magic Friend.mp3
-rw-rw-r-- 1 simon simon 6,4M oct 10 16:00 04 - Contrast.mp3
-rw-rw-r-- 1 simon simon 7,9M oct 10 16:00 05 - Desire.mp3
-rw-rw-r-- 1 simon simon 5,7M oct 10 16:00 06 - Pacific Walk.mp3
-rw-rw-r-- 1 simon simon 7,2M oct 10 16:00 07 - Workaholic.mp3
-rw-rw-r-- 1 simon simon 7,6M oct 10 16:00 08 - Rougher Than The Average.mp3
-rw-rw-r-- 1 simon simon 6,5M oct 10 16:00 09 - Delight.mp3
-rw-rw-r-- 1 simon simon 7,6M oct 10 16:00 10 - Eternally Yours.mp3
-rw-rw-r-- 1 simon simon 7,1M oct 10 16:00 11 - Twilight Zone (Rave Version Edit).mp3
-rw-rw-r-- 1 simon simon 6,4M oct 10 16:00 12 - Get Ready For This (Vocal Version).mp3
-rw-rw-r-- 1 simon simon 5,9M oct 10 16:00 13 - The Magic Friend (Instrumental).mp3
-rw-rw-r-- 1 simon simon 7,2M oct 10 16:00 14 - Workaholic (Instrumental).mp3
-rw-rw-r-- 1 simon simon 30K oct 10 16:00 cover.jpg
```
What is the duplicate of that?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `beetsplug/duplicates.py`
Content:
```
1 # This file is part of beets.
2 # Copyright 2015, Pedro Silva.
3 #
4 # Permission is hereby granted, free of charge, to any person obtaining
5 # a copy of this software and associated documentation files (the
6 # "Software"), to deal in the Software without restriction, including
7 # without limitation the rights to use, copy, modify, merge, publish,
8 # distribute, sublicense, and/or sell copies of the Software, and to
9 # permit persons to whom the Software is furnished to do so, subject to
10 # the following conditions:
11 #
12 # The above copyright notice and this permission notice shall be
13 # included in all copies or substantial portions of the Software.
14
15 """List duplicate tracks or albums.
16 """
17 from __future__ import (division, absolute_import, print_function,
18 unicode_literals)
19
20 import shlex
21
22 from beets.plugins import BeetsPlugin
23 from beets.ui import decargs, print_, vararg_callback, Subcommand, UserError
24 from beets.util import command_output, displayable_path, subprocess
25
26 PLUGIN = 'duplicates'
27
28
29 def _process_item(item, lib, copy=False, move=False, delete=False,
30 tag=False, fmt=''):
31 """Process Item `item` in `lib`.
32 """
33 if copy:
34 item.move(basedir=copy, copy=True)
35 item.store()
36 if move:
37 item.move(basedir=move, copy=False)
38 item.store()
39 if delete:
40 item.remove(delete=True)
41 if tag:
42 try:
43 k, v = tag.split('=')
44 except:
45 raise UserError('%s: can\'t parse k=v tag: %s' % (PLUGIN, tag))
46 setattr(k, v)
47 item.store()
48 print_(format(item, fmt))
49
50
51 def _checksum(item, prog, log):
52 """Run external `prog` on file path associated with `item`, cache
53 output as flexattr on a key that is the name of the program, and
54 return the key, checksum tuple.
55 """
56 args = [p.format(file=item.path) for p in shlex.split(prog)]
57 key = args[0]
58 checksum = getattr(item, key, False)
59 if not checksum:
60 log.debug(u'{0}: key {1} on item {2} not cached: computing checksum',
61 PLUGIN, key, displayable_path(item.path))
62 try:
63 checksum = command_output(args)
64 setattr(item, key, checksum)
65 item.store()
66 log.debug(u'{0}: computed checksum for {1} using {2}',
67 PLUGIN, item.title, key)
68 except subprocess.CalledProcessError as e:
69 log.debug(u'{0}: failed to checksum {1}: {2}',
70 PLUGIN, displayable_path(item.path), e)
71 else:
72 log.debug(u'{0}: key {1} on item {2} cached: not computing checksum',
73 PLUGIN, key, displayable_path(item.path))
74 return key, checksum
75
76
77 def _group_by(objs, keys, log):
78 """Return a dictionary with keys arbitrary concatenations of attributes and
79 values lists of objects (Albums or Items) with those keys.
80 """
81 import collections
82 counts = collections.defaultdict(list)
83 for obj in objs:
84 values = [getattr(obj, k, None) for k in keys]
85 values = [v for v in values if v not in (None, '')]
86 if values:
87 key = '\001'.join(values)
88 counts[key].append(obj)
89 else:
90 log.debug(u'{0}: all keys {1} on item {2} are null: skipping',
91 PLUGIN, keys, displayable_path(obj.path))
92
93 return counts
94
95
96 def _duplicates(objs, keys, full, log):
97 """Generate triples of keys, duplicate counts, and constituent objects.
98 """
99 offset = 0 if full else 1
100 for k, objs in _group_by(objs, keys, log).iteritems():
101 if len(objs) > 1:
102 yield (k, len(objs) - offset, objs[offset:])
103
104
105 class DuplicatesPlugin(BeetsPlugin):
106 """List duplicate tracks or albums
107 """
108 def __init__(self):
109 super(DuplicatesPlugin, self).__init__()
110
111 self.config.add({
112 'format': '',
113 'count': False,
114 'album': False,
115 'full': False,
116 'path': False,
117 'keys': ['mb_trackid', 'mb_albumid'],
118 'checksum': None,
119 'copy': False,
120 'move': False,
121 'delete': False,
122 'tag': False,
123 })
124
125 self._command = Subcommand('duplicates',
126 help=__doc__,
127 aliases=['dup'])
128 self._command.parser.add_option('-c', '--count', dest='count',
129 action='store_true',
130 help='show duplicate counts')
131
132 self._command.parser.add_option('-C', '--checksum', dest='checksum',
133 action='store', metavar='PROG',
134 help='report duplicates based on'
135 ' arbitrary command')
136
137 self._command.parser.add_option('-d', '--delete', dest='delete',
138 action='store_true',
139 help='delete items from library and '
140 'disk')
141
142 self._command.parser.add_option('-F', '--full', dest='full',
143 action='store_true',
144 help='show all versions of duplicate'
145 ' tracks or albums')
146
147 self._command.parser.add_option('-k', '--keys', dest='keys',
148 action='callback', metavar='KEY1 KEY2',
149 callback=vararg_callback,
150 help='report duplicates based on keys')
151
152 self._command.parser.add_option('-m', '--move', dest='move',
153 action='store', metavar='DEST',
154 help='move items to dest')
155
156 self._command.parser.add_option('-o', '--copy', dest='copy',
157 action='store', metavar='DEST',
158 help='copy items to dest')
159
160 self._command.parser.add_option('-t', '--tag', dest='tag',
161 action='store',
162 help='tag matched items with \'k=v\''
163 ' attribute')
164 self._command.parser.add_all_common_options()
165
166 def commands(self):
167
168 def _dup(lib, opts, args):
169 self.config.set_args(opts)
170 fmt = self.config['format'].get()
171 album = self.config['album'].get(bool)
172 full = self.config['full'].get(bool)
173 keys = self.config['keys'].get()
174 checksum = self.config['checksum'].get()
175 copy = self.config['copy'].get()
176 move = self.config['move'].get()
177 delete = self.config['delete'].get(bool)
178 tag = self.config['tag'].get()
179
180 if album:
181 keys = ['mb_albumid']
182 items = lib.albums(decargs(args))
183 else:
184 items = lib.items(decargs(args))
185
186 if self.config['path']:
187 fmt = '$path'
188
189 # Default format string for count mode.
190 if self.config['count'] and not fmt:
191 if album:
192 fmt = '$albumartist - $album'
193 else:
194 fmt = '$albumartist - $album - $title'
195 fmt += ': {0}'
196
197 if checksum:
198 if not isinstance(checksum, basestring):
199 raise UserError(
200 'duplicates: "checksum" option must be a command'
201 )
202 for i in items:
203 k, _ = self._checksum(i, checksum, self._log)
204 keys = [k]
205
206 for obj_id, obj_count, objs in _duplicates(items,
207 keys=keys,
208 full=full,
209 log=self._log):
210 if obj_id: # Skip empty IDs.
211 for o in objs:
212 _process_item(o, lib,
213 copy=copy,
214 move=move,
215 delete=delete,
216 tag=tag,
217 fmt=fmt.format(obj_count))
218
219 self._command.func = _dup
220 return [self._command]
221
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/beetsplug/duplicates.py b/beetsplug/duplicates.py
--- a/beetsplug/duplicates.py
+++ b/beetsplug/duplicates.py
@@ -74,30 +74,37 @@
return key, checksum
-def _group_by(objs, keys, log):
+def _group_by(objs, keys, strict, log):
"""Return a dictionary with keys arbitrary concatenations of attributes and
values lists of objects (Albums or Items) with those keys.
+
+ If strict, all attributes must be defined for a duplicate match.
"""
import collections
counts = collections.defaultdict(list)
for obj in objs:
values = [getattr(obj, k, None) for k in keys]
values = [v for v in values if v not in (None, '')]
- if values:
+ if strict and len(values) < len(keys):
+ log.debug(u'{0}: some keys {1} on item {2} are null or empty: '
+ 'skipping',
+ PLUGIN, keys, displayable_path(obj.path))
+ elif (not strict and not len(values)):
+ log.debug(u'{0}: all keys {1} on item {2} are null or empty: '
+ 'skipping',
+ PLUGIN, keys, displayable_path(obj.path))
+ else:
key = '\001'.join(values)
counts[key].append(obj)
- else:
- log.debug(u'{0}: all keys {1} on item {2} are null: skipping',
- PLUGIN, keys, displayable_path(obj.path))
return counts
-def _duplicates(objs, keys, full, log):
+def _duplicates(objs, keys, full, strict, log):
"""Generate triples of keys, duplicate counts, and constituent objects.
"""
offset = 0 if full else 1
- for k, objs in _group_by(objs, keys, log).iteritems():
+ for k, objs in _group_by(objs, keys, strict, log).iteritems():
if len(objs) > 1:
yield (k, len(objs) - offset, objs[offset:])
@@ -113,6 +120,7 @@
'count': False,
'album': False,
'full': False,
+ 'strict': False,
'path': False,
'keys': ['mb_trackid', 'mb_albumid'],
'checksum': None,
@@ -144,6 +152,11 @@
help='show all versions of duplicate'
' tracks or albums')
+ self._command.parser.add_option('-s', '--strict', dest='strict',
+ action='store_true',
+ help='report duplicates only if all'
+ ' attributes are set')
+
self._command.parser.add_option('-k', '--keys', dest='keys',
action='callback', metavar='KEY1 KEY2',
callback=vararg_callback,
@@ -170,6 +183,7 @@
fmt = self.config['format'].get()
album = self.config['album'].get(bool)
full = self.config['full'].get(bool)
+ strict = self.config['strict'].get(bool)
keys = self.config['keys'].get()
checksum = self.config['checksum'].get()
copy = self.config['copy'].get()
@@ -206,6 +220,7 @@
for obj_id, obj_count, objs in _duplicates(items,
keys=keys,
full=full,
+ strict=strict,
log=self._log):
if obj_id: # Skip empty IDs.
for o in objs:
|
{"golden_diff": "diff --git a/beetsplug/duplicates.py b/beetsplug/duplicates.py\n--- a/beetsplug/duplicates.py\n+++ b/beetsplug/duplicates.py\n@@ -74,30 +74,37 @@\n return key, checksum\n \n \n-def _group_by(objs, keys, log):\n+def _group_by(objs, keys, strict, log):\n \"\"\"Return a dictionary with keys arbitrary concatenations of attributes and\n values lists of objects (Albums or Items) with those keys.\n+\n+ If strict, all attributes must be defined for a duplicate match.\n \"\"\"\n import collections\n counts = collections.defaultdict(list)\n for obj in objs:\n values = [getattr(obj, k, None) for k in keys]\n values = [v for v in values if v not in (None, '')]\n- if values:\n+ if strict and len(values) < len(keys):\n+ log.debug(u'{0}: some keys {1} on item {2} are null or empty: '\n+ 'skipping',\n+ PLUGIN, keys, displayable_path(obj.path))\n+ elif (not strict and not len(values)):\n+ log.debug(u'{0}: all keys {1} on item {2} are null or empty: '\n+ 'skipping',\n+ PLUGIN, keys, displayable_path(obj.path))\n+ else:\n key = '\\001'.join(values)\n counts[key].append(obj)\n- else:\n- log.debug(u'{0}: all keys {1} on item {2} are null: skipping',\n- PLUGIN, keys, displayable_path(obj.path))\n \n return counts\n \n \n-def _duplicates(objs, keys, full, log):\n+def _duplicates(objs, keys, full, strict, log):\n \"\"\"Generate triples of keys, duplicate counts, and constituent objects.\n \"\"\"\n offset = 0 if full else 1\n- for k, objs in _group_by(objs, keys, log).iteritems():\n+ for k, objs in _group_by(objs, keys, strict, log).iteritems():\n if len(objs) > 1:\n yield (k, len(objs) - offset, objs[offset:])\n \n@@ -113,6 +120,7 @@\n 'count': False,\n 'album': False,\n 'full': False,\n+ 'strict': False,\n 'path': False,\n 'keys': ['mb_trackid', 'mb_albumid'],\n 'checksum': None,\n@@ -144,6 +152,11 @@\n help='show all versions of duplicate'\n ' tracks or albums')\n \n+ self._command.parser.add_option('-s', '--strict', dest='strict',\n+ action='store_true',\n+ help='report duplicates only if all'\n+ ' attributes are set')\n+\n self._command.parser.add_option('-k', '--keys', dest='keys',\n action='callback', metavar='KEY1 KEY2',\n callback=vararg_callback,\n@@ -170,6 +183,7 @@\n fmt = self.config['format'].get()\n album = self.config['album'].get(bool)\n full = self.config['full'].get(bool)\n+ strict = self.config['strict'].get(bool)\n keys = self.config['keys'].get()\n checksum = self.config['checksum'].get()\n copy = self.config['copy'].get()\n@@ -206,6 +220,7 @@\n for obj_id, obj_count, objs in _duplicates(items,\n keys=keys,\n full=full,\n+ strict=strict,\n log=self._log):\n if obj_id: # Skip empty IDs.\n for o in objs:\n", "issue": "duplicates: Do not report duplicates when only `mb_albumid` is duplicated\n```\n$ beet duplicates -F\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/03 - The Magic Friend.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/04 - Contrast.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/05 - Desire.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/06 - Pacific Walk.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/07 - Workaholic.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/08 - Rougher Than The Average.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/09 - Delight.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/10 - Eternally Yours.mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/11 - Twilight Zone (Rave Version Edit).mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/12 - Get Ready For This (Vocal Version).mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/13 - The Magic Friend (Instrumental).mp3\n/home/simon/M\u00fasica/2 Unlimited/Get Ready/14 - Workaholic (Instrumental).mp3\n```\n\n```\n$ ll M\u00fasica/2\\ Unlimited/Get\\ Ready/\ntotal 83M\n-rw-rw-r-- 1 simon simon 7,5M oct 10 16:00 03 - The Magic Friend.mp3\n-rw-rw-r-- 1 simon simon 6,4M oct 10 16:00 04 - Contrast.mp3\n-rw-rw-r-- 1 simon simon 7,9M oct 10 16:00 05 - Desire.mp3\n-rw-rw-r-- 1 simon simon 5,7M oct 10 16:00 06 - Pacific Walk.mp3\n-rw-rw-r-- 1 simon simon 7,2M oct 10 16:00 07 - Workaholic.mp3\n-rw-rw-r-- 1 simon simon 7,6M oct 10 16:00 08 - Rougher Than The Average.mp3\n-rw-rw-r-- 1 simon simon 6,5M oct 10 16:00 09 - Delight.mp3\n-rw-rw-r-- 1 simon simon 7,6M oct 10 16:00 10 - Eternally Yours.mp3\n-rw-rw-r-- 1 simon simon 7,1M oct 10 16:00 11 - Twilight Zone (Rave Version Edit).mp3\n-rw-rw-r-- 1 simon simon 6,4M oct 10 16:00 12 - Get Ready For This (Vocal Version).mp3\n-rw-rw-r-- 1 simon simon 5,9M oct 10 16:00 13 - The Magic Friend (Instrumental).mp3\n-rw-rw-r-- 1 simon simon 7,2M oct 10 16:00 14 - Workaholic (Instrumental).mp3\n-rw-rw-r-- 1 simon simon 30K oct 10 16:00 cover.jpg\n```\n\nWhat is the duplicate of that?\n\n", "before_files": [{"content": "# This file is part of beets.\n# Copyright 2015, Pedro Silva.\n#\n# Permission is hereby granted, free of charge, to any person obtaining\n# a copy of this software and associated documentation files (the\n# \"Software\"), to deal in the Software without restriction, including\n# without limitation the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the Software, and to\n# permit persons to whom the Software is furnished to do so, subject to\n# the following conditions:\n#\n# The above copyright notice and this permission notice shall be\n# included in all copies or substantial portions of the Software.\n\n\"\"\"List duplicate tracks or albums.\n\"\"\"\nfrom __future__ import (division, absolute_import, print_function,\n unicode_literals)\n\nimport shlex\n\nfrom beets.plugins import BeetsPlugin\nfrom beets.ui import decargs, print_, vararg_callback, Subcommand, UserError\nfrom beets.util import command_output, displayable_path, subprocess\n\nPLUGIN = 'duplicates'\n\n\ndef _process_item(item, lib, copy=False, move=False, delete=False,\n tag=False, fmt=''):\n \"\"\"Process Item `item` in `lib`.\n \"\"\"\n if copy:\n item.move(basedir=copy, copy=True)\n item.store()\n if move:\n item.move(basedir=move, copy=False)\n item.store()\n if delete:\n item.remove(delete=True)\n if tag:\n try:\n k, v = tag.split('=')\n except:\n raise UserError('%s: can\\'t parse k=v tag: %s' % (PLUGIN, tag))\n setattr(k, v)\n item.store()\n print_(format(item, fmt))\n\n\ndef _checksum(item, prog, log):\n \"\"\"Run external `prog` on file path associated with `item`, cache\n output as flexattr on a key that is the name of the program, and\n return the key, checksum tuple.\n \"\"\"\n args = [p.format(file=item.path) for p in shlex.split(prog)]\n key = args[0]\n checksum = getattr(item, key, False)\n if not checksum:\n log.debug(u'{0}: key {1} on item {2} not cached: computing checksum',\n PLUGIN, key, displayable_path(item.path))\n try:\n checksum = command_output(args)\n setattr(item, key, checksum)\n item.store()\n log.debug(u'{0}: computed checksum for {1} using {2}',\n PLUGIN, item.title, key)\n except subprocess.CalledProcessError as e:\n log.debug(u'{0}: failed to checksum {1}: {2}',\n PLUGIN, displayable_path(item.path), e)\n else:\n log.debug(u'{0}: key {1} on item {2} cached: not computing checksum',\n PLUGIN, key, displayable_path(item.path))\n return key, checksum\n\n\ndef _group_by(objs, keys, log):\n \"\"\"Return a dictionary with keys arbitrary concatenations of attributes and\n values lists of objects (Albums or Items) with those keys.\n \"\"\"\n import collections\n counts = collections.defaultdict(list)\n for obj in objs:\n values = [getattr(obj, k, None) for k in keys]\n values = [v for v in values if v not in (None, '')]\n if values:\n key = '\\001'.join(values)\n counts[key].append(obj)\n else:\n log.debug(u'{0}: all keys {1} on item {2} are null: skipping',\n PLUGIN, keys, displayable_path(obj.path))\n\n return counts\n\n\ndef _duplicates(objs, keys, full, log):\n \"\"\"Generate triples of keys, duplicate counts, and constituent objects.\n \"\"\"\n offset = 0 if full else 1\n for k, objs in _group_by(objs, keys, log).iteritems():\n if len(objs) > 1:\n yield (k, len(objs) - offset, objs[offset:])\n\n\nclass DuplicatesPlugin(BeetsPlugin):\n \"\"\"List duplicate tracks or albums\n \"\"\"\n def __init__(self):\n super(DuplicatesPlugin, self).__init__()\n\n self.config.add({\n 'format': '',\n 'count': False,\n 'album': False,\n 'full': False,\n 'path': False,\n 'keys': ['mb_trackid', 'mb_albumid'],\n 'checksum': None,\n 'copy': False,\n 'move': False,\n 'delete': False,\n 'tag': False,\n })\n\n self._command = Subcommand('duplicates',\n help=__doc__,\n aliases=['dup'])\n self._command.parser.add_option('-c', '--count', dest='count',\n action='store_true',\n help='show duplicate counts')\n\n self._command.parser.add_option('-C', '--checksum', dest='checksum',\n action='store', metavar='PROG',\n help='report duplicates based on'\n ' arbitrary command')\n\n self._command.parser.add_option('-d', '--delete', dest='delete',\n action='store_true',\n help='delete items from library and '\n 'disk')\n\n self._command.parser.add_option('-F', '--full', dest='full',\n action='store_true',\n help='show all versions of duplicate'\n ' tracks or albums')\n\n self._command.parser.add_option('-k', '--keys', dest='keys',\n action='callback', metavar='KEY1 KEY2',\n callback=vararg_callback,\n help='report duplicates based on keys')\n\n self._command.parser.add_option('-m', '--move', dest='move',\n action='store', metavar='DEST',\n help='move items to dest')\n\n self._command.parser.add_option('-o', '--copy', dest='copy',\n action='store', metavar='DEST',\n help='copy items to dest')\n\n self._command.parser.add_option('-t', '--tag', dest='tag',\n action='store',\n help='tag matched items with \\'k=v\\''\n ' attribute')\n self._command.parser.add_all_common_options()\n\n def commands(self):\n\n def _dup(lib, opts, args):\n self.config.set_args(opts)\n fmt = self.config['format'].get()\n album = self.config['album'].get(bool)\n full = self.config['full'].get(bool)\n keys = self.config['keys'].get()\n checksum = self.config['checksum'].get()\n copy = self.config['copy'].get()\n move = self.config['move'].get()\n delete = self.config['delete'].get(bool)\n tag = self.config['tag'].get()\n\n if album:\n keys = ['mb_albumid']\n items = lib.albums(decargs(args))\n else:\n items = lib.items(decargs(args))\n\n if self.config['path']:\n fmt = '$path'\n\n # Default format string for count mode.\n if self.config['count'] and not fmt:\n if album:\n fmt = '$albumartist - $album'\n else:\n fmt = '$albumartist - $album - $title'\n fmt += ': {0}'\n\n if checksum:\n if not isinstance(checksum, basestring):\n raise UserError(\n 'duplicates: \"checksum\" option must be a command'\n )\n for i in items:\n k, _ = self._checksum(i, checksum, self._log)\n keys = [k]\n\n for obj_id, obj_count, objs in _duplicates(items,\n keys=keys,\n full=full,\n log=self._log):\n if obj_id: # Skip empty IDs.\n for o in objs:\n _process_item(o, lib,\n copy=copy,\n move=move,\n delete=delete,\n tag=tag,\n fmt=fmt.format(obj_count))\n\n self._command.func = _dup\n return [self._command]\n", "path": "beetsplug/duplicates.py"}], "after_files": [{"content": "# This file is part of beets.\n# Copyright 2015, Pedro Silva.\n#\n# Permission is hereby granted, free of charge, to any person obtaining\n# a copy of this software and associated documentation files (the\n# \"Software\"), to deal in the Software without restriction, including\n# without limitation the rights to use, copy, modify, merge, publish,\n# distribute, sublicense, and/or sell copies of the Software, and to\n# permit persons to whom the Software is furnished to do so, subject to\n# the following conditions:\n#\n# The above copyright notice and this permission notice shall be\n# included in all copies or substantial portions of the Software.\n\n\"\"\"List duplicate tracks or albums.\n\"\"\"\nfrom __future__ import (division, absolute_import, print_function,\n unicode_literals)\n\nimport shlex\n\nfrom beets.plugins import BeetsPlugin\nfrom beets.ui import decargs, print_, vararg_callback, Subcommand, UserError\nfrom beets.util import command_output, displayable_path, subprocess\n\nPLUGIN = 'duplicates'\n\n\ndef _process_item(item, lib, copy=False, move=False, delete=False,\n tag=False, fmt=''):\n \"\"\"Process Item `item` in `lib`.\n \"\"\"\n if copy:\n item.move(basedir=copy, copy=True)\n item.store()\n if move:\n item.move(basedir=move, copy=False)\n item.store()\n if delete:\n item.remove(delete=True)\n if tag:\n try:\n k, v = tag.split('=')\n except:\n raise UserError('%s: can\\'t parse k=v tag: %s' % (PLUGIN, tag))\n setattr(k, v)\n item.store()\n print_(format(item, fmt))\n\n\ndef _checksum(item, prog, log):\n \"\"\"Run external `prog` on file path associated with `item`, cache\n output as flexattr on a key that is the name of the program, and\n return the key, checksum tuple.\n \"\"\"\n args = [p.format(file=item.path) for p in shlex.split(prog)]\n key = args[0]\n checksum = getattr(item, key, False)\n if not checksum:\n log.debug(u'{0}: key {1} on item {2} not cached: computing checksum',\n PLUGIN, key, displayable_path(item.path))\n try:\n checksum = command_output(args)\n setattr(item, key, checksum)\n item.store()\n log.debug(u'{0}: computed checksum for {1} using {2}',\n PLUGIN, item.title, key)\n except subprocess.CalledProcessError as e:\n log.debug(u'{0}: failed to checksum {1}: {2}',\n PLUGIN, displayable_path(item.path), e)\n else:\n log.debug(u'{0}: key {1} on item {2} cached: not computing checksum',\n PLUGIN, key, displayable_path(item.path))\n return key, checksum\n\n\ndef _group_by(objs, keys, strict, log):\n \"\"\"Return a dictionary with keys arbitrary concatenations of attributes and\n values lists of objects (Albums or Items) with those keys.\n\n If strict, all attributes must be defined for a duplicate match.\n \"\"\"\n import collections\n counts = collections.defaultdict(list)\n for obj in objs:\n values = [getattr(obj, k, None) for k in keys]\n values = [v for v in values if v not in (None, '')]\n if strict and len(values) < len(keys):\n log.debug(u'{0}: some keys {1} on item {2} are null or empty: '\n 'skipping',\n PLUGIN, keys, displayable_path(obj.path))\n elif (not strict and not len(values)):\n log.debug(u'{0}: all keys {1} on item {2} are null or empty: '\n 'skipping',\n PLUGIN, keys, displayable_path(obj.path))\n else:\n key = '\\001'.join(values)\n counts[key].append(obj)\n\n return counts\n\n\ndef _duplicates(objs, keys, full, strict, log):\n \"\"\"Generate triples of keys, duplicate counts, and constituent objects.\n \"\"\"\n offset = 0 if full else 1\n for k, objs in _group_by(objs, keys, strict, log).iteritems():\n if len(objs) > 1:\n yield (k, len(objs) - offset, objs[offset:])\n\n\nclass DuplicatesPlugin(BeetsPlugin):\n \"\"\"List duplicate tracks or albums\n \"\"\"\n def __init__(self):\n super(DuplicatesPlugin, self).__init__()\n\n self.config.add({\n 'format': '',\n 'count': False,\n 'album': False,\n 'full': False,\n 'strict': False,\n 'path': False,\n 'keys': ['mb_trackid', 'mb_albumid'],\n 'checksum': None,\n 'copy': False,\n 'move': False,\n 'delete': False,\n 'tag': False,\n })\n\n self._command = Subcommand('duplicates',\n help=__doc__,\n aliases=['dup'])\n self._command.parser.add_option('-c', '--count', dest='count',\n action='store_true',\n help='show duplicate counts')\n\n self._command.parser.add_option('-C', '--checksum', dest='checksum',\n action='store', metavar='PROG',\n help='report duplicates based on'\n ' arbitrary command')\n\n self._command.parser.add_option('-d', '--delete', dest='delete',\n action='store_true',\n help='delete items from library and '\n 'disk')\n\n self._command.parser.add_option('-F', '--full', dest='full',\n action='store_true',\n help='show all versions of duplicate'\n ' tracks or albums')\n\n self._command.parser.add_option('-s', '--strict', dest='strict',\n action='store_true',\n help='report duplicates only if all'\n ' attributes are set')\n\n self._command.parser.add_option('-k', '--keys', dest='keys',\n action='callback', metavar='KEY1 KEY2',\n callback=vararg_callback,\n help='report duplicates based on keys')\n\n self._command.parser.add_option('-m', '--move', dest='move',\n action='store', metavar='DEST',\n help='move items to dest')\n\n self._command.parser.add_option('-o', '--copy', dest='copy',\n action='store', metavar='DEST',\n help='copy items to dest')\n\n self._command.parser.add_option('-t', '--tag', dest='tag',\n action='store',\n help='tag matched items with \\'k=v\\''\n ' attribute')\n self._command.parser.add_all_common_options()\n\n def commands(self):\n\n def _dup(lib, opts, args):\n self.config.set_args(opts)\n fmt = self.config['format'].get()\n album = self.config['album'].get(bool)\n full = self.config['full'].get(bool)\n strict = self.config['strict'].get(bool)\n keys = self.config['keys'].get()\n checksum = self.config['checksum'].get()\n copy = self.config['copy'].get()\n move = self.config['move'].get()\n delete = self.config['delete'].get(bool)\n tag = self.config['tag'].get()\n\n if album:\n keys = ['mb_albumid']\n items = lib.albums(decargs(args))\n else:\n items = lib.items(decargs(args))\n\n if self.config['path']:\n fmt = '$path'\n\n # Default format string for count mode.\n if self.config['count'] and not fmt:\n if album:\n fmt = '$albumartist - $album'\n else:\n fmt = '$albumartist - $album - $title'\n fmt += ': {0}'\n\n if checksum:\n if not isinstance(checksum, basestring):\n raise UserError(\n 'duplicates: \"checksum\" option must be a command'\n )\n for i in items:\n k, _ = self._checksum(i, checksum, self._log)\n keys = [k]\n\n for obj_id, obj_count, objs in _duplicates(items,\n keys=keys,\n full=full,\n strict=strict,\n log=self._log):\n if obj_id: # Skip empty IDs.\n for o in objs:\n _process_item(o, lib,\n copy=copy,\n move=move,\n delete=delete,\n tag=tag,\n fmt=fmt.format(obj_count))\n\n self._command.func = _dup\n return [self._command]\n", "path": "beetsplug/duplicates.py"}]}
| 3,358 | 815 |
gh_patches_debug_12859
|
rasdani/github-patches
|
git_diff
|
secdev__scapy-2556
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
plain_str not consistent between Py2 and Py3
**Python 2 (expected result)**
```python
>>> plain_str(b'\x3e\xb4\xef\x2c\x11')
'>\xb4\xef,\x11'
```
**Python 3 (unexpected result)**
```python
>>> plain_str(b'\x3e\xb4\xef\x2c\x11')
'>,\x11'
```
**Cause** is `errors="ignore"`
```python
def plain_str(x):
"""Convert basic byte objects to str"""
if isinstance(x, bytes):
return x.decode(errors="ignore")
return str(x)
```
It just ignores values that are not in a valid range for the codec.
**Possible fix**
```python
def plain_str(x):
"""Convert basic byte objects to str"""
if isinstance(x, bytes):
return x.decode(errors="backslashreplace")
return str(x)
```
But this breaks compatibility for Python 3.4
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `scapy/compat.py`
Content:
```
1 # This file is part of Scapy
2 # See http://www.secdev.org/projects/scapy for more information
3 # Copyright (C) Philippe Biondi <[email protected]>
4 # Copyright (C) Gabriel Potter <[email protected]>
5 # This program is published under a GPLv2 license
6
7 """
8 Python 2 and 3 link classes.
9 """
10
11 from __future__ import absolute_import
12 import base64
13 import binascii
14 import gzip
15 import struct
16
17 import scapy.modules.six as six
18
19 ###########
20 # Python3 #
21 ###########
22
23
24 def lambda_tuple_converter(func):
25 """
26 Converts a Python 2 function as
27 lambda (x,y): x + y
28 In the Python 3 format:
29 lambda x,y : x + y
30 """
31 if func is not None and func.__code__.co_argcount == 1:
32 return lambda *args: func(args[0] if len(args) == 1 else args)
33 else:
34 return func
35
36
37 if six.PY2:
38 bytes_encode = plain_str = str
39 chb = lambda x: x if isinstance(x, str) else chr(x)
40 orb = ord
41
42 def raw(x):
43 """Builds a packet and returns its bytes representation.
44 This function is and always be cross-version compatible"""
45 if hasattr(x, "__bytes__"):
46 return x.__bytes__()
47 return bytes(x)
48 else:
49 def raw(x):
50 """Builds a packet and returns its bytes representation.
51 This function is and always be cross-version compatible"""
52 return bytes(x)
53
54 def bytes_encode(x):
55 """Ensure that the given object is bytes.
56 If the parameter is a packet, raw() should be preferred.
57 """
58 if isinstance(x, str):
59 return x.encode()
60 return bytes(x)
61
62 def plain_str(x):
63 """Convert basic byte objects to str"""
64 if isinstance(x, bytes):
65 return x.decode(errors="ignore")
66 return str(x)
67
68 def chb(x):
69 """Same than chr() but encode as bytes."""
70 return struct.pack("!B", x)
71
72 def orb(x):
73 """Return ord(x) when not already an int."""
74 if isinstance(x, int):
75 return x
76 return ord(x)
77
78
79 def bytes_hex(x):
80 """Hexify a str or a bytes object"""
81 return binascii.b2a_hex(bytes_encode(x))
82
83
84 def hex_bytes(x):
85 """De-hexify a str or a byte object"""
86 return binascii.a2b_hex(bytes_encode(x))
87
88
89 def base64_bytes(x):
90 """Turn base64 into bytes"""
91 if six.PY2:
92 return base64.decodestring(x)
93 return base64.decodebytes(bytes_encode(x))
94
95
96 def bytes_base64(x):
97 """Turn bytes into base64"""
98 if six.PY2:
99 return base64.encodestring(x).replace('\n', '')
100 return base64.encodebytes(bytes_encode(x)).replace(b'\n', b'')
101
102
103 if six.PY2:
104 from StringIO import StringIO
105
106 def gzip_decompress(x):
107 """Decompress using gzip"""
108 with gzip.GzipFile(fileobj=StringIO(x), mode='rb') as fdesc:
109 return fdesc.read()
110
111 def gzip_compress(x):
112 """Compress using gzip"""
113 buf = StringIO()
114 with gzip.GzipFile(fileobj=buf, mode='wb') as fdesc:
115 fdesc.write(x)
116 return buf.getvalue()
117 else:
118 gzip_decompress = gzip.decompress
119 gzip_compress = gzip.compress
120
121 # Typing compatibility
122
123 try:
124 # Only required if using mypy-lang for static typing
125 from typing import Optional, List, Union, Callable, Any, AnyStr, Tuple, \
126 Sized, Dict, Pattern, cast
127 except ImportError:
128 # Let's make some fake ones.
129
130 def cast(_type, obj):
131 return obj
132
133 class _FakeType(object):
134 # make the objects subscriptable indefinetly
135 def __getitem__(self, item):
136 return _FakeType()
137
138 Optional = _FakeType()
139 Union = _FakeType()
140 Callable = _FakeType()
141 List = _FakeType()
142 Dict = _FakeType()
143 Any = _FakeType()
144 AnyStr = _FakeType()
145 Tuple = _FakeType()
146 Pattern = _FakeType()
147
148 class Sized(object):
149 pass
150
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/scapy/compat.py b/scapy/compat.py
--- a/scapy/compat.py
+++ b/scapy/compat.py
@@ -59,11 +59,19 @@
return x.encode()
return bytes(x)
- def plain_str(x):
- """Convert basic byte objects to str"""
- if isinstance(x, bytes):
- return x.decode(errors="ignore")
- return str(x)
+ if six.PY34:
+ def plain_str(x):
+ """Convert basic byte objects to str"""
+ if isinstance(x, bytes):
+ return x.decode(errors="ignore")
+ return str(x)
+ else:
+ # Python 3.5+
+ def plain_str(x):
+ """Convert basic byte objects to str"""
+ if isinstance(x, bytes):
+ return x.decode(errors="backslashreplace")
+ return str(x)
def chb(x):
"""Same than chr() but encode as bytes."""
|
{"golden_diff": "diff --git a/scapy/compat.py b/scapy/compat.py\n--- a/scapy/compat.py\n+++ b/scapy/compat.py\n@@ -59,11 +59,19 @@\n return x.encode()\n return bytes(x)\n \n- def plain_str(x):\n- \"\"\"Convert basic byte objects to str\"\"\"\n- if isinstance(x, bytes):\n- return x.decode(errors=\"ignore\")\n- return str(x)\n+ if six.PY34:\n+ def plain_str(x):\n+ \"\"\"Convert basic byte objects to str\"\"\"\n+ if isinstance(x, bytes):\n+ return x.decode(errors=\"ignore\")\n+ return str(x)\n+ else:\n+ # Python 3.5+\n+ def plain_str(x):\n+ \"\"\"Convert basic byte objects to str\"\"\"\n+ if isinstance(x, bytes):\n+ return x.decode(errors=\"backslashreplace\")\n+ return str(x)\n \n def chb(x):\n \"\"\"Same than chr() but encode as bytes.\"\"\"\n", "issue": "plain_str not consistent between Py2 and Py3\n**Python 2 (expected result)**\r\n```python\r\n>>> plain_str(b'\\x3e\\xb4\\xef\\x2c\\x11')\r\n'>\\xb4\\xef,\\x11'\r\n```\r\n\r\n**Python 3 (unexpected result)**\r\n```python\r\n>>> plain_str(b'\\x3e\\xb4\\xef\\x2c\\x11') \r\n'>,\\x11'\r\n```\r\n\r\n**Cause** is `errors=\"ignore\"`\r\n```python\r\ndef plain_str(x): \r\n \"\"\"Convert basic byte objects to str\"\"\"\r\n if isinstance(x, bytes): \r\n return x.decode(errors=\"ignore\") \r\n return str(x) \r\n```\r\n\r\nIt just ignores values that are not in a valid range for the codec.\r\n\r\n**Possible fix**\r\n```python\r\ndef plain_str(x): \r\n \"\"\"Convert basic byte objects to str\"\"\"\r\n if isinstance(x, bytes): \r\n return x.decode(errors=\"backslashreplace\") \r\n return str(x) \r\n``` \r\nBut this breaks compatibility for Python 3.4\r\n\r\n\n", "before_files": [{"content": "# This file is part of Scapy\n# See http://www.secdev.org/projects/scapy for more information\n# Copyright (C) Philippe Biondi <[email protected]>\n# Copyright (C) Gabriel Potter <[email protected]>\n# This program is published under a GPLv2 license\n\n\"\"\"\nPython 2 and 3 link classes.\n\"\"\"\n\nfrom __future__ import absolute_import\nimport base64\nimport binascii\nimport gzip\nimport struct\n\nimport scapy.modules.six as six\n\n###########\n# Python3 #\n###########\n\n\ndef lambda_tuple_converter(func):\n \"\"\"\n Converts a Python 2 function as\n lambda (x,y): x + y\n In the Python 3 format:\n lambda x,y : x + y\n \"\"\"\n if func is not None and func.__code__.co_argcount == 1:\n return lambda *args: func(args[0] if len(args) == 1 else args)\n else:\n return func\n\n\nif six.PY2:\n bytes_encode = plain_str = str\n chb = lambda x: x if isinstance(x, str) else chr(x)\n orb = ord\n\n def raw(x):\n \"\"\"Builds a packet and returns its bytes representation.\n This function is and always be cross-version compatible\"\"\"\n if hasattr(x, \"__bytes__\"):\n return x.__bytes__()\n return bytes(x)\nelse:\n def raw(x):\n \"\"\"Builds a packet and returns its bytes representation.\n This function is and always be cross-version compatible\"\"\"\n return bytes(x)\n\n def bytes_encode(x):\n \"\"\"Ensure that the given object is bytes.\n If the parameter is a packet, raw() should be preferred.\n \"\"\"\n if isinstance(x, str):\n return x.encode()\n return bytes(x)\n\n def plain_str(x):\n \"\"\"Convert basic byte objects to str\"\"\"\n if isinstance(x, bytes):\n return x.decode(errors=\"ignore\")\n return str(x)\n\n def chb(x):\n \"\"\"Same than chr() but encode as bytes.\"\"\"\n return struct.pack(\"!B\", x)\n\n def orb(x):\n \"\"\"Return ord(x) when not already an int.\"\"\"\n if isinstance(x, int):\n return x\n return ord(x)\n\n\ndef bytes_hex(x):\n \"\"\"Hexify a str or a bytes object\"\"\"\n return binascii.b2a_hex(bytes_encode(x))\n\n\ndef hex_bytes(x):\n \"\"\"De-hexify a str or a byte object\"\"\"\n return binascii.a2b_hex(bytes_encode(x))\n\n\ndef base64_bytes(x):\n \"\"\"Turn base64 into bytes\"\"\"\n if six.PY2:\n return base64.decodestring(x)\n return base64.decodebytes(bytes_encode(x))\n\n\ndef bytes_base64(x):\n \"\"\"Turn bytes into base64\"\"\"\n if six.PY2:\n return base64.encodestring(x).replace('\\n', '')\n return base64.encodebytes(bytes_encode(x)).replace(b'\\n', b'')\n\n\nif six.PY2:\n from StringIO import StringIO\n\n def gzip_decompress(x):\n \"\"\"Decompress using gzip\"\"\"\n with gzip.GzipFile(fileobj=StringIO(x), mode='rb') as fdesc:\n return fdesc.read()\n\n def gzip_compress(x):\n \"\"\"Compress using gzip\"\"\"\n buf = StringIO()\n with gzip.GzipFile(fileobj=buf, mode='wb') as fdesc:\n fdesc.write(x)\n return buf.getvalue()\nelse:\n gzip_decompress = gzip.decompress\n gzip_compress = gzip.compress\n\n# Typing compatibility\n\ntry:\n # Only required if using mypy-lang for static typing\n from typing import Optional, List, Union, Callable, Any, AnyStr, Tuple, \\\n Sized, Dict, Pattern, cast\nexcept ImportError:\n # Let's make some fake ones.\n\n def cast(_type, obj):\n return obj\n\n class _FakeType(object):\n # make the objects subscriptable indefinetly\n def __getitem__(self, item):\n return _FakeType()\n\n Optional = _FakeType()\n Union = _FakeType()\n Callable = _FakeType()\n List = _FakeType()\n Dict = _FakeType()\n Any = _FakeType()\n AnyStr = _FakeType()\n Tuple = _FakeType()\n Pattern = _FakeType()\n\n class Sized(object):\n pass\n", "path": "scapy/compat.py"}], "after_files": [{"content": "# This file is part of Scapy\n# See http://www.secdev.org/projects/scapy for more information\n# Copyright (C) Philippe Biondi <[email protected]>\n# Copyright (C) Gabriel Potter <[email protected]>\n# This program is published under a GPLv2 license\n\n\"\"\"\nPython 2 and 3 link classes.\n\"\"\"\n\nfrom __future__ import absolute_import\nimport base64\nimport binascii\nimport gzip\nimport struct\n\nimport scapy.modules.six as six\n\n###########\n# Python3 #\n###########\n\n\ndef lambda_tuple_converter(func):\n \"\"\"\n Converts a Python 2 function as\n lambda (x,y): x + y\n In the Python 3 format:\n lambda x,y : x + y\n \"\"\"\n if func is not None and func.__code__.co_argcount == 1:\n return lambda *args: func(args[0] if len(args) == 1 else args)\n else:\n return func\n\n\nif six.PY2:\n bytes_encode = plain_str = str\n chb = lambda x: x if isinstance(x, str) else chr(x)\n orb = ord\n\n def raw(x):\n \"\"\"Builds a packet and returns its bytes representation.\n This function is and always be cross-version compatible\"\"\"\n if hasattr(x, \"__bytes__\"):\n return x.__bytes__()\n return bytes(x)\nelse:\n def raw(x):\n \"\"\"Builds a packet and returns its bytes representation.\n This function is and always be cross-version compatible\"\"\"\n return bytes(x)\n\n def bytes_encode(x):\n \"\"\"Ensure that the given object is bytes.\n If the parameter is a packet, raw() should be preferred.\n \"\"\"\n if isinstance(x, str):\n return x.encode()\n return bytes(x)\n\n if six.PY34:\n def plain_str(x):\n \"\"\"Convert basic byte objects to str\"\"\"\n if isinstance(x, bytes):\n return x.decode(errors=\"ignore\")\n return str(x)\n else:\n # Python 3.5+\n def plain_str(x):\n \"\"\"Convert basic byte objects to str\"\"\"\n if isinstance(x, bytes):\n return x.decode(errors=\"backslashreplace\")\n return str(x)\n\n def chb(x):\n \"\"\"Same than chr() but encode as bytes.\"\"\"\n return struct.pack(\"!B\", x)\n\n def orb(x):\n \"\"\"Return ord(x) when not already an int.\"\"\"\n if isinstance(x, int):\n return x\n return ord(x)\n\n\ndef bytes_hex(x):\n \"\"\"Hexify a str or a bytes object\"\"\"\n return binascii.b2a_hex(bytes_encode(x))\n\n\ndef hex_bytes(x):\n \"\"\"De-hexify a str or a byte object\"\"\"\n return binascii.a2b_hex(bytes_encode(x))\n\n\ndef base64_bytes(x):\n \"\"\"Turn base64 into bytes\"\"\"\n if six.PY2:\n return base64.decodestring(x)\n return base64.decodebytes(bytes_encode(x))\n\n\ndef bytes_base64(x):\n \"\"\"Turn bytes into base64\"\"\"\n if six.PY2:\n return base64.encodestring(x).replace('\\n', '')\n return base64.encodebytes(bytes_encode(x)).replace(b'\\n', b'')\n\n\nif six.PY2:\n from StringIO import StringIO\n\n def gzip_decompress(x):\n \"\"\"Decompress using gzip\"\"\"\n with gzip.GzipFile(fileobj=StringIO(x), mode='rb') as fdesc:\n return fdesc.read()\n\n def gzip_compress(x):\n \"\"\"Compress using gzip\"\"\"\n buf = StringIO()\n with gzip.GzipFile(fileobj=buf, mode='wb') as fdesc:\n fdesc.write(x)\n return buf.getvalue()\nelse:\n gzip_decompress = gzip.decompress\n gzip_compress = gzip.compress\n\n# Typing compatibility\n\ntry:\n # Only required if using mypy-lang for static typing\n from typing import Optional, List, Union, Callable, Any, AnyStr, Tuple, \\\n Sized, Dict, Pattern, cast\nexcept ImportError:\n # Let's make some fake ones.\n\n def cast(_type, obj):\n return obj\n\n class _FakeType(object):\n # make the objects subscriptable indefinetly\n def __getitem__(self, item):\n return _FakeType()\n\n Optional = _FakeType()\n Union = _FakeType()\n Callable = _FakeType()\n List = _FakeType()\n Dict = _FakeType()\n Any = _FakeType()\n AnyStr = _FakeType()\n Tuple = _FakeType()\n Pattern = _FakeType()\n\n class Sized(object):\n pass\n", "path": "scapy/compat.py"}]}
| 1,792 | 219 |
gh_patches_debug_36045
|
rasdani/github-patches
|
git_diff
|
pydantic__pydantic-3176
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
underscore_attrs_are_private and validate_all not compatible with validate_arguments
Output of `python -c "import pydantic.utils; print(pydantic.utils.version_info())"`:
```
pydantic version: 1.8.2
pydantic compiled: True
install path: /home/xxx/.conda/envs/py38/lib/python3.8/site-packages/pydantic
python version: 3.8.5 (default, Sep 4 2020, 07:30:14) [GCC 7.3.0]
platform: Linux-4.18.0-80.el8.x86_64-x86_64-with-glibc2.10
optional deps. installed: ['typing-extensions']
```
`underscore_attrs_are_private` not work with `validate_arguments` just like it not work with `create_model` #3134 .
When I define a function decorated by `validate_arguments` with `validate_all` been set `True`, It will raise `ValidationError`:
```py
from datetime import datetime
from pydantic import validate_arguments, Field
@validate_arguments(config={'validate_all': True})
def foo(dt: datetime = Field(default_factory=lambda : 42)):
print(type(dt))
foo()
```
```
---------------------------------------------------------------------------
ValidationError Traceback (most recent call last)
<ipython-input-4-6c97ede6fd47> in <module>
7 print(type(dt))
8
----> 9 foo()
~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.validate_arguments.validate.wrapper_function()
~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.ValidatedFunction.call()
~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.ValidatedFunction.init_model_instance()
~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/main.cpython-38-x86_64-linux-gnu.so in pydantic.main.BaseModel.__init__()
ValidationError: 3 validation errors for Foo
v__duplicate_kwargs
object of type 'NoneType' has no len() (type=type_error)
args
object of type 'NoneType' has no len() (type=type_error)
kwargs
object of type 'NoneType' has no len() (type=type_error)
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pydantic/decorator.py`
Content:
```
1 from functools import wraps
2 from typing import TYPE_CHECKING, Any, Callable, Dict, List, Mapping, Optional, Tuple, Type, TypeVar, Union, overload
3
4 from . import validator
5 from .config import Extra
6 from .errors import ConfigError
7 from .main import BaseModel, create_model
8 from .typing import get_all_type_hints
9 from .utils import to_camel
10
11 __all__ = ('validate_arguments',)
12
13 if TYPE_CHECKING:
14 from .typing import AnyCallable
15
16 AnyCallableT = TypeVar('AnyCallableT', bound=AnyCallable)
17 ConfigType = Union[None, Type[Any], Dict[str, Any]]
18
19
20 @overload
21 def validate_arguments(func: None = None, *, config: 'ConfigType' = None) -> Callable[['AnyCallableT'], 'AnyCallableT']:
22 ...
23
24
25 @overload
26 def validate_arguments(func: 'AnyCallableT') -> 'AnyCallableT':
27 ...
28
29
30 def validate_arguments(func: Optional['AnyCallableT'] = None, *, config: 'ConfigType' = None) -> Any:
31 """
32 Decorator to validate the arguments passed to a function.
33 """
34
35 def validate(_func: 'AnyCallable') -> 'AnyCallable':
36 vd = ValidatedFunction(_func, config)
37
38 @wraps(_func)
39 def wrapper_function(*args: Any, **kwargs: Any) -> Any:
40 return vd.call(*args, **kwargs)
41
42 wrapper_function.vd = vd # type: ignore
43 wrapper_function.validate = vd.init_model_instance # type: ignore
44 wrapper_function.raw_function = vd.raw_function # type: ignore
45 wrapper_function.model = vd.model # type: ignore
46 return wrapper_function
47
48 if func:
49 return validate(func)
50 else:
51 return validate
52
53
54 ALT_V_ARGS = 'v__args'
55 ALT_V_KWARGS = 'v__kwargs'
56 V_POSITIONAL_ONLY_NAME = 'v__positional_only'
57 V_DUPLICATE_KWARGS = 'v__duplicate_kwargs'
58
59
60 class ValidatedFunction:
61 def __init__(self, function: 'AnyCallableT', config: 'ConfigType'): # noqa C901
62 from inspect import Parameter, signature
63
64 parameters: Mapping[str, Parameter] = signature(function).parameters
65
66 if parameters.keys() & {ALT_V_ARGS, ALT_V_KWARGS, V_POSITIONAL_ONLY_NAME, V_DUPLICATE_KWARGS}:
67 raise ConfigError(
68 f'"{ALT_V_ARGS}", "{ALT_V_KWARGS}", "{V_POSITIONAL_ONLY_NAME}" and "{V_DUPLICATE_KWARGS}" '
69 f'are not permitted as argument names when using the "{validate_arguments.__name__}" decorator'
70 )
71
72 self.raw_function = function
73 self.arg_mapping: Dict[int, str] = {}
74 self.positional_only_args = set()
75 self.v_args_name = 'args'
76 self.v_kwargs_name = 'kwargs'
77
78 type_hints = get_all_type_hints(function)
79 takes_args = False
80 takes_kwargs = False
81 fields: Dict[str, Tuple[Any, Any]] = {}
82 for i, (name, p) in enumerate(parameters.items()):
83 if p.annotation is p.empty:
84 annotation = Any
85 else:
86 annotation = type_hints[name]
87
88 default = ... if p.default is p.empty else p.default
89 if p.kind == Parameter.POSITIONAL_ONLY:
90 self.arg_mapping[i] = name
91 fields[name] = annotation, default
92 fields[V_POSITIONAL_ONLY_NAME] = List[str], None
93 self.positional_only_args.add(name)
94 elif p.kind == Parameter.POSITIONAL_OR_KEYWORD:
95 self.arg_mapping[i] = name
96 fields[name] = annotation, default
97 fields[V_DUPLICATE_KWARGS] = List[str], None
98 elif p.kind == Parameter.KEYWORD_ONLY:
99 fields[name] = annotation, default
100 elif p.kind == Parameter.VAR_POSITIONAL:
101 self.v_args_name = name
102 fields[name] = Tuple[annotation, ...], None
103 takes_args = True
104 else:
105 assert p.kind == Parameter.VAR_KEYWORD, p.kind
106 self.v_kwargs_name = name
107 fields[name] = Dict[str, annotation], None # type: ignore
108 takes_kwargs = True
109
110 # these checks avoid a clash between "args" and a field with that name
111 if not takes_args and self.v_args_name in fields:
112 self.v_args_name = ALT_V_ARGS
113
114 # same with "kwargs"
115 if not takes_kwargs and self.v_kwargs_name in fields:
116 self.v_kwargs_name = ALT_V_KWARGS
117
118 if not takes_args:
119 # we add the field so validation below can raise the correct exception
120 fields[self.v_args_name] = List[Any], None
121
122 if not takes_kwargs:
123 # same with kwargs
124 fields[self.v_kwargs_name] = Dict[Any, Any], None
125
126 self.create_model(fields, takes_args, takes_kwargs, config)
127
128 def init_model_instance(self, *args: Any, **kwargs: Any) -> BaseModel:
129 values = self.build_values(args, kwargs)
130 return self.model(**values)
131
132 def call(self, *args: Any, **kwargs: Any) -> Any:
133 m = self.init_model_instance(*args, **kwargs)
134 return self.execute(m)
135
136 def build_values(self, args: Tuple[Any, ...], kwargs: Dict[str, Any]) -> Dict[str, Any]:
137 values: Dict[str, Any] = {}
138 if args:
139 arg_iter = enumerate(args)
140 while True:
141 try:
142 i, a = next(arg_iter)
143 except StopIteration:
144 break
145 arg_name = self.arg_mapping.get(i)
146 if arg_name is not None:
147 values[arg_name] = a
148 else:
149 values[self.v_args_name] = [a] + [a for _, a in arg_iter]
150 break
151
152 var_kwargs = {}
153 wrong_positional_args = []
154 duplicate_kwargs = []
155 non_var_fields = set(self.model.__fields__) - {self.v_args_name, self.v_kwargs_name}
156 for k, v in kwargs.items():
157 if k in non_var_fields:
158 if k in self.positional_only_args:
159 wrong_positional_args.append(k)
160 if k in values:
161 duplicate_kwargs.append(k)
162 values[k] = v
163 else:
164 var_kwargs[k] = v
165
166 if var_kwargs:
167 values[self.v_kwargs_name] = var_kwargs
168 if wrong_positional_args:
169 values[V_POSITIONAL_ONLY_NAME] = wrong_positional_args
170 if duplicate_kwargs:
171 values[V_DUPLICATE_KWARGS] = duplicate_kwargs
172 return values
173
174 def execute(self, m: BaseModel) -> Any:
175 d = {k: v for k, v in m._iter() if k in m.__fields_set__ or m.__fields__[k].default_factory}
176 var_kwargs = d.pop(self.v_kwargs_name, {})
177
178 if self.v_args_name in d:
179 args_: List[Any] = []
180 in_kwargs = False
181 kwargs = {}
182 for name, value in d.items():
183 if in_kwargs:
184 kwargs[name] = value
185 elif name == self.v_args_name:
186 args_ += value
187 in_kwargs = True
188 else:
189 args_.append(value)
190 return self.raw_function(*args_, **kwargs, **var_kwargs)
191 elif self.positional_only_args:
192 args_ = []
193 kwargs = {}
194 for name, value in d.items():
195 if name in self.positional_only_args:
196 args_.append(value)
197 else:
198 kwargs[name] = value
199 return self.raw_function(*args_, **kwargs, **var_kwargs)
200 else:
201 return self.raw_function(**d, **var_kwargs)
202
203 def create_model(self, fields: Dict[str, Any], takes_args: bool, takes_kwargs: bool, config: 'ConfigType') -> None:
204 pos_args = len(self.arg_mapping)
205
206 class CustomConfig:
207 pass
208
209 if not TYPE_CHECKING: # pragma: no branch
210 if isinstance(config, dict):
211 CustomConfig = type('Config', (), config) # noqa: F811
212 elif config is not None:
213 CustomConfig = config # noqa: F811
214
215 if hasattr(CustomConfig, 'fields') or hasattr(CustomConfig, 'alias_generator'):
216 raise ConfigError(
217 'Setting the "fields" and "alias_generator" property on custom Config for '
218 '@validate_arguments is not yet supported, please remove.'
219 )
220
221 class DecoratorBaseModel(BaseModel):
222 @validator(self.v_args_name, check_fields=False, allow_reuse=True)
223 def check_args(cls, v: List[Any]) -> List[Any]:
224 if takes_args:
225 return v
226
227 raise TypeError(f'{pos_args} positional arguments expected but {pos_args + len(v)} given')
228
229 @validator(self.v_kwargs_name, check_fields=False, allow_reuse=True)
230 def check_kwargs(cls, v: Dict[str, Any]) -> Dict[str, Any]:
231 if takes_kwargs:
232 return v
233
234 plural = '' if len(v) == 1 else 's'
235 keys = ', '.join(map(repr, v.keys()))
236 raise TypeError(f'unexpected keyword argument{plural}: {keys}')
237
238 @validator(V_POSITIONAL_ONLY_NAME, check_fields=False, allow_reuse=True)
239 def check_positional_only(cls, v: List[str]) -> None:
240 plural = '' if len(v) == 1 else 's'
241 keys = ', '.join(map(repr, v))
242 raise TypeError(f'positional-only argument{plural} passed as keyword argument{plural}: {keys}')
243
244 @validator(V_DUPLICATE_KWARGS, check_fields=False, allow_reuse=True)
245 def check_duplicate_kwargs(cls, v: List[str]) -> None:
246 plural = '' if len(v) == 1 else 's'
247 keys = ', '.join(map(repr, v))
248 raise TypeError(f'multiple values for argument{plural}: {keys}')
249
250 class Config(CustomConfig):
251 extra = Extra.forbid
252
253 self.model = create_model(to_camel(self.raw_function.__name__), __base__=DecoratorBaseModel, **fields)
254
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pydantic/decorator.py b/pydantic/decorator.py
--- a/pydantic/decorator.py
+++ b/pydantic/decorator.py
@@ -220,15 +220,15 @@
class DecoratorBaseModel(BaseModel):
@validator(self.v_args_name, check_fields=False, allow_reuse=True)
- def check_args(cls, v: List[Any]) -> List[Any]:
- if takes_args:
+ def check_args(cls, v: Optional[List[Any]]) -> Optional[List[Any]]:
+ if takes_args or v is None:
return v
raise TypeError(f'{pos_args} positional arguments expected but {pos_args + len(v)} given')
@validator(self.v_kwargs_name, check_fields=False, allow_reuse=True)
- def check_kwargs(cls, v: Dict[str, Any]) -> Dict[str, Any]:
- if takes_kwargs:
+ def check_kwargs(cls, v: Optional[Dict[str, Any]]) -> Optional[Dict[str, Any]]:
+ if takes_kwargs or v is None:
return v
plural = '' if len(v) == 1 else 's'
@@ -236,13 +236,19 @@
raise TypeError(f'unexpected keyword argument{plural}: {keys}')
@validator(V_POSITIONAL_ONLY_NAME, check_fields=False, allow_reuse=True)
- def check_positional_only(cls, v: List[str]) -> None:
+ def check_positional_only(cls, v: Optional[List[str]]) -> None:
+ if v is None:
+ return
+
plural = '' if len(v) == 1 else 's'
keys = ', '.join(map(repr, v))
raise TypeError(f'positional-only argument{plural} passed as keyword argument{plural}: {keys}')
@validator(V_DUPLICATE_KWARGS, check_fields=False, allow_reuse=True)
- def check_duplicate_kwargs(cls, v: List[str]) -> None:
+ def check_duplicate_kwargs(cls, v: Optional[List[str]]) -> None:
+ if v is None:
+ return
+
plural = '' if len(v) == 1 else 's'
keys = ', '.join(map(repr, v))
raise TypeError(f'multiple values for argument{plural}: {keys}')
|
{"golden_diff": "diff --git a/pydantic/decorator.py b/pydantic/decorator.py\n--- a/pydantic/decorator.py\n+++ b/pydantic/decorator.py\n@@ -220,15 +220,15 @@\n \n class DecoratorBaseModel(BaseModel):\n @validator(self.v_args_name, check_fields=False, allow_reuse=True)\n- def check_args(cls, v: List[Any]) -> List[Any]:\n- if takes_args:\n+ def check_args(cls, v: Optional[List[Any]]) -> Optional[List[Any]]:\n+ if takes_args or v is None:\n return v\n \n raise TypeError(f'{pos_args} positional arguments expected but {pos_args + len(v)} given')\n \n @validator(self.v_kwargs_name, check_fields=False, allow_reuse=True)\n- def check_kwargs(cls, v: Dict[str, Any]) -> Dict[str, Any]:\n- if takes_kwargs:\n+ def check_kwargs(cls, v: Optional[Dict[str, Any]]) -> Optional[Dict[str, Any]]:\n+ if takes_kwargs or v is None:\n return v\n \n plural = '' if len(v) == 1 else 's'\n@@ -236,13 +236,19 @@\n raise TypeError(f'unexpected keyword argument{plural}: {keys}')\n \n @validator(V_POSITIONAL_ONLY_NAME, check_fields=False, allow_reuse=True)\n- def check_positional_only(cls, v: List[str]) -> None:\n+ def check_positional_only(cls, v: Optional[List[str]]) -> None:\n+ if v is None:\n+ return\n+\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'positional-only argument{plural} passed as keyword argument{plural}: {keys}')\n \n @validator(V_DUPLICATE_KWARGS, check_fields=False, allow_reuse=True)\n- def check_duplicate_kwargs(cls, v: List[str]) -> None:\n+ def check_duplicate_kwargs(cls, v: Optional[List[str]]) -> None:\n+ if v is None:\n+ return\n+\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'multiple values for argument{plural}: {keys}')\n", "issue": "underscore_attrs_are_private and validate_all not compatible with validate_arguments\nOutput of `python -c \"import pydantic.utils; print(pydantic.utils.version_info())\"`:\r\n```\r\n pydantic version: 1.8.2\r\n pydantic compiled: True\r\n install path: /home/xxx/.conda/envs/py38/lib/python3.8/site-packages/pydantic\r\n python version: 3.8.5 (default, Sep 4 2020, 07:30:14) [GCC 7.3.0]\r\n platform: Linux-4.18.0-80.el8.x86_64-x86_64-with-glibc2.10\r\n optional deps. installed: ['typing-extensions']\r\n```\r\n`underscore_attrs_are_private` not work with `validate_arguments` just like it not work with `create_model` #3134 .\r\n\r\nWhen I define a function decorated by `validate_arguments` with `validate_all` been set `True`, It will raise `ValidationError`:\r\n\r\n```py\r\nfrom datetime import datetime\r\nfrom pydantic import validate_arguments, Field\r\n\r\n\r\n@validate_arguments(config={'validate_all': True})\r\ndef foo(dt: datetime = Field(default_factory=lambda : 42)):\r\n print(type(dt))\r\n\r\nfoo()\r\n```\r\n```\r\n---------------------------------------------------------------------------\r\nValidationError Traceback (most recent call last)\r\n<ipython-input-4-6c97ede6fd47> in <module>\r\n 7 print(type(dt))\r\n 8 \r\n----> 9 foo()\r\n\r\n~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.validate_arguments.validate.wrapper_function()\r\n\r\n~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.ValidatedFunction.call()\r\n\r\n~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/decorator.cpython-38-x86_64-linux-gnu.so in pydantic.decorator.ValidatedFunction.init_model_instance()\r\n\r\n~/.conda/envs/py38/lib/python3.8/site-packages/pydantic/main.cpython-38-x86_64-linux-gnu.so in pydantic.main.BaseModel.__init__()\r\n\r\nValidationError: 3 validation errors for Foo\r\nv__duplicate_kwargs\r\n object of type 'NoneType' has no len() (type=type_error)\r\nargs\r\n object of type 'NoneType' has no len() (type=type_error)\r\nkwargs\r\n object of type 'NoneType' has no len() (type=type_error)\r\n```\r\n\n", "before_files": [{"content": "from functools import wraps\nfrom typing import TYPE_CHECKING, Any, Callable, Dict, List, Mapping, Optional, Tuple, Type, TypeVar, Union, overload\n\nfrom . import validator\nfrom .config import Extra\nfrom .errors import ConfigError\nfrom .main import BaseModel, create_model\nfrom .typing import get_all_type_hints\nfrom .utils import to_camel\n\n__all__ = ('validate_arguments',)\n\nif TYPE_CHECKING:\n from .typing import AnyCallable\n\n AnyCallableT = TypeVar('AnyCallableT', bound=AnyCallable)\n ConfigType = Union[None, Type[Any], Dict[str, Any]]\n\n\n@overload\ndef validate_arguments(func: None = None, *, config: 'ConfigType' = None) -> Callable[['AnyCallableT'], 'AnyCallableT']:\n ...\n\n\n@overload\ndef validate_arguments(func: 'AnyCallableT') -> 'AnyCallableT':\n ...\n\n\ndef validate_arguments(func: Optional['AnyCallableT'] = None, *, config: 'ConfigType' = None) -> Any:\n \"\"\"\n Decorator to validate the arguments passed to a function.\n \"\"\"\n\n def validate(_func: 'AnyCallable') -> 'AnyCallable':\n vd = ValidatedFunction(_func, config)\n\n @wraps(_func)\n def wrapper_function(*args: Any, **kwargs: Any) -> Any:\n return vd.call(*args, **kwargs)\n\n wrapper_function.vd = vd # type: ignore\n wrapper_function.validate = vd.init_model_instance # type: ignore\n wrapper_function.raw_function = vd.raw_function # type: ignore\n wrapper_function.model = vd.model # type: ignore\n return wrapper_function\n\n if func:\n return validate(func)\n else:\n return validate\n\n\nALT_V_ARGS = 'v__args'\nALT_V_KWARGS = 'v__kwargs'\nV_POSITIONAL_ONLY_NAME = 'v__positional_only'\nV_DUPLICATE_KWARGS = 'v__duplicate_kwargs'\n\n\nclass ValidatedFunction:\n def __init__(self, function: 'AnyCallableT', config: 'ConfigType'): # noqa C901\n from inspect import Parameter, signature\n\n parameters: Mapping[str, Parameter] = signature(function).parameters\n\n if parameters.keys() & {ALT_V_ARGS, ALT_V_KWARGS, V_POSITIONAL_ONLY_NAME, V_DUPLICATE_KWARGS}:\n raise ConfigError(\n f'\"{ALT_V_ARGS}\", \"{ALT_V_KWARGS}\", \"{V_POSITIONAL_ONLY_NAME}\" and \"{V_DUPLICATE_KWARGS}\" '\n f'are not permitted as argument names when using the \"{validate_arguments.__name__}\" decorator'\n )\n\n self.raw_function = function\n self.arg_mapping: Dict[int, str] = {}\n self.positional_only_args = set()\n self.v_args_name = 'args'\n self.v_kwargs_name = 'kwargs'\n\n type_hints = get_all_type_hints(function)\n takes_args = False\n takes_kwargs = False\n fields: Dict[str, Tuple[Any, Any]] = {}\n for i, (name, p) in enumerate(parameters.items()):\n if p.annotation is p.empty:\n annotation = Any\n else:\n annotation = type_hints[name]\n\n default = ... if p.default is p.empty else p.default\n if p.kind == Parameter.POSITIONAL_ONLY:\n self.arg_mapping[i] = name\n fields[name] = annotation, default\n fields[V_POSITIONAL_ONLY_NAME] = List[str], None\n self.positional_only_args.add(name)\n elif p.kind == Parameter.POSITIONAL_OR_KEYWORD:\n self.arg_mapping[i] = name\n fields[name] = annotation, default\n fields[V_DUPLICATE_KWARGS] = List[str], None\n elif p.kind == Parameter.KEYWORD_ONLY:\n fields[name] = annotation, default\n elif p.kind == Parameter.VAR_POSITIONAL:\n self.v_args_name = name\n fields[name] = Tuple[annotation, ...], None\n takes_args = True\n else:\n assert p.kind == Parameter.VAR_KEYWORD, p.kind\n self.v_kwargs_name = name\n fields[name] = Dict[str, annotation], None # type: ignore\n takes_kwargs = True\n\n # these checks avoid a clash between \"args\" and a field with that name\n if not takes_args and self.v_args_name in fields:\n self.v_args_name = ALT_V_ARGS\n\n # same with \"kwargs\"\n if not takes_kwargs and self.v_kwargs_name in fields:\n self.v_kwargs_name = ALT_V_KWARGS\n\n if not takes_args:\n # we add the field so validation below can raise the correct exception\n fields[self.v_args_name] = List[Any], None\n\n if not takes_kwargs:\n # same with kwargs\n fields[self.v_kwargs_name] = Dict[Any, Any], None\n\n self.create_model(fields, takes_args, takes_kwargs, config)\n\n def init_model_instance(self, *args: Any, **kwargs: Any) -> BaseModel:\n values = self.build_values(args, kwargs)\n return self.model(**values)\n\n def call(self, *args: Any, **kwargs: Any) -> Any:\n m = self.init_model_instance(*args, **kwargs)\n return self.execute(m)\n\n def build_values(self, args: Tuple[Any, ...], kwargs: Dict[str, Any]) -> Dict[str, Any]:\n values: Dict[str, Any] = {}\n if args:\n arg_iter = enumerate(args)\n while True:\n try:\n i, a = next(arg_iter)\n except StopIteration:\n break\n arg_name = self.arg_mapping.get(i)\n if arg_name is not None:\n values[arg_name] = a\n else:\n values[self.v_args_name] = [a] + [a for _, a in arg_iter]\n break\n\n var_kwargs = {}\n wrong_positional_args = []\n duplicate_kwargs = []\n non_var_fields = set(self.model.__fields__) - {self.v_args_name, self.v_kwargs_name}\n for k, v in kwargs.items():\n if k in non_var_fields:\n if k in self.positional_only_args:\n wrong_positional_args.append(k)\n if k in values:\n duplicate_kwargs.append(k)\n values[k] = v\n else:\n var_kwargs[k] = v\n\n if var_kwargs:\n values[self.v_kwargs_name] = var_kwargs\n if wrong_positional_args:\n values[V_POSITIONAL_ONLY_NAME] = wrong_positional_args\n if duplicate_kwargs:\n values[V_DUPLICATE_KWARGS] = duplicate_kwargs\n return values\n\n def execute(self, m: BaseModel) -> Any:\n d = {k: v for k, v in m._iter() if k in m.__fields_set__ or m.__fields__[k].default_factory}\n var_kwargs = d.pop(self.v_kwargs_name, {})\n\n if self.v_args_name in d:\n args_: List[Any] = []\n in_kwargs = False\n kwargs = {}\n for name, value in d.items():\n if in_kwargs:\n kwargs[name] = value\n elif name == self.v_args_name:\n args_ += value\n in_kwargs = True\n else:\n args_.append(value)\n return self.raw_function(*args_, **kwargs, **var_kwargs)\n elif self.positional_only_args:\n args_ = []\n kwargs = {}\n for name, value in d.items():\n if name in self.positional_only_args:\n args_.append(value)\n else:\n kwargs[name] = value\n return self.raw_function(*args_, **kwargs, **var_kwargs)\n else:\n return self.raw_function(**d, **var_kwargs)\n\n def create_model(self, fields: Dict[str, Any], takes_args: bool, takes_kwargs: bool, config: 'ConfigType') -> None:\n pos_args = len(self.arg_mapping)\n\n class CustomConfig:\n pass\n\n if not TYPE_CHECKING: # pragma: no branch\n if isinstance(config, dict):\n CustomConfig = type('Config', (), config) # noqa: F811\n elif config is not None:\n CustomConfig = config # noqa: F811\n\n if hasattr(CustomConfig, 'fields') or hasattr(CustomConfig, 'alias_generator'):\n raise ConfigError(\n 'Setting the \"fields\" and \"alias_generator\" property on custom Config for '\n '@validate_arguments is not yet supported, please remove.'\n )\n\n class DecoratorBaseModel(BaseModel):\n @validator(self.v_args_name, check_fields=False, allow_reuse=True)\n def check_args(cls, v: List[Any]) -> List[Any]:\n if takes_args:\n return v\n\n raise TypeError(f'{pos_args} positional arguments expected but {pos_args + len(v)} given')\n\n @validator(self.v_kwargs_name, check_fields=False, allow_reuse=True)\n def check_kwargs(cls, v: Dict[str, Any]) -> Dict[str, Any]:\n if takes_kwargs:\n return v\n\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v.keys()))\n raise TypeError(f'unexpected keyword argument{plural}: {keys}')\n\n @validator(V_POSITIONAL_ONLY_NAME, check_fields=False, allow_reuse=True)\n def check_positional_only(cls, v: List[str]) -> None:\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'positional-only argument{plural} passed as keyword argument{plural}: {keys}')\n\n @validator(V_DUPLICATE_KWARGS, check_fields=False, allow_reuse=True)\n def check_duplicate_kwargs(cls, v: List[str]) -> None:\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'multiple values for argument{plural}: {keys}')\n\n class Config(CustomConfig):\n extra = Extra.forbid\n\n self.model = create_model(to_camel(self.raw_function.__name__), __base__=DecoratorBaseModel, **fields)\n", "path": "pydantic/decorator.py"}], "after_files": [{"content": "from functools import wraps\nfrom typing import TYPE_CHECKING, Any, Callable, Dict, List, Mapping, Optional, Tuple, Type, TypeVar, Union, overload\n\nfrom . import validator\nfrom .config import Extra\nfrom .errors import ConfigError\nfrom .main import BaseModel, create_model\nfrom .typing import get_all_type_hints\nfrom .utils import to_camel\n\n__all__ = ('validate_arguments',)\n\nif TYPE_CHECKING:\n from .typing import AnyCallable\n\n AnyCallableT = TypeVar('AnyCallableT', bound=AnyCallable)\n ConfigType = Union[None, Type[Any], Dict[str, Any]]\n\n\n@overload\ndef validate_arguments(func: None = None, *, config: 'ConfigType' = None) -> Callable[['AnyCallableT'], 'AnyCallableT']:\n ...\n\n\n@overload\ndef validate_arguments(func: 'AnyCallableT') -> 'AnyCallableT':\n ...\n\n\ndef validate_arguments(func: Optional['AnyCallableT'] = None, *, config: 'ConfigType' = None) -> Any:\n \"\"\"\n Decorator to validate the arguments passed to a function.\n \"\"\"\n\n def validate(_func: 'AnyCallable') -> 'AnyCallable':\n vd = ValidatedFunction(_func, config)\n\n @wraps(_func)\n def wrapper_function(*args: Any, **kwargs: Any) -> Any:\n return vd.call(*args, **kwargs)\n\n wrapper_function.vd = vd # type: ignore\n wrapper_function.validate = vd.init_model_instance # type: ignore\n wrapper_function.raw_function = vd.raw_function # type: ignore\n wrapper_function.model = vd.model # type: ignore\n return wrapper_function\n\n if func:\n return validate(func)\n else:\n return validate\n\n\nALT_V_ARGS = 'v__args'\nALT_V_KWARGS = 'v__kwargs'\nV_POSITIONAL_ONLY_NAME = 'v__positional_only'\nV_DUPLICATE_KWARGS = 'v__duplicate_kwargs'\n\n\nclass ValidatedFunction:\n def __init__(self, function: 'AnyCallableT', config: 'ConfigType'): # noqa C901\n from inspect import Parameter, signature\n\n parameters: Mapping[str, Parameter] = signature(function).parameters\n\n if parameters.keys() & {ALT_V_ARGS, ALT_V_KWARGS, V_POSITIONAL_ONLY_NAME, V_DUPLICATE_KWARGS}:\n raise ConfigError(\n f'\"{ALT_V_ARGS}\", \"{ALT_V_KWARGS}\", \"{V_POSITIONAL_ONLY_NAME}\" and \"{V_DUPLICATE_KWARGS}\" '\n f'are not permitted as argument names when using the \"{validate_arguments.__name__}\" decorator'\n )\n\n self.raw_function = function\n self.arg_mapping: Dict[int, str] = {}\n self.positional_only_args = set()\n self.v_args_name = 'args'\n self.v_kwargs_name = 'kwargs'\n\n type_hints = get_all_type_hints(function)\n takes_args = False\n takes_kwargs = False\n fields: Dict[str, Tuple[Any, Any]] = {}\n for i, (name, p) in enumerate(parameters.items()):\n if p.annotation is p.empty:\n annotation = Any\n else:\n annotation = type_hints[name]\n\n default = ... if p.default is p.empty else p.default\n if p.kind == Parameter.POSITIONAL_ONLY:\n self.arg_mapping[i] = name\n fields[name] = annotation, default\n fields[V_POSITIONAL_ONLY_NAME] = List[str], None\n self.positional_only_args.add(name)\n elif p.kind == Parameter.POSITIONAL_OR_KEYWORD:\n self.arg_mapping[i] = name\n fields[name] = annotation, default\n fields[V_DUPLICATE_KWARGS] = List[str], None\n elif p.kind == Parameter.KEYWORD_ONLY:\n fields[name] = annotation, default\n elif p.kind == Parameter.VAR_POSITIONAL:\n self.v_args_name = name\n fields[name] = Tuple[annotation, ...], None\n takes_args = True\n else:\n assert p.kind == Parameter.VAR_KEYWORD, p.kind\n self.v_kwargs_name = name\n fields[name] = Dict[str, annotation], None # type: ignore\n takes_kwargs = True\n\n # these checks avoid a clash between \"args\" and a field with that name\n if not takes_args and self.v_args_name in fields:\n self.v_args_name = ALT_V_ARGS\n\n # same with \"kwargs\"\n if not takes_kwargs and self.v_kwargs_name in fields:\n self.v_kwargs_name = ALT_V_KWARGS\n\n if not takes_args:\n # we add the field so validation below can raise the correct exception\n fields[self.v_args_name] = List[Any], None\n\n if not takes_kwargs:\n # same with kwargs\n fields[self.v_kwargs_name] = Dict[Any, Any], None\n\n self.create_model(fields, takes_args, takes_kwargs, config)\n\n def init_model_instance(self, *args: Any, **kwargs: Any) -> BaseModel:\n values = self.build_values(args, kwargs)\n return self.model(**values)\n\n def call(self, *args: Any, **kwargs: Any) -> Any:\n m = self.init_model_instance(*args, **kwargs)\n return self.execute(m)\n\n def build_values(self, args: Tuple[Any, ...], kwargs: Dict[str, Any]) -> Dict[str, Any]:\n values: Dict[str, Any] = {}\n if args:\n arg_iter = enumerate(args)\n while True:\n try:\n i, a = next(arg_iter)\n except StopIteration:\n break\n arg_name = self.arg_mapping.get(i)\n if arg_name is not None:\n values[arg_name] = a\n else:\n values[self.v_args_name] = [a] + [a for _, a in arg_iter]\n break\n\n var_kwargs = {}\n wrong_positional_args = []\n duplicate_kwargs = []\n non_var_fields = set(self.model.__fields__) - {self.v_args_name, self.v_kwargs_name}\n for k, v in kwargs.items():\n if k in non_var_fields:\n if k in self.positional_only_args:\n wrong_positional_args.append(k)\n if k in values:\n duplicate_kwargs.append(k)\n values[k] = v\n else:\n var_kwargs[k] = v\n\n if var_kwargs:\n values[self.v_kwargs_name] = var_kwargs\n if wrong_positional_args:\n values[V_POSITIONAL_ONLY_NAME] = wrong_positional_args\n if duplicate_kwargs:\n values[V_DUPLICATE_KWARGS] = duplicate_kwargs\n return values\n\n def execute(self, m: BaseModel) -> Any:\n d = {k: v for k, v in m._iter() if k in m.__fields_set__ or m.__fields__[k].default_factory}\n var_kwargs = d.pop(self.v_kwargs_name, {})\n\n if self.v_args_name in d:\n args_: List[Any] = []\n in_kwargs = False\n kwargs = {}\n for name, value in d.items():\n if in_kwargs:\n kwargs[name] = value\n elif name == self.v_args_name:\n args_ += value\n in_kwargs = True\n else:\n args_.append(value)\n return self.raw_function(*args_, **kwargs, **var_kwargs)\n elif self.positional_only_args:\n args_ = []\n kwargs = {}\n for name, value in d.items():\n if name in self.positional_only_args:\n args_.append(value)\n else:\n kwargs[name] = value\n return self.raw_function(*args_, **kwargs, **var_kwargs)\n else:\n return self.raw_function(**d, **var_kwargs)\n\n def create_model(self, fields: Dict[str, Any], takes_args: bool, takes_kwargs: bool, config: 'ConfigType') -> None:\n pos_args = len(self.arg_mapping)\n\n class CustomConfig:\n pass\n\n if not TYPE_CHECKING: # pragma: no branch\n if isinstance(config, dict):\n CustomConfig = type('Config', (), config) # noqa: F811\n elif config is not None:\n CustomConfig = config # noqa: F811\n\n if hasattr(CustomConfig, 'fields') or hasattr(CustomConfig, 'alias_generator'):\n raise ConfigError(\n 'Setting the \"fields\" and \"alias_generator\" property on custom Config for '\n '@validate_arguments is not yet supported, please remove.'\n )\n\n class DecoratorBaseModel(BaseModel):\n @validator(self.v_args_name, check_fields=False, allow_reuse=True)\n def check_args(cls, v: Optional[List[Any]]) -> Optional[List[Any]]:\n if takes_args or v is None:\n return v\n\n raise TypeError(f'{pos_args} positional arguments expected but {pos_args + len(v)} given')\n\n @validator(self.v_kwargs_name, check_fields=False, allow_reuse=True)\n def check_kwargs(cls, v: Optional[Dict[str, Any]]) -> Optional[Dict[str, Any]]:\n if takes_kwargs or v is None:\n return v\n\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v.keys()))\n raise TypeError(f'unexpected keyword argument{plural}: {keys}')\n\n @validator(V_POSITIONAL_ONLY_NAME, check_fields=False, allow_reuse=True)\n def check_positional_only(cls, v: Optional[List[str]]) -> None:\n if v is None:\n return\n\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'positional-only argument{plural} passed as keyword argument{plural}: {keys}')\n\n @validator(V_DUPLICATE_KWARGS, check_fields=False, allow_reuse=True)\n def check_duplicate_kwargs(cls, v: Optional[List[str]]) -> None:\n if v is None:\n return\n\n plural = '' if len(v) == 1 else 's'\n keys = ', '.join(map(repr, v))\n raise TypeError(f'multiple values for argument{plural}: {keys}')\n\n class Config(CustomConfig):\n extra = Extra.forbid\n\n self.model = create_model(to_camel(self.raw_function.__name__), __base__=DecoratorBaseModel, **fields)\n", "path": "pydantic/decorator.py"}]}
| 3,730 | 507 |
gh_patches_debug_25839
|
rasdani/github-patches
|
git_diff
|
python-discord__site-650
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add an endpoint to fetch user information from a username.
We'd need this endpoint to be able to implement features requested in [this](https://github.com/python-discord/bot/issues/1757) bot issue.
```py
from django.shortcuts import get_object_or_404
@action(
detail=False,
methods=["GET"],
url_path=r"username/(?P<name>[^#]{,32})#(?P<discriminator>[0-9]{4})"
)
def get_by_username(self, request, name, discriminator):
user = get_object_or_404(User, name=name, discriminator=discriminator)
return Response(UserSerializer(user).data, status=status.HTTP_200_OK)
```
This addition to the user view set seems to make GET requests to URLs like `bot/users/username/Qwerty#0001` work , but there might be a more ideal way to do this.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pydis_site/apps/api/viewsets/bot/user.py`
Content:
```
1 import typing
2 from collections import OrderedDict
3
4 from django.db.models import Q
5 from rest_framework import status
6 from rest_framework.decorators import action
7 from rest_framework.pagination import PageNumberPagination
8 from rest_framework.request import Request
9 from rest_framework.response import Response
10 from rest_framework.serializers import ModelSerializer
11 from rest_framework.viewsets import ModelViewSet
12
13 from pydis_site.apps.api.models.bot.infraction import Infraction
14 from pydis_site.apps.api.models.bot.metricity import Metricity, NotFoundError
15 from pydis_site.apps.api.models.bot.user import User
16 from pydis_site.apps.api.serializers import UserSerializer
17
18
19 class UserListPagination(PageNumberPagination):
20 """Custom pagination class for the User Model."""
21
22 page_size = 2500
23 page_size_query_param = "page_size"
24
25 def get_next_page_number(self) -> typing.Optional[int]:
26 """Get the next page number."""
27 if not self.page.has_next():
28 return None
29 page_number = self.page.next_page_number()
30 return page_number
31
32 def get_previous_page_number(self) -> typing.Optional[int]:
33 """Get the previous page number."""
34 if not self.page.has_previous():
35 return None
36
37 page_number = self.page.previous_page_number()
38 return page_number
39
40 def get_paginated_response(self, data: list) -> Response:
41 """Override method to send modified response."""
42 return Response(OrderedDict([
43 ('count', self.page.paginator.count),
44 ('next_page_no', self.get_next_page_number()),
45 ('previous_page_no', self.get_previous_page_number()),
46 ('results', data)
47 ]))
48
49
50 class UserViewSet(ModelViewSet):
51 """
52 View providing CRUD operations on Discord users through the bot.
53
54 ## Routes
55 ### GET /bot/users
56 Returns all users currently known with pagination.
57
58 #### Response format
59 >>> {
60 ... 'count': 95000,
61 ... 'next_page_no': "2",
62 ... 'previous_page_no': None,
63 ... 'results': [
64 ... {
65 ... 'id': 409107086526644234,
66 ... 'name': "Python",
67 ... 'discriminator': 4329,
68 ... 'roles': [
69 ... 352427296948486144,
70 ... 270988689419665409,
71 ... 277546923144249364,
72 ... 458226699344019457
73 ... ],
74 ... 'in_guild': True
75 ... },
76 ... ]
77 ... }
78
79 #### Optional Query Parameters
80 - page_size: number of Users in one page, defaults to 10,000
81 - page: page number
82
83 #### Status codes
84 - 200: returned on success
85
86 ### GET /bot/users/<snowflake:int>
87 Gets a single user by ID.
88
89 #### Response format
90 >>> {
91 ... 'id': 409107086526644234,
92 ... 'name': "Python",
93 ... 'discriminator': 4329,
94 ... 'roles': [
95 ... 352427296948486144,
96 ... 270988689419665409,
97 ... 277546923144249364,
98 ... 458226699344019457
99 ... ],
100 ... 'in_guild': True
101 ... }
102
103 #### Status codes
104 - 200: returned on success
105 - 404: if a user with the given `snowflake` could not be found
106
107 ### GET /bot/users/<snowflake:int>/metricity_data
108 Gets metricity data for a single user by ID.
109
110 #### Response format
111 >>> {
112 ... "joined_at": "2020-10-06T21:54:23.540766",
113 ... "total_messages": 2,
114 ... "voice_banned": False,
115 ... "activity_blocks": 1
116 ...}
117
118 #### Status codes
119 - 200: returned on success
120 - 404: if a user with the given `snowflake` could not be found
121
122 ### GET /bot/users/<snowflake:int>/metricity_review_data
123 Gets metricity data for a single user's review by ID.
124
125 #### Response format
126 >>> {
127 ... 'joined_at': '2020-08-26T08:09:43.507000',
128 ... 'top_channel_activity': [['off-topic', 15],
129 ... ['talent-pool', 4],
130 ... ['defcon', 2]],
131 ... 'total_messages': 22
132 ... }
133
134 #### Status codes
135 - 200: returned on success
136 - 404: if a user with the given `snowflake` could not be found
137
138 ### POST /bot/users
139 Adds a single or multiple new users.
140 The roles attached to the user(s) must be roles known by the site.
141 Users that already exist in the database will be skipped.
142
143 #### Request body
144 >>> {
145 ... 'id': int,
146 ... 'name': str,
147 ... 'discriminator': int,
148 ... 'roles': List[int],
149 ... 'in_guild': bool
150 ... }
151
152 Alternatively, request users can be POSTed as a list of above objects,
153 in which case multiple users will be created at once. In this case,
154 the response is an empty list.
155
156 #### Status codes
157 - 201: returned on success
158 - 400: if one of the given roles does not exist, or one of the given fields is invalid
159 - 400: if multiple user objects with the same id are given
160
161 ### PUT /bot/users/<snowflake:int>
162 Update the user with the given `snowflake`.
163 All fields in the request body are required.
164
165 #### Request body
166 >>> {
167 ... 'id': int,
168 ... 'name': str,
169 ... 'discriminator': int,
170 ... 'roles': List[int],
171 ... 'in_guild': bool
172 ... }
173
174 #### Status codes
175 - 200: returned on success
176 - 400: if the request body was invalid, see response body for details
177 - 404: if the user with the given `snowflake` could not be found
178
179 ### PATCH /bot/users/<snowflake:int>
180 Update the user with the given `snowflake`.
181 All fields in the request body are optional.
182
183 #### Request body
184 >>> {
185 ... 'id': int,
186 ... 'name': str,
187 ... 'discriminator': int,
188 ... 'roles': List[int],
189 ... 'in_guild': bool
190 ... }
191
192 #### Status codes
193 - 200: returned on success
194 - 400: if the request body was invalid, see response body for details
195 - 404: if the user with the given `snowflake` could not be found
196
197 ### BULK PATCH /bot/users/bulk_patch
198 Update users with the given `ids` and `details`.
199 `id` field and at least one other field is mandatory.
200
201 #### Request body
202 >>> [
203 ... {
204 ... 'id': int,
205 ... 'name': str,
206 ... 'discriminator': int,
207 ... 'roles': List[int],
208 ... 'in_guild': bool
209 ... },
210 ... {
211 ... 'id': int,
212 ... 'name': str,
213 ... 'discriminator': int,
214 ... 'roles': List[int],
215 ... 'in_guild': bool
216 ... },
217 ... ]
218
219 #### Status codes
220 - 200: returned on success
221 - 400: if the request body was invalid, see response body for details
222 - 400: if multiple user objects with the same id are given
223 - 404: if the user with the given id does not exist
224
225 ### DELETE /bot/users/<snowflake:int>
226 Deletes the user with the given `snowflake`.
227
228 #### Status codes
229 - 204: returned on success
230 - 404: if a user with the given `snowflake` does not exist
231 """
232
233 serializer_class = UserSerializer
234 queryset = User.objects.all().order_by("id")
235 pagination_class = UserListPagination
236
237 def get_serializer(self, *args, **kwargs) -> ModelSerializer:
238 """Set Serializer many attribute to True if request body contains a list."""
239 if isinstance(kwargs.get('data', {}), list):
240 kwargs['many'] = True
241
242 return super().get_serializer(*args, **kwargs)
243
244 @action(detail=False, methods=["PATCH"], name='user-bulk-patch')
245 def bulk_patch(self, request: Request) -> Response:
246 """Update multiple User objects in a single request."""
247 serializer = self.get_serializer(
248 instance=self.get_queryset(),
249 data=request.data,
250 many=True,
251 partial=True
252 )
253
254 serializer.is_valid(raise_exception=True)
255 serializer.save()
256
257 return Response(serializer.data, status=status.HTTP_200_OK)
258
259 @action(detail=True)
260 def metricity_data(self, request: Request, pk: str = None) -> Response:
261 """Request handler for metricity_data endpoint."""
262 user = self.get_object()
263
264 has_voice_infraction = Infraction.objects.filter(
265 Q(user__id=user.id, active=True),
266 Q(type="voice_ban") | Q(type="voice_mute")
267 ).exists()
268
269 with Metricity() as metricity:
270 try:
271 data = metricity.user(user.id)
272
273 data["total_messages"] = metricity.total_messages(user.id)
274 data["activity_blocks"] = metricity.total_message_blocks(user.id)
275
276 data["voice_gate_blocked"] = has_voice_infraction
277 return Response(data, status=status.HTTP_200_OK)
278 except NotFoundError:
279 return Response(dict(detail="User not found in metricity"),
280 status=status.HTTP_404_NOT_FOUND)
281
282 @action(detail=True)
283 def metricity_review_data(self, request: Request, pk: str = None) -> Response:
284 """Request handler for metricity_review_data endpoint."""
285 user = self.get_object()
286
287 with Metricity() as metricity:
288 try:
289 data = metricity.user(user.id)
290 data["total_messages"] = metricity.total_messages(user.id)
291 data["top_channel_activity"] = metricity.top_channel_activity(user.id)
292 return Response(data, status=status.HTTP_200_OK)
293 except NotFoundError:
294 return Response(dict(detail="User not found in metricity"),
295 status=status.HTTP_404_NOT_FOUND)
296
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pydis_site/apps/api/viewsets/bot/user.py b/pydis_site/apps/api/viewsets/bot/user.py
--- a/pydis_site/apps/api/viewsets/bot/user.py
+++ b/pydis_site/apps/api/viewsets/bot/user.py
@@ -2,6 +2,7 @@
from collections import OrderedDict
from django.db.models import Q
+from django_filters.rest_framework import DjangoFilterBackend
from rest_framework import status
from rest_framework.decorators import action
from rest_framework.pagination import PageNumberPagination
@@ -77,6 +78,8 @@
... }
#### Optional Query Parameters
+ - username: username to search for
+ - discriminator: discriminator to search for
- page_size: number of Users in one page, defaults to 10,000
- page: page number
@@ -233,6 +236,8 @@
serializer_class = UserSerializer
queryset = User.objects.all().order_by("id")
pagination_class = UserListPagination
+ filter_backends = (DjangoFilterBackend,)
+ filter_fields = ('name', 'discriminator')
def get_serializer(self, *args, **kwargs) -> ModelSerializer:
"""Set Serializer many attribute to True if request body contains a list."""
|
{"golden_diff": "diff --git a/pydis_site/apps/api/viewsets/bot/user.py b/pydis_site/apps/api/viewsets/bot/user.py\n--- a/pydis_site/apps/api/viewsets/bot/user.py\n+++ b/pydis_site/apps/api/viewsets/bot/user.py\n@@ -2,6 +2,7 @@\n from collections import OrderedDict\n \n from django.db.models import Q\n+from django_filters.rest_framework import DjangoFilterBackend\n from rest_framework import status\n from rest_framework.decorators import action\n from rest_framework.pagination import PageNumberPagination\n@@ -77,6 +78,8 @@\n ... }\n \n #### Optional Query Parameters\n+ - username: username to search for\n+ - discriminator: discriminator to search for\n - page_size: number of Users in one page, defaults to 10,000\n - page: page number\n \n@@ -233,6 +236,8 @@\n serializer_class = UserSerializer\n queryset = User.objects.all().order_by(\"id\")\n pagination_class = UserListPagination\n+ filter_backends = (DjangoFilterBackend,)\n+ filter_fields = ('name', 'discriminator')\n \n def get_serializer(self, *args, **kwargs) -> ModelSerializer:\n \"\"\"Set Serializer many attribute to True if request body contains a list.\"\"\"\n", "issue": "Add an endpoint to fetch user information from a username.\nWe'd need this endpoint to be able to implement features requested in [this](https://github.com/python-discord/bot/issues/1757) bot issue.\r\n\r\n```py\r\nfrom django.shortcuts import get_object_or_404\r\n\r\n @action(\r\n detail=False,\r\n methods=[\"GET\"],\r\n url_path=r\"username/(?P<name>[^#]{,32})#(?P<discriminator>[0-9]{4})\"\r\n )\r\n def get_by_username(self, request, name, discriminator):\r\n user = get_object_or_404(User, name=name, discriminator=discriminator)\r\n return Response(UserSerializer(user).data, status=status.HTTP_200_OK)\r\n```\r\n\r\nThis addition to the user view set seems to make GET requests to URLs like `bot/users/username/Qwerty#0001` work , but there might be a more ideal way to do this.\n", "before_files": [{"content": "import typing\nfrom collections import OrderedDict\n\nfrom django.db.models import Q\nfrom rest_framework import status\nfrom rest_framework.decorators import action\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.request import Request\nfrom rest_framework.response import Response\nfrom rest_framework.serializers import ModelSerializer\nfrom rest_framework.viewsets import ModelViewSet\n\nfrom pydis_site.apps.api.models.bot.infraction import Infraction\nfrom pydis_site.apps.api.models.bot.metricity import Metricity, NotFoundError\nfrom pydis_site.apps.api.models.bot.user import User\nfrom pydis_site.apps.api.serializers import UserSerializer\n\n\nclass UserListPagination(PageNumberPagination):\n \"\"\"Custom pagination class for the User Model.\"\"\"\n\n page_size = 2500\n page_size_query_param = \"page_size\"\n\n def get_next_page_number(self) -> typing.Optional[int]:\n \"\"\"Get the next page number.\"\"\"\n if not self.page.has_next():\n return None\n page_number = self.page.next_page_number()\n return page_number\n\n def get_previous_page_number(self) -> typing.Optional[int]:\n \"\"\"Get the previous page number.\"\"\"\n if not self.page.has_previous():\n return None\n\n page_number = self.page.previous_page_number()\n return page_number\n\n def get_paginated_response(self, data: list) -> Response:\n \"\"\"Override method to send modified response.\"\"\"\n return Response(OrderedDict([\n ('count', self.page.paginator.count),\n ('next_page_no', self.get_next_page_number()),\n ('previous_page_no', self.get_previous_page_number()),\n ('results', data)\n ]))\n\n\nclass UserViewSet(ModelViewSet):\n \"\"\"\n View providing CRUD operations on Discord users through the bot.\n\n ## Routes\n ### GET /bot/users\n Returns all users currently known with pagination.\n\n #### Response format\n >>> {\n ... 'count': 95000,\n ... 'next_page_no': \"2\",\n ... 'previous_page_no': None,\n ... 'results': [\n ... {\n ... 'id': 409107086526644234,\n ... 'name': \"Python\",\n ... 'discriminator': 4329,\n ... 'roles': [\n ... 352427296948486144,\n ... 270988689419665409,\n ... 277546923144249364,\n ... 458226699344019457\n ... ],\n ... 'in_guild': True\n ... },\n ... ]\n ... }\n\n #### Optional Query Parameters\n - page_size: number of Users in one page, defaults to 10,000\n - page: page number\n\n #### Status codes\n - 200: returned on success\n\n ### GET /bot/users/<snowflake:int>\n Gets a single user by ID.\n\n #### Response format\n >>> {\n ... 'id': 409107086526644234,\n ... 'name': \"Python\",\n ... 'discriminator': 4329,\n ... 'roles': [\n ... 352427296948486144,\n ... 270988689419665409,\n ... 277546923144249364,\n ... 458226699344019457\n ... ],\n ... 'in_guild': True\n ... }\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### GET /bot/users/<snowflake:int>/metricity_data\n Gets metricity data for a single user by ID.\n\n #### Response format\n >>> {\n ... \"joined_at\": \"2020-10-06T21:54:23.540766\",\n ... \"total_messages\": 2,\n ... \"voice_banned\": False,\n ... \"activity_blocks\": 1\n ...}\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### GET /bot/users/<snowflake:int>/metricity_review_data\n Gets metricity data for a single user's review by ID.\n\n #### Response format\n >>> {\n ... 'joined_at': '2020-08-26T08:09:43.507000',\n ... 'top_channel_activity': [['off-topic', 15],\n ... ['talent-pool', 4],\n ... ['defcon', 2]],\n ... 'total_messages': 22\n ... }\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### POST /bot/users\n Adds a single or multiple new users.\n The roles attached to the user(s) must be roles known by the site.\n Users that already exist in the database will be skipped.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n Alternatively, request users can be POSTed as a list of above objects,\n in which case multiple users will be created at once. In this case,\n the response is an empty list.\n\n #### Status codes\n - 201: returned on success\n - 400: if one of the given roles does not exist, or one of the given fields is invalid\n - 400: if multiple user objects with the same id are given\n\n ### PUT /bot/users/<snowflake:int>\n Update the user with the given `snowflake`.\n All fields in the request body are required.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 404: if the user with the given `snowflake` could not be found\n\n ### PATCH /bot/users/<snowflake:int>\n Update the user with the given `snowflake`.\n All fields in the request body are optional.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 404: if the user with the given `snowflake` could not be found\n\n ### BULK PATCH /bot/users/bulk_patch\n Update users with the given `ids` and `details`.\n `id` field and at least one other field is mandatory.\n\n #### Request body\n >>> [\n ... {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... },\n ... {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... },\n ... ]\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 400: if multiple user objects with the same id are given\n - 404: if the user with the given id does not exist\n\n ### DELETE /bot/users/<snowflake:int>\n Deletes the user with the given `snowflake`.\n\n #### Status codes\n - 204: returned on success\n - 404: if a user with the given `snowflake` does not exist\n \"\"\"\n\n serializer_class = UserSerializer\n queryset = User.objects.all().order_by(\"id\")\n pagination_class = UserListPagination\n\n def get_serializer(self, *args, **kwargs) -> ModelSerializer:\n \"\"\"Set Serializer many attribute to True if request body contains a list.\"\"\"\n if isinstance(kwargs.get('data', {}), list):\n kwargs['many'] = True\n\n return super().get_serializer(*args, **kwargs)\n\n @action(detail=False, methods=[\"PATCH\"], name='user-bulk-patch')\n def bulk_patch(self, request: Request) -> Response:\n \"\"\"Update multiple User objects in a single request.\"\"\"\n serializer = self.get_serializer(\n instance=self.get_queryset(),\n data=request.data,\n many=True,\n partial=True\n )\n\n serializer.is_valid(raise_exception=True)\n serializer.save()\n\n return Response(serializer.data, status=status.HTTP_200_OK)\n\n @action(detail=True)\n def metricity_data(self, request: Request, pk: str = None) -> Response:\n \"\"\"Request handler for metricity_data endpoint.\"\"\"\n user = self.get_object()\n\n has_voice_infraction = Infraction.objects.filter(\n Q(user__id=user.id, active=True),\n Q(type=\"voice_ban\") | Q(type=\"voice_mute\")\n ).exists()\n\n with Metricity() as metricity:\n try:\n data = metricity.user(user.id)\n\n data[\"total_messages\"] = metricity.total_messages(user.id)\n data[\"activity_blocks\"] = metricity.total_message_blocks(user.id)\n\n data[\"voice_gate_blocked\"] = has_voice_infraction\n return Response(data, status=status.HTTP_200_OK)\n except NotFoundError:\n return Response(dict(detail=\"User not found in metricity\"),\n status=status.HTTP_404_NOT_FOUND)\n\n @action(detail=True)\n def metricity_review_data(self, request: Request, pk: str = None) -> Response:\n \"\"\"Request handler for metricity_review_data endpoint.\"\"\"\n user = self.get_object()\n\n with Metricity() as metricity:\n try:\n data = metricity.user(user.id)\n data[\"total_messages\"] = metricity.total_messages(user.id)\n data[\"top_channel_activity\"] = metricity.top_channel_activity(user.id)\n return Response(data, status=status.HTTP_200_OK)\n except NotFoundError:\n return Response(dict(detail=\"User not found in metricity\"),\n status=status.HTTP_404_NOT_FOUND)\n", "path": "pydis_site/apps/api/viewsets/bot/user.py"}], "after_files": [{"content": "import typing\nfrom collections import OrderedDict\n\nfrom django.db.models import Q\nfrom django_filters.rest_framework import DjangoFilterBackend\nfrom rest_framework import status\nfrom rest_framework.decorators import action\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.request import Request\nfrom rest_framework.response import Response\nfrom rest_framework.serializers import ModelSerializer\nfrom rest_framework.viewsets import ModelViewSet\n\nfrom pydis_site.apps.api.models.bot.infraction import Infraction\nfrom pydis_site.apps.api.models.bot.metricity import Metricity, NotFoundError\nfrom pydis_site.apps.api.models.bot.user import User\nfrom pydis_site.apps.api.serializers import UserSerializer\n\n\nclass UserListPagination(PageNumberPagination):\n \"\"\"Custom pagination class for the User Model.\"\"\"\n\n page_size = 2500\n page_size_query_param = \"page_size\"\n\n def get_next_page_number(self) -> typing.Optional[int]:\n \"\"\"Get the next page number.\"\"\"\n if not self.page.has_next():\n return None\n page_number = self.page.next_page_number()\n return page_number\n\n def get_previous_page_number(self) -> typing.Optional[int]:\n \"\"\"Get the previous page number.\"\"\"\n if not self.page.has_previous():\n return None\n\n page_number = self.page.previous_page_number()\n return page_number\n\n def get_paginated_response(self, data: list) -> Response:\n \"\"\"Override method to send modified response.\"\"\"\n return Response(OrderedDict([\n ('count', self.page.paginator.count),\n ('next_page_no', self.get_next_page_number()),\n ('previous_page_no', self.get_previous_page_number()),\n ('results', data)\n ]))\n\n\nclass UserViewSet(ModelViewSet):\n \"\"\"\n View providing CRUD operations on Discord users through the bot.\n\n ## Routes\n ### GET /bot/users\n Returns all users currently known with pagination.\n\n #### Response format\n >>> {\n ... 'count': 95000,\n ... 'next_page_no': \"2\",\n ... 'previous_page_no': None,\n ... 'results': [\n ... {\n ... 'id': 409107086526644234,\n ... 'name': \"Python\",\n ... 'discriminator': 4329,\n ... 'roles': [\n ... 352427296948486144,\n ... 270988689419665409,\n ... 277546923144249364,\n ... 458226699344019457\n ... ],\n ... 'in_guild': True\n ... },\n ... ]\n ... }\n\n #### Optional Query Parameters\n - username: username to search for\n - discriminator: discriminator to search for\n - page_size: number of Users in one page, defaults to 10,000\n - page: page number\n\n #### Status codes\n - 200: returned on success\n\n ### GET /bot/users/<snowflake:int>\n Gets a single user by ID.\n\n #### Response format\n >>> {\n ... 'id': 409107086526644234,\n ... 'name': \"Python\",\n ... 'discriminator': 4329,\n ... 'roles': [\n ... 352427296948486144,\n ... 270988689419665409,\n ... 277546923144249364,\n ... 458226699344019457\n ... ],\n ... 'in_guild': True\n ... }\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### GET /bot/users/<snowflake:int>/metricity_data\n Gets metricity data for a single user by ID.\n\n #### Response format\n >>> {\n ... \"joined_at\": \"2020-10-06T21:54:23.540766\",\n ... \"total_messages\": 2,\n ... \"voice_banned\": False,\n ... \"activity_blocks\": 1\n ...}\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### GET /bot/users/<snowflake:int>/metricity_review_data\n Gets metricity data for a single user's review by ID.\n\n #### Response format\n >>> {\n ... 'joined_at': '2020-08-26T08:09:43.507000',\n ... 'top_channel_activity': [['off-topic', 15],\n ... ['talent-pool', 4],\n ... ['defcon', 2]],\n ... 'total_messages': 22\n ... }\n\n #### Status codes\n - 200: returned on success\n - 404: if a user with the given `snowflake` could not be found\n\n ### POST /bot/users\n Adds a single or multiple new users.\n The roles attached to the user(s) must be roles known by the site.\n Users that already exist in the database will be skipped.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n Alternatively, request users can be POSTed as a list of above objects,\n in which case multiple users will be created at once. In this case,\n the response is an empty list.\n\n #### Status codes\n - 201: returned on success\n - 400: if one of the given roles does not exist, or one of the given fields is invalid\n - 400: if multiple user objects with the same id are given\n\n ### PUT /bot/users/<snowflake:int>\n Update the user with the given `snowflake`.\n All fields in the request body are required.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 404: if the user with the given `snowflake` could not be found\n\n ### PATCH /bot/users/<snowflake:int>\n Update the user with the given `snowflake`.\n All fields in the request body are optional.\n\n #### Request body\n >>> {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... }\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 404: if the user with the given `snowflake` could not be found\n\n ### BULK PATCH /bot/users/bulk_patch\n Update users with the given `ids` and `details`.\n `id` field and at least one other field is mandatory.\n\n #### Request body\n >>> [\n ... {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... },\n ... {\n ... 'id': int,\n ... 'name': str,\n ... 'discriminator': int,\n ... 'roles': List[int],\n ... 'in_guild': bool\n ... },\n ... ]\n\n #### Status codes\n - 200: returned on success\n - 400: if the request body was invalid, see response body for details\n - 400: if multiple user objects with the same id are given\n - 404: if the user with the given id does not exist\n\n ### DELETE /bot/users/<snowflake:int>\n Deletes the user with the given `snowflake`.\n\n #### Status codes\n - 204: returned on success\n - 404: if a user with the given `snowflake` does not exist\n \"\"\"\n\n serializer_class = UserSerializer\n queryset = User.objects.all().order_by(\"id\")\n pagination_class = UserListPagination\n filter_backends = (DjangoFilterBackend,)\n filter_fields = ('name', 'discriminator')\n\n def get_serializer(self, *args, **kwargs) -> ModelSerializer:\n \"\"\"Set Serializer many attribute to True if request body contains a list.\"\"\"\n if isinstance(kwargs.get('data', {}), list):\n kwargs['many'] = True\n\n return super().get_serializer(*args, **kwargs)\n\n @action(detail=False, methods=[\"PATCH\"], name='user-bulk-patch')\n def bulk_patch(self, request: Request) -> Response:\n \"\"\"Update multiple User objects in a single request.\"\"\"\n serializer = self.get_serializer(\n instance=self.get_queryset(),\n data=request.data,\n many=True,\n partial=True\n )\n\n serializer.is_valid(raise_exception=True)\n serializer.save()\n\n return Response(serializer.data, status=status.HTTP_200_OK)\n\n @action(detail=True)\n def metricity_data(self, request: Request, pk: str = None) -> Response:\n \"\"\"Request handler for metricity_data endpoint.\"\"\"\n user = self.get_object()\n\n has_voice_infraction = Infraction.objects.filter(\n Q(user__id=user.id, active=True),\n Q(type=\"voice_ban\") | Q(type=\"voice_mute\")\n ).exists()\n\n with Metricity() as metricity:\n try:\n data = metricity.user(user.id)\n\n data[\"total_messages\"] = metricity.total_messages(user.id)\n data[\"activity_blocks\"] = metricity.total_message_blocks(user.id)\n\n data[\"voice_gate_blocked\"] = has_voice_infraction\n return Response(data, status=status.HTTP_200_OK)\n except NotFoundError:\n return Response(dict(detail=\"User not found in metricity\"),\n status=status.HTTP_404_NOT_FOUND)\n\n @action(detail=True)\n def metricity_review_data(self, request: Request, pk: str = None) -> Response:\n \"\"\"Request handler for metricity_review_data endpoint.\"\"\"\n user = self.get_object()\n\n with Metricity() as metricity:\n try:\n data = metricity.user(user.id)\n data[\"total_messages\"] = metricity.total_messages(user.id)\n data[\"top_channel_activity\"] = metricity.top_channel_activity(user.id)\n return Response(data, status=status.HTTP_200_OK)\n except NotFoundError:\n return Response(dict(detail=\"User not found in metricity\"),\n status=status.HTTP_404_NOT_FOUND)\n", "path": "pydis_site/apps/api/viewsets/bot/user.py"}]}
| 3,776 | 279 |
gh_patches_debug_8552
|
rasdani/github-patches
|
git_diff
|
Gallopsled__pwntools-323
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Importing shellcraft submodules doesn't work
```
$ python
>>> import pwnlib.shellcraft.arm
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: No module named arm
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pwnlib/shellcraft/__init__.py`
Content:
```
1 from types import ModuleType
2 import sys, os, re
3 from . import internal
4 from ..context import context
5
6 class module(ModuleType):
7 def __init__(self, name, directory):
8 super(module, self).__init__(name)
9
10 # Insert nice properties
11 self.__dict__.update({
12 '__file__': __file__,
13 '__package__': __package__,
14 '__path__': __path__,
15 })
16
17 # Save the shellcode directory
18 self._dir = directory
19
20 # Find the absolute path of the directory
21 self._absdir = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'templates', self._dir)
22
23 # Get the docstring
24 with open(os.path.join(self._absdir, "__doc__")) as fd:
25 self.__doc__ = fd.read()
26
27 # Insert into the module list
28 sys.modules[self.__name__] = self
29
30 def __lazyinit__(self):
31
32 # Create a dictionary of submodules
33 self._submodules = {}
34 self._shellcodes = {}
35 for name in os.listdir(self._absdir):
36 path = os.path.join(self._absdir, name)
37 if os.path.isdir(path):
38 self._submodules[name] = module(self.__name__ + '.' + name, os.path.join(self._dir, name))
39 elif os.path.isfile(path) and name != '__doc__' and name[0] != '.':
40 funcname, _ext = os.path.splitext(name)
41 if not re.match('^[a-zA-Z][a-zA-Z0-9_]*$', funcname):
42 raise ValueError("found illegal filename, %r" % name)
43 self._shellcodes[funcname] = name
44
45 # Put the submodules into toplevel
46 self.__dict__.update(self._submodules)
47
48 # These are exported
49 self.__all__ = sorted(self._shellcodes.keys() + self._submodules.keys())
50
51 # Make sure this is not called again
52 self.__lazyinit__ = None
53
54 def __getattr__(self, key):
55 self.__lazyinit__ and self.__lazyinit__()
56
57 # Maybe the lazyinit added it
58 if key in self.__dict__:
59 return self.__dict__[key]
60
61 # This function lazy-loads the shellcodes
62 if key in self._shellcodes:
63 real = internal.make_function(key, self._shellcodes[key], self._dir)
64 setattr(self, key, real)
65 return real
66
67 for m in self._context_modules():
68 try:
69 return getattr(m, key)
70 except AttributeError:
71 pass
72
73 raise AttributeError("'module' object has no attribute '%s'" % key)
74
75 def __dir__(self):
76 # This function lists the available submodules, available shellcodes
77 # and potentially shellcodes available in submodules that should be
78 # avilable because of the context
79 self.__lazyinit__ and self.__lazyinit__()
80
81 result = list(self._submodules.keys())
82 result.extend(('__file__', '__package__', '__path__',
83 '__all__', '__name__'))
84 result.extend(self.__shellcodes__())
85
86 return result
87
88 def _context_modules(self):
89 self.__lazyinit__ and self.__lazyinit__()
90 for k, m in self._submodules.items():
91 if k in [context.arch, context.os]:
92 yield m
93
94 def __shellcodes__(self):
95 self.__lazyinit__ and self.__lazyinit__()
96 result = self._shellcodes.keys()
97 for m in self._context_modules():
98 result.extend(m.__shellcodes__())
99 return result
100
101 # To prevent garbage collection
102 tether = sys.modules[__name__]
103
104 # Create the module structure
105 module(__name__, '')
106
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pwnlib/shellcraft/__init__.py b/pwnlib/shellcraft/__init__.py
--- a/pwnlib/shellcraft/__init__.py
+++ b/pwnlib/shellcraft/__init__.py
@@ -102,4 +102,22 @@
tether = sys.modules[__name__]
# Create the module structure
-module(__name__, '')
+shellcraft = module(__name__, '')
+
+class LazyImporter:
+ def find_module(self, fullname, path):
+ if not fullname.startswith('pwnlib.shellcraft.'):
+ return None
+
+ parts = fullname.split('.')[2:]
+ cur = shellcraft
+ for part in parts:
+ cur = getattr(cur, part, None)
+ if not isinstance(cur, ModuleType):
+ return None
+
+ return self
+
+ def load_module(self, fullname):
+ return sys.modules[fullname]
+sys.meta_path.append(LazyImporter())
|
{"golden_diff": "diff --git a/pwnlib/shellcraft/__init__.py b/pwnlib/shellcraft/__init__.py\n--- a/pwnlib/shellcraft/__init__.py\n+++ b/pwnlib/shellcraft/__init__.py\n@@ -102,4 +102,22 @@\n tether = sys.modules[__name__]\n \n # Create the module structure\n-module(__name__, '')\n+shellcraft = module(__name__, '')\n+\n+class LazyImporter:\n+ def find_module(self, fullname, path):\n+ if not fullname.startswith('pwnlib.shellcraft.'):\n+ return None\n+\n+ parts = fullname.split('.')[2:]\n+ cur = shellcraft\n+ for part in parts:\n+ cur = getattr(cur, part, None)\n+ if not isinstance(cur, ModuleType):\n+ return None\n+\n+ return self\n+\n+ def load_module(self, fullname):\n+ return sys.modules[fullname]\n+sys.meta_path.append(LazyImporter())\n", "issue": "Importing shellcraft submodules doesn't work\n```\n$ python\n>>> import pwnlib.shellcraft.arm\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in <module>\nImportError: No module named arm\n```\n\n", "before_files": [{"content": "from types import ModuleType\nimport sys, os, re\nfrom . import internal\nfrom ..context import context\n\nclass module(ModuleType):\n def __init__(self, name, directory):\n super(module, self).__init__(name)\n\n # Insert nice properties\n self.__dict__.update({\n '__file__': __file__,\n '__package__': __package__,\n '__path__': __path__,\n })\n\n # Save the shellcode directory\n self._dir = directory\n\n # Find the absolute path of the directory\n self._absdir = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'templates', self._dir)\n\n # Get the docstring\n with open(os.path.join(self._absdir, \"__doc__\")) as fd:\n self.__doc__ = fd.read()\n\n # Insert into the module list\n sys.modules[self.__name__] = self\n\n def __lazyinit__(self):\n\n # Create a dictionary of submodules\n self._submodules = {}\n self._shellcodes = {}\n for name in os.listdir(self._absdir):\n path = os.path.join(self._absdir, name)\n if os.path.isdir(path):\n self._submodules[name] = module(self.__name__ + '.' + name, os.path.join(self._dir, name))\n elif os.path.isfile(path) and name != '__doc__' and name[0] != '.':\n funcname, _ext = os.path.splitext(name)\n if not re.match('^[a-zA-Z][a-zA-Z0-9_]*$', funcname):\n raise ValueError(\"found illegal filename, %r\" % name)\n self._shellcodes[funcname] = name\n\n # Put the submodules into toplevel\n self.__dict__.update(self._submodules)\n\n # These are exported\n self.__all__ = sorted(self._shellcodes.keys() + self._submodules.keys())\n\n # Make sure this is not called again\n self.__lazyinit__ = None\n\n def __getattr__(self, key):\n self.__lazyinit__ and self.__lazyinit__()\n\n # Maybe the lazyinit added it\n if key in self.__dict__:\n return self.__dict__[key]\n\n # This function lazy-loads the shellcodes\n if key in self._shellcodes:\n real = internal.make_function(key, self._shellcodes[key], self._dir)\n setattr(self, key, real)\n return real\n\n for m in self._context_modules():\n try:\n return getattr(m, key)\n except AttributeError:\n pass\n\n raise AttributeError(\"'module' object has no attribute '%s'\" % key)\n\n def __dir__(self):\n # This function lists the available submodules, available shellcodes\n # and potentially shellcodes available in submodules that should be\n # avilable because of the context\n self.__lazyinit__ and self.__lazyinit__()\n\n result = list(self._submodules.keys())\n result.extend(('__file__', '__package__', '__path__',\n '__all__', '__name__'))\n result.extend(self.__shellcodes__())\n\n return result\n\n def _context_modules(self):\n self.__lazyinit__ and self.__lazyinit__()\n for k, m in self._submodules.items():\n if k in [context.arch, context.os]:\n yield m\n\n def __shellcodes__(self):\n self.__lazyinit__ and self.__lazyinit__()\n result = self._shellcodes.keys()\n for m in self._context_modules():\n result.extend(m.__shellcodes__())\n return result\n\n# To prevent garbage collection\ntether = sys.modules[__name__]\n\n# Create the module structure\nmodule(__name__, '')\n", "path": "pwnlib/shellcraft/__init__.py"}], "after_files": [{"content": "from types import ModuleType\nimport sys, os, re\nfrom . import internal\nfrom ..context import context\n\nclass module(ModuleType):\n def __init__(self, name, directory):\n super(module, self).__init__(name)\n\n # Insert nice properties\n self.__dict__.update({\n '__file__': __file__,\n '__package__': __package__,\n '__path__': __path__,\n })\n\n # Save the shellcode directory\n self._dir = directory\n\n # Find the absolute path of the directory\n self._absdir = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'templates', self._dir)\n\n # Get the docstring\n with open(os.path.join(self._absdir, \"__doc__\")) as fd:\n self.__doc__ = fd.read()\n\n # Insert into the module list\n sys.modules[self.__name__] = self\n\n def __lazyinit__(self):\n\n # Create a dictionary of submodules\n self._submodules = {}\n self._shellcodes = {}\n for name in os.listdir(self._absdir):\n path = os.path.join(self._absdir, name)\n if os.path.isdir(path):\n self._submodules[name] = module(self.__name__ + '.' + name, os.path.join(self._dir, name))\n elif os.path.isfile(path) and name != '__doc__' and name[0] != '.':\n funcname, _ext = os.path.splitext(name)\n if not re.match('^[a-zA-Z][a-zA-Z0-9_]*$', funcname):\n raise ValueError(\"found illegal filename, %r\" % name)\n self._shellcodes[funcname] = name\n\n # Put the submodules into toplevel\n self.__dict__.update(self._submodules)\n\n # These are exported\n self.__all__ = sorted(self._shellcodes.keys() + self._submodules.keys())\n\n # Make sure this is not called again\n self.__lazyinit__ = None\n\n def __getattr__(self, key):\n self.__lazyinit__ and self.__lazyinit__()\n\n # Maybe the lazyinit added it\n if key in self.__dict__:\n return self.__dict__[key]\n\n # This function lazy-loads the shellcodes\n if key in self._shellcodes:\n real = internal.make_function(key, self._shellcodes[key], self._dir)\n setattr(self, key, real)\n return real\n\n for m in self._context_modules():\n try:\n return getattr(m, key)\n except AttributeError:\n pass\n\n raise AttributeError(\"'module' object has no attribute '%s'\" % key)\n\n def __dir__(self):\n # This function lists the available submodules, available shellcodes\n # and potentially shellcodes available in submodules that should be\n # avilable because of the context\n self.__lazyinit__ and self.__lazyinit__()\n\n result = list(self._submodules.keys())\n result.extend(('__file__', '__package__', '__path__',\n '__all__', '__name__'))\n result.extend(self.__shellcodes__())\n\n return result\n\n def _context_modules(self):\n self.__lazyinit__ and self.__lazyinit__()\n for k, m in self._submodules.items():\n if k in [context.arch, context.os]:\n yield m\n\n def __shellcodes__(self):\n self.__lazyinit__ and self.__lazyinit__()\n result = self._shellcodes.keys()\n for m in self._context_modules():\n result.extend(m.__shellcodes__())\n return result\n\n# To prevent garbage collection\ntether = sys.modules[__name__]\n\n# Create the module structure\nshellcraft = module(__name__, '')\n\nclass LazyImporter:\n def find_module(self, fullname, path):\n if not fullname.startswith('pwnlib.shellcraft.'):\n return None\n\n parts = fullname.split('.')[2:]\n cur = shellcraft\n for part in parts:\n cur = getattr(cur, part, None)\n if not isinstance(cur, ModuleType):\n return None\n\n return self\n\n def load_module(self, fullname):\n return sys.modules[fullname]\nsys.meta_path.append(LazyImporter())\n", "path": "pwnlib/shellcraft/__init__.py"}]}
| 1,349 | 215 |
gh_patches_debug_26773
|
rasdani/github-patches
|
git_diff
|
optuna__optuna-4940
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Fix `Test` warnings from `BoTorch`
### Motivation
Resolve following warnings from test:
```
tests/terminator_tests/improvement_tests/gp_tests/test_botorch.py: 4 warnings
tests/visualization_tests/test_terminator_improvement.py: 46 warnings
/opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/botorch/fit.py:139: DeprecationWarning:
`fit_gpytorch_model` is marked for deprecation, consider using `fit_gpytorch_mll` instead.
```
### Suggestion
Look for the proper way to fix this warning (Hopefully, just replacing `fit_gpytorch_model` with `fit_gpytorch_mll`).
### Additional context (optional)
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `optuna/terminator/improvement/gp/botorch.py`
Content:
```
1 from __future__ import annotations
2
3 from typing import Optional
4
5 import numpy as np
6
7 from optuna._imports import try_import
8 from optuna.distributions import _is_distribution_log
9 from optuna.distributions import CategoricalDistribution
10 from optuna.distributions import FloatDistribution
11 from optuna.distributions import IntDistribution
12 from optuna.search_space import intersection_search_space
13 from optuna.terminator.improvement.gp.base import BaseGaussianProcess
14 from optuna.trial._frozen import FrozenTrial
15 from optuna.trial._state import TrialState
16
17
18 with try_import() as _imports:
19 from botorch.fit import fit_gpytorch_model
20 from botorch.models import SingleTaskGP
21 from botorch.models.transforms import Normalize
22 from botorch.models.transforms import Standardize
23 import gpytorch
24 import torch
25
26 __all__ = [
27 "fit_gpytorch_model",
28 "SingleTaskGP",
29 "Normalize",
30 "Standardize",
31 "gpytorch",
32 "torch",
33 ]
34
35
36 class _BoTorchGaussianProcess(BaseGaussianProcess):
37 def __init__(self) -> None:
38 _imports.check()
39
40 self._gp: Optional[SingleTaskGP] = None
41
42 def fit(
43 self,
44 trials: list[FrozenTrial],
45 ) -> None:
46 self._trials = trials
47
48 x, bounds = _convert_trials_to_tensors(trials)
49
50 n_params = x.shape[1]
51
52 y = torch.tensor([trial.value for trial in trials], dtype=torch.float64)
53 y = torch.unsqueeze(y, 1)
54
55 self._gp = SingleTaskGP(
56 x,
57 y,
58 input_transform=Normalize(d=n_params, bounds=bounds),
59 outcome_transform=Standardize(m=1),
60 )
61
62 mll = gpytorch.mlls.ExactMarginalLogLikelihood(self._gp.likelihood, self._gp)
63
64 fit_gpytorch_model(mll)
65
66 def predict_mean_std(
67 self,
68 trials: list[FrozenTrial],
69 ) -> tuple[np.ndarray, np.ndarray]:
70 assert self._gp is not None
71
72 x, _ = _convert_trials_to_tensors(trials)
73
74 with torch.no_grad(), gpytorch.settings.fast_pred_var():
75 posterior = self._gp.posterior(x)
76 mean = posterior.mean
77 variance = posterior.variance
78 std = variance.sqrt()
79
80 return mean.detach().numpy(), std.detach().numpy()
81
82
83 def _convert_trials_to_tensors(trials: list[FrozenTrial]) -> tuple[torch.Tensor, torch.Tensor]:
84 """Convert a list of FrozenTrial objects to tensors inputs and bounds.
85
86 This function assumes the following condition for input trials:
87 - any categorical param is converted to a float or int one;
88 - log is unscaled for any float/int distribution;
89 - the state is COMPLETE for any trial;
90 - direction is MINIMIZE for any trial.
91 """
92 search_space = intersection_search_space(trials)
93 sorted_params = sorted(search_space.keys())
94
95 x = []
96 for trial in trials:
97 assert trial.state == TrialState.COMPLETE
98 x_row = []
99 for param in sorted_params:
100 distribution = search_space[param]
101
102 assert not _is_distribution_log(distribution)
103 assert not isinstance(distribution, CategoricalDistribution)
104
105 param_value = float(trial.params[param])
106 x_row.append(param_value)
107
108 x.append(x_row)
109
110 min_bounds = []
111 max_bounds = []
112 for param, distribution in search_space.items():
113 assert isinstance(distribution, (FloatDistribution, IntDistribution))
114 min_bounds.append(distribution.low)
115 max_bounds.append(distribution.high)
116 bounds = [min_bounds, max_bounds]
117
118 return torch.tensor(x, dtype=torch.float64), torch.tensor(bounds, dtype=torch.float64)
119
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/optuna/terminator/improvement/gp/botorch.py b/optuna/terminator/improvement/gp/botorch.py
--- a/optuna/terminator/improvement/gp/botorch.py
+++ b/optuna/terminator/improvement/gp/botorch.py
@@ -3,6 +3,7 @@
from typing import Optional
import numpy as np
+from packaging import version
from optuna._imports import try_import
from optuna.distributions import _is_distribution_log
@@ -16,15 +17,20 @@
with try_import() as _imports:
- from botorch.fit import fit_gpytorch_model
+ import botorch
from botorch.models import SingleTaskGP
from botorch.models.transforms import Normalize
from botorch.models.transforms import Standardize
import gpytorch
import torch
+ if version.parse(botorch.version.version) < version.parse("0.8.0"):
+ from botorch.fit import fit_gpytorch_model as fit_gpytorch_mll
+ else:
+ from botorch.fit import fit_gpytorch_mll
+
__all__ = [
- "fit_gpytorch_model",
+ "fit_gpytorch_mll",
"SingleTaskGP",
"Normalize",
"Standardize",
@@ -61,7 +67,7 @@
mll = gpytorch.mlls.ExactMarginalLogLikelihood(self._gp.likelihood, self._gp)
- fit_gpytorch_model(mll)
+ fit_gpytorch_mll(mll)
def predict_mean_std(
self,
|
{"golden_diff": "diff --git a/optuna/terminator/improvement/gp/botorch.py b/optuna/terminator/improvement/gp/botorch.py\n--- a/optuna/terminator/improvement/gp/botorch.py\n+++ b/optuna/terminator/improvement/gp/botorch.py\n@@ -3,6 +3,7 @@\n from typing import Optional\n \n import numpy as np\n+from packaging import version\n \n from optuna._imports import try_import\n from optuna.distributions import _is_distribution_log\n@@ -16,15 +17,20 @@\n \n \n with try_import() as _imports:\n- from botorch.fit import fit_gpytorch_model\n+ import botorch\n from botorch.models import SingleTaskGP\n from botorch.models.transforms import Normalize\n from botorch.models.transforms import Standardize\n import gpytorch\n import torch\n \n+ if version.parse(botorch.version.version) < version.parse(\"0.8.0\"):\n+ from botorch.fit import fit_gpytorch_model as fit_gpytorch_mll\n+ else:\n+ from botorch.fit import fit_gpytorch_mll\n+\n __all__ = [\n- \"fit_gpytorch_model\",\n+ \"fit_gpytorch_mll\",\n \"SingleTaskGP\",\n \"Normalize\",\n \"Standardize\",\n@@ -61,7 +67,7 @@\n \n mll = gpytorch.mlls.ExactMarginalLogLikelihood(self._gp.likelihood, self._gp)\n \n- fit_gpytorch_model(mll)\n+ fit_gpytorch_mll(mll)\n \n def predict_mean_std(\n self,\n", "issue": "Fix `Test` warnings from `BoTorch`\n### Motivation\n\nResolve following warnings from test:\r\n```\r\ntests/terminator_tests/improvement_tests/gp_tests/test_botorch.py: 4 warnings\r\ntests/visualization_tests/test_terminator_improvement.py: 46 warnings\r\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/botorch/fit.py:139: DeprecationWarning:\r\n \r\n `fit_gpytorch_model` is marked for deprecation, consider using `fit_gpytorch_mll` instead.\r\n```\n\n### Suggestion\n\nLook for the proper way to fix this warning (Hopefully, just replacing `fit_gpytorch_model` with `fit_gpytorch_mll`).\n\n### Additional context (optional)\n\n_No response_\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom typing import Optional\n\nimport numpy as np\n\nfrom optuna._imports import try_import\nfrom optuna.distributions import _is_distribution_log\nfrom optuna.distributions import CategoricalDistribution\nfrom optuna.distributions import FloatDistribution\nfrom optuna.distributions import IntDistribution\nfrom optuna.search_space import intersection_search_space\nfrom optuna.terminator.improvement.gp.base import BaseGaussianProcess\nfrom optuna.trial._frozen import FrozenTrial\nfrom optuna.trial._state import TrialState\n\n\nwith try_import() as _imports:\n from botorch.fit import fit_gpytorch_model\n from botorch.models import SingleTaskGP\n from botorch.models.transforms import Normalize\n from botorch.models.transforms import Standardize\n import gpytorch\n import torch\n\n__all__ = [\n \"fit_gpytorch_model\",\n \"SingleTaskGP\",\n \"Normalize\",\n \"Standardize\",\n \"gpytorch\",\n \"torch\",\n]\n\n\nclass _BoTorchGaussianProcess(BaseGaussianProcess):\n def __init__(self) -> None:\n _imports.check()\n\n self._gp: Optional[SingleTaskGP] = None\n\n def fit(\n self,\n trials: list[FrozenTrial],\n ) -> None:\n self._trials = trials\n\n x, bounds = _convert_trials_to_tensors(trials)\n\n n_params = x.shape[1]\n\n y = torch.tensor([trial.value for trial in trials], dtype=torch.float64)\n y = torch.unsqueeze(y, 1)\n\n self._gp = SingleTaskGP(\n x,\n y,\n input_transform=Normalize(d=n_params, bounds=bounds),\n outcome_transform=Standardize(m=1),\n )\n\n mll = gpytorch.mlls.ExactMarginalLogLikelihood(self._gp.likelihood, self._gp)\n\n fit_gpytorch_model(mll)\n\n def predict_mean_std(\n self,\n trials: list[FrozenTrial],\n ) -> tuple[np.ndarray, np.ndarray]:\n assert self._gp is not None\n\n x, _ = _convert_trials_to_tensors(trials)\n\n with torch.no_grad(), gpytorch.settings.fast_pred_var():\n posterior = self._gp.posterior(x)\n mean = posterior.mean\n variance = posterior.variance\n std = variance.sqrt()\n\n return mean.detach().numpy(), std.detach().numpy()\n\n\ndef _convert_trials_to_tensors(trials: list[FrozenTrial]) -> tuple[torch.Tensor, torch.Tensor]:\n \"\"\"Convert a list of FrozenTrial objects to tensors inputs and bounds.\n\n This function assumes the following condition for input trials:\n - any categorical param is converted to a float or int one;\n - log is unscaled for any float/int distribution;\n - the state is COMPLETE for any trial;\n - direction is MINIMIZE for any trial.\n \"\"\"\n search_space = intersection_search_space(trials)\n sorted_params = sorted(search_space.keys())\n\n x = []\n for trial in trials:\n assert trial.state == TrialState.COMPLETE\n x_row = []\n for param in sorted_params:\n distribution = search_space[param]\n\n assert not _is_distribution_log(distribution)\n assert not isinstance(distribution, CategoricalDistribution)\n\n param_value = float(trial.params[param])\n x_row.append(param_value)\n\n x.append(x_row)\n\n min_bounds = []\n max_bounds = []\n for param, distribution in search_space.items():\n assert isinstance(distribution, (FloatDistribution, IntDistribution))\n min_bounds.append(distribution.low)\n max_bounds.append(distribution.high)\n bounds = [min_bounds, max_bounds]\n\n return torch.tensor(x, dtype=torch.float64), torch.tensor(bounds, dtype=torch.float64)\n", "path": "optuna/terminator/improvement/gp/botorch.py"}], "after_files": [{"content": "from __future__ import annotations\n\nfrom typing import Optional\n\nimport numpy as np\nfrom packaging import version\n\nfrom optuna._imports import try_import\nfrom optuna.distributions import _is_distribution_log\nfrom optuna.distributions import CategoricalDistribution\nfrom optuna.distributions import FloatDistribution\nfrom optuna.distributions import IntDistribution\nfrom optuna.search_space import intersection_search_space\nfrom optuna.terminator.improvement.gp.base import BaseGaussianProcess\nfrom optuna.trial._frozen import FrozenTrial\nfrom optuna.trial._state import TrialState\n\n\nwith try_import() as _imports:\n import botorch\n from botorch.models import SingleTaskGP\n from botorch.models.transforms import Normalize\n from botorch.models.transforms import Standardize\n import gpytorch\n import torch\n\n if version.parse(botorch.version.version) < version.parse(\"0.8.0\"):\n from botorch.fit import fit_gpytorch_model as fit_gpytorch_mll\n else:\n from botorch.fit import fit_gpytorch_mll\n\n__all__ = [\n \"fit_gpytorch_mll\",\n \"SingleTaskGP\",\n \"Normalize\",\n \"Standardize\",\n \"gpytorch\",\n \"torch\",\n]\n\n\nclass _BoTorchGaussianProcess(BaseGaussianProcess):\n def __init__(self) -> None:\n _imports.check()\n\n self._gp: Optional[SingleTaskGP] = None\n\n def fit(\n self,\n trials: list[FrozenTrial],\n ) -> None:\n self._trials = trials\n\n x, bounds = _convert_trials_to_tensors(trials)\n\n n_params = x.shape[1]\n\n y = torch.tensor([trial.value for trial in trials], dtype=torch.float64)\n y = torch.unsqueeze(y, 1)\n\n self._gp = SingleTaskGP(\n x,\n y,\n input_transform=Normalize(d=n_params, bounds=bounds),\n outcome_transform=Standardize(m=1),\n )\n\n mll = gpytorch.mlls.ExactMarginalLogLikelihood(self._gp.likelihood, self._gp)\n\n fit_gpytorch_mll(mll)\n\n def predict_mean_std(\n self,\n trials: list[FrozenTrial],\n ) -> tuple[np.ndarray, np.ndarray]:\n assert self._gp is not None\n\n x, _ = _convert_trials_to_tensors(trials)\n\n with torch.no_grad(), gpytorch.settings.fast_pred_var():\n posterior = self._gp.posterior(x)\n mean = posterior.mean\n variance = posterior.variance\n std = variance.sqrt()\n\n return mean.detach().numpy(), std.detach().numpy()\n\n\ndef _convert_trials_to_tensors(trials: list[FrozenTrial]) -> tuple[torch.Tensor, torch.Tensor]:\n \"\"\"Convert a list of FrozenTrial objects to tensors inputs and bounds.\n\n This function assumes the following condition for input trials:\n - any categorical param is converted to a float or int one;\n - log is unscaled for any float/int distribution;\n - the state is COMPLETE for any trial;\n - direction is MINIMIZE for any trial.\n \"\"\"\n search_space = intersection_search_space(trials)\n sorted_params = sorted(search_space.keys())\n\n x = []\n for trial in trials:\n assert trial.state == TrialState.COMPLETE\n x_row = []\n for param in sorted_params:\n distribution = search_space[param]\n\n assert not _is_distribution_log(distribution)\n assert not isinstance(distribution, CategoricalDistribution)\n\n param_value = float(trial.params[param])\n x_row.append(param_value)\n\n x.append(x_row)\n\n min_bounds = []\n max_bounds = []\n for param, distribution in search_space.items():\n assert isinstance(distribution, (FloatDistribution, IntDistribution))\n min_bounds.append(distribution.low)\n max_bounds.append(distribution.high)\n bounds = [min_bounds, max_bounds]\n\n return torch.tensor(x, dtype=torch.float64), torch.tensor(bounds, dtype=torch.float64)\n", "path": "optuna/terminator/improvement/gp/botorch.py"}]}
| 1,523 | 373 |
gh_patches_debug_40268
|
rasdani/github-patches
|
git_diff
|
Textualize__textual-2607
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Placeholder cycle per app
The placeholder stores the current color in a classvar.
This means that a second app will start where the previous app left off. And may return a different color from the first iteration.
Not much of an issue in practice, but in the docs the colors change every screenshot. I think we need a separate cycle per app. Suggest we have an independent sequence for each app instance.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `docs/examples/styles/width_comparison.py`
Content:
```
1 from textual.app import App
2 from textual.containers import Horizontal
3 from textual.widgets import Placeholder, Label, Static
4
5
6 class Ruler(Static):
7 def compose(self):
8 ruler_text = "····•" * 100
9 yield Label(ruler_text)
10
11
12 class HeightComparisonApp(App):
13 def compose(self):
14 yield Horizontal(
15 Placeholder(id="cells"), # (1)!
16 Placeholder(id="percent"),
17 Placeholder(id="w"),
18 Placeholder(id="h"),
19 Placeholder(id="vw"),
20 Placeholder(id="vh"),
21 Placeholder(id="auto"),
22 Placeholder(id="fr1"),
23 Placeholder(id="fr3"),
24 )
25 yield Ruler()
26
27
28 app = HeightComparisonApp(css_path="width_comparison.css")
29
```
Path: `src/textual/widgets/_placeholder.py`
Content:
```
1 """Provides a Textual placeholder widget; useful when designing an app's layout."""
2
3 from __future__ import annotations
4
5 from itertools import cycle
6
7 from rich.console import RenderableType
8 from typing_extensions import Literal, Self
9
10 from .. import events
11 from ..css._error_tools import friendly_list
12 from ..reactive import Reactive, reactive
13 from ..widget import Widget
14
15 PlaceholderVariant = Literal["default", "size", "text"]
16 """The different variants of placeholder."""
17
18 _VALID_PLACEHOLDER_VARIANTS_ORDERED: list[PlaceholderVariant] = [
19 "default",
20 "size",
21 "text",
22 ]
23 _VALID_PLACEHOLDER_VARIANTS: set[PlaceholderVariant] = set(
24 _VALID_PLACEHOLDER_VARIANTS_ORDERED
25 )
26 _PLACEHOLDER_BACKGROUND_COLORS = [
27 "#881177",
28 "#aa3355",
29 "#cc6666",
30 "#ee9944",
31 "#eedd00",
32 "#99dd55",
33 "#44dd88",
34 "#22ccbb",
35 "#00bbcc",
36 "#0099cc",
37 "#3366bb",
38 "#663399",
39 ]
40 _LOREM_IPSUM_PLACEHOLDER_TEXT = "Lorem ipsum dolor sit amet, consectetur adipiscing elit. Etiam feugiat ac elit sit amet accumsan. Suspendisse bibendum nec libero quis gravida. Phasellus id eleifend ligula. Nullam imperdiet sem tellus, sed vehicula nisl faucibus sit amet. Praesent iaculis tempor ultricies. Sed lacinia, tellus id rutrum lacinia, sapien sapien congue mauris, sit amet pellentesque quam quam vel nisl. Curabitur vulputate erat pellentesque mauris posuere, non dictum risus mattis."
41
42
43 class InvalidPlaceholderVariant(Exception):
44 """Raised when an invalid Placeholder variant is set."""
45
46
47 class Placeholder(Widget):
48 """A simple placeholder widget to use before you build your custom widgets.
49
50 This placeholder has a couple of variants that show different data.
51 Clicking the placeholder cycles through the available variants, but a placeholder
52 can also be initialised in a specific variant.
53
54 The variants available are:
55
56 | Variant | Placeholder shows |
57 |---------|------------------------------------------------|
58 | default | Identifier label or the ID of the placeholder. |
59 | size | Size of the placeholder. |
60 | text | Lorem Ipsum text. |
61 """
62
63 DEFAULT_CSS = """
64 Placeholder {
65 content-align: center middle;
66 overflow: hidden;
67 color: $text;
68 }
69 Placeholder.-text {
70 padding: 1;
71 }
72 """
73
74 # Consecutive placeholders get assigned consecutive colors.
75 _COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)
76 _SIZE_RENDER_TEMPLATE = "[b]{} x {}[/b]"
77
78 variant: Reactive[PlaceholderVariant] = reactive[PlaceholderVariant]("default")
79
80 _renderables: dict[PlaceholderVariant, str]
81
82 @classmethod
83 def reset_color_cycle(cls) -> None:
84 """Reset the placeholder background color cycle."""
85 cls._COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)
86
87 def __init__(
88 self,
89 label: str | None = None,
90 variant: PlaceholderVariant = "default",
91 *,
92 name: str | None = None,
93 id: str | None = None,
94 classes: str | None = None,
95 ) -> None:
96 """Create a Placeholder widget.
97
98 Args:
99 label: The label to identify the placeholder.
100 If no label is present, uses the placeholder ID instead.
101 variant: The variant of the placeholder.
102 name: The name of the placeholder.
103 id: The ID of the placeholder in the DOM.
104 classes: A space separated string with the CSS classes
105 of the placeholder, if any.
106 """
107 # Create and cache renderables for all the variants.
108 self._renderables = {
109 "default": label if label else f"#{id}" if id else "Placeholder",
110 "size": "",
111 "text": "\n\n".join(_LOREM_IPSUM_PLACEHOLDER_TEXT for _ in range(5)),
112 }
113
114 super().__init__(name=name, id=id, classes=classes)
115
116 self.styles.background = f"{next(Placeholder._COLORS)} 50%"
117
118 self.variant = self.validate_variant(variant)
119 """The current variant of the placeholder."""
120
121 # Set a cycle through the variants with the correct starting point.
122 self._variants_cycle = cycle(_VALID_PLACEHOLDER_VARIANTS_ORDERED)
123 while next(self._variants_cycle) != self.variant:
124 pass
125
126 def render(self) -> RenderableType:
127 """Render the placeholder.
128
129 Returns:
130 The value to render.
131 """
132 return self._renderables[self.variant]
133
134 def cycle_variant(self) -> Self:
135 """Get the next variant in the cycle.
136
137 Returns:
138 The `Placeholder` instance.
139 """
140 self.variant = next(self._variants_cycle)
141 return self
142
143 def watch_variant(
144 self, old_variant: PlaceholderVariant, variant: PlaceholderVariant
145 ) -> None:
146 self.remove_class(f"-{old_variant}")
147 self.add_class(f"-{variant}")
148
149 def validate_variant(self, variant: PlaceholderVariant) -> PlaceholderVariant:
150 """Validate the variant to which the placeholder was set."""
151 if variant not in _VALID_PLACEHOLDER_VARIANTS:
152 raise InvalidPlaceholderVariant(
153 "Valid placeholder variants are "
154 + f"{friendly_list(_VALID_PLACEHOLDER_VARIANTS)}"
155 )
156 return variant
157
158 async def _on_click(self, _: events.Click) -> None:
159 """Click handler to cycle through the placeholder variants."""
160 self.cycle_variant()
161
162 def _on_resize(self, event: events.Resize) -> None:
163 """Update the placeholder "size" variant with the new placeholder size."""
164 self._renderables["size"] = self._SIZE_RENDER_TEMPLATE.format(*event.size)
165 if self.variant == "size":
166 self.refresh()
167
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/docs/examples/styles/width_comparison.py b/docs/examples/styles/width_comparison.py
--- a/docs/examples/styles/width_comparison.py
+++ b/docs/examples/styles/width_comparison.py
@@ -1,6 +1,6 @@
from textual.app import App
from textual.containers import Horizontal
-from textual.widgets import Placeholder, Label, Static
+from textual.widgets import Label, Placeholder, Static
class Ruler(Static):
@@ -9,7 +9,7 @@
yield Label(ruler_text)
-class HeightComparisonApp(App):
+class WidthComparisonApp(App):
def compose(self):
yield Horizontal(
Placeholder(id="cells"), # (1)!
@@ -25,4 +25,6 @@
yield Ruler()
-app = HeightComparisonApp(css_path="width_comparison.css")
+app = WidthComparisonApp(css_path="width_comparison.css")
+if __name__ == "__main__":
+ app.run()
diff --git a/src/textual/widgets/_placeholder.py b/src/textual/widgets/_placeholder.py
--- a/src/textual/widgets/_placeholder.py
+++ b/src/textual/widgets/_placeholder.py
@@ -3,10 +3,14 @@
from __future__ import annotations
from itertools import cycle
+from typing import Iterator
+from weakref import WeakKeyDictionary
from rich.console import RenderableType
from typing_extensions import Literal, Self
+from textual.app import App
+
from .. import events
from ..css._error_tools import friendly_list
from ..reactive import Reactive, reactive
@@ -72,18 +76,13 @@
"""
# Consecutive placeholders get assigned consecutive colors.
- _COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)
+ _COLORS: WeakKeyDictionary[App, Iterator[str]] = WeakKeyDictionary()
_SIZE_RENDER_TEMPLATE = "[b]{} x {}[/b]"
variant: Reactive[PlaceholderVariant] = reactive[PlaceholderVariant]("default")
_renderables: dict[PlaceholderVariant, str]
- @classmethod
- def reset_color_cycle(cls) -> None:
- """Reset the placeholder background color cycle."""
- cls._COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)
-
def __init__(
self,
label: str | None = None,
@@ -113,8 +112,6 @@
super().__init__(name=name, id=id, classes=classes)
- self.styles.background = f"{next(Placeholder._COLORS)} 50%"
-
self.variant = self.validate_variant(variant)
"""The current variant of the placeholder."""
@@ -123,6 +120,13 @@
while next(self._variants_cycle) != self.variant:
pass
+ def on_mount(self) -> None:
+ """Set the color for this placeholder."""
+ colors = Placeholder._COLORS.setdefault(
+ self.app, cycle(_PLACEHOLDER_BACKGROUND_COLORS)
+ )
+ self.styles.background = f"{next(colors)} 50%"
+
def render(self) -> RenderableType:
"""Render the placeholder.
|
{"golden_diff": "diff --git a/docs/examples/styles/width_comparison.py b/docs/examples/styles/width_comparison.py\n--- a/docs/examples/styles/width_comparison.py\n+++ b/docs/examples/styles/width_comparison.py\n@@ -1,6 +1,6 @@\n from textual.app import App\n from textual.containers import Horizontal\n-from textual.widgets import Placeholder, Label, Static\n+from textual.widgets import Label, Placeholder, Static\n \n \n class Ruler(Static):\n@@ -9,7 +9,7 @@\n yield Label(ruler_text)\n \n \n-class HeightComparisonApp(App):\n+class WidthComparisonApp(App):\n def compose(self):\n yield Horizontal(\n Placeholder(id=\"cells\"), # (1)!\n@@ -25,4 +25,6 @@\n yield Ruler()\n \n \n-app = HeightComparisonApp(css_path=\"width_comparison.css\")\n+app = WidthComparisonApp(css_path=\"width_comparison.css\")\n+if __name__ == \"__main__\":\n+ app.run()\ndiff --git a/src/textual/widgets/_placeholder.py b/src/textual/widgets/_placeholder.py\n--- a/src/textual/widgets/_placeholder.py\n+++ b/src/textual/widgets/_placeholder.py\n@@ -3,10 +3,14 @@\n from __future__ import annotations\n \n from itertools import cycle\n+from typing import Iterator\n+from weakref import WeakKeyDictionary\n \n from rich.console import RenderableType\n from typing_extensions import Literal, Self\n \n+from textual.app import App\n+\n from .. import events\n from ..css._error_tools import friendly_list\n from ..reactive import Reactive, reactive\n@@ -72,18 +76,13 @@\n \"\"\"\n \n # Consecutive placeholders get assigned consecutive colors.\n- _COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n+ _COLORS: WeakKeyDictionary[App, Iterator[str]] = WeakKeyDictionary()\n _SIZE_RENDER_TEMPLATE = \"[b]{} x {}[/b]\"\n \n variant: Reactive[PlaceholderVariant] = reactive[PlaceholderVariant](\"default\")\n \n _renderables: dict[PlaceholderVariant, str]\n \n- @classmethod\n- def reset_color_cycle(cls) -> None:\n- \"\"\"Reset the placeholder background color cycle.\"\"\"\n- cls._COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n-\n def __init__(\n self,\n label: str | None = None,\n@@ -113,8 +112,6 @@\n \n super().__init__(name=name, id=id, classes=classes)\n \n- self.styles.background = f\"{next(Placeholder._COLORS)} 50%\"\n-\n self.variant = self.validate_variant(variant)\n \"\"\"The current variant of the placeholder.\"\"\"\n \n@@ -123,6 +120,13 @@\n while next(self._variants_cycle) != self.variant:\n pass\n \n+ def on_mount(self) -> None:\n+ \"\"\"Set the color for this placeholder.\"\"\"\n+ colors = Placeholder._COLORS.setdefault(\n+ self.app, cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n+ )\n+ self.styles.background = f\"{next(colors)} 50%\"\n+\n def render(self) -> RenderableType:\n \"\"\"Render the placeholder.\n", "issue": "Placeholder cycle per app\nThe placeholder stores the current color in a classvar.\n\nThis means that a second app will start where the previous app left off. And may return a different color from the first iteration.\n\nNot much of an issue in practice, but in the docs the colors change every screenshot. I think we need a separate cycle per app. Suggest we have an independent sequence for each app instance.\n", "before_files": [{"content": "from textual.app import App\nfrom textual.containers import Horizontal\nfrom textual.widgets import Placeholder, Label, Static\n\n\nclass Ruler(Static):\n def compose(self):\n ruler_text = \"\u00b7\u00b7\u00b7\u00b7\u2022\" * 100\n yield Label(ruler_text)\n\n\nclass HeightComparisonApp(App):\n def compose(self):\n yield Horizontal(\n Placeholder(id=\"cells\"), # (1)!\n Placeholder(id=\"percent\"),\n Placeholder(id=\"w\"),\n Placeholder(id=\"h\"),\n Placeholder(id=\"vw\"),\n Placeholder(id=\"vh\"),\n Placeholder(id=\"auto\"),\n Placeholder(id=\"fr1\"),\n Placeholder(id=\"fr3\"),\n )\n yield Ruler()\n\n\napp = HeightComparisonApp(css_path=\"width_comparison.css\")\n", "path": "docs/examples/styles/width_comparison.py"}, {"content": "\"\"\"Provides a Textual placeholder widget; useful when designing an app's layout.\"\"\"\n\nfrom __future__ import annotations\n\nfrom itertools import cycle\n\nfrom rich.console import RenderableType\nfrom typing_extensions import Literal, Self\n\nfrom .. import events\nfrom ..css._error_tools import friendly_list\nfrom ..reactive import Reactive, reactive\nfrom ..widget import Widget\n\nPlaceholderVariant = Literal[\"default\", \"size\", \"text\"]\n\"\"\"The different variants of placeholder.\"\"\"\n\n_VALID_PLACEHOLDER_VARIANTS_ORDERED: list[PlaceholderVariant] = [\n \"default\",\n \"size\",\n \"text\",\n]\n_VALID_PLACEHOLDER_VARIANTS: set[PlaceholderVariant] = set(\n _VALID_PLACEHOLDER_VARIANTS_ORDERED\n)\n_PLACEHOLDER_BACKGROUND_COLORS = [\n \"#881177\",\n \"#aa3355\",\n \"#cc6666\",\n \"#ee9944\",\n \"#eedd00\",\n \"#99dd55\",\n \"#44dd88\",\n \"#22ccbb\",\n \"#00bbcc\",\n \"#0099cc\",\n \"#3366bb\",\n \"#663399\",\n]\n_LOREM_IPSUM_PLACEHOLDER_TEXT = \"Lorem ipsum dolor sit amet, consectetur adipiscing elit. Etiam feugiat ac elit sit amet accumsan. Suspendisse bibendum nec libero quis gravida. Phasellus id eleifend ligula. Nullam imperdiet sem tellus, sed vehicula nisl faucibus sit amet. Praesent iaculis tempor ultricies. Sed lacinia, tellus id rutrum lacinia, sapien sapien congue mauris, sit amet pellentesque quam quam vel nisl. Curabitur vulputate erat pellentesque mauris posuere, non dictum risus mattis.\"\n\n\nclass InvalidPlaceholderVariant(Exception):\n \"\"\"Raised when an invalid Placeholder variant is set.\"\"\"\n\n\nclass Placeholder(Widget):\n \"\"\"A simple placeholder widget to use before you build your custom widgets.\n\n This placeholder has a couple of variants that show different data.\n Clicking the placeholder cycles through the available variants, but a placeholder\n can also be initialised in a specific variant.\n\n The variants available are:\n\n | Variant | Placeholder shows |\n |---------|------------------------------------------------|\n | default | Identifier label or the ID of the placeholder. |\n | size | Size of the placeholder. |\n | text | Lorem Ipsum text. |\n \"\"\"\n\n DEFAULT_CSS = \"\"\"\n Placeholder {\n content-align: center middle;\n overflow: hidden;\n color: $text;\n }\n Placeholder.-text {\n padding: 1;\n }\n \"\"\"\n\n # Consecutive placeholders get assigned consecutive colors.\n _COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n _SIZE_RENDER_TEMPLATE = \"[b]{} x {}[/b]\"\n\n variant: Reactive[PlaceholderVariant] = reactive[PlaceholderVariant](\"default\")\n\n _renderables: dict[PlaceholderVariant, str]\n\n @classmethod\n def reset_color_cycle(cls) -> None:\n \"\"\"Reset the placeholder background color cycle.\"\"\"\n cls._COLORS = cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n\n def __init__(\n self,\n label: str | None = None,\n variant: PlaceholderVariant = \"default\",\n *,\n name: str | None = None,\n id: str | None = None,\n classes: str | None = None,\n ) -> None:\n \"\"\"Create a Placeholder widget.\n\n Args:\n label: The label to identify the placeholder.\n If no label is present, uses the placeholder ID instead.\n variant: The variant of the placeholder.\n name: The name of the placeholder.\n id: The ID of the placeholder in the DOM.\n classes: A space separated string with the CSS classes\n of the placeholder, if any.\n \"\"\"\n # Create and cache renderables for all the variants.\n self._renderables = {\n \"default\": label if label else f\"#{id}\" if id else \"Placeholder\",\n \"size\": \"\",\n \"text\": \"\\n\\n\".join(_LOREM_IPSUM_PLACEHOLDER_TEXT for _ in range(5)),\n }\n\n super().__init__(name=name, id=id, classes=classes)\n\n self.styles.background = f\"{next(Placeholder._COLORS)} 50%\"\n\n self.variant = self.validate_variant(variant)\n \"\"\"The current variant of the placeholder.\"\"\"\n\n # Set a cycle through the variants with the correct starting point.\n self._variants_cycle = cycle(_VALID_PLACEHOLDER_VARIANTS_ORDERED)\n while next(self._variants_cycle) != self.variant:\n pass\n\n def render(self) -> RenderableType:\n \"\"\"Render the placeholder.\n\n Returns:\n The value to render.\n \"\"\"\n return self._renderables[self.variant]\n\n def cycle_variant(self) -> Self:\n \"\"\"Get the next variant in the cycle.\n\n Returns:\n The `Placeholder` instance.\n \"\"\"\n self.variant = next(self._variants_cycle)\n return self\n\n def watch_variant(\n self, old_variant: PlaceholderVariant, variant: PlaceholderVariant\n ) -> None:\n self.remove_class(f\"-{old_variant}\")\n self.add_class(f\"-{variant}\")\n\n def validate_variant(self, variant: PlaceholderVariant) -> PlaceholderVariant:\n \"\"\"Validate the variant to which the placeholder was set.\"\"\"\n if variant not in _VALID_PLACEHOLDER_VARIANTS:\n raise InvalidPlaceholderVariant(\n \"Valid placeholder variants are \"\n + f\"{friendly_list(_VALID_PLACEHOLDER_VARIANTS)}\"\n )\n return variant\n\n async def _on_click(self, _: events.Click) -> None:\n \"\"\"Click handler to cycle through the placeholder variants.\"\"\"\n self.cycle_variant()\n\n def _on_resize(self, event: events.Resize) -> None:\n \"\"\"Update the placeholder \"size\" variant with the new placeholder size.\"\"\"\n self._renderables[\"size\"] = self._SIZE_RENDER_TEMPLATE.format(*event.size)\n if self.variant == \"size\":\n self.refresh()\n", "path": "src/textual/widgets/_placeholder.py"}], "after_files": [{"content": "from textual.app import App\nfrom textual.containers import Horizontal\nfrom textual.widgets import Label, Placeholder, Static\n\n\nclass Ruler(Static):\n def compose(self):\n ruler_text = \"\u00b7\u00b7\u00b7\u00b7\u2022\" * 100\n yield Label(ruler_text)\n\n\nclass WidthComparisonApp(App):\n def compose(self):\n yield Horizontal(\n Placeholder(id=\"cells\"), # (1)!\n Placeholder(id=\"percent\"),\n Placeholder(id=\"w\"),\n Placeholder(id=\"h\"),\n Placeholder(id=\"vw\"),\n Placeholder(id=\"vh\"),\n Placeholder(id=\"auto\"),\n Placeholder(id=\"fr1\"),\n Placeholder(id=\"fr3\"),\n )\n yield Ruler()\n\n\napp = WidthComparisonApp(css_path=\"width_comparison.css\")\nif __name__ == \"__main__\":\n app.run()\n", "path": "docs/examples/styles/width_comparison.py"}, {"content": "\"\"\"Provides a Textual placeholder widget; useful when designing an app's layout.\"\"\"\n\nfrom __future__ import annotations\n\nfrom itertools import cycle\nfrom typing import Iterator\nfrom weakref import WeakKeyDictionary\n\nfrom rich.console import RenderableType\nfrom typing_extensions import Literal, Self\n\nfrom textual.app import App\n\nfrom .. import events\nfrom ..css._error_tools import friendly_list\nfrom ..reactive import Reactive, reactive\nfrom ..widget import Widget\n\nPlaceholderVariant = Literal[\"default\", \"size\", \"text\"]\n\"\"\"The different variants of placeholder.\"\"\"\n\n_VALID_PLACEHOLDER_VARIANTS_ORDERED: list[PlaceholderVariant] = [\n \"default\",\n \"size\",\n \"text\",\n]\n_VALID_PLACEHOLDER_VARIANTS: set[PlaceholderVariant] = set(\n _VALID_PLACEHOLDER_VARIANTS_ORDERED\n)\n_PLACEHOLDER_BACKGROUND_COLORS = [\n \"#881177\",\n \"#aa3355\",\n \"#cc6666\",\n \"#ee9944\",\n \"#eedd00\",\n \"#99dd55\",\n \"#44dd88\",\n \"#22ccbb\",\n \"#00bbcc\",\n \"#0099cc\",\n \"#3366bb\",\n \"#663399\",\n]\n_LOREM_IPSUM_PLACEHOLDER_TEXT = \"Lorem ipsum dolor sit amet, consectetur adipiscing elit. Etiam feugiat ac elit sit amet accumsan. Suspendisse bibendum nec libero quis gravida. Phasellus id eleifend ligula. Nullam imperdiet sem tellus, sed vehicula nisl faucibus sit amet. Praesent iaculis tempor ultricies. Sed lacinia, tellus id rutrum lacinia, sapien sapien congue mauris, sit amet pellentesque quam quam vel nisl. Curabitur vulputate erat pellentesque mauris posuere, non dictum risus mattis.\"\n\n\nclass InvalidPlaceholderVariant(Exception):\n \"\"\"Raised when an invalid Placeholder variant is set.\"\"\"\n\n\nclass Placeholder(Widget):\n \"\"\"A simple placeholder widget to use before you build your custom widgets.\n\n This placeholder has a couple of variants that show different data.\n Clicking the placeholder cycles through the available variants, but a placeholder\n can also be initialised in a specific variant.\n\n The variants available are:\n\n | Variant | Placeholder shows |\n |---------|------------------------------------------------|\n | default | Identifier label or the ID of the placeholder. |\n | size | Size of the placeholder. |\n | text | Lorem Ipsum text. |\n \"\"\"\n\n DEFAULT_CSS = \"\"\"\n Placeholder {\n content-align: center middle;\n overflow: hidden;\n color: $text;\n }\n Placeholder.-text {\n padding: 1;\n }\n \"\"\"\n\n # Consecutive placeholders get assigned consecutive colors.\n _COLORS: WeakKeyDictionary[App, Iterator[str]] = WeakKeyDictionary()\n _SIZE_RENDER_TEMPLATE = \"[b]{} x {}[/b]\"\n\n variant: Reactive[PlaceholderVariant] = reactive[PlaceholderVariant](\"default\")\n\n _renderables: dict[PlaceholderVariant, str]\n\n def __init__(\n self,\n label: str | None = None,\n variant: PlaceholderVariant = \"default\",\n *,\n name: str | None = None,\n id: str | None = None,\n classes: str | None = None,\n ) -> None:\n \"\"\"Create a Placeholder widget.\n\n Args:\n label: The label to identify the placeholder.\n If no label is present, uses the placeholder ID instead.\n variant: The variant of the placeholder.\n name: The name of the placeholder.\n id: The ID of the placeholder in the DOM.\n classes: A space separated string with the CSS classes\n of the placeholder, if any.\n \"\"\"\n # Create and cache renderables for all the variants.\n self._renderables = {\n \"default\": label if label else f\"#{id}\" if id else \"Placeholder\",\n \"size\": \"\",\n \"text\": \"\\n\\n\".join(_LOREM_IPSUM_PLACEHOLDER_TEXT for _ in range(5)),\n }\n\n super().__init__(name=name, id=id, classes=classes)\n\n self.variant = self.validate_variant(variant)\n \"\"\"The current variant of the placeholder.\"\"\"\n\n # Set a cycle through the variants with the correct starting point.\n self._variants_cycle = cycle(_VALID_PLACEHOLDER_VARIANTS_ORDERED)\n while next(self._variants_cycle) != self.variant:\n pass\n\n def on_mount(self) -> None:\n \"\"\"Set the color for this placeholder.\"\"\"\n colors = Placeholder._COLORS.setdefault(\n self.app, cycle(_PLACEHOLDER_BACKGROUND_COLORS)\n )\n self.styles.background = f\"{next(colors)} 50%\"\n\n def render(self) -> RenderableType:\n \"\"\"Render the placeholder.\n\n Returns:\n The value to render.\n \"\"\"\n return self._renderables[self.variant]\n\n def cycle_variant(self) -> Self:\n \"\"\"Get the next variant in the cycle.\n\n Returns:\n The `Placeholder` instance.\n \"\"\"\n self.variant = next(self._variants_cycle)\n return self\n\n def watch_variant(\n self, old_variant: PlaceholderVariant, variant: PlaceholderVariant\n ) -> None:\n self.remove_class(f\"-{old_variant}\")\n self.add_class(f\"-{variant}\")\n\n def validate_variant(self, variant: PlaceholderVariant) -> PlaceholderVariant:\n \"\"\"Validate the variant to which the placeholder was set.\"\"\"\n if variant not in _VALID_PLACEHOLDER_VARIANTS:\n raise InvalidPlaceholderVariant(\n \"Valid placeholder variants are \"\n + f\"{friendly_list(_VALID_PLACEHOLDER_VARIANTS)}\"\n )\n return variant\n\n async def _on_click(self, _: events.Click) -> None:\n \"\"\"Click handler to cycle through the placeholder variants.\"\"\"\n self.cycle_variant()\n\n def _on_resize(self, event: events.Resize) -> None:\n \"\"\"Update the placeholder \"size\" variant with the new placeholder size.\"\"\"\n self._renderables[\"size\"] = self._SIZE_RENDER_TEMPLATE.format(*event.size)\n if self.variant == \"size\":\n self.refresh()\n", "path": "src/textual/widgets/_placeholder.py"}]}
| 2,271 | 668 |
gh_patches_debug_24273
|
rasdani/github-patches
|
git_diff
|
open-telemetry__opentelemetry-python-2727
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
exception in prometheus exporter AttributeError: 'frozenset' object has no attribute 'items'
Hi,
Using the code sample from https://github.com/open-telemetry/opentelemetry-python/blob/v1.12.0rc1/exporter/opentelemetry-exporter-prometheus/src/opentelemetry/exporter/prometheus/__init__.py
```
from prometheus_client import start_http_server
from opentelemetry.exporter.prometheus import PrometheusMetricReader
from opentelemetry.metrics import get_meter_provider, set_meter_provider
from opentelemetry.sdk.metrics import MeterProvider
import random
start_http_server(port=8000, addr="localhost")
prefix = "MyAppPrefix"
reader = PrometheusMetricReader(prefix)
set_meter_provider(MeterProvider(metric_readers=[reader]))
meter = get_meter_provider().get_meter("myapp", "0.1.2")
counter = meter.create_counter(
"requests",
"requests",
"number of requests",
)
labels = {"environment": "staging"}
counter.add(25, labels)
```
when accessing http://localhost:8000 and exception is thrown
` File "/Users/m_652923/.pyenv/versions/3.8.9/lib/python3.8/wsgiref/handlers.py", line 137, in run
self.result = application(self.environ, self.start_response)
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py", line 128, in prometheus_app
status, headers, output = _bake_output(registry, accept_header, accept_encoding_header, params, disable_compression)
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py", line 104, in _bake_output
output = encoder(registry)
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py", line 197, in generate_latest
for metric in registry.collect():
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/registry.py", line 97, in collect
yield from collector.collect()
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/opentelemetry/exporter/prometheus/__init__.py", line 166, in collect
self._translate_to_prometheus(
File "/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/opentelemetry/exporter/prometheus/__init__.py", line 204, in _translate_to_prometheus
for key, value in number_data_point.attributes.items():
AttributeError: 'frozenset' object has no attribute 'items'`
```
$pip freeze
backoff==1.11.1
certifi==2022.5.18.1
charset-normalizer==2.0.12
Deprecated==1.2.13
googleapis-common-protos==1.56.1
grpcio==1.46.3
idna==3.3
opentelemetry-api==1.12.0rc1
opentelemetry-exporter-otlp==1.12.0rc1
opentelemetry-exporter-otlp-proto-grpc==1.12.0rc1
opentelemetry-exporter-otlp-proto-http==1.12.0rc1
opentelemetry-exporter-prometheus==1.12.0rc1
opentelemetry-proto==1.12.0rc1
opentelemetry-sdk==1.12.0rc1
opentelemetry-semantic-conventions==0.31b0
prometheus-client==0.14.1
protobuf==3.20.1
requests==2.27.1
six==1.16.0
typing_extensions==4.2.0
urllib3==1.26.9
wrapt==1.14.1
```
This also happens when I try an `up_down_counter`
```
gauge = meter.create_up_down_counter(
'requests_for_endpoints_ms',
'millis',
'Requests for endpoints in milliseconds'
)
res = random.choice(results)
gauge.add(random.randint(10, 40), {'endpoint': res['endpoint']})
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py`
Content:
```
1 # Copyright The OpenTelemetry Authors
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15
16 from logging import getLogger
17 from threading import Lock
18 from typing import Dict, Iterable
19
20 from opentelemetry.metrics import Instrument
21 from opentelemetry.sdk.metrics._internal.aggregation import (
22 Aggregation,
23 DefaultAggregation,
24 _Aggregation,
25 _SumAggregation,
26 )
27 from opentelemetry.sdk.metrics._internal.export import AggregationTemporality
28 from opentelemetry.sdk.metrics._internal.measurement import Measurement
29 from opentelemetry.sdk.metrics._internal.point import DataPointT
30 from opentelemetry.sdk.metrics._internal.view import View
31 from opentelemetry.util._time import _time_ns
32
33 _logger = getLogger(__name__)
34
35
36 class _ViewInstrumentMatch:
37 def __init__(
38 self,
39 view: View,
40 instrument: Instrument,
41 instrument_class_aggregation: Dict[type, Aggregation],
42 ):
43 self._start_time_unix_nano = _time_ns()
44 self._view = view
45 self._instrument = instrument
46 self._attributes_aggregation: Dict[frozenset, _Aggregation] = {}
47 self._lock = Lock()
48 self._instrument_class_aggregation = instrument_class_aggregation
49 self._name = self._view._name or self._instrument.name
50 self._description = (
51 self._view._description or self._instrument.description
52 )
53 if not isinstance(self._view._aggregation, DefaultAggregation):
54 self._aggregation = self._view._aggregation._create_aggregation(
55 self._instrument, None, 0
56 )
57 else:
58 self._aggregation = self._instrument_class_aggregation[
59 self._instrument.__class__
60 ]._create_aggregation(self._instrument, None, 0)
61
62 def conflicts(self, other: "_ViewInstrumentMatch") -> bool:
63 # pylint: disable=protected-access
64
65 result = (
66 self._name == other._name
67 and self._instrument.unit == other._instrument.unit
68 # The aggregation class is being used here instead of data point
69 # type since they are functionally equivalent.
70 and self._aggregation.__class__ == other._aggregation.__class__
71 )
72 if isinstance(self._aggregation, _SumAggregation):
73 result = (
74 result
75 and self._aggregation._instrument_is_monotonic
76 == other._aggregation._instrument_is_monotonic
77 and self._aggregation._instrument_temporality
78 == other._aggregation._instrument_temporality
79 )
80
81 return result
82
83 # pylint: disable=protected-access
84 def consume_measurement(self, measurement: Measurement) -> None:
85
86 if self._view._attribute_keys is not None:
87
88 attributes = {}
89
90 for key, value in (measurement.attributes or {}).items():
91 if key in self._view._attribute_keys:
92 attributes[key] = value
93 elif measurement.attributes is not None:
94 attributes = measurement.attributes
95 else:
96 attributes = {}
97
98 attributes = frozenset(attributes.items())
99
100 if attributes not in self._attributes_aggregation:
101 with self._lock:
102 if attributes not in self._attributes_aggregation:
103 if not isinstance(
104 self._view._aggregation, DefaultAggregation
105 ):
106 aggregation = (
107 self._view._aggregation._create_aggregation(
108 self._instrument,
109 attributes,
110 self._start_time_unix_nano,
111 )
112 )
113 else:
114 aggregation = self._instrument_class_aggregation[
115 self._instrument.__class__
116 ]._create_aggregation(
117 self._instrument,
118 attributes,
119 self._start_time_unix_nano,
120 )
121 self._attributes_aggregation[attributes] = aggregation
122
123 self._attributes_aggregation[attributes].aggregate(measurement)
124
125 def collect(
126 self,
127 aggregation_temporality: AggregationTemporality,
128 collection_start_nanos: int,
129 ) -> Iterable[DataPointT]:
130
131 with self._lock:
132 for aggregation in self._attributes_aggregation.values():
133 yield aggregation.collect(
134 aggregation_temporality, collection_start_nanos
135 )
136
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py
@@ -95,11 +95,11 @@
else:
attributes = {}
- attributes = frozenset(attributes.items())
+ aggr_key = frozenset(attributes.items())
- if attributes not in self._attributes_aggregation:
+ if aggr_key not in self._attributes_aggregation:
with self._lock:
- if attributes not in self._attributes_aggregation:
+ if aggr_key not in self._attributes_aggregation:
if not isinstance(
self._view._aggregation, DefaultAggregation
):
@@ -118,9 +118,9 @@
attributes,
self._start_time_unix_nano,
)
- self._attributes_aggregation[attributes] = aggregation
+ self._attributes_aggregation[aggr_key] = aggregation
- self._attributes_aggregation[attributes].aggregate(measurement)
+ self._attributes_aggregation[aggr_key].aggregate(measurement)
def collect(
self,
|
{"golden_diff": "diff --git a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n--- a/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n+++ b/opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py\n@@ -95,11 +95,11 @@\n else:\n attributes = {}\n \n- attributes = frozenset(attributes.items())\n+ aggr_key = frozenset(attributes.items())\n \n- if attributes not in self._attributes_aggregation:\n+ if aggr_key not in self._attributes_aggregation:\n with self._lock:\n- if attributes not in self._attributes_aggregation:\n+ if aggr_key not in self._attributes_aggregation:\n if not isinstance(\n self._view._aggregation, DefaultAggregation\n ):\n@@ -118,9 +118,9 @@\n attributes,\n self._start_time_unix_nano,\n )\n- self._attributes_aggregation[attributes] = aggregation\n+ self._attributes_aggregation[aggr_key] = aggregation\n \n- self._attributes_aggregation[attributes].aggregate(measurement)\n+ self._attributes_aggregation[aggr_key].aggregate(measurement)\n \n def collect(\n self,\n", "issue": "exception in prometheus exporter AttributeError: 'frozenset' object has no attribute 'items'\nHi,\r\nUsing the code sample from https://github.com/open-telemetry/opentelemetry-python/blob/v1.12.0rc1/exporter/opentelemetry-exporter-prometheus/src/opentelemetry/exporter/prometheus/__init__.py \r\n\r\n```\r\nfrom prometheus_client import start_http_server\r\nfrom opentelemetry.exporter.prometheus import PrometheusMetricReader\r\nfrom opentelemetry.metrics import get_meter_provider, set_meter_provider\r\nfrom opentelemetry.sdk.metrics import MeterProvider\r\nimport random\r\n\r\nstart_http_server(port=8000, addr=\"localhost\")\r\n\r\nprefix = \"MyAppPrefix\"\r\nreader = PrometheusMetricReader(prefix)\r\n\r\nset_meter_provider(MeterProvider(metric_readers=[reader]))\r\nmeter = get_meter_provider().get_meter(\"myapp\", \"0.1.2\")\r\ncounter = meter.create_counter(\r\n \"requests\",\r\n \"requests\",\r\n \"number of requests\",\r\n)\r\nlabels = {\"environment\": \"staging\"}\r\ncounter.add(25, labels)\r\n```\r\n\r\nwhen accessing http://localhost:8000 and exception is thrown\r\n` File \"/Users/m_652923/.pyenv/versions/3.8.9/lib/python3.8/wsgiref/handlers.py\", line 137, in run\r\n self.result = application(self.environ, self.start_response)\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py\", line 128, in prometheus_app\r\n status, headers, output = _bake_output(registry, accept_header, accept_encoding_header, params, disable_compression)\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py\", line 104, in _bake_output\r\n output = encoder(registry)\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/exposition.py\", line 197, in generate_latest\r\n for metric in registry.collect():\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/prometheus_client/registry.py\", line 97, in collect\r\n yield from collector.collect()\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/opentelemetry/exporter/prometheus/__init__.py\", line 166, in collect\r\n self._translate_to_prometheus(\r\n File \"/Users/m_652923/.pyenv/versions/otel/lib/python3.8/site-packages/opentelemetry/exporter/prometheus/__init__.py\", line 204, in _translate_to_prometheus\r\n for key, value in number_data_point.attributes.items():\r\nAttributeError: 'frozenset' object has no attribute 'items'`\r\n\r\n```\r\n$pip freeze\r\nbackoff==1.11.1\r\ncertifi==2022.5.18.1\r\ncharset-normalizer==2.0.12\r\nDeprecated==1.2.13\r\ngoogleapis-common-protos==1.56.1\r\ngrpcio==1.46.3\r\nidna==3.3\r\nopentelemetry-api==1.12.0rc1\r\nopentelemetry-exporter-otlp==1.12.0rc1\r\nopentelemetry-exporter-otlp-proto-grpc==1.12.0rc1\r\nopentelemetry-exporter-otlp-proto-http==1.12.0rc1\r\nopentelemetry-exporter-prometheus==1.12.0rc1\r\nopentelemetry-proto==1.12.0rc1\r\nopentelemetry-sdk==1.12.0rc1\r\nopentelemetry-semantic-conventions==0.31b0\r\nprometheus-client==0.14.1\r\nprotobuf==3.20.1\r\nrequests==2.27.1\r\nsix==1.16.0\r\ntyping_extensions==4.2.0\r\nurllib3==1.26.9\r\nwrapt==1.14.1\r\n```\r\nThis also happens when I try an `up_down_counter`\r\n```\r\ngauge = meter.create_up_down_counter(\r\n 'requests_for_endpoints_ms',\r\n 'millis',\r\n 'Requests for endpoints in milliseconds'\r\n )\r\n\r\nres = random.choice(results)\r\ngauge.add(random.randint(10, 40), {'endpoint': res['endpoint']})\r\n```\n", "before_files": [{"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nfrom logging import getLogger\nfrom threading import Lock\nfrom typing import Dict, Iterable\n\nfrom opentelemetry.metrics import Instrument\nfrom opentelemetry.sdk.metrics._internal.aggregation import (\n Aggregation,\n DefaultAggregation,\n _Aggregation,\n _SumAggregation,\n)\nfrom opentelemetry.sdk.metrics._internal.export import AggregationTemporality\nfrom opentelemetry.sdk.metrics._internal.measurement import Measurement\nfrom opentelemetry.sdk.metrics._internal.point import DataPointT\nfrom opentelemetry.sdk.metrics._internal.view import View\nfrom opentelemetry.util._time import _time_ns\n\n_logger = getLogger(__name__)\n\n\nclass _ViewInstrumentMatch:\n def __init__(\n self,\n view: View,\n instrument: Instrument,\n instrument_class_aggregation: Dict[type, Aggregation],\n ):\n self._start_time_unix_nano = _time_ns()\n self._view = view\n self._instrument = instrument\n self._attributes_aggregation: Dict[frozenset, _Aggregation] = {}\n self._lock = Lock()\n self._instrument_class_aggregation = instrument_class_aggregation\n self._name = self._view._name or self._instrument.name\n self._description = (\n self._view._description or self._instrument.description\n )\n if not isinstance(self._view._aggregation, DefaultAggregation):\n self._aggregation = self._view._aggregation._create_aggregation(\n self._instrument, None, 0\n )\n else:\n self._aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(self._instrument, None, 0)\n\n def conflicts(self, other: \"_ViewInstrumentMatch\") -> bool:\n # pylint: disable=protected-access\n\n result = (\n self._name == other._name\n and self._instrument.unit == other._instrument.unit\n # The aggregation class is being used here instead of data point\n # type since they are functionally equivalent.\n and self._aggregation.__class__ == other._aggregation.__class__\n )\n if isinstance(self._aggregation, _SumAggregation):\n result = (\n result\n and self._aggregation._instrument_is_monotonic\n == other._aggregation._instrument_is_monotonic\n and self._aggregation._instrument_temporality\n == other._aggregation._instrument_temporality\n )\n\n return result\n\n # pylint: disable=protected-access\n def consume_measurement(self, measurement: Measurement) -> None:\n\n if self._view._attribute_keys is not None:\n\n attributes = {}\n\n for key, value in (measurement.attributes or {}).items():\n if key in self._view._attribute_keys:\n attributes[key] = value\n elif measurement.attributes is not None:\n attributes = measurement.attributes\n else:\n attributes = {}\n\n attributes = frozenset(attributes.items())\n\n if attributes not in self._attributes_aggregation:\n with self._lock:\n if attributes not in self._attributes_aggregation:\n if not isinstance(\n self._view._aggregation, DefaultAggregation\n ):\n aggregation = (\n self._view._aggregation._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n )\n else:\n aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n self._attributes_aggregation[attributes] = aggregation\n\n self._attributes_aggregation[attributes].aggregate(measurement)\n\n def collect(\n self,\n aggregation_temporality: AggregationTemporality,\n collection_start_nanos: int,\n ) -> Iterable[DataPointT]:\n\n with self._lock:\n for aggregation in self._attributes_aggregation.values():\n yield aggregation.collect(\n aggregation_temporality, collection_start_nanos\n )\n", "path": "opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py"}], "after_files": [{"content": "# Copyright The OpenTelemetry Authors\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nfrom logging import getLogger\nfrom threading import Lock\nfrom typing import Dict, Iterable\n\nfrom opentelemetry.metrics import Instrument\nfrom opentelemetry.sdk.metrics._internal.aggregation import (\n Aggregation,\n DefaultAggregation,\n _Aggregation,\n _SumAggregation,\n)\nfrom opentelemetry.sdk.metrics._internal.export import AggregationTemporality\nfrom opentelemetry.sdk.metrics._internal.measurement import Measurement\nfrom opentelemetry.sdk.metrics._internal.point import DataPointT\nfrom opentelemetry.sdk.metrics._internal.view import View\nfrom opentelemetry.util._time import _time_ns\n\n_logger = getLogger(__name__)\n\n\nclass _ViewInstrumentMatch:\n def __init__(\n self,\n view: View,\n instrument: Instrument,\n instrument_class_aggregation: Dict[type, Aggregation],\n ):\n self._start_time_unix_nano = _time_ns()\n self._view = view\n self._instrument = instrument\n self._attributes_aggregation: Dict[frozenset, _Aggregation] = {}\n self._lock = Lock()\n self._instrument_class_aggregation = instrument_class_aggregation\n self._name = self._view._name or self._instrument.name\n self._description = (\n self._view._description or self._instrument.description\n )\n if not isinstance(self._view._aggregation, DefaultAggregation):\n self._aggregation = self._view._aggregation._create_aggregation(\n self._instrument, None, 0\n )\n else:\n self._aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(self._instrument, None, 0)\n\n def conflicts(self, other: \"_ViewInstrumentMatch\") -> bool:\n # pylint: disable=protected-access\n\n result = (\n self._name == other._name\n and self._instrument.unit == other._instrument.unit\n # The aggregation class is being used here instead of data point\n # type since they are functionally equivalent.\n and self._aggregation.__class__ == other._aggregation.__class__\n )\n if isinstance(self._aggregation, _SumAggregation):\n result = (\n result\n and self._aggregation._instrument_is_monotonic\n == other._aggregation._instrument_is_monotonic\n and self._aggregation._instrument_temporality\n == other._aggregation._instrument_temporality\n )\n\n return result\n\n # pylint: disable=protected-access\n def consume_measurement(self, measurement: Measurement) -> None:\n\n if self._view._attribute_keys is not None:\n\n attributes = {}\n\n for key, value in (measurement.attributes or {}).items():\n if key in self._view._attribute_keys:\n attributes[key] = value\n elif measurement.attributes is not None:\n attributes = measurement.attributes\n else:\n attributes = {}\n\n aggr_key = frozenset(attributes.items())\n\n if aggr_key not in self._attributes_aggregation:\n with self._lock:\n if aggr_key not in self._attributes_aggregation:\n if not isinstance(\n self._view._aggregation, DefaultAggregation\n ):\n aggregation = (\n self._view._aggregation._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n )\n else:\n aggregation = self._instrument_class_aggregation[\n self._instrument.__class__\n ]._create_aggregation(\n self._instrument,\n attributes,\n self._start_time_unix_nano,\n )\n self._attributes_aggregation[aggr_key] = aggregation\n\n self._attributes_aggregation[aggr_key].aggregate(measurement)\n\n def collect(\n self,\n aggregation_temporality: AggregationTemporality,\n collection_start_nanos: int,\n ) -> Iterable[DataPointT]:\n\n with self._lock:\n for aggregation in self._attributes_aggregation.values():\n yield aggregation.collect(\n aggregation_temporality, collection_start_nanos\n )\n", "path": "opentelemetry-sdk/src/opentelemetry/sdk/metrics/_internal/_view_instrument_match.py"}]}
| 2,580 | 304 |
gh_patches_debug_17033
|
rasdani/github-patches
|
git_diff
|
HybirdCorp__creme_crm-233
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
management command 'geolocation --stat' failed
really not a blocking bug, but seems very easy to fix:
```
self.sysout(f'{Town.objects.count()} town(s) in database.')
TypeError: sysout() missing 1 required positional argument: 'visible'
```
there is just a missing argument in self.sysout()
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `creme/geolocation/management/commands/geolocation.py`
Content:
```
1 # -*- coding: utf-8 -*-
2
3 ################################################################################
4 # Creme is a free/open-source Customer Relationship Management software
5 # Copyright (C) 2015-2021 Hybird
6 #
7 # This program is free software: you can redistribute it and/or modify
8 # it under the terms of the GNU Affero General Public License as published by
9 # the Free Software Foundation, either version 3 of the License, or
10 # (at your option) any later version.
11 #
12 # This program is distributed in the hope that it will be useful,
13 # but WITHOUT ANY WARRANTY; without even the implied warranty of
14 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 # GNU Affero General Public License for more details.
16 #
17 # You should have received a copy of the GNU Affero General Public License
18 # along with this program. If not, see <http://www.gnu.org/licenses/>.
19 ################################################################################
20
21 import csv
22 import io
23 import logging
24 from functools import partial
25 from urllib.request import urlopen
26 from zipfile import ZipFile
27
28 from django.conf import settings
29 from django.core.management.base import BaseCommand
30 from django.db import transaction
31 from django.template.defaultfilters import slugify
32
33 from creme.creme_core.utils.chunktools import iter_as_chunk
34 from creme.creme_core.utils.collections import OrderedSet
35 from creme.creme_core.utils.url import parse_path
36 from creme.persons import get_address_model
37
38 from ...models import GeoAddress, Town
39
40 logger = logging.getLogger(__name__)
41
42
43 class CSVPopulatorError(Exception):
44 pass
45
46
47 class CSVPopulator:
48 class ProtocolError(CSVPopulatorError):
49 pass
50
51 class ReadError(CSVPopulatorError):
52 pass
53
54 class ParseError(CSVPopulatorError):
55 pass
56
57 class Context:
58 def __init__(self, defaults):
59 self.line = 1
60 self.defaults = defaults
61
62 def __init__(self, columns, defaults=None, chunksize=50):
63 """Constructor.
64 @param columns: Names of columns to extract from CSV file.
65 Raises an error if a column is neither in file nor in defaults.
66 @param defaults: dict of default values.
67 @param chunksize: Number of lines in same transaction.
68 By default sqlite supports 999 entries for each transaction,
69 so use 999/fields as max chunksize value.
70 """
71 self.columns = columns
72 self.defaults = defaults or {}
73 self.chunksize = chunksize
74
75 def _get_source_file(self, url_info):
76 if url_info.scheme in {'file', ''}:
77 self.info(f'Reading database from {url_info.geturl()}...')
78 return open(url_info.path, 'rb')
79 elif url_info.scheme in {'http', 'https'}:
80 self.info(f'Downloading database from {url_info.geturl()}...')
81 return urlopen(url_info.geturl())
82 else:
83 raise self.ProtocolError(
84 f'Unable to open CSV data from {url_info.geturl()} : '
85 f'unsupported protocol.'
86 )
87
88 def _mapper(self, header):
89 columns = self.columns
90 defaults = self.defaults
91
92 column_keys = OrderedSet(h.lower() for h in columns) # TODO: OrderedFrozenSet
93 row_keys = frozenset(k.lower() for k in header)
94
95 missings = []
96 constants = {}
97 indices = [(key, index) for index, key in enumerate(header) if key in column_keys]
98
99 for key in column_keys:
100 if key not in row_keys:
101 try:
102 constants[key] = defaults[key]
103 except KeyError:
104 missings.append(key)
105
106 if missings:
107 raise self.ParseError(
108 f"Following columns are missing and haven't got any default value : {missings}"
109 )
110
111 def _aux(row):
112 data = {key: row[index] or defaults.get(key) for key, index in indices}
113 data.update(constants)
114 return data
115
116 return _aux
117
118 def create(self, row, context):
119 raise NotImplementedError
120
121 def save(self, entries, context):
122 raise NotImplementedError
123
124 def pre(self, rows, context):
125 pass
126
127 def post(self, entries, context):
128 pass
129
130 def line_error(self, e, row, context):
131 pass
132
133 def chunk_error(self, e, rows, context):
134 pass
135
136 def info(self, message):
137 logger.info(message)
138
139 def populate(self, source):
140 if isinstance(source, str):
141 try:
142 url_info = parse_path(source)
143
144 with self._get_source_file(url_info) as bytes_input:
145 if url_info.path.endswith('.zip'):
146 archive = ZipFile(
147 bytes_input
148 if bytes_input.seekable() else
149 io.BytesIO(bytes_input.read())
150 )
151
152 with archive.open(archive.namelist()[0]) as zipped_bytes_input:
153 self._populate_from_bytes(zipped_bytes_input)
154 else:
155 self._populate_from_bytes(bytes_input)
156 except CSVPopulatorError:
157 raise
158 except Exception as e:
159 raise self.ReadError(f'Unable to open CSV data from {source} : {e}') from e
160 elif hasattr(source, '__iter__'):
161 self._populate_from_lines(iter(source))
162 else:
163 raise ValueError('The source must be a path or an iterable.')
164
165 def _populate_from_bytes(self, bytes_input):
166 with io.TextIOWrapper(bytes_input) as wrapped_bytes_input:
167 self._populate_from_lines(csv.reader(wrapped_bytes_input))
168
169 def _populate_from_lines(self, lines):
170 mapper = self._mapper(next(lines))
171 context = self.Context(self.defaults)
172
173 for rows in iter_as_chunk(lines, self.chunksize):
174 entries = []
175
176 if mapper:
177 rows = [mapper(row) for row in rows]
178
179 try:
180 self.pre(rows, context)
181
182 for row in rows:
183 try:
184 entries.extend(self.create(row, context))
185 except Exception as e:
186 self.line_error(e, row, context)
187
188 context.line += 1
189
190 self.save(entries, context)
191 self.post(entries, context)
192 except Exception as e:
193 self.chunk_error(e, rows, context)
194
195 def sync(self, model, entries, build_pk):
196 created = []
197 updated = []
198
199 for t in entries:
200 pk = build_pk(t)
201
202 if not pk:
203 created.append(t)
204 else:
205 t.pk = pk
206 updated.append(t)
207
208 with transaction.atomic():
209 model.objects.bulk_create(created)
210
211 # TODO: bulk_update() ?
212 for entry in updated:
213 entry.save(force_update=True)
214
215
216 class CSVTownPopulator(CSVPopulator):
217 def __init__(self, defaults=None, chunksize=100):
218 super().__init__(['title', 'zipcode', 'latitude', 'longitude', 'country'],
219 defaults=defaults, chunksize=chunksize,
220 )
221
222 def line_error(self, e, row, context):
223 logger.error(' invalid data (line %d) : %s', context.line, e)
224
225 def chunk_error(self, e, rows, context):
226 logger.error(' invalid data chunk : %s', e)
227
228 def create(self, row, context):
229 zipcodes = row['zipcode'].split('-')
230
231 name = row['title']
232 latitude = row['latitude']
233 longitude = row['longitude']
234
235 slug = slugify(name)
236 country = row['country']
237
238 build_town = partial(Town, country=country)
239
240 return [build_town(name=name,
241 slug=slug,
242 zipcode=zipcode,
243 latitude=latitude,
244 longitude=longitude,
245 ) for zipcode in zipcodes
246 ]
247
248 def save(self, entries, context):
249 get_existing_pk_by_zipcode = dict(
250 Town.objects.filter(
251 zipcode__in=(t.zipcode for t in entries),
252 slug__in=(t.slug for t in entries),
253 ).values_list('zipcode', 'pk')
254 ).get
255
256 self.sync(
257 Town, entries,
258 lambda town: get_existing_pk_by_zipcode(town.zipcode),
259 )
260
261
262 class Command(BaseCommand):
263 def add_arguments(self, parser):
264 add_argument = parser.add_argument
265 add_argument(
266 '-p', '--populate', action='store_true', dest='populate',
267 help='Populate addresses', default=False,
268 )
269 add_argument(
270 '-s', '--stat', action='store_true', dest='stats',
271 help='Display geolocation database stats', default=False,
272 )
273 add_argument(
274 '-i', '--import', action='store_true', dest='import', default=False,
275 help='Import towns configured in GEOLOCATION_TOWNS setting',
276 )
277
278 def sysout(self, message, visible):
279 if visible:
280 self.stdout.write(message)
281
282 def syserr(self, message):
283 self.stderr.write(message)
284
285 def populate_addresses(self, verbosity=0):
286 self.sysout('Populate geolocation information of addresses...', verbosity > 0)
287 GeoAddress.populate_geoaddresses(get_address_model().objects.exclude(zipcode='', city=''))
288
289 def import_town_database(self, url, defaults):
290 try:
291 CSVTownPopulator(defaults=defaults).populate(url)
292 except Exception as e:
293 self.syserr(str(e))
294
295 def import_town_all(self, verbosity=0):
296 self.sysout('Importing Towns database...', verbosity > 0)
297
298 for url, defaults in settings.GEOLOCATION_TOWNS:
299 self.sysout(url, verbosity > 1)
300 self.import_town_database(url, defaults)
301
302 def print_stats(self):
303 self.sysout(f'{Town.objects.count()} town(s) in database.')
304
305 def handle(self, *args, **options):
306 populate = options.get('populate')
307 stats = options.get('stats')
308 imports = options.get('import')
309 verbosity = options.get('verbosity')
310
311 if stats:
312 self.print_stats()
313
314 if imports:
315 self.import_town_all(verbosity)
316
317 if populate:
318 self.populate_addresses(verbosity)
319
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/creme/geolocation/management/commands/geolocation.py b/creme/geolocation/management/commands/geolocation.py
--- a/creme/geolocation/management/commands/geolocation.py
+++ b/creme/geolocation/management/commands/geolocation.py
@@ -299,8 +299,8 @@
self.sysout(url, verbosity > 1)
self.import_town_database(url, defaults)
- def print_stats(self):
- self.sysout(f'{Town.objects.count()} town(s) in database.')
+ def print_stats(self, verbosity=0):
+ self.sysout(f'{Town.objects.count()} town(s) in database.', verbosity > 0)
def handle(self, *args, **options):
populate = options.get('populate')
@@ -309,7 +309,7 @@
verbosity = options.get('verbosity')
if stats:
- self.print_stats()
+ self.print_stats(verbosity)
if imports:
self.import_town_all(verbosity)
|
{"golden_diff": "diff --git a/creme/geolocation/management/commands/geolocation.py b/creme/geolocation/management/commands/geolocation.py\n--- a/creme/geolocation/management/commands/geolocation.py\n+++ b/creme/geolocation/management/commands/geolocation.py\n@@ -299,8 +299,8 @@\n self.sysout(url, verbosity > 1)\n self.import_town_database(url, defaults)\n \n- def print_stats(self):\n- self.sysout(f'{Town.objects.count()} town(s) in database.')\n+ def print_stats(self, verbosity=0):\n+ self.sysout(f'{Town.objects.count()} town(s) in database.', verbosity > 0)\n \n def handle(self, *args, **options):\n populate = options.get('populate')\n@@ -309,7 +309,7 @@\n verbosity = options.get('verbosity')\n \n if stats:\n- self.print_stats()\n+ self.print_stats(verbosity)\n \n if imports:\n self.import_town_all(verbosity)\n", "issue": "management command 'geolocation --stat' failed\nreally not a blocking bug, but seems very easy to fix:\r\n```\r\n self.sysout(f'{Town.objects.count()} town(s) in database.')\r\nTypeError: sysout() missing 1 required positional argument: 'visible'\r\n```\r\n\r\nthere is just a missing argument in self.sysout()\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n################################################################################\n# Creme is a free/open-source Customer Relationship Management software\n# Copyright (C) 2015-2021 Hybird\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n################################################################################\n\nimport csv\nimport io\nimport logging\nfrom functools import partial\nfrom urllib.request import urlopen\nfrom zipfile import ZipFile\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.db import transaction\nfrom django.template.defaultfilters import slugify\n\nfrom creme.creme_core.utils.chunktools import iter_as_chunk\nfrom creme.creme_core.utils.collections import OrderedSet\nfrom creme.creme_core.utils.url import parse_path\nfrom creme.persons import get_address_model\n\nfrom ...models import GeoAddress, Town\n\nlogger = logging.getLogger(__name__)\n\n\nclass CSVPopulatorError(Exception):\n pass\n\n\nclass CSVPopulator:\n class ProtocolError(CSVPopulatorError):\n pass\n\n class ReadError(CSVPopulatorError):\n pass\n\n class ParseError(CSVPopulatorError):\n pass\n\n class Context:\n def __init__(self, defaults):\n self.line = 1\n self.defaults = defaults\n\n def __init__(self, columns, defaults=None, chunksize=50):\n \"\"\"Constructor.\n @param columns: Names of columns to extract from CSV file.\n Raises an error if a column is neither in file nor in defaults.\n @param defaults: dict of default values.\n @param chunksize: Number of lines in same transaction.\n By default sqlite supports 999 entries for each transaction,\n so use 999/fields as max chunksize value.\n \"\"\"\n self.columns = columns\n self.defaults = defaults or {}\n self.chunksize = chunksize\n\n def _get_source_file(self, url_info):\n if url_info.scheme in {'file', ''}:\n self.info(f'Reading database from {url_info.geturl()}...')\n return open(url_info.path, 'rb')\n elif url_info.scheme in {'http', 'https'}:\n self.info(f'Downloading database from {url_info.geturl()}...')\n return urlopen(url_info.geturl())\n else:\n raise self.ProtocolError(\n f'Unable to open CSV data from {url_info.geturl()} : '\n f'unsupported protocol.'\n )\n\n def _mapper(self, header):\n columns = self.columns\n defaults = self.defaults\n\n column_keys = OrderedSet(h.lower() for h in columns) # TODO: OrderedFrozenSet\n row_keys = frozenset(k.lower() for k in header)\n\n missings = []\n constants = {}\n indices = [(key, index) for index, key in enumerate(header) if key in column_keys]\n\n for key in column_keys:\n if key not in row_keys:\n try:\n constants[key] = defaults[key]\n except KeyError:\n missings.append(key)\n\n if missings:\n raise self.ParseError(\n f\"Following columns are missing and haven't got any default value : {missings}\"\n )\n\n def _aux(row):\n data = {key: row[index] or defaults.get(key) for key, index in indices}\n data.update(constants)\n return data\n\n return _aux\n\n def create(self, row, context):\n raise NotImplementedError\n\n def save(self, entries, context):\n raise NotImplementedError\n\n def pre(self, rows, context):\n pass\n\n def post(self, entries, context):\n pass\n\n def line_error(self, e, row, context):\n pass\n\n def chunk_error(self, e, rows, context):\n pass\n\n def info(self, message):\n logger.info(message)\n\n def populate(self, source):\n if isinstance(source, str):\n try:\n url_info = parse_path(source)\n\n with self._get_source_file(url_info) as bytes_input:\n if url_info.path.endswith('.zip'):\n archive = ZipFile(\n bytes_input\n if bytes_input.seekable() else\n io.BytesIO(bytes_input.read())\n )\n\n with archive.open(archive.namelist()[0]) as zipped_bytes_input:\n self._populate_from_bytes(zipped_bytes_input)\n else:\n self._populate_from_bytes(bytes_input)\n except CSVPopulatorError:\n raise\n except Exception as e:\n raise self.ReadError(f'Unable to open CSV data from {source} : {e}') from e\n elif hasattr(source, '__iter__'):\n self._populate_from_lines(iter(source))\n else:\n raise ValueError('The source must be a path or an iterable.')\n\n def _populate_from_bytes(self, bytes_input):\n with io.TextIOWrapper(bytes_input) as wrapped_bytes_input:\n self._populate_from_lines(csv.reader(wrapped_bytes_input))\n\n def _populate_from_lines(self, lines):\n mapper = self._mapper(next(lines))\n context = self.Context(self.defaults)\n\n for rows in iter_as_chunk(lines, self.chunksize):\n entries = []\n\n if mapper:\n rows = [mapper(row) for row in rows]\n\n try:\n self.pre(rows, context)\n\n for row in rows:\n try:\n entries.extend(self.create(row, context))\n except Exception as e:\n self.line_error(e, row, context)\n\n context.line += 1\n\n self.save(entries, context)\n self.post(entries, context)\n except Exception as e:\n self.chunk_error(e, rows, context)\n\n def sync(self, model, entries, build_pk):\n created = []\n updated = []\n\n for t in entries:\n pk = build_pk(t)\n\n if not pk:\n created.append(t)\n else:\n t.pk = pk\n updated.append(t)\n\n with transaction.atomic():\n model.objects.bulk_create(created)\n\n # TODO: bulk_update() ?\n for entry in updated:\n entry.save(force_update=True)\n\n\nclass CSVTownPopulator(CSVPopulator):\n def __init__(self, defaults=None, chunksize=100):\n super().__init__(['title', 'zipcode', 'latitude', 'longitude', 'country'],\n defaults=defaults, chunksize=chunksize,\n )\n\n def line_error(self, e, row, context):\n logger.error(' invalid data (line %d) : %s', context.line, e)\n\n def chunk_error(self, e, rows, context):\n logger.error(' invalid data chunk : %s', e)\n\n def create(self, row, context):\n zipcodes = row['zipcode'].split('-')\n\n name = row['title']\n latitude = row['latitude']\n longitude = row['longitude']\n\n slug = slugify(name)\n country = row['country']\n\n build_town = partial(Town, country=country)\n\n return [build_town(name=name,\n slug=slug,\n zipcode=zipcode,\n latitude=latitude,\n longitude=longitude,\n ) for zipcode in zipcodes\n ]\n\n def save(self, entries, context):\n get_existing_pk_by_zipcode = dict(\n Town.objects.filter(\n zipcode__in=(t.zipcode for t in entries),\n slug__in=(t.slug for t in entries),\n ).values_list('zipcode', 'pk')\n ).get\n\n self.sync(\n Town, entries,\n lambda town: get_existing_pk_by_zipcode(town.zipcode),\n )\n\n\nclass Command(BaseCommand):\n def add_arguments(self, parser):\n add_argument = parser.add_argument\n add_argument(\n '-p', '--populate', action='store_true', dest='populate',\n help='Populate addresses', default=False,\n )\n add_argument(\n '-s', '--stat', action='store_true', dest='stats',\n help='Display geolocation database stats', default=False,\n )\n add_argument(\n '-i', '--import', action='store_true', dest='import', default=False,\n help='Import towns configured in GEOLOCATION_TOWNS setting',\n )\n\n def sysout(self, message, visible):\n if visible:\n self.stdout.write(message)\n\n def syserr(self, message):\n self.stderr.write(message)\n\n def populate_addresses(self, verbosity=0):\n self.sysout('Populate geolocation information of addresses...', verbosity > 0)\n GeoAddress.populate_geoaddresses(get_address_model().objects.exclude(zipcode='', city=''))\n\n def import_town_database(self, url, defaults):\n try:\n CSVTownPopulator(defaults=defaults).populate(url)\n except Exception as e:\n self.syserr(str(e))\n\n def import_town_all(self, verbosity=0):\n self.sysout('Importing Towns database...', verbosity > 0)\n\n for url, defaults in settings.GEOLOCATION_TOWNS:\n self.sysout(url, verbosity > 1)\n self.import_town_database(url, defaults)\n\n def print_stats(self):\n self.sysout(f'{Town.objects.count()} town(s) in database.')\n\n def handle(self, *args, **options):\n populate = options.get('populate')\n stats = options.get('stats')\n imports = options.get('import')\n verbosity = options.get('verbosity')\n\n if stats:\n self.print_stats()\n\n if imports:\n self.import_town_all(verbosity)\n\n if populate:\n self.populate_addresses(verbosity)\n", "path": "creme/geolocation/management/commands/geolocation.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\n################################################################################\n# Creme is a free/open-source Customer Relationship Management software\n# Copyright (C) 2015-2021 Hybird\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see <http://www.gnu.org/licenses/>.\n################################################################################\n\nimport csv\nimport io\nimport logging\nfrom functools import partial\nfrom urllib.request import urlopen\nfrom zipfile import ZipFile\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.db import transaction\nfrom django.template.defaultfilters import slugify\n\nfrom creme.creme_core.utils.chunktools import iter_as_chunk\nfrom creme.creme_core.utils.collections import OrderedSet\nfrom creme.creme_core.utils.url import parse_path\nfrom creme.persons import get_address_model\n\nfrom ...models import GeoAddress, Town\n\nlogger = logging.getLogger(__name__)\n\n\nclass CSVPopulatorError(Exception):\n pass\n\n\nclass CSVPopulator:\n class ProtocolError(CSVPopulatorError):\n pass\n\n class ReadError(CSVPopulatorError):\n pass\n\n class ParseError(CSVPopulatorError):\n pass\n\n class Context:\n def __init__(self, defaults):\n self.line = 1\n self.defaults = defaults\n\n def __init__(self, columns, defaults=None, chunksize=50):\n \"\"\"Constructor.\n @param columns: Names of columns to extract from CSV file.\n Raises an error if a column is neither in file nor in defaults.\n @param defaults: dict of default values.\n @param chunksize: Number of lines in same transaction.\n By default sqlite supports 999 entries for each transaction,\n so use 999/fields as max chunksize value.\n \"\"\"\n self.columns = columns\n self.defaults = defaults or {}\n self.chunksize = chunksize\n\n def _get_source_file(self, url_info):\n if url_info.scheme in {'file', ''}:\n self.info(f'Reading database from {url_info.geturl()}...')\n return open(url_info.path, 'rb')\n elif url_info.scheme in {'http', 'https'}:\n self.info(f'Downloading database from {url_info.geturl()}...')\n return urlopen(url_info.geturl())\n else:\n raise self.ProtocolError(\n f'Unable to open CSV data from {url_info.geturl()} : '\n f'unsupported protocol.'\n )\n\n def _mapper(self, header):\n columns = self.columns\n defaults = self.defaults\n\n column_keys = OrderedSet(h.lower() for h in columns) # TODO: OrderedFrozenSet\n row_keys = frozenset(k.lower() for k in header)\n\n missings = []\n constants = {}\n indices = [(key, index) for index, key in enumerate(header) if key in column_keys]\n\n for key in column_keys:\n if key not in row_keys:\n try:\n constants[key] = defaults[key]\n except KeyError:\n missings.append(key)\n\n if missings:\n raise self.ParseError(\n f\"Following columns are missing and haven't got any default value : {missings}\"\n )\n\n def _aux(row):\n data = {key: row[index] or defaults.get(key) for key, index in indices}\n data.update(constants)\n return data\n\n return _aux\n\n def create(self, row, context):\n raise NotImplementedError\n\n def save(self, entries, context):\n raise NotImplementedError\n\n def pre(self, rows, context):\n pass\n\n def post(self, entries, context):\n pass\n\n def line_error(self, e, row, context):\n pass\n\n def chunk_error(self, e, rows, context):\n pass\n\n def info(self, message):\n logger.info(message)\n\n def populate(self, source):\n if isinstance(source, str):\n try:\n url_info = parse_path(source)\n\n with self._get_source_file(url_info) as bytes_input:\n if url_info.path.endswith('.zip'):\n archive = ZipFile(\n bytes_input\n if bytes_input.seekable() else\n io.BytesIO(bytes_input.read())\n )\n\n with archive.open(archive.namelist()[0]) as zipped_bytes_input:\n self._populate_from_bytes(zipped_bytes_input)\n else:\n self._populate_from_bytes(bytes_input)\n except CSVPopulatorError:\n raise\n except Exception as e:\n raise self.ReadError(f'Unable to open CSV data from {source} : {e}') from e\n elif hasattr(source, '__iter__'):\n self._populate_from_lines(iter(source))\n else:\n raise ValueError('The source must be a path or an iterable.')\n\n def _populate_from_bytes(self, bytes_input):\n with io.TextIOWrapper(bytes_input) as wrapped_bytes_input:\n self._populate_from_lines(csv.reader(wrapped_bytes_input))\n\n def _populate_from_lines(self, lines):\n mapper = self._mapper(next(lines))\n context = self.Context(self.defaults)\n\n for rows in iter_as_chunk(lines, self.chunksize):\n entries = []\n\n if mapper:\n rows = [mapper(row) for row in rows]\n\n try:\n self.pre(rows, context)\n\n for row in rows:\n try:\n entries.extend(self.create(row, context))\n except Exception as e:\n self.line_error(e, row, context)\n\n context.line += 1\n\n self.save(entries, context)\n self.post(entries, context)\n except Exception as e:\n self.chunk_error(e, rows, context)\n\n def sync(self, model, entries, build_pk):\n created = []\n updated = []\n\n for t in entries:\n pk = build_pk(t)\n\n if not pk:\n created.append(t)\n else:\n t.pk = pk\n updated.append(t)\n\n with transaction.atomic():\n model.objects.bulk_create(created)\n\n # TODO: bulk_update() ?\n for entry in updated:\n entry.save(force_update=True)\n\n\nclass CSVTownPopulator(CSVPopulator):\n def __init__(self, defaults=None, chunksize=100):\n super().__init__(['title', 'zipcode', 'latitude', 'longitude', 'country'],\n defaults=defaults, chunksize=chunksize,\n )\n\n def line_error(self, e, row, context):\n logger.error(' invalid data (line %d) : %s', context.line, e)\n\n def chunk_error(self, e, rows, context):\n logger.error(' invalid data chunk : %s', e)\n\n def create(self, row, context):\n zipcodes = row['zipcode'].split('-')\n\n name = row['title']\n latitude = row['latitude']\n longitude = row['longitude']\n\n slug = slugify(name)\n country = row['country']\n\n build_town = partial(Town, country=country)\n\n return [build_town(name=name,\n slug=slug,\n zipcode=zipcode,\n latitude=latitude,\n longitude=longitude,\n ) for zipcode in zipcodes\n ]\n\n def save(self, entries, context):\n get_existing_pk_by_zipcode = dict(\n Town.objects.filter(\n zipcode__in=(t.zipcode for t in entries),\n slug__in=(t.slug for t in entries),\n ).values_list('zipcode', 'pk')\n ).get\n\n self.sync(\n Town, entries,\n lambda town: get_existing_pk_by_zipcode(town.zipcode),\n )\n\n\nclass Command(BaseCommand):\n def add_arguments(self, parser):\n add_argument = parser.add_argument\n add_argument(\n '-p', '--populate', action='store_true', dest='populate',\n help='Populate addresses', default=False,\n )\n add_argument(\n '-s', '--stat', action='store_true', dest='stats',\n help='Display geolocation database stats', default=False,\n )\n add_argument(\n '-i', '--import', action='store_true', dest='import', default=False,\n help='Import towns configured in GEOLOCATION_TOWNS setting',\n )\n\n def sysout(self, message, visible):\n if visible:\n self.stdout.write(message)\n\n def syserr(self, message):\n self.stderr.write(message)\n\n def populate_addresses(self, verbosity=0):\n self.sysout('Populate geolocation information of addresses...', verbosity > 0)\n GeoAddress.populate_geoaddresses(get_address_model().objects.exclude(zipcode='', city=''))\n\n def import_town_database(self, url, defaults):\n try:\n CSVTownPopulator(defaults=defaults).populate(url)\n except Exception as e:\n self.syserr(str(e))\n\n def import_town_all(self, verbosity=0):\n self.sysout('Importing Towns database...', verbosity > 0)\n\n for url, defaults in settings.GEOLOCATION_TOWNS:\n self.sysout(url, verbosity > 1)\n self.import_town_database(url, defaults)\n\n def print_stats(self, verbosity=0):\n self.sysout(f'{Town.objects.count()} town(s) in database.', verbosity > 0)\n\n def handle(self, *args, **options):\n populate = options.get('populate')\n stats = options.get('stats')\n imports = options.get('import')\n verbosity = options.get('verbosity')\n\n if stats:\n self.print_stats(verbosity)\n\n if imports:\n self.import_town_all(verbosity)\n\n if populate:\n self.populate_addresses(verbosity)\n", "path": "creme/geolocation/management/commands/geolocation.py"}]}
| 3,378 | 228 |
gh_patches_debug_8875
|
rasdani/github-patches
|
git_diff
|
microsoft__botbuilder-python-1401
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add tests for SkillHttpClient
see dotnet and javascript imp
[enhancement]
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py`
Content:
```
1 # Copyright (c) Microsoft Corporation. All rights reserved.
2 # Licensed under the MIT License.
3
4 from logging import Logger
5
6 from botbuilder.core import InvokeResponse
7 from botbuilder.integration.aiohttp import BotFrameworkHttpClient
8 from botbuilder.core.skills import (
9 ConversationIdFactoryBase,
10 SkillConversationIdFactoryOptions,
11 BotFrameworkSkill,
12 )
13 from botbuilder.schema import Activity
14 from botframework.connector.auth import (
15 AuthenticationConstants,
16 ChannelProvider,
17 GovernmentConstants,
18 SimpleCredentialProvider,
19 )
20
21
22 class SkillHttpClient(BotFrameworkHttpClient):
23 def __init__(
24 self,
25 credential_provider: SimpleCredentialProvider,
26 skill_conversation_id_factory: ConversationIdFactoryBase,
27 channel_provider: ChannelProvider = None,
28 logger: Logger = None,
29 ):
30 if not skill_conversation_id_factory:
31 raise TypeError(
32 "SkillHttpClient(): skill_conversation_id_factory can't be None"
33 )
34
35 super().__init__(credential_provider)
36
37 self._skill_conversation_id_factory = skill_conversation_id_factory
38 self._channel_provider = channel_provider
39
40 async def post_activity_to_skill(
41 self,
42 from_bot_id: str,
43 to_skill: BotFrameworkSkill,
44 service_url: str,
45 activity: Activity,
46 originating_audience: str = None,
47 ) -> InvokeResponse:
48
49 if originating_audience is None:
50 originating_audience = (
51 GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
52 if self._channel_provider is not None
53 and self._channel_provider.IsGovernment()
54 else AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
55 )
56
57 options = SkillConversationIdFactoryOptions(
58 from_bot_oauth_scope=originating_audience,
59 from_bot_id=from_bot_id,
60 activity=activity,
61 bot_framework_skill=to_skill,
62 )
63
64 skill_conversation_id = await self._skill_conversation_id_factory.create_skill_conversation_id(
65 options
66 )
67
68 return await super().post_activity(
69 from_bot_id,
70 to_skill.app_id,
71 to_skill.skill_endpoint,
72 service_url,
73 skill_conversation_id,
74 activity,
75 )
76
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py b/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py
--- a/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py
+++ b/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py
@@ -50,7 +50,7 @@
originating_audience = (
GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
if self._channel_provider is not None
- and self._channel_provider.IsGovernment()
+ and self._channel_provider.is_government()
else AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE
)
|
{"golden_diff": "diff --git a/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py b/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py\n--- a/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py\n+++ b/libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py\n@@ -50,7 +50,7 @@\n originating_audience = (\n GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n if self._channel_provider is not None\n- and self._channel_provider.IsGovernment()\n+ and self._channel_provider.is_government()\n else AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n )\n", "issue": "Add tests for SkillHttpClient\nsee dotnet and javascript imp\r\n\r\n[enhancement]\r\n\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\nfrom logging import Logger\n\nfrom botbuilder.core import InvokeResponse\nfrom botbuilder.integration.aiohttp import BotFrameworkHttpClient\nfrom botbuilder.core.skills import (\n ConversationIdFactoryBase,\n SkillConversationIdFactoryOptions,\n BotFrameworkSkill,\n)\nfrom botbuilder.schema import Activity\nfrom botframework.connector.auth import (\n AuthenticationConstants,\n ChannelProvider,\n GovernmentConstants,\n SimpleCredentialProvider,\n)\n\n\nclass SkillHttpClient(BotFrameworkHttpClient):\n def __init__(\n self,\n credential_provider: SimpleCredentialProvider,\n skill_conversation_id_factory: ConversationIdFactoryBase,\n channel_provider: ChannelProvider = None,\n logger: Logger = None,\n ):\n if not skill_conversation_id_factory:\n raise TypeError(\n \"SkillHttpClient(): skill_conversation_id_factory can't be None\"\n )\n\n super().__init__(credential_provider)\n\n self._skill_conversation_id_factory = skill_conversation_id_factory\n self._channel_provider = channel_provider\n\n async def post_activity_to_skill(\n self,\n from_bot_id: str,\n to_skill: BotFrameworkSkill,\n service_url: str,\n activity: Activity,\n originating_audience: str = None,\n ) -> InvokeResponse:\n\n if originating_audience is None:\n originating_audience = (\n GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n if self._channel_provider is not None\n and self._channel_provider.IsGovernment()\n else AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n )\n\n options = SkillConversationIdFactoryOptions(\n from_bot_oauth_scope=originating_audience,\n from_bot_id=from_bot_id,\n activity=activity,\n bot_framework_skill=to_skill,\n )\n\n skill_conversation_id = await self._skill_conversation_id_factory.create_skill_conversation_id(\n options\n )\n\n return await super().post_activity(\n from_bot_id,\n to_skill.app_id,\n to_skill.skill_endpoint,\n service_url,\n skill_conversation_id,\n activity,\n )\n", "path": "libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py"}], "after_files": [{"content": "# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License.\n\nfrom logging import Logger\n\nfrom botbuilder.core import InvokeResponse\nfrom botbuilder.integration.aiohttp import BotFrameworkHttpClient\nfrom botbuilder.core.skills import (\n ConversationIdFactoryBase,\n SkillConversationIdFactoryOptions,\n BotFrameworkSkill,\n)\nfrom botbuilder.schema import Activity\nfrom botframework.connector.auth import (\n AuthenticationConstants,\n ChannelProvider,\n GovernmentConstants,\n SimpleCredentialProvider,\n)\n\n\nclass SkillHttpClient(BotFrameworkHttpClient):\n def __init__(\n self,\n credential_provider: SimpleCredentialProvider,\n skill_conversation_id_factory: ConversationIdFactoryBase,\n channel_provider: ChannelProvider = None,\n logger: Logger = None,\n ):\n if not skill_conversation_id_factory:\n raise TypeError(\n \"SkillHttpClient(): skill_conversation_id_factory can't be None\"\n )\n\n super().__init__(credential_provider)\n\n self._skill_conversation_id_factory = skill_conversation_id_factory\n self._channel_provider = channel_provider\n\n async def post_activity_to_skill(\n self,\n from_bot_id: str,\n to_skill: BotFrameworkSkill,\n service_url: str,\n activity: Activity,\n originating_audience: str = None,\n ) -> InvokeResponse:\n\n if originating_audience is None:\n originating_audience = (\n GovernmentConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n if self._channel_provider is not None\n and self._channel_provider.is_government()\n else AuthenticationConstants.TO_CHANNEL_FROM_BOT_OAUTH_SCOPE\n )\n\n options = SkillConversationIdFactoryOptions(\n from_bot_oauth_scope=originating_audience,\n from_bot_id=from_bot_id,\n activity=activity,\n bot_framework_skill=to_skill,\n )\n\n skill_conversation_id = await self._skill_conversation_id_factory.create_skill_conversation_id(\n options\n )\n\n return await super().post_activity(\n from_bot_id,\n to_skill.app_id,\n to_skill.skill_endpoint,\n service_url,\n skill_conversation_id,\n activity,\n )\n", "path": "libraries/botbuilder-integration-aiohttp/botbuilder/integration/aiohttp/skills/skill_http_client.py"}]}
| 894 | 189 |
gh_patches_debug_1421
|
rasdani/github-patches
|
git_diff
|
beeware__toga-2582
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Dialog windows are not modal
### Describe the bug
Create a dialog like this:
```
async def on_button_test(widget):
await self.main_window.info_dialog(title="Dialog", message="An Info Dialog")
```
You can now click on the main window behind the dialog and the main window is activated although the dialog is still floating above the window. This allows the dialog to be shown again or other actions to be triggered although the user should be blocked in the dialog.
Either change the behavior of dialogs to be modal or add a "modal={True|False}" property to "info_dialog" (and all similar functions) to enforce this.
### Steps to reproduce
1) Create self.main_window.info_dialog
2) Click main window
3) Main window is activated and can be interacted with.
### Expected behavior
Main window can not be activated (dialog keeps focus)
### Screenshots
_No response_
### Environment
Linux (GenToo). Toga in virtual environment.
### Logs
_No response_
### Additional context
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `gtk/src/toga_gtk/dialogs.py`
Content:
```
1 from abc import ABC
2 from pathlib import Path
3
4 from .libs import Gtk
5
6
7 class BaseDialog(ABC):
8 def __init__(self, interface):
9 self.interface = interface
10 self.interface._impl = self
11
12
13 class MessageDialog(BaseDialog):
14 def __init__(
15 self,
16 interface,
17 title,
18 message_type,
19 buttons,
20 success_result=None,
21 **kwargs,
22 ):
23 super().__init__(interface=interface)
24 self.success_result = success_result
25
26 self.native = Gtk.MessageDialog(
27 transient_for=interface.window._impl.native,
28 flags=0,
29 message_type=message_type,
30 buttons=buttons,
31 text=title,
32 )
33 self.build_dialog(**kwargs)
34
35 self.native.connect("response", self.gtk_response)
36 self.native.show()
37
38 def build_dialog(self, message):
39 self.native.format_secondary_text(message)
40
41 def gtk_response(self, dialog, response):
42 if self.success_result:
43 result = response == self.success_result
44 else:
45 result = None
46
47 self.interface.set_result(result)
48
49 self.native.destroy()
50
51
52 class InfoDialog(MessageDialog):
53 def __init__(self, interface, title, message):
54 super().__init__(
55 interface=interface,
56 title=title,
57 message=message,
58 message_type=Gtk.MessageType.INFO,
59 buttons=Gtk.ButtonsType.OK,
60 )
61
62
63 class QuestionDialog(MessageDialog):
64 def __init__(self, interface, title, message):
65 super().__init__(
66 interface=interface,
67 title=title,
68 message=message,
69 message_type=Gtk.MessageType.QUESTION,
70 buttons=Gtk.ButtonsType.YES_NO,
71 success_result=Gtk.ResponseType.YES,
72 )
73
74
75 class ConfirmDialog(MessageDialog):
76 def __init__(self, interface, title, message):
77 super().__init__(
78 interface=interface,
79 title=title,
80 message=message,
81 message_type=Gtk.MessageType.WARNING,
82 buttons=Gtk.ButtonsType.OK_CANCEL,
83 success_result=Gtk.ResponseType.OK,
84 )
85
86
87 class ErrorDialog(MessageDialog):
88 def __init__(self, interface, title, message):
89 super().__init__(
90 interface=interface,
91 title=title,
92 message=message,
93 message_type=Gtk.MessageType.ERROR,
94 buttons=Gtk.ButtonsType.CANCEL,
95 )
96
97
98 class StackTraceDialog(MessageDialog):
99 def __init__(self, interface, title, **kwargs):
100 super().__init__(
101 interface=interface,
102 title=title,
103 message_type=Gtk.MessageType.ERROR,
104 buttons=(
105 Gtk.ButtonsType.CANCEL if kwargs.get("retry") else Gtk.ButtonsType.OK
106 ),
107 success_result=Gtk.ResponseType.OK if kwargs.get("retry") else None,
108 **kwargs,
109 )
110
111 def build_dialog(self, message, content, retry):
112 container = self.native.get_message_area()
113
114 self.native.format_secondary_text(message)
115
116 # Create a scrolling readonly text area, in monospace font, to contain the stack trace.
117 buffer = Gtk.TextBuffer()
118 buffer.set_text(content)
119
120 trace = Gtk.TextView()
121 trace.set_buffer(buffer)
122 trace.set_wrap_mode(Gtk.WrapMode.WORD_CHAR)
123 trace.set_property("editable", False)
124 trace.set_property("cursor-visible", False)
125
126 trace.get_style_context().add_class("toga")
127 trace.get_style_context().add_class("stacktrace")
128 trace.get_style_context().add_class("dialog")
129
130 style_provider = Gtk.CssProvider()
131 style_provider.load_from_data(b".toga.stacktrace {font-family: monospace;}")
132
133 trace.get_style_context().add_provider(
134 style_provider,
135 Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION,
136 )
137
138 scroll = Gtk.ScrolledWindow()
139 scroll.set_policy(Gtk.PolicyType.NEVER, Gtk.PolicyType.AUTOMATIC)
140 scroll.set_size_request(500, 200)
141 scroll.add(trace)
142
143 container.pack_end(scroll, False, False, 0)
144
145 container.show_all()
146
147 # If this is a retry dialog, add a retry button (which maps to OK).
148 if retry:
149 self.native.add_button("Retry", Gtk.ResponseType.OK)
150
151
152 class FileDialog(BaseDialog):
153 def __init__(
154 self,
155 interface,
156 title,
157 filename,
158 initial_directory,
159 file_types,
160 multiple_select,
161 action,
162 ok_icon,
163 ):
164 super().__init__(interface=interface)
165
166 self.native = Gtk.FileChooserDialog(
167 transient_for=interface.window._impl.native,
168 title=title,
169 action=action,
170 )
171 self.native.add_button(Gtk.STOCK_CANCEL, Gtk.ResponseType.CANCEL)
172 self.native.add_button(ok_icon, Gtk.ResponseType.OK)
173
174 if filename:
175 self.native.set_current_name(filename)
176
177 if initial_directory:
178 self.native.set_current_folder(str(initial_directory))
179
180 if file_types:
181 for file_type in file_types:
182 filter_filetype = Gtk.FileFilter()
183 filter_filetype.set_name("." + file_type + " files")
184 filter_filetype.add_pattern("*." + file_type)
185 self.native.add_filter(filter_filetype)
186
187 self.multiple_select = multiple_select
188 if self.multiple_select:
189 self.native.set_select_multiple(True)
190
191 self.native.connect("response", self.gtk_response)
192 self.native.show()
193
194 # Provided as a stub that can be mocked in test conditions
195 def selected_path(self):
196 return self.native.get_filename()
197
198 # Provided as a stub that can be mocked in test conditions
199 def selected_paths(self):
200 return self.native.get_filenames()
201
202 def gtk_response(self, dialog, response):
203 if response == Gtk.ResponseType.OK:
204 if self.multiple_select:
205 result = [Path(filename) for filename in self.selected_paths()]
206 else:
207 result = Path(self.selected_path())
208 else:
209 result = None
210
211 self.interface.set_result(result)
212
213 self.native.destroy()
214
215
216 class SaveFileDialog(FileDialog):
217 def __init__(
218 self,
219 interface,
220 title,
221 filename,
222 initial_directory,
223 file_types=None,
224 ):
225 super().__init__(
226 interface=interface,
227 title=title,
228 filename=filename,
229 initial_directory=initial_directory,
230 file_types=file_types,
231 multiple_select=False,
232 action=Gtk.FileChooserAction.SAVE,
233 ok_icon=Gtk.STOCK_SAVE,
234 )
235
236
237 class OpenFileDialog(FileDialog):
238 def __init__(
239 self,
240 interface,
241 title,
242 initial_directory,
243 file_types,
244 multiple_select,
245 ):
246 super().__init__(
247 interface=interface,
248 title=title,
249 filename=None,
250 initial_directory=initial_directory,
251 file_types=file_types,
252 multiple_select=multiple_select,
253 action=Gtk.FileChooserAction.OPEN,
254 ok_icon=Gtk.STOCK_OPEN,
255 )
256
257
258 class SelectFolderDialog(FileDialog):
259 def __init__(
260 self,
261 interface,
262 title,
263 initial_directory,
264 multiple_select,
265 ):
266 super().__init__(
267 interface=interface,
268 title=title,
269 filename=None,
270 initial_directory=initial_directory,
271 file_types=None,
272 multiple_select=multiple_select,
273 action=Gtk.FileChooserAction.SELECT_FOLDER,
274 ok_icon=Gtk.STOCK_OPEN,
275 )
276
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/gtk/src/toga_gtk/dialogs.py b/gtk/src/toga_gtk/dialogs.py
--- a/gtk/src/toga_gtk/dialogs.py
+++ b/gtk/src/toga_gtk/dialogs.py
@@ -30,6 +30,7 @@
buttons=buttons,
text=title,
)
+ self.native.set_modal(True)
self.build_dialog(**kwargs)
self.native.connect("response", self.gtk_response)
|
{"golden_diff": "diff --git a/gtk/src/toga_gtk/dialogs.py b/gtk/src/toga_gtk/dialogs.py\n--- a/gtk/src/toga_gtk/dialogs.py\n+++ b/gtk/src/toga_gtk/dialogs.py\n@@ -30,6 +30,7 @@\n buttons=buttons,\n text=title,\n )\n+ self.native.set_modal(True)\n self.build_dialog(**kwargs)\n \n self.native.connect(\"response\", self.gtk_response)\n", "issue": "Dialog windows are not modal\n### Describe the bug\n\nCreate a dialog like this:\r\n```\r\nasync def on_button_test(widget):\r\n await self.main_window.info_dialog(title=\"Dialog\", message=\"An Info Dialog\")\r\n```\r\nYou can now click on the main window behind the dialog and the main window is activated although the dialog is still floating above the window. This allows the dialog to be shown again or other actions to be triggered although the user should be blocked in the dialog.\r\n\r\nEither change the behavior of dialogs to be modal or add a \"modal={True|False}\" property to \"info_dialog\" (and all similar functions) to enforce this.\n\n### Steps to reproduce\n\n1) Create self.main_window.info_dialog\r\n2) Click main window\r\n3) Main window is activated and can be interacted with. \n\n### Expected behavior\n\nMain window can not be activated (dialog keeps focus)\n\n### Screenshots\n\n_No response_\n\n### Environment\n\nLinux (GenToo). Toga in virtual environment.\n\n### Logs\n\n_No response_\n\n### Additional context\n\n_No response_\n", "before_files": [{"content": "from abc import ABC\nfrom pathlib import Path\n\nfrom .libs import Gtk\n\n\nclass BaseDialog(ABC):\n def __init__(self, interface):\n self.interface = interface\n self.interface._impl = self\n\n\nclass MessageDialog(BaseDialog):\n def __init__(\n self,\n interface,\n title,\n message_type,\n buttons,\n success_result=None,\n **kwargs,\n ):\n super().__init__(interface=interface)\n self.success_result = success_result\n\n self.native = Gtk.MessageDialog(\n transient_for=interface.window._impl.native,\n flags=0,\n message_type=message_type,\n buttons=buttons,\n text=title,\n )\n self.build_dialog(**kwargs)\n\n self.native.connect(\"response\", self.gtk_response)\n self.native.show()\n\n def build_dialog(self, message):\n self.native.format_secondary_text(message)\n\n def gtk_response(self, dialog, response):\n if self.success_result:\n result = response == self.success_result\n else:\n result = None\n\n self.interface.set_result(result)\n\n self.native.destroy()\n\n\nclass InfoDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.INFO,\n buttons=Gtk.ButtonsType.OK,\n )\n\n\nclass QuestionDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.QUESTION,\n buttons=Gtk.ButtonsType.YES_NO,\n success_result=Gtk.ResponseType.YES,\n )\n\n\nclass ConfirmDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.WARNING,\n buttons=Gtk.ButtonsType.OK_CANCEL,\n success_result=Gtk.ResponseType.OK,\n )\n\n\nclass ErrorDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.ERROR,\n buttons=Gtk.ButtonsType.CANCEL,\n )\n\n\nclass StackTraceDialog(MessageDialog):\n def __init__(self, interface, title, **kwargs):\n super().__init__(\n interface=interface,\n title=title,\n message_type=Gtk.MessageType.ERROR,\n buttons=(\n Gtk.ButtonsType.CANCEL if kwargs.get(\"retry\") else Gtk.ButtonsType.OK\n ),\n success_result=Gtk.ResponseType.OK if kwargs.get(\"retry\") else None,\n **kwargs,\n )\n\n def build_dialog(self, message, content, retry):\n container = self.native.get_message_area()\n\n self.native.format_secondary_text(message)\n\n # Create a scrolling readonly text area, in monospace font, to contain the stack trace.\n buffer = Gtk.TextBuffer()\n buffer.set_text(content)\n\n trace = Gtk.TextView()\n trace.set_buffer(buffer)\n trace.set_wrap_mode(Gtk.WrapMode.WORD_CHAR)\n trace.set_property(\"editable\", False)\n trace.set_property(\"cursor-visible\", False)\n\n trace.get_style_context().add_class(\"toga\")\n trace.get_style_context().add_class(\"stacktrace\")\n trace.get_style_context().add_class(\"dialog\")\n\n style_provider = Gtk.CssProvider()\n style_provider.load_from_data(b\".toga.stacktrace {font-family: monospace;}\")\n\n trace.get_style_context().add_provider(\n style_provider,\n Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION,\n )\n\n scroll = Gtk.ScrolledWindow()\n scroll.set_policy(Gtk.PolicyType.NEVER, Gtk.PolicyType.AUTOMATIC)\n scroll.set_size_request(500, 200)\n scroll.add(trace)\n\n container.pack_end(scroll, False, False, 0)\n\n container.show_all()\n\n # If this is a retry dialog, add a retry button (which maps to OK).\n if retry:\n self.native.add_button(\"Retry\", Gtk.ResponseType.OK)\n\n\nclass FileDialog(BaseDialog):\n def __init__(\n self,\n interface,\n title,\n filename,\n initial_directory,\n file_types,\n multiple_select,\n action,\n ok_icon,\n ):\n super().__init__(interface=interface)\n\n self.native = Gtk.FileChooserDialog(\n transient_for=interface.window._impl.native,\n title=title,\n action=action,\n )\n self.native.add_button(Gtk.STOCK_CANCEL, Gtk.ResponseType.CANCEL)\n self.native.add_button(ok_icon, Gtk.ResponseType.OK)\n\n if filename:\n self.native.set_current_name(filename)\n\n if initial_directory:\n self.native.set_current_folder(str(initial_directory))\n\n if file_types:\n for file_type in file_types:\n filter_filetype = Gtk.FileFilter()\n filter_filetype.set_name(\".\" + file_type + \" files\")\n filter_filetype.add_pattern(\"*.\" + file_type)\n self.native.add_filter(filter_filetype)\n\n self.multiple_select = multiple_select\n if self.multiple_select:\n self.native.set_select_multiple(True)\n\n self.native.connect(\"response\", self.gtk_response)\n self.native.show()\n\n # Provided as a stub that can be mocked in test conditions\n def selected_path(self):\n return self.native.get_filename()\n\n # Provided as a stub that can be mocked in test conditions\n def selected_paths(self):\n return self.native.get_filenames()\n\n def gtk_response(self, dialog, response):\n if response == Gtk.ResponseType.OK:\n if self.multiple_select:\n result = [Path(filename) for filename in self.selected_paths()]\n else:\n result = Path(self.selected_path())\n else:\n result = None\n\n self.interface.set_result(result)\n\n self.native.destroy()\n\n\nclass SaveFileDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n filename,\n initial_directory,\n file_types=None,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=filename,\n initial_directory=initial_directory,\n file_types=file_types,\n multiple_select=False,\n action=Gtk.FileChooserAction.SAVE,\n ok_icon=Gtk.STOCK_SAVE,\n )\n\n\nclass OpenFileDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n initial_directory,\n file_types,\n multiple_select,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=None,\n initial_directory=initial_directory,\n file_types=file_types,\n multiple_select=multiple_select,\n action=Gtk.FileChooserAction.OPEN,\n ok_icon=Gtk.STOCK_OPEN,\n )\n\n\nclass SelectFolderDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n initial_directory,\n multiple_select,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=None,\n initial_directory=initial_directory,\n file_types=None,\n multiple_select=multiple_select,\n action=Gtk.FileChooserAction.SELECT_FOLDER,\n ok_icon=Gtk.STOCK_OPEN,\n )\n", "path": "gtk/src/toga_gtk/dialogs.py"}], "after_files": [{"content": "from abc import ABC\nfrom pathlib import Path\n\nfrom .libs import Gtk\n\n\nclass BaseDialog(ABC):\n def __init__(self, interface):\n self.interface = interface\n self.interface._impl = self\n\n\nclass MessageDialog(BaseDialog):\n def __init__(\n self,\n interface,\n title,\n message_type,\n buttons,\n success_result=None,\n **kwargs,\n ):\n super().__init__(interface=interface)\n self.success_result = success_result\n\n self.native = Gtk.MessageDialog(\n transient_for=interface.window._impl.native,\n flags=0,\n message_type=message_type,\n buttons=buttons,\n text=title,\n )\n self.native.set_modal(True)\n self.build_dialog(**kwargs)\n\n self.native.connect(\"response\", self.gtk_response)\n self.native.show()\n\n def build_dialog(self, message):\n self.native.format_secondary_text(message)\n\n def gtk_response(self, dialog, response):\n if self.success_result:\n result = response == self.success_result\n else:\n result = None\n\n self.interface.set_result(result)\n\n self.native.destroy()\n\n\nclass InfoDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.INFO,\n buttons=Gtk.ButtonsType.OK,\n )\n\n\nclass QuestionDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.QUESTION,\n buttons=Gtk.ButtonsType.YES_NO,\n success_result=Gtk.ResponseType.YES,\n )\n\n\nclass ConfirmDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.WARNING,\n buttons=Gtk.ButtonsType.OK_CANCEL,\n success_result=Gtk.ResponseType.OK,\n )\n\n\nclass ErrorDialog(MessageDialog):\n def __init__(self, interface, title, message):\n super().__init__(\n interface=interface,\n title=title,\n message=message,\n message_type=Gtk.MessageType.ERROR,\n buttons=Gtk.ButtonsType.CANCEL,\n )\n\n\nclass StackTraceDialog(MessageDialog):\n def __init__(self, interface, title, **kwargs):\n super().__init__(\n interface=interface,\n title=title,\n message_type=Gtk.MessageType.ERROR,\n buttons=(\n Gtk.ButtonsType.CANCEL if kwargs.get(\"retry\") else Gtk.ButtonsType.OK\n ),\n success_result=Gtk.ResponseType.OK if kwargs.get(\"retry\") else None,\n **kwargs,\n )\n\n def build_dialog(self, message, content, retry):\n container = self.native.get_message_area()\n\n self.native.format_secondary_text(message)\n\n # Create a scrolling readonly text area, in monospace font, to contain the stack trace.\n buffer = Gtk.TextBuffer()\n buffer.set_text(content)\n\n trace = Gtk.TextView()\n trace.set_buffer(buffer)\n trace.set_wrap_mode(Gtk.WrapMode.WORD_CHAR)\n trace.set_property(\"editable\", False)\n trace.set_property(\"cursor-visible\", False)\n\n trace.get_style_context().add_class(\"toga\")\n trace.get_style_context().add_class(\"stacktrace\")\n trace.get_style_context().add_class(\"dialog\")\n\n style_provider = Gtk.CssProvider()\n style_provider.load_from_data(b\".toga.stacktrace {font-family: monospace;}\")\n\n trace.get_style_context().add_provider(\n style_provider,\n Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION,\n )\n\n scroll = Gtk.ScrolledWindow()\n scroll.set_policy(Gtk.PolicyType.NEVER, Gtk.PolicyType.AUTOMATIC)\n scroll.set_size_request(500, 200)\n scroll.add(trace)\n\n container.pack_end(scroll, False, False, 0)\n\n container.show_all()\n\n # If this is a retry dialog, add a retry button (which maps to OK).\n if retry:\n self.native.add_button(\"Retry\", Gtk.ResponseType.OK)\n\n\nclass FileDialog(BaseDialog):\n def __init__(\n self,\n interface,\n title,\n filename,\n initial_directory,\n file_types,\n multiple_select,\n action,\n ok_icon,\n ):\n super().__init__(interface=interface)\n\n self.native = Gtk.FileChooserDialog(\n transient_for=interface.window._impl.native,\n title=title,\n action=action,\n )\n self.native.add_button(Gtk.STOCK_CANCEL, Gtk.ResponseType.CANCEL)\n self.native.add_button(ok_icon, Gtk.ResponseType.OK)\n\n if filename:\n self.native.set_current_name(filename)\n\n if initial_directory:\n self.native.set_current_folder(str(initial_directory))\n\n if file_types:\n for file_type in file_types:\n filter_filetype = Gtk.FileFilter()\n filter_filetype.set_name(\".\" + file_type + \" files\")\n filter_filetype.add_pattern(\"*.\" + file_type)\n self.native.add_filter(filter_filetype)\n\n self.multiple_select = multiple_select\n if self.multiple_select:\n self.native.set_select_multiple(True)\n\n self.native.connect(\"response\", self.gtk_response)\n self.native.show()\n\n # Provided as a stub that can be mocked in test conditions\n def selected_path(self):\n return self.native.get_filename()\n\n # Provided as a stub that can be mocked in test conditions\n def selected_paths(self):\n return self.native.get_filenames()\n\n def gtk_response(self, dialog, response):\n if response == Gtk.ResponseType.OK:\n if self.multiple_select:\n result = [Path(filename) for filename in self.selected_paths()]\n else:\n result = Path(self.selected_path())\n else:\n result = None\n\n self.interface.set_result(result)\n\n self.native.destroy()\n\n\nclass SaveFileDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n filename,\n initial_directory,\n file_types=None,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=filename,\n initial_directory=initial_directory,\n file_types=file_types,\n multiple_select=False,\n action=Gtk.FileChooserAction.SAVE,\n ok_icon=Gtk.STOCK_SAVE,\n )\n\n\nclass OpenFileDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n initial_directory,\n file_types,\n multiple_select,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=None,\n initial_directory=initial_directory,\n file_types=file_types,\n multiple_select=multiple_select,\n action=Gtk.FileChooserAction.OPEN,\n ok_icon=Gtk.STOCK_OPEN,\n )\n\n\nclass SelectFolderDialog(FileDialog):\n def __init__(\n self,\n interface,\n title,\n initial_directory,\n multiple_select,\n ):\n super().__init__(\n interface=interface,\n title=title,\n filename=None,\n initial_directory=initial_directory,\n file_types=None,\n multiple_select=multiple_select,\n action=Gtk.FileChooserAction.SELECT_FOLDER,\n ok_icon=Gtk.STOCK_OPEN,\n )\n", "path": "gtk/src/toga_gtk/dialogs.py"}]}
| 2,736 | 103 |
gh_patches_debug_35082
|
rasdani/github-patches
|
git_diff
|
cookiecutter__cookiecutter-480
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Instructions for local development are incomplete.
When executing the line:
```
python setup.py test
```
as directed, I get the following error:
```
running test
Traceback (most recent call last):
File "setup.py", line 111, in <module>
tests_require=test_requirements
File "/usr/lib/python2.7/distutils/core.py", line 151, in setup
dist.run_commands()
File "/usr/lib/python2.7/distutils/dist.py", line 953, in run_commands
self.run_command(cmd)
File "/usr/lib/python2.7/distutils/dist.py", line 972, in run_command
cmd_obj.run()
File "setup.py", line 60, in run
import pytest
ImportError: No module named pytest
```
I'm not sure what you are supposed to do to install the 'tests_require' dependencies.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `cookiecutter/compat.py`
Content:
```
1 import os
2 import sys
3
4 PY3 = sys.version_info[0] == 3
5 OLD_PY2 = sys.version_info[:2] < (2, 7)
6
7
8 if PY3: # pragma: no cover
9 input_str = 'builtins.input'
10 iteritems = lambda d: iter(d.items())
11 from unittest.mock import patch
12 from io import StringIO
13
14
15 else: # pragma: no cover
16 from __builtin__ import raw_input
17 input = raw_input
18 input_str = '__builtin__.raw_input'
19 iteritems = lambda d: d.iteritems()
20 from mock import patch
21 from cStringIO import StringIO
22
23
24 if PY3: # Forced testing
25
26 from shutil import which
27
28 else: # Forced testing
29
30 def is_exe(program):
31 """
32 Returns whether or not a file is an executable.
33 """
34 return os.path.isfile(program) and os.access(program, os.X_OK)
35
36 def which(cmd, mode=os.F_OK | os.X_OK, path=None):
37 """Given a command, mode, and a PATH string, return the path which
38 conforms to the given mode on the PATH, or None if there is no such
39 file.
40 `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
41 of os.environ.get("PATH"), or can be overridden with a custom search
42 path.
43
44 Note: This function was backported from the Python 3 source code.
45 """
46 # Check that a given file can be accessed with the correct mode.
47 # Additionally check that `file` is not a directory, as on Windows
48 # directories pass the os.access check.
49 def _access_check(fn, mode):
50 return (os.path.exists(fn) and os.access(fn, mode) and
51 not os.path.isdir(fn))
52
53 # If we're given a path with a directory part, look it up directly
54 # rather than referring to PATH directories. This includes checking
55 # relative to the current directory, e.g. ./script
56 if os.path.dirname(cmd):
57 if _access_check(cmd, mode):
58 return cmd
59 return None
60
61 if path is None:
62 path = os.environ.get("PATH", os.defpath)
63 if not path:
64 return None
65 path = path.split(os.pathsep)
66
67 if sys.platform == "win32":
68 # The current directory takes precedence on Windows.
69 if os.curdir not in path:
70 path.insert(0, os.curdir)
71
72 # PATHEXT is necessary to check on Windows.
73 pathext = os.environ.get("PATHEXT", "").split(os.pathsep)
74 # See if the given file matches any of the expected path
75 # extensions. This will allow us to short circuit when given
76 # "python.exe". If it does match, only test that one, otherwise we
77 # have to try others.
78 if any(cmd.lower().endswith(ext.lower()) for ext in pathext):
79 files = [cmd]
80 else:
81 files = [cmd + ext for ext in pathext]
82 else:
83 # On other platforms you don't have things like PATHEXT to tell you
84 # what file suffixes are executable, so just pass on cmd as-is.
85 files = [cmd]
86
87 seen = set()
88 for dir in path:
89 normdir = os.path.normcase(dir)
90 if normdir not in seen:
91 seen.add(normdir)
92 for thefile in files:
93 name = os.path.join(dir, thefile)
94 if _access_check(name, mode):
95 return name
96 return None
97
98
99 def is_string(obj):
100 """Determine if an object is a string."""
101 return isinstance(obj, str if PY3 else basestring)
102
103
104 _hush_pyflakes = (patch, StringIO, which)
105
```
Path: `setup.py`
Content:
```
1 #!/usr/bin/env python
2
3 import os
4 import sys
5
6 try:
7 from setuptools import setup, Command
8 except ImportError:
9 from distutils.core import setup, Command
10
11 version = "1.0.0"
12
13 if sys.argv[-1] == 'publish':
14 os.system('python setup.py sdist upload')
15 os.system('python setup.py bdist_wheel upload')
16 sys.exit()
17
18 if sys.argv[-1] == 'tag':
19 os.system("git tag -a %s -m 'version %s'" % (version, version))
20 os.system("git push --tags")
21 sys.exit()
22
23 with open('README.rst') as readme_file:
24 readme = readme_file.read()
25
26 with open('HISTORY.rst') as history_file:
27 history = history_file.read().replace('.. :changelog:', '')
28
29 requirements = [
30 'binaryornot>=0.2.0',
31 'jinja2>=2.7',
32 'PyYAML>=3.10',
33 'click<5.0'
34 ]
35
36 test_requirements = [
37 'pytest'
38 ]
39
40 # Add Python 2.7-specific dependencies
41 if sys.version < '3':
42 requirements.append('mock')
43
44 # There are no Python 3-specific dependencies to add
45
46 long_description = readme + '\n\n' + history
47
48 if sys.argv[-1] == 'readme':
49 print(long_description)
50 sys.exit()
51
52
53 class PyTest(Command):
54 user_options = [('pytest-args=', 'a', "Arguments to pass to py.test")]
55
56 def initialize_options(self):
57 self.pytest_args = []
58
59 def finalize_options(self):
60 pass
61
62 def run(self):
63 import pytest
64 errno = pytest.main(self.pytest_args)
65 sys.exit(errno)
66
67
68 setup(
69 name='cookiecutter',
70 version=version,
71 description=('A command-line utility that creates projects from project '
72 'templates, e.g. creating a Python package project from a '
73 'Python package project template.'),
74 long_description=long_description,
75 author='Audrey Roy',
76 author_email='[email protected]',
77 url='https://github.com/audreyr/cookiecutter',
78 packages=[
79 'cookiecutter',
80 ],
81 package_dir={'cookiecutter': 'cookiecutter'},
82 entry_points={
83 'console_scripts': [
84 'cookiecutter = cookiecutter.cli:main',
85 ]
86 },
87 include_package_data=True,
88 install_requires=requirements,
89 license='BSD',
90 zip_safe=False,
91 classifiers=[
92 'Development Status :: 3 - Alpha',
93 'Environment :: Console',
94 'Intended Audience :: Developers',
95 'Natural Language :: English',
96 'License :: OSI Approved :: BSD License',
97 'Programming Language :: Python',
98 'Programming Language :: Python :: 2',
99 'Programming Language :: Python :: 2.7',
100 'Programming Language :: Python :: 3',
101 'Programming Language :: Python :: 3.3',
102 'Programming Language :: Python :: 3.4',
103 'Programming Language :: Python :: Implementation :: CPython',
104 'Programming Language :: Python :: Implementation :: PyPy',
105 'Topic :: Software Development',
106 ],
107 keywords=(
108 'cookiecutter, Python, projects, project templates, Jinja2, '
109 'skeleton, scaffolding, project directory, setup.py, package, '
110 'packaging'
111 ),
112 cmdclass = {'test': PyTest},
113 test_suite='tests',
114 tests_require=test_requirements
115 )
116
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/cookiecutter/compat.py b/cookiecutter/compat.py
--- a/cookiecutter/compat.py
+++ b/cookiecutter/compat.py
@@ -8,7 +8,6 @@
if PY3: # pragma: no cover
input_str = 'builtins.input'
iteritems = lambda d: iter(d.items())
- from unittest.mock import patch
from io import StringIO
@@ -17,7 +16,6 @@
input = raw_input
input_str = '__builtin__.raw_input'
iteritems = lambda d: d.iteritems()
- from mock import patch
from cStringIO import StringIO
@@ -101,4 +99,4 @@
return isinstance(obj, str if PY3 else basestring)
-_hush_pyflakes = (patch, StringIO, which)
+_hush_pyflakes = (StringIO, which)
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -4,9 +4,9 @@
import sys
try:
- from setuptools import setup, Command
+ from setuptools import setup
except ImportError:
- from distutils.core import setup, Command
+ from distutils.core import setup
version = "1.0.0"
@@ -33,16 +33,6 @@
'click<5.0'
]
-test_requirements = [
- 'pytest'
-]
-
-# Add Python 2.7-specific dependencies
-if sys.version < '3':
- requirements.append('mock')
-
-# There are no Python 3-specific dependencies to add
-
long_description = readme + '\n\n' + history
if sys.argv[-1] == 'readme':
@@ -50,21 +40,6 @@
sys.exit()
-class PyTest(Command):
- user_options = [('pytest-args=', 'a', "Arguments to pass to py.test")]
-
- def initialize_options(self):
- self.pytest_args = []
-
- def finalize_options(self):
- pass
-
- def run(self):
- import pytest
- errno = pytest.main(self.pytest_args)
- sys.exit(errno)
-
-
setup(
name='cookiecutter',
version=version,
@@ -109,7 +84,4 @@
'skeleton, scaffolding, project directory, setup.py, package, '
'packaging'
),
- cmdclass = {'test': PyTest},
- test_suite='tests',
- tests_require=test_requirements
)
|
{"golden_diff": "diff --git a/cookiecutter/compat.py b/cookiecutter/compat.py\n--- a/cookiecutter/compat.py\n+++ b/cookiecutter/compat.py\n@@ -8,7 +8,6 @@\n if PY3: # pragma: no cover\n input_str = 'builtins.input'\n iteritems = lambda d: iter(d.items())\n- from unittest.mock import patch\n from io import StringIO\n \n \n@@ -17,7 +16,6 @@\n input = raw_input\n input_str = '__builtin__.raw_input'\n iteritems = lambda d: d.iteritems()\n- from mock import patch\n from cStringIO import StringIO\n \n \n@@ -101,4 +99,4 @@\n return isinstance(obj, str if PY3 else basestring)\n \n \n-_hush_pyflakes = (patch, StringIO, which)\n+_hush_pyflakes = (StringIO, which)\ndiff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -4,9 +4,9 @@\n import sys\n \n try:\n- from setuptools import setup, Command\n+ from setuptools import setup\n except ImportError:\n- from distutils.core import setup, Command\n+ from distutils.core import setup\n \n version = \"1.0.0\"\n \n@@ -33,16 +33,6 @@\n 'click<5.0'\n ]\n \n-test_requirements = [\n- 'pytest'\n-]\n-\n-# Add Python 2.7-specific dependencies\n-if sys.version < '3':\n- requirements.append('mock')\n-\n-# There are no Python 3-specific dependencies to add\n-\n long_description = readme + '\\n\\n' + history\n \n if sys.argv[-1] == 'readme':\n@@ -50,21 +40,6 @@\n sys.exit()\n \n \n-class PyTest(Command):\n- user_options = [('pytest-args=', 'a', \"Arguments to pass to py.test\")]\n-\n- def initialize_options(self):\n- self.pytest_args = []\n-\n- def finalize_options(self):\n- pass\n-\n- def run(self):\n- import pytest\n- errno = pytest.main(self.pytest_args)\n- sys.exit(errno)\n-\n-\n setup(\n name='cookiecutter',\n version=version,\n@@ -109,7 +84,4 @@\n 'skeleton, scaffolding, project directory, setup.py, package, '\n 'packaging'\n ),\n- cmdclass = {'test': PyTest},\n- test_suite='tests',\n- tests_require=test_requirements\n )\n", "issue": "Instructions for local development are incomplete.\nWhen executing the line:\n\n```\npython setup.py test\n```\n\nas directed, I get the following error:\n\n```\nrunning test\nTraceback (most recent call last):\n File \"setup.py\", line 111, in <module>\n tests_require=test_requirements\n File \"/usr/lib/python2.7/distutils/core.py\", line 151, in setup\n dist.run_commands()\n File \"/usr/lib/python2.7/distutils/dist.py\", line 953, in run_commands\n self.run_command(cmd)\n File \"/usr/lib/python2.7/distutils/dist.py\", line 972, in run_command\n cmd_obj.run()\n File \"setup.py\", line 60, in run\n import pytest\nImportError: No module named pytest\n```\n\nI'm not sure what you are supposed to do to install the 'tests_require' dependencies.\n\n", "before_files": [{"content": "import os\nimport sys\n\nPY3 = sys.version_info[0] == 3\nOLD_PY2 = sys.version_info[:2] < (2, 7)\n\n\nif PY3: # pragma: no cover\n input_str = 'builtins.input'\n iteritems = lambda d: iter(d.items())\n from unittest.mock import patch\n from io import StringIO\n\n\nelse: # pragma: no cover\n from __builtin__ import raw_input\n input = raw_input\n input_str = '__builtin__.raw_input'\n iteritems = lambda d: d.iteritems()\n from mock import patch\n from cStringIO import StringIO\n\n\nif PY3: # Forced testing\n\n from shutil import which\n\nelse: # Forced testing\n\n def is_exe(program):\n \"\"\"\n Returns whether or not a file is an executable.\n \"\"\"\n return os.path.isfile(program) and os.access(program, os.X_OK)\n\n def which(cmd, mode=os.F_OK | os.X_OK, path=None):\n \"\"\"Given a command, mode, and a PATH string, return the path which\n conforms to the given mode on the PATH, or None if there is no such\n file.\n `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result\n of os.environ.get(\"PATH\"), or can be overridden with a custom search\n path.\n\n Note: This function was backported from the Python 3 source code.\n \"\"\"\n # Check that a given file can be accessed with the correct mode.\n # Additionally check that `file` is not a directory, as on Windows\n # directories pass the os.access check.\n def _access_check(fn, mode):\n return (os.path.exists(fn) and os.access(fn, mode) and\n not os.path.isdir(fn))\n\n # If we're given a path with a directory part, look it up directly\n # rather than referring to PATH directories. This includes checking\n # relative to the current directory, e.g. ./script\n if os.path.dirname(cmd):\n if _access_check(cmd, mode):\n return cmd\n return None\n\n if path is None:\n path = os.environ.get(\"PATH\", os.defpath)\n if not path:\n return None\n path = path.split(os.pathsep)\n\n if sys.platform == \"win32\":\n # The current directory takes precedence on Windows.\n if os.curdir not in path:\n path.insert(0, os.curdir)\n\n # PATHEXT is necessary to check on Windows.\n pathext = os.environ.get(\"PATHEXT\", \"\").split(os.pathsep)\n # See if the given file matches any of the expected path\n # extensions. This will allow us to short circuit when given\n # \"python.exe\". If it does match, only test that one, otherwise we\n # have to try others.\n if any(cmd.lower().endswith(ext.lower()) for ext in pathext):\n files = [cmd]\n else:\n files = [cmd + ext for ext in pathext]\n else:\n # On other platforms you don't have things like PATHEXT to tell you\n # what file suffixes are executable, so just pass on cmd as-is.\n files = [cmd]\n\n seen = set()\n for dir in path:\n normdir = os.path.normcase(dir)\n if normdir not in seen:\n seen.add(normdir)\n for thefile in files:\n name = os.path.join(dir, thefile)\n if _access_check(name, mode):\n return name\n return None\n\n\ndef is_string(obj):\n \"\"\"Determine if an object is a string.\"\"\"\n return isinstance(obj, str if PY3 else basestring)\n\n\n_hush_pyflakes = (patch, StringIO, which)\n", "path": "cookiecutter/compat.py"}, {"content": "#!/usr/bin/env python\n\nimport os\nimport sys\n\ntry:\n from setuptools import setup, Command\nexcept ImportError:\n from distutils.core import setup, Command\n\nversion = \"1.0.0\"\n\nif sys.argv[-1] == 'publish':\n os.system('python setup.py sdist upload')\n os.system('python setup.py bdist_wheel upload')\n sys.exit()\n\nif sys.argv[-1] == 'tag':\n os.system(\"git tag -a %s -m 'version %s'\" % (version, version))\n os.system(\"git push --tags\")\n sys.exit()\n\nwith open('README.rst') as readme_file:\n readme = readme_file.read()\n\nwith open('HISTORY.rst') as history_file:\n history = history_file.read().replace('.. :changelog:', '')\n\nrequirements = [\n 'binaryornot>=0.2.0',\n 'jinja2>=2.7',\n 'PyYAML>=3.10',\n 'click<5.0'\n]\n\ntest_requirements = [\n 'pytest'\n]\n\n# Add Python 2.7-specific dependencies\nif sys.version < '3':\n requirements.append('mock')\n\n# There are no Python 3-specific dependencies to add\n\nlong_description = readme + '\\n\\n' + history\n\nif sys.argv[-1] == 'readme':\n print(long_description)\n sys.exit()\n\n\nclass PyTest(Command):\n user_options = [('pytest-args=', 'a', \"Arguments to pass to py.test\")]\n\n def initialize_options(self):\n self.pytest_args = []\n\n def finalize_options(self):\n pass\n\n def run(self):\n import pytest\n errno = pytest.main(self.pytest_args)\n sys.exit(errno)\n\n\nsetup(\n name='cookiecutter',\n version=version,\n description=('A command-line utility that creates projects from project '\n 'templates, e.g. creating a Python package project from a '\n 'Python package project template.'),\n long_description=long_description,\n author='Audrey Roy',\n author_email='[email protected]',\n url='https://github.com/audreyr/cookiecutter',\n packages=[\n 'cookiecutter',\n ],\n package_dir={'cookiecutter': 'cookiecutter'},\n entry_points={\n 'console_scripts': [\n 'cookiecutter = cookiecutter.cli:main',\n ]\n },\n include_package_data=True,\n install_requires=requirements,\n license='BSD',\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Environment :: Console',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: BSD License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Topic :: Software Development',\n ],\n keywords=(\n 'cookiecutter, Python, projects, project templates, Jinja2, '\n 'skeleton, scaffolding, project directory, setup.py, package, '\n 'packaging'\n ),\n cmdclass = {'test': PyTest},\n test_suite='tests',\n tests_require=test_requirements\n)\n", "path": "setup.py"}], "after_files": [{"content": "import os\nimport sys\n\nPY3 = sys.version_info[0] == 3\nOLD_PY2 = sys.version_info[:2] < (2, 7)\n\n\nif PY3: # pragma: no cover\n input_str = 'builtins.input'\n iteritems = lambda d: iter(d.items())\n from io import StringIO\n\n\nelse: # pragma: no cover\n from __builtin__ import raw_input\n input = raw_input\n input_str = '__builtin__.raw_input'\n iteritems = lambda d: d.iteritems()\n from cStringIO import StringIO\n\n\nif PY3: # Forced testing\n\n from shutil import which\n\nelse: # Forced testing\n\n def is_exe(program):\n \"\"\"\n Returns whether or not a file is an executable.\n \"\"\"\n return os.path.isfile(program) and os.access(program, os.X_OK)\n\n def which(cmd, mode=os.F_OK | os.X_OK, path=None):\n \"\"\"Given a command, mode, and a PATH string, return the path which\n conforms to the given mode on the PATH, or None if there is no such\n file.\n `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result\n of os.environ.get(\"PATH\"), or can be overridden with a custom search\n path.\n\n Note: This function was backported from the Python 3 source code.\n \"\"\"\n # Check that a given file can be accessed with the correct mode.\n # Additionally check that `file` is not a directory, as on Windows\n # directories pass the os.access check.\n def _access_check(fn, mode):\n return (os.path.exists(fn) and os.access(fn, mode) and\n not os.path.isdir(fn))\n\n # If we're given a path with a directory part, look it up directly\n # rather than referring to PATH directories. This includes checking\n # relative to the current directory, e.g. ./script\n if os.path.dirname(cmd):\n if _access_check(cmd, mode):\n return cmd\n return None\n\n if path is None:\n path = os.environ.get(\"PATH\", os.defpath)\n if not path:\n return None\n path = path.split(os.pathsep)\n\n if sys.platform == \"win32\":\n # The current directory takes precedence on Windows.\n if os.curdir not in path:\n path.insert(0, os.curdir)\n\n # PATHEXT is necessary to check on Windows.\n pathext = os.environ.get(\"PATHEXT\", \"\").split(os.pathsep)\n # See if the given file matches any of the expected path\n # extensions. This will allow us to short circuit when given\n # \"python.exe\". If it does match, only test that one, otherwise we\n # have to try others.\n if any(cmd.lower().endswith(ext.lower()) for ext in pathext):\n files = [cmd]\n else:\n files = [cmd + ext for ext in pathext]\n else:\n # On other platforms you don't have things like PATHEXT to tell you\n # what file suffixes are executable, so just pass on cmd as-is.\n files = [cmd]\n\n seen = set()\n for dir in path:\n normdir = os.path.normcase(dir)\n if normdir not in seen:\n seen.add(normdir)\n for thefile in files:\n name = os.path.join(dir, thefile)\n if _access_check(name, mode):\n return name\n return None\n\n\ndef is_string(obj):\n \"\"\"Determine if an object is a string.\"\"\"\n return isinstance(obj, str if PY3 else basestring)\n\n\n_hush_pyflakes = (StringIO, which)\n", "path": "cookiecutter/compat.py"}, {"content": "#!/usr/bin/env python\n\nimport os\nimport sys\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nversion = \"1.0.0\"\n\nif sys.argv[-1] == 'publish':\n os.system('python setup.py sdist upload')\n os.system('python setup.py bdist_wheel upload')\n sys.exit()\n\nif sys.argv[-1] == 'tag':\n os.system(\"git tag -a %s -m 'version %s'\" % (version, version))\n os.system(\"git push --tags\")\n sys.exit()\n\nwith open('README.rst') as readme_file:\n readme = readme_file.read()\n\nwith open('HISTORY.rst') as history_file:\n history = history_file.read().replace('.. :changelog:', '')\n\nrequirements = [\n 'binaryornot>=0.2.0',\n 'jinja2>=2.7',\n 'PyYAML>=3.10',\n 'click<5.0'\n]\n\nlong_description = readme + '\\n\\n' + history\n\nif sys.argv[-1] == 'readme':\n print(long_description)\n sys.exit()\n\n\nsetup(\n name='cookiecutter',\n version=version,\n description=('A command-line utility that creates projects from project '\n 'templates, e.g. creating a Python package project from a '\n 'Python package project template.'),\n long_description=long_description,\n author='Audrey Roy',\n author_email='[email protected]',\n url='https://github.com/audreyr/cookiecutter',\n packages=[\n 'cookiecutter',\n ],\n package_dir={'cookiecutter': 'cookiecutter'},\n entry_points={\n 'console_scripts': [\n 'cookiecutter = cookiecutter.cli:main',\n ]\n },\n include_package_data=True,\n install_requires=requirements,\n license='BSD',\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Environment :: Console',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: BSD License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Topic :: Software Development',\n ],\n keywords=(\n 'cookiecutter, Python, projects, project templates, Jinja2, '\n 'skeleton, scaffolding, project directory, setup.py, package, '\n 'packaging'\n ),\n)\n", "path": "setup.py"}]}
| 2,507 | 566 |
gh_patches_debug_41513
|
rasdani/github-patches
|
git_diff
|
e-valuation__EvaP-663
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Reward point redemption error message
When users try to redeem 0 reward points they get the error message "You don't have enough reward points". The error message should be something like "You must select the number of reward points you want to redeem" instead.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `evap/rewards/tools.py`
Content:
```
1 from django.conf import settings
2 from django.contrib import messages
3 from django.db import transaction
4 from django.utils.translation import ugettext as _
5 from django.dispatch import receiver
6
7 from django.contrib.auth.decorators import login_required
8 from evap.evaluation.models import Course
9
10 from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation
11
12 @login_required
13 @transaction.atomic
14 def save_redemptions(request, redemptions):
15 total_points_available = reward_points_of_user(request.user)
16 total_points_redeemed = sum(redemptions.values())
17
18 if total_points_redeemed == 0 or total_points_redeemed > total_points_available:
19 return False
20
21 for event_id in redemptions:
22 if redemptions[event_id] > 0:
23 redemption = RewardPointRedemption(
24 user_profile=request.user,
25 value=redemptions[event_id],
26 event=RewardPointRedemptionEvent.objects.get(id=event_id)
27 )
28 redemption.save()
29 return True
30
31
32 def can_user_use_reward_points(user):
33 return not user.is_external and user.is_participant
34
35
36 def reward_points_of_user(user):
37 reward_point_grantings = RewardPointGranting.objects.filter(user_profile=user)
38 reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=user)
39
40 count = 0
41 for granting in reward_point_grantings:
42 count += granting.value
43 for redemption in reward_point_redemptions:
44 count -= redemption.value
45
46 return count
47
48
49 @receiver(Course.course_evaluated)
50 def grant_reward_points(sender, **kwargs):
51 # grant reward points if all conditions are fulfilled
52
53 request = kwargs['request']
54 semester = kwargs['semester']
55 if not can_user_use_reward_points(request.user):
56 return
57 # has the semester been activated for reward points?
58 if not is_semester_activated(semester):
59 return
60 # does the user have at least one required course in this semester?
61 required_courses = Course.objects.filter(participants=request.user, semester=semester, is_required_for_reward=True)
62 if not required_courses.exists():
63 return
64 # does the user not participate in any more required courses in this semester?
65 if required_courses.exclude(voters=request.user).exists():
66 return
67 # did the user not already get reward points for this semester?
68 if not RewardPointGranting.objects.filter(user_profile=request.user, semester=semester):
69 granting = RewardPointGranting(user_profile=request.user, semester=semester, value=settings.REWARD_POINTS_PER_SEMESTER)
70 granting.save()
71 messages.success(request, _("You just have earned reward points for this semester because you evaluated all your courses. Thank you very much!"))
72
73
74 def is_semester_activated(semester):
75 try:
76 activation = SemesterActivation.objects.get(semester=semester)
77 return activation.is_active
78 except SemesterActivation.DoesNotExist:
79 return False
80
```
Path: `evap/rewards/views.py`
Content:
```
1 from django.contrib import messages
2 from django.shortcuts import get_object_or_404, redirect, render
3 from django.utils.translation import ugettext as _
4 from django.utils.translation import get_language
5 from django.http import HttpResponse
6 from datetime import datetime
7 from operator import attrgetter
8
9 from evap.evaluation.auth import reward_user_required, staff_required
10 from evap.evaluation.models import Semester, Course
11
12 from evap.staff.views import semester_view
13
14 from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation
15 from evap.rewards.tools import save_redemptions, reward_points_of_user, can_user_use_reward_points
16 from evap.rewards.forms import RewardPointRedemptionEventForm
17 from evap.rewards.exporters import ExcelExporter
18
19 @reward_user_required
20 def index(request):
21 if request.method == 'POST':
22 redemptions = {}
23 for key, value in request.POST.items():
24 if(key.startswith('points-')):
25 event_id = int(key.rpartition('-')[2])
26 redemptions[event_id] = int(value)
27
28 if save_redemptions(request, redemptions):
29 messages.success(request, _("You successfully redeemed your points."))
30 else:
31 messages.warning(request, _("You don't have enough reward points."))
32
33 total_points_available = reward_points_of_user(request.user)
34 reward_point_grantings = RewardPointGranting.objects.filter(user_profile=request.user)
35 reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=request.user)
36 events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now())
37 events = sorted(events, key=lambda event: event.date)
38
39 reward_point_actions=[]
40 for granting in reward_point_grantings:
41 reward_point_actions.append((granting.granting_time, _('Reward for') + ' ' + granting.semester.name, granting.value, ''))
42 for redemption in reward_point_redemptions:
43 reward_point_actions.append((redemption.redemption_time, redemption.event.name, '', redemption.value))
44
45 reward_point_actions.sort(key=lambda action: action[0], reverse=True)
46
47 template_data = dict(
48 reward_point_actions=reward_point_actions,
49 total_points_available=total_points_available,
50 events=events,
51 point_selection=[x for x in range(0,total_points_available+1)])
52 return render(request, "rewards_index.html", template_data)
53
54
55 @staff_required
56 def semester_reward_points(request, semester_id):
57 semester = get_object_or_404(Semester, id=semester_id)
58 courses = Course.objects.filter(semester=semester)
59 participants = set()
60 for course in courses:
61 for participant in course.participants.all():
62 if can_user_use_reward_points(participant):
63 participants.add(participant)
64 participants = sorted(participants, key=attrgetter('last_name', 'first_name'))
65
66 data = []
67 for participant in participants:
68 number_of_required_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=True).count()
69 number_of_required_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=True).count()
70 number_of_optional_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=False).count()
71 number_of_optional_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=False).count()
72 earned_reward_points = RewardPointGranting.objects.filter(semester=semester, user_profile=participant).exists()
73 data.append((participant, number_of_required_courses_voted_for, number_of_required_courses,
74 number_of_optional_courses_voted_for, number_of_optional_courses, earned_reward_points))
75
76 template_data = dict(semester=semester, data=data, disable_breadcrumb_semester=False)
77 return render(request, "rewards_semester_reward_points_view.html", template_data)
78
79
80 @staff_required
81 def reward_point_redemption_events(request):
82 upcoming_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now()).order_by('date')
83 past_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__lt=datetime.now()).order_by('-date')
84 template_data = dict(upcoming_events=upcoming_events, past_events=past_events)
85 return render(request, "rewards_reward_point_redemption_events.html", template_data)
86
87
88 @staff_required
89 def reward_point_redemption_event_create(request):
90 event = RewardPointRedemptionEvent()
91 form = RewardPointRedemptionEventForm(request.POST or None, instance=event)
92
93 if form.is_valid():
94 form.save()
95 messages.success(request, _("Successfully created event."))
96 return redirect('rewards:reward_point_redemption_events')
97 else:
98 return render(request, "rewards_reward_point_redemption_event_form.html", dict(form=form))
99
100
101 @staff_required
102 def reward_point_redemption_event_edit(request, event_id):
103 event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)
104 form = RewardPointRedemptionEventForm(request.POST or None, instance=event)
105
106 if form.is_valid():
107 event = form.save()
108
109 messages.success(request, _("Successfully updated event."))
110 return redirect('rewards:reward_point_redemption_events')
111 else:
112 return render(request, "rewards_reward_point_redemption_event_form.html", dict(event=event, form=form))
113
114
115 @staff_required
116 def reward_point_redemption_event_delete(request, event_id):
117 event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)
118
119 if event.can_delete:
120 if request.method == 'POST':
121 event.delete()
122 return redirect('rewards:reward_point_redemption_events')
123 else:
124 return render(request, "rewards_reward_point_redemption_event_delete.html", dict(event=event))
125 else:
126 messages.warning(request, _("This event cannot be deleted because some users already redeemed points for it."))
127 return redirect('rewards:reward_point_redemption_events')
128
129
130 @staff_required
131 def reward_point_redemption_event_export(request, event_id):
132 event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)
133
134 filename = _("RewardPoints")+"-%s-%s-%s.xls" % (event.date, event.name, get_language())
135
136 response = HttpResponse(content_type="application/vnd.ms-excel")
137 response["Content-Disposition"] = "attachment; filename=\"%s\"" % filename
138
139 ExcelExporter(event.redemptions_by_user()).export(response)
140
141 return response
142
143
144 @staff_required
145 def semester_activation(request, semester_id, active):
146 if active == 'on':
147 active = True
148 else:
149 active = False
150
151 try:
152 activation = SemesterActivation.objects.filter(semester=Semester.objects.get(id=semester_id)).get()
153 activation.is_active = active
154 except SemesterActivation.DoesNotExist:
155 activation = SemesterActivation(semester=Semester.objects.get(id=semester_id), is_active=active)
156 activation.save()
157
158 return semester_view(request=request, semester_id=semester_id)
159
```
Path: `evap/rewards/models.py`
Content:
```
1 from django.utils.translation import ugettext_lazy as _
2 from django.db import models
3
4 from collections import OrderedDict
5
6
7 class RewardPointRedemptionEvent(models.Model):
8 name = models.CharField(max_length=1024, verbose_name=_("event name"))
9 date = models.DateField(verbose_name=_("event date"))
10 redeem_end_date = models.DateField(verbose_name=_("redemption end date"))
11
12 @property
13 def can_delete(self):
14 if RewardPointRedemption.objects.filter(event=self).exists():
15 return False
16 return True
17
18 def redemptions_by_user(self):
19 redemptions = self.reward_point_redemptions.order_by('user_profile')
20 redemptions_dict = OrderedDict()
21 for redemption in redemptions:
22 if not redemption.user_profile in redemptions_dict:
23 redemptions_dict[redemption.user_profile] = 0
24 redemptions_dict[redemption.user_profile] += redemption.value
25 return redemptions_dict
26
27 class RewardPointGranting(models.Model):
28 user_profile = models.ForeignKey('evaluation.UserProfile', related_name="reward_point_grantings")
29 semester = models.ForeignKey('evaluation.Semester', related_name="reward_point_grantings", blank=True, null=True)
30 granting_time = models.DateTimeField(verbose_name=_("granting time"), auto_now_add=True)
31 value = models.IntegerField(verbose_name=_("value"), default=0)
32
33 class RewardPointRedemption(models.Model):
34 user_profile = models.ForeignKey('evaluation.UserProfile', related_name="reward_point_redemptions")
35 redemption_time = models.DateTimeField(verbose_name=_("redemption time"), auto_now_add=True)
36 value = models.IntegerField(verbose_name=_("value"), default=0)
37 event = models.ForeignKey(RewardPointRedemptionEvent, related_name="reward_point_redemptions")
38
39 class SemesterActivation(models.Model):
40 semester = models.OneToOneField('evaluation.Semester', related_name='rewards_active')
41 is_active = models.BooleanField(default=False)
42
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/evap/rewards/models.py b/evap/rewards/models.py
--- a/evap/rewards/models.py
+++ b/evap/rewards/models.py
@@ -3,6 +3,13 @@
from collections import OrderedDict
+class NoPointsSelected(Exception):
+ """An attempt has been made to redeem <= 0 points."""
+ pass
+
+class NotEnoughPoints(Exception):
+ """An attempt has been made to redeem more points than available."""
+ pass
class RewardPointRedemptionEvent(models.Model):
name = models.CharField(max_length=1024, verbose_name=_("event name"))
diff --git a/evap/rewards/tools.py b/evap/rewards/tools.py
--- a/evap/rewards/tools.py
+++ b/evap/rewards/tools.py
@@ -7,7 +7,7 @@
from django.contrib.auth.decorators import login_required
from evap.evaluation.models import Course
-from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation
+from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints
@login_required
@transaction.atomic
@@ -15,8 +15,11 @@
total_points_available = reward_points_of_user(request.user)
total_points_redeemed = sum(redemptions.values())
- if total_points_redeemed == 0 or total_points_redeemed > total_points_available:
- return False
+ if total_points_redeemed <= 0:
+ raise NoPointsSelected(_("You cannot redeem 0 points."))
+
+ if total_points_redeemed > total_points_available:
+ raise NotEnoughPoints(_("You don't have enough reward points."))
for event_id in redemptions:
if redemptions[event_id] > 0:
@@ -26,8 +29,6 @@
event=RewardPointRedemptionEvent.objects.get(id=event_id)
)
redemption.save()
- return True
-
def can_user_use_reward_points(user):
return not user.is_external and user.is_participant
diff --git a/evap/rewards/views.py b/evap/rewards/views.py
--- a/evap/rewards/views.py
+++ b/evap/rewards/views.py
@@ -11,7 +11,7 @@
from evap.staff.views import semester_view
-from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation
+from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints
from evap.rewards.tools import save_redemptions, reward_points_of_user, can_user_use_reward_points
from evap.rewards.forms import RewardPointRedemptionEventForm
from evap.rewards.exporters import ExcelExporter
@@ -25,10 +25,11 @@
event_id = int(key.rpartition('-')[2])
redemptions[event_id] = int(value)
- if save_redemptions(request, redemptions):
+ try:
+ save_redemptions(request, redemptions)
messages.success(request, _("You successfully redeemed your points."))
- else:
- messages.warning(request, _("You don't have enough reward points."))
+ except (NoPointsSelected, NotEnoughPoints) as error:
+ messages.warning(request, error)
total_points_available = reward_points_of_user(request.user)
reward_point_grantings = RewardPointGranting.objects.filter(user_profile=request.user)
|
{"golden_diff": "diff --git a/evap/rewards/models.py b/evap/rewards/models.py\n--- a/evap/rewards/models.py\n+++ b/evap/rewards/models.py\n@@ -3,6 +3,13 @@\n \n from collections import OrderedDict\n \n+class NoPointsSelected(Exception):\n+ \"\"\"An attempt has been made to redeem <= 0 points.\"\"\"\n+ pass\n+\n+class NotEnoughPoints(Exception):\n+ \"\"\"An attempt has been made to redeem more points than available.\"\"\"\n+ pass\n \n class RewardPointRedemptionEvent(models.Model):\n name = models.CharField(max_length=1024, verbose_name=_(\"event name\"))\ndiff --git a/evap/rewards/tools.py b/evap/rewards/tools.py\n--- a/evap/rewards/tools.py\n+++ b/evap/rewards/tools.py\n@@ -7,7 +7,7 @@\n from django.contrib.auth.decorators import login_required\n from evap.evaluation.models import Course\n \n-from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation\n+from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints\n \n @login_required\n @transaction.atomic\n@@ -15,8 +15,11 @@\n total_points_available = reward_points_of_user(request.user)\n total_points_redeemed = sum(redemptions.values())\n \n- if total_points_redeemed == 0 or total_points_redeemed > total_points_available:\n- return False\n+ if total_points_redeemed <= 0:\n+ raise NoPointsSelected(_(\"You cannot redeem 0 points.\"))\n+\n+ if total_points_redeemed > total_points_available:\n+ raise NotEnoughPoints(_(\"You don't have enough reward points.\"))\n \n for event_id in redemptions:\n if redemptions[event_id] > 0:\n@@ -26,8 +29,6 @@\n event=RewardPointRedemptionEvent.objects.get(id=event_id)\n )\n redemption.save()\n- return True\n-\n \n def can_user_use_reward_points(user):\n return not user.is_external and user.is_participant\ndiff --git a/evap/rewards/views.py b/evap/rewards/views.py\n--- a/evap/rewards/views.py\n+++ b/evap/rewards/views.py\n@@ -11,7 +11,7 @@\n \n from evap.staff.views import semester_view\n \n-from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation\n+from evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints\n from evap.rewards.tools import save_redemptions, reward_points_of_user, can_user_use_reward_points\n from evap.rewards.forms import RewardPointRedemptionEventForm\n from evap.rewards.exporters import ExcelExporter\n@@ -25,10 +25,11 @@\n event_id = int(key.rpartition('-')[2])\n redemptions[event_id] = int(value)\n \n- if save_redemptions(request, redemptions):\n+ try:\n+ save_redemptions(request, redemptions)\n messages.success(request, _(\"You successfully redeemed your points.\"))\n- else:\n- messages.warning(request, _(\"You don't have enough reward points.\"))\n+ except (NoPointsSelected, NotEnoughPoints) as error:\n+ messages.warning(request, error)\n \n total_points_available = reward_points_of_user(request.user)\n reward_point_grantings = RewardPointGranting.objects.filter(user_profile=request.user)\n", "issue": "Reward point redemption error message\nWhen users try to redeem 0 reward points they get the error message \"You don't have enough reward points\". The error message should be something like \"You must select the number of reward points you want to redeem\" instead.\n\n", "before_files": [{"content": "from django.conf import settings\nfrom django.contrib import messages\nfrom django.db import transaction\nfrom django.utils.translation import ugettext as _\nfrom django.dispatch import receiver\n\nfrom django.contrib.auth.decorators import login_required\nfrom evap.evaluation.models import Course\n\nfrom evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation\n\n@login_required\[email protected]\ndef save_redemptions(request, redemptions):\n total_points_available = reward_points_of_user(request.user)\n total_points_redeemed = sum(redemptions.values())\n\n if total_points_redeemed == 0 or total_points_redeemed > total_points_available:\n return False\n\n for event_id in redemptions:\n if redemptions[event_id] > 0:\n redemption = RewardPointRedemption(\n user_profile=request.user,\n value=redemptions[event_id],\n event=RewardPointRedemptionEvent.objects.get(id=event_id)\n )\n redemption.save()\n return True\n\n\ndef can_user_use_reward_points(user):\n return not user.is_external and user.is_participant\n\n\ndef reward_points_of_user(user):\n reward_point_grantings = RewardPointGranting.objects.filter(user_profile=user)\n reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=user)\n\n count = 0\n for granting in reward_point_grantings:\n count += granting.value\n for redemption in reward_point_redemptions:\n count -= redemption.value\n\n return count\n\n\n@receiver(Course.course_evaluated)\ndef grant_reward_points(sender, **kwargs):\n # grant reward points if all conditions are fulfilled\n\n request = kwargs['request']\n semester = kwargs['semester']\n if not can_user_use_reward_points(request.user):\n return\n # has the semester been activated for reward points?\n if not is_semester_activated(semester):\n return\n # does the user have at least one required course in this semester?\n required_courses = Course.objects.filter(participants=request.user, semester=semester, is_required_for_reward=True)\n if not required_courses.exists():\n return\n # does the user not participate in any more required courses in this semester?\n if required_courses.exclude(voters=request.user).exists():\n return\n # did the user not already get reward points for this semester?\n if not RewardPointGranting.objects.filter(user_profile=request.user, semester=semester):\n granting = RewardPointGranting(user_profile=request.user, semester=semester, value=settings.REWARD_POINTS_PER_SEMESTER)\n granting.save()\n messages.success(request, _(\"You just have earned reward points for this semester because you evaluated all your courses. Thank you very much!\"))\n\n\ndef is_semester_activated(semester):\n try:\n activation = SemesterActivation.objects.get(semester=semester)\n return activation.is_active\n except SemesterActivation.DoesNotExist:\n return False\n", "path": "evap/rewards/tools.py"}, {"content": "from django.contrib import messages\nfrom django.shortcuts import get_object_or_404, redirect, render\nfrom django.utils.translation import ugettext as _\nfrom django.utils.translation import get_language\nfrom django.http import HttpResponse\nfrom datetime import datetime\nfrom operator import attrgetter\n\nfrom evap.evaluation.auth import reward_user_required, staff_required\nfrom evap.evaluation.models import Semester, Course\n\nfrom evap.staff.views import semester_view\n\nfrom evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation\nfrom evap.rewards.tools import save_redemptions, reward_points_of_user, can_user_use_reward_points\nfrom evap.rewards.forms import RewardPointRedemptionEventForm\nfrom evap.rewards.exporters import ExcelExporter\n\n@reward_user_required\ndef index(request):\n if request.method == 'POST':\n redemptions = {}\n for key, value in request.POST.items():\n if(key.startswith('points-')):\n event_id = int(key.rpartition('-')[2])\n redemptions[event_id] = int(value)\n\n if save_redemptions(request, redemptions):\n messages.success(request, _(\"You successfully redeemed your points.\"))\n else:\n messages.warning(request, _(\"You don't have enough reward points.\"))\n\n total_points_available = reward_points_of_user(request.user)\n reward_point_grantings = RewardPointGranting.objects.filter(user_profile=request.user)\n reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=request.user)\n events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now())\n events = sorted(events, key=lambda event: event.date)\n\n reward_point_actions=[]\n for granting in reward_point_grantings:\n reward_point_actions.append((granting.granting_time, _('Reward for') + ' ' + granting.semester.name, granting.value, ''))\n for redemption in reward_point_redemptions:\n reward_point_actions.append((redemption.redemption_time, redemption.event.name, '', redemption.value))\n\n reward_point_actions.sort(key=lambda action: action[0], reverse=True)\n\n template_data = dict(\n reward_point_actions=reward_point_actions,\n total_points_available=total_points_available,\n events=events,\n point_selection=[x for x in range(0,total_points_available+1)])\n return render(request, \"rewards_index.html\", template_data)\n\n\n@staff_required\ndef semester_reward_points(request, semester_id):\n semester = get_object_or_404(Semester, id=semester_id)\n courses = Course.objects.filter(semester=semester)\n participants = set()\n for course in courses:\n for participant in course.participants.all():\n if can_user_use_reward_points(participant):\n participants.add(participant)\n participants = sorted(participants, key=attrgetter('last_name', 'first_name'))\n\n data = []\n for participant in participants:\n number_of_required_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=True).count()\n number_of_required_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=True).count()\n number_of_optional_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=False).count()\n number_of_optional_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=False).count()\n earned_reward_points = RewardPointGranting.objects.filter(semester=semester, user_profile=participant).exists()\n data.append((participant, number_of_required_courses_voted_for, number_of_required_courses,\n number_of_optional_courses_voted_for, number_of_optional_courses, earned_reward_points))\n\n template_data = dict(semester=semester, data=data, disable_breadcrumb_semester=False)\n return render(request, \"rewards_semester_reward_points_view.html\", template_data)\n\n\n@staff_required\ndef reward_point_redemption_events(request):\n upcoming_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now()).order_by('date')\n past_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__lt=datetime.now()).order_by('-date')\n template_data = dict(upcoming_events=upcoming_events, past_events=past_events)\n return render(request, \"rewards_reward_point_redemption_events.html\", template_data)\n\n\n@staff_required\ndef reward_point_redemption_event_create(request):\n event = RewardPointRedemptionEvent()\n form = RewardPointRedemptionEventForm(request.POST or None, instance=event)\n\n if form.is_valid():\n form.save()\n messages.success(request, _(\"Successfully created event.\"))\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_form.html\", dict(form=form))\n\n\n@staff_required\ndef reward_point_redemption_event_edit(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n form = RewardPointRedemptionEventForm(request.POST or None, instance=event)\n\n if form.is_valid():\n event = form.save()\n\n messages.success(request, _(\"Successfully updated event.\"))\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_form.html\", dict(event=event, form=form))\n\n\n@staff_required\ndef reward_point_redemption_event_delete(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n\n if event.can_delete:\n if request.method == 'POST':\n event.delete()\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_delete.html\", dict(event=event))\n else:\n messages.warning(request, _(\"This event cannot be deleted because some users already redeemed points for it.\"))\n return redirect('rewards:reward_point_redemption_events')\n\n\n@staff_required\ndef reward_point_redemption_event_export(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n\n filename = _(\"RewardPoints\")+\"-%s-%s-%s.xls\" % (event.date, event.name, get_language())\n\n response = HttpResponse(content_type=\"application/vnd.ms-excel\")\n response[\"Content-Disposition\"] = \"attachment; filename=\\\"%s\\\"\" % filename\n\n ExcelExporter(event.redemptions_by_user()).export(response)\n\n return response\n\n\n@staff_required\ndef semester_activation(request, semester_id, active):\n if active == 'on':\n active = True\n else:\n active = False\n\n try:\n activation = SemesterActivation.objects.filter(semester=Semester.objects.get(id=semester_id)).get()\n activation.is_active = active\n except SemesterActivation.DoesNotExist:\n activation = SemesterActivation(semester=Semester.objects.get(id=semester_id), is_active=active)\n activation.save()\n\n return semester_view(request=request, semester_id=semester_id)\n", "path": "evap/rewards/views.py"}, {"content": "from django.utils.translation import ugettext_lazy as _\nfrom django.db import models\n\nfrom collections import OrderedDict\n\n\nclass RewardPointRedemptionEvent(models.Model):\n name = models.CharField(max_length=1024, verbose_name=_(\"event name\"))\n date = models.DateField(verbose_name=_(\"event date\"))\n redeem_end_date = models.DateField(verbose_name=_(\"redemption end date\"))\n\n @property\n def can_delete(self):\n if RewardPointRedemption.objects.filter(event=self).exists():\n return False\n return True\n\n def redemptions_by_user(self):\n redemptions = self.reward_point_redemptions.order_by('user_profile')\n redemptions_dict = OrderedDict()\n for redemption in redemptions:\n if not redemption.user_profile in redemptions_dict:\n redemptions_dict[redemption.user_profile] = 0\n redemptions_dict[redemption.user_profile] += redemption.value\n return redemptions_dict\n\nclass RewardPointGranting(models.Model):\n user_profile = models.ForeignKey('evaluation.UserProfile', related_name=\"reward_point_grantings\")\n semester = models.ForeignKey('evaluation.Semester', related_name=\"reward_point_grantings\", blank=True, null=True)\n granting_time = models.DateTimeField(verbose_name=_(\"granting time\"), auto_now_add=True)\n value = models.IntegerField(verbose_name=_(\"value\"), default=0)\n\nclass RewardPointRedemption(models.Model):\n user_profile = models.ForeignKey('evaluation.UserProfile', related_name=\"reward_point_redemptions\")\n redemption_time = models.DateTimeField(verbose_name=_(\"redemption time\"), auto_now_add=True)\n value = models.IntegerField(verbose_name=_(\"value\"), default=0)\n event = models.ForeignKey(RewardPointRedemptionEvent, related_name=\"reward_point_redemptions\")\n\nclass SemesterActivation(models.Model):\n semester = models.OneToOneField('evaluation.Semester', related_name='rewards_active')\n is_active = models.BooleanField(default=False)\n", "path": "evap/rewards/models.py"}], "after_files": [{"content": "from django.conf import settings\nfrom django.contrib import messages\nfrom django.db import transaction\nfrom django.utils.translation import ugettext as _\nfrom django.dispatch import receiver\n\nfrom django.contrib.auth.decorators import login_required\nfrom evap.evaluation.models import Course\n\nfrom evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints\n\n@login_required\[email protected]\ndef save_redemptions(request, redemptions):\n total_points_available = reward_points_of_user(request.user)\n total_points_redeemed = sum(redemptions.values())\n\n if total_points_redeemed <= 0:\n raise NoPointsSelected(_(\"You cannot redeem 0 points.\"))\n\n if total_points_redeemed > total_points_available:\n raise NotEnoughPoints(_(\"You don't have enough reward points.\"))\n\n for event_id in redemptions:\n if redemptions[event_id] > 0:\n redemption = RewardPointRedemption(\n user_profile=request.user,\n value=redemptions[event_id],\n event=RewardPointRedemptionEvent.objects.get(id=event_id)\n )\n redemption.save()\n\ndef can_user_use_reward_points(user):\n return not user.is_external and user.is_participant\n\n\ndef reward_points_of_user(user):\n reward_point_grantings = RewardPointGranting.objects.filter(user_profile=user)\n reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=user)\n\n count = 0\n for granting in reward_point_grantings:\n count += granting.value\n for redemption in reward_point_redemptions:\n count -= redemption.value\n\n return count\n\n\n@receiver(Course.course_evaluated)\ndef grant_reward_points(sender, **kwargs):\n # grant reward points if all conditions are fulfilled\n\n request = kwargs['request']\n semester = kwargs['semester']\n if not can_user_use_reward_points(request.user):\n return\n # has the semester been activated for reward points?\n if not is_semester_activated(semester):\n return\n # does the user have at least one required course in this semester?\n required_courses = Course.objects.filter(participants=request.user, semester=semester, is_required_for_reward=True)\n if not required_courses.exists():\n return\n # does the user not participate in any more required courses in this semester?\n if required_courses.exclude(voters=request.user).exists():\n return\n # did the user not already get reward points for this semester?\n if not RewardPointGranting.objects.filter(user_profile=request.user, semester=semester):\n granting = RewardPointGranting(user_profile=request.user, semester=semester, value=settings.REWARD_POINTS_PER_SEMESTER)\n granting.save()\n messages.success(request, _(\"You just have earned reward points for this semester because you evaluated all your courses. Thank you very much!\"))\n\n\ndef is_semester_activated(semester):\n try:\n activation = SemesterActivation.objects.get(semester=semester)\n return activation.is_active\n except SemesterActivation.DoesNotExist:\n return False\n", "path": "evap/rewards/tools.py"}, {"content": "from django.contrib import messages\nfrom django.shortcuts import get_object_or_404, redirect, render\nfrom django.utils.translation import ugettext as _\nfrom django.utils.translation import get_language\nfrom django.http import HttpResponse\nfrom datetime import datetime\nfrom operator import attrgetter\n\nfrom evap.evaluation.auth import reward_user_required, staff_required\nfrom evap.evaluation.models import Semester, Course\n\nfrom evap.staff.views import semester_view\n\nfrom evap.rewards.models import RewardPointGranting, RewardPointRedemption, RewardPointRedemptionEvent, SemesterActivation, NoPointsSelected, NotEnoughPoints\nfrom evap.rewards.tools import save_redemptions, reward_points_of_user, can_user_use_reward_points\nfrom evap.rewards.forms import RewardPointRedemptionEventForm\nfrom evap.rewards.exporters import ExcelExporter\n\n@reward_user_required\ndef index(request):\n if request.method == 'POST':\n redemptions = {}\n for key, value in request.POST.items():\n if(key.startswith('points-')):\n event_id = int(key.rpartition('-')[2])\n redemptions[event_id] = int(value)\n\n try:\n save_redemptions(request, redemptions)\n messages.success(request, _(\"You successfully redeemed your points.\"))\n except (NoPointsSelected, NotEnoughPoints) as error:\n messages.warning(request, error)\n\n total_points_available = reward_points_of_user(request.user)\n reward_point_grantings = RewardPointGranting.objects.filter(user_profile=request.user)\n reward_point_redemptions = RewardPointRedemption.objects.filter(user_profile=request.user)\n events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now())\n events = sorted(events, key=lambda event: event.date)\n\n reward_point_actions=[]\n for granting in reward_point_grantings:\n reward_point_actions.append((granting.granting_time, _('Reward for') + ' ' + granting.semester.name, granting.value, ''))\n for redemption in reward_point_redemptions:\n reward_point_actions.append((redemption.redemption_time, redemption.event.name, '', redemption.value))\n\n reward_point_actions.sort(key=lambda action: action[0], reverse=True)\n\n template_data = dict(\n reward_point_actions=reward_point_actions,\n total_points_available=total_points_available,\n events=events,\n point_selection=[x for x in range(0,total_points_available+1)])\n return render(request, \"rewards_index.html\", template_data)\n\n\n@staff_required\ndef semester_reward_points(request, semester_id):\n semester = get_object_or_404(Semester, id=semester_id)\n courses = Course.objects.filter(semester=semester)\n participants = set()\n for course in courses:\n for participant in course.participants.all():\n if can_user_use_reward_points(participant):\n participants.add(participant)\n participants = sorted(participants, key=attrgetter('last_name', 'first_name'))\n\n data = []\n for participant in participants:\n number_of_required_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=True).count()\n number_of_required_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=True).count()\n number_of_optional_courses = Course.objects.filter(semester=semester, participants=participant, is_required_for_reward=False).count()\n number_of_optional_courses_voted_for = Course.objects.filter(semester=semester, voters=participant, is_required_for_reward=False).count()\n earned_reward_points = RewardPointGranting.objects.filter(semester=semester, user_profile=participant).exists()\n data.append((participant, number_of_required_courses_voted_for, number_of_required_courses,\n number_of_optional_courses_voted_for, number_of_optional_courses, earned_reward_points))\n\n template_data = dict(semester=semester, data=data, disable_breadcrumb_semester=False)\n return render(request, \"rewards_semester_reward_points_view.html\", template_data)\n\n\n@staff_required\ndef reward_point_redemption_events(request):\n upcoming_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__gte=datetime.now()).order_by('date')\n past_events = RewardPointRedemptionEvent.objects.filter(redeem_end_date__lt=datetime.now()).order_by('-date')\n template_data = dict(upcoming_events=upcoming_events, past_events=past_events)\n return render(request, \"rewards_reward_point_redemption_events.html\", template_data)\n\n\n@staff_required\ndef reward_point_redemption_event_create(request):\n event = RewardPointRedemptionEvent()\n form = RewardPointRedemptionEventForm(request.POST or None, instance=event)\n\n if form.is_valid():\n form.save()\n messages.success(request, _(\"Successfully created event.\"))\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_form.html\", dict(form=form))\n\n\n@staff_required\ndef reward_point_redemption_event_edit(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n form = RewardPointRedemptionEventForm(request.POST or None, instance=event)\n\n if form.is_valid():\n event = form.save()\n\n messages.success(request, _(\"Successfully updated event.\"))\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_form.html\", dict(event=event, form=form))\n\n\n@staff_required\ndef reward_point_redemption_event_delete(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n\n if event.can_delete:\n if request.method == 'POST':\n event.delete()\n return redirect('rewards:reward_point_redemption_events')\n else:\n return render(request, \"rewards_reward_point_redemption_event_delete.html\", dict(event=event))\n else:\n messages.warning(request, _(\"This event cannot be deleted because some users already redeemed points for it.\"))\n return redirect('rewards:reward_point_redemption_events')\n\n\n@staff_required\ndef reward_point_redemption_event_export(request, event_id):\n event = get_object_or_404(RewardPointRedemptionEvent, id=event_id)\n\n filename = _(\"RewardPoints\")+\"-%s-%s-%s.xls\" % (event.date, event.name, get_language())\n\n response = HttpResponse(content_type=\"application/vnd.ms-excel\")\n response[\"Content-Disposition\"] = \"attachment; filename=\\\"%s\\\"\" % filename\n\n ExcelExporter(event.redemptions_by_user()).export(response)\n\n return response\n\n\n@staff_required\ndef semester_activation(request, semester_id, active):\n if active == 'on':\n active = True\n else:\n active = False\n\n try:\n activation = SemesterActivation.objects.filter(semester=Semester.objects.get(id=semester_id)).get()\n activation.is_active = active\n except SemesterActivation.DoesNotExist:\n activation = SemesterActivation(semester=Semester.objects.get(id=semester_id), is_active=active)\n activation.save()\n\n return semester_view(request=request, semester_id=semester_id)\n", "path": "evap/rewards/views.py"}, {"content": "from django.utils.translation import ugettext_lazy as _\nfrom django.db import models\n\nfrom collections import OrderedDict\n\nclass NoPointsSelected(Exception):\n \"\"\"An attempt has been made to redeem <= 0 points.\"\"\"\n pass\n\nclass NotEnoughPoints(Exception):\n \"\"\"An attempt has been made to redeem more points than available.\"\"\"\n pass\n\nclass RewardPointRedemptionEvent(models.Model):\n name = models.CharField(max_length=1024, verbose_name=_(\"event name\"))\n date = models.DateField(verbose_name=_(\"event date\"))\n redeem_end_date = models.DateField(verbose_name=_(\"redemption end date\"))\n\n @property\n def can_delete(self):\n if RewardPointRedemption.objects.filter(event=self).exists():\n return False\n return True\n\n def redemptions_by_user(self):\n redemptions = self.reward_point_redemptions.order_by('user_profile')\n redemptions_dict = OrderedDict()\n for redemption in redemptions:\n if not redemption.user_profile in redemptions_dict:\n redemptions_dict[redemption.user_profile] = 0\n redemptions_dict[redemption.user_profile] += redemption.value\n return redemptions_dict\n\nclass RewardPointGranting(models.Model):\n user_profile = models.ForeignKey('evaluation.UserProfile', related_name=\"reward_point_grantings\")\n semester = models.ForeignKey('evaluation.Semester', related_name=\"reward_point_grantings\", blank=True, null=True)\n granting_time = models.DateTimeField(verbose_name=_(\"granting time\"), auto_now_add=True)\n value = models.IntegerField(verbose_name=_(\"value\"), default=0)\n\nclass RewardPointRedemption(models.Model):\n user_profile = models.ForeignKey('evaluation.UserProfile', related_name=\"reward_point_redemptions\")\n redemption_time = models.DateTimeField(verbose_name=_(\"redemption time\"), auto_now_add=True)\n value = models.IntegerField(verbose_name=_(\"value\"), default=0)\n event = models.ForeignKey(RewardPointRedemptionEvent, related_name=\"reward_point_redemptions\")\n\nclass SemesterActivation(models.Model):\n semester = models.OneToOneField('evaluation.Semester', related_name='rewards_active')\n is_active = models.BooleanField(default=False)\n", "path": "evap/rewards/models.py"}]}
| 3,487 | 804 |
gh_patches_debug_18075
|
rasdani/github-patches
|
git_diff
|
meltano__meltano-8169
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
feature: Set the meltano.yml path in a directory different from the project
### Feature scope
Configuration (settings parsing, validation, etc.)
### Description
I really tried to make a docker container with meltano to run in AWS Lambda, but I had a lot of problems because in my case I needed to change meltano.yml in run time, but only possible to change files in AWS Lambda if this file is in /tmp.
I believe could be a great feature if it is possible to set meltano.yml path by a environment variable. I tried to change this meltano source code but I had problems with the plugins installations. These changes are beyond my skills.
So I propose this new feature like other projects have (DBT as an example), where it is possible set a new configuration file or set a path for the configuration file.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/meltano/cli/add.py`
Content:
```
1 """Plugin Add CLI."""
2
3 from __future__ import annotations
4
5 import typing as t
6 from pathlib import Path
7 from urllib.parse import urlparse
8
9 import click
10 import requests
11
12 from meltano.cli.params import pass_project
13 from meltano.cli.utils import (
14 CliError,
15 PartialInstrumentedCmd,
16 add_plugin,
17 add_required_plugins,
18 check_dependencies_met,
19 install_plugins,
20 )
21 from meltano.core.plugin import PluginRef, PluginType
22 from meltano.core.plugin_install_service import PluginInstallReason
23 from meltano.core.project_add_service import ProjectAddService
24 from meltano.core.tracking.contexts import CliEvent, PluginsTrackingContext
25 from meltano.core.yaml import yaml
26
27 if t.TYPE_CHECKING:
28 from meltano.core.plugin.project_plugin import ProjectPlugin
29 from meltano.core.project import Project
30 from meltano.core.tracking import Tracker
31
32
33 def _load_yaml_from_ref(_ctx, _param, value: str | None) -> dict:
34 if not value:
35 return
36
37 try:
38 url = urlparse(value)
39 if url.scheme and url.netloc:
40 response = requests.get(value, timeout=10)
41 response.raise_for_status()
42 content = response.text
43 else:
44 content = Path(value).read_text()
45
46 except (ValueError, FileNotFoundError, IsADirectoryError) as e:
47 raise click.BadParameter(e) from e
48
49 return yaml.load(content) or {}
50
51
52 @click.command( # noqa: WPS238
53 cls=PartialInstrumentedCmd,
54 short_help="Add a plugin to your project.",
55 )
56 @click.argument("plugin_type", type=click.Choice(PluginType.cli_arguments()))
57 @click.argument("plugin_name", nargs=-1, required=True)
58 @click.option(
59 "--inherit-from",
60 help=(
61 "Add a plugin inheriting from an existing plugin in the project"
62 " or a discoverable plugin identified, by name."
63 ),
64 )
65 @click.option(
66 "--variant",
67 help="Add a specific (non-default) variant of the identified discoverable plugin.",
68 )
69 @click.option(
70 "--as",
71 "as_name",
72 help=(
73 "Shorthand for '--inherit-from', that can be used to add a "
74 "discoverable plugin to your project with a different name. "
75 "Usage:\b\n\nadd <type> <inherit-from> --as <name>"
76 ),
77 )
78 @click.option(
79 "--from-ref",
80 "plugin_yaml",
81 callback=_load_yaml_from_ref,
82 help="Reference a plugin defintion to add from.",
83 )
84 @click.option(
85 "--python",
86 help=(
87 "The Python version to use for the plugin. Only applies to base plugins which "
88 "have Python virtual environments, rather than inherited plugins which use the "
89 "virtual environment of their base plugin, or plugins that run in a container."
90 ),
91 )
92 @click.option(
93 "--custom",
94 is_flag=True,
95 help=(
96 "Add a custom plugin. The command will prompt you for the package's "
97 "base plugin description metadata."
98 ),
99 )
100 @click.option(
101 "--no-install",
102 is_flag=True,
103 help="Do not install the plugin after adding it to the project.",
104 )
105 @pass_project()
106 @click.pass_context
107 def add( # noqa: WPS238
108 ctx,
109 project: Project,
110 plugin_type: str,
111 plugin_name: str,
112 inherit_from: str | None = None,
113 variant: str | None = None,
114 as_name: str | None = None,
115 plugin_yaml: dict | None = None,
116 python: str | None = None,
117 **flags,
118 ):
119 """
120 Add a plugin to your project.
121
122 \b\nRead more at https://docs.meltano.com/reference/command-line-interface#add
123 """
124 tracker: Tracker = ctx.obj["tracker"]
125
126 plugin_type = PluginType.from_cli_argument(plugin_type)
127 plugin_names = plugin_name # nargs=-1
128
129 if as_name:
130 # `add <type> <inherit-from> --as <name>``
131 # is equivalent to:
132 # `add <type> <name> --inherit-from <inherit-from>``
133 inherit_from = plugin_names[0]
134 plugin_names = [as_name]
135
136 if flags["custom"] and plugin_type in {
137 PluginType.TRANSFORMS,
138 PluginType.ORCHESTRATORS,
139 }:
140 tracker.track_command_event(CliEvent.aborted)
141 raise CliError(f"--custom is not supported for {plugin_type}")
142
143 plugin_refs = [
144 PluginRef(plugin_type=plugin_type, name=name) for name in plugin_names
145 ]
146 dependencies_met, err = check_dependencies_met(
147 plugin_refs=plugin_refs,
148 plugins_service=project.plugins,
149 )
150 if not dependencies_met:
151 tracker.track_command_event(CliEvent.aborted)
152 raise CliError(f"Failed to install plugin(s): {err}")
153
154 add_service = ProjectAddService(project)
155
156 plugins: list[ProjectPlugin] = []
157 for plugin in plugin_names:
158 try:
159 plugins.append(
160 add_plugin(
161 plugin_type,
162 plugin,
163 python=python,
164 inherit_from=inherit_from,
165 variant=variant,
166 custom=flags["custom"],
167 add_service=add_service,
168 plugin_yaml=plugin_yaml,
169 ),
170 )
171 except Exception:
172 # if the plugin is not known to meltano send what information we do have
173 tracker.add_contexts(
174 PluginsTrackingContext([(plugin, None) for plugin in plugins]),
175 )
176 tracker.track_command_event(CliEvent.aborted)
177 raise
178
179 required_plugins = add_required_plugins(
180 plugins,
181 add_service=add_service,
182 )
183 plugins.extend(required_plugins)
184 tracker.add_contexts(
185 PluginsTrackingContext([(candidate, None) for candidate in plugins]),
186 )
187 tracker.track_command_event(CliEvent.inflight)
188
189 if not flags.get("no_install"):
190 success = install_plugins(project, plugins, reason=PluginInstallReason.ADD)
191
192 if not success:
193 tracker.track_command_event(CliEvent.failed)
194 raise CliError("Failed to install plugin(s)")
195
196 _print_plugins(plugins)
197 tracker.track_command_event(CliEvent.completed)
198
199
200 def _print_plugins(plugins):
201 printed_empty_line = False
202 for plugin in plugins:
203 docs_url = plugin.docs or plugin.repo
204 if not docs_url:
205 continue
206
207 if not printed_empty_line:
208 click.echo()
209 printed_empty_line = True
210
211 click.echo(
212 f"To learn more about {plugin.type.descriptor} '{plugin.name}', "
213 f"visit {docs_url}",
214 )
215
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/meltano/cli/add.py b/src/meltano/cli/add.py
--- a/src/meltano/cli/add.py
+++ b/src/meltano/cli/add.py
@@ -102,6 +102,11 @@
is_flag=True,
help="Do not install the plugin after adding it to the project.",
)
[email protected](
+ "--force-install",
+ is_flag=True,
+ help="Ignore the required Python version declared by the plugins.",
+)
@pass_project()
@click.pass_context
def add( # noqa: WPS238
@@ -187,7 +192,12 @@
tracker.track_command_event(CliEvent.inflight)
if not flags.get("no_install"):
- success = install_plugins(project, plugins, reason=PluginInstallReason.ADD)
+ success = install_plugins(
+ project,
+ plugins,
+ reason=PluginInstallReason.ADD,
+ force=flags.get("force_install", False),
+ )
if not success:
tracker.track_command_event(CliEvent.failed)
|
{"golden_diff": "diff --git a/src/meltano/cli/add.py b/src/meltano/cli/add.py\n--- a/src/meltano/cli/add.py\n+++ b/src/meltano/cli/add.py\n@@ -102,6 +102,11 @@\n is_flag=True,\n help=\"Do not install the plugin after adding it to the project.\",\n )\[email protected](\n+ \"--force-install\",\n+ is_flag=True,\n+ help=\"Ignore the required Python version declared by the plugins.\",\n+)\n @pass_project()\n @click.pass_context\n def add( # noqa: WPS238\n@@ -187,7 +192,12 @@\n tracker.track_command_event(CliEvent.inflight)\n \n if not flags.get(\"no_install\"):\n- success = install_plugins(project, plugins, reason=PluginInstallReason.ADD)\n+ success = install_plugins(\n+ project,\n+ plugins,\n+ reason=PluginInstallReason.ADD,\n+ force=flags.get(\"force_install\", False),\n+ )\n \n if not success:\n tracker.track_command_event(CliEvent.failed)\n", "issue": "feature: Set the meltano.yml path in a directory different from the project \n### Feature scope\n\nConfiguration (settings parsing, validation, etc.)\n\n### Description\n\nI really tried to make a docker container with meltano to run in AWS Lambda, but I had a lot of problems because in my case I needed to change meltano.yml in run time, but only possible to change files in AWS Lambda if this file is in /tmp.\r\n\r\nI believe could be a great feature if it is possible to set meltano.yml path by a environment variable. I tried to change this meltano source code but I had problems with the plugins installations. These changes are beyond my skills. \r\n\r\nSo I propose this new feature like other projects have (DBT as an example), where it is possible set a new configuration file or set a path for the configuration file.\n", "before_files": [{"content": "\"\"\"Plugin Add CLI.\"\"\"\n\nfrom __future__ import annotations\n\nimport typing as t\nfrom pathlib import Path\nfrom urllib.parse import urlparse\n\nimport click\nimport requests\n\nfrom meltano.cli.params import pass_project\nfrom meltano.cli.utils import (\n CliError,\n PartialInstrumentedCmd,\n add_plugin,\n add_required_plugins,\n check_dependencies_met,\n install_plugins,\n)\nfrom meltano.core.plugin import PluginRef, PluginType\nfrom meltano.core.plugin_install_service import PluginInstallReason\nfrom meltano.core.project_add_service import ProjectAddService\nfrom meltano.core.tracking.contexts import CliEvent, PluginsTrackingContext\nfrom meltano.core.yaml import yaml\n\nif t.TYPE_CHECKING:\n from meltano.core.plugin.project_plugin import ProjectPlugin\n from meltano.core.project import Project\n from meltano.core.tracking import Tracker\n\n\ndef _load_yaml_from_ref(_ctx, _param, value: str | None) -> dict:\n if not value:\n return\n\n try:\n url = urlparse(value)\n if url.scheme and url.netloc:\n response = requests.get(value, timeout=10)\n response.raise_for_status()\n content = response.text\n else:\n content = Path(value).read_text()\n\n except (ValueError, FileNotFoundError, IsADirectoryError) as e:\n raise click.BadParameter(e) from e\n\n return yaml.load(content) or {}\n\n\[email protected]( # noqa: WPS238\n cls=PartialInstrumentedCmd,\n short_help=\"Add a plugin to your project.\",\n)\[email protected](\"plugin_type\", type=click.Choice(PluginType.cli_arguments()))\[email protected](\"plugin_name\", nargs=-1, required=True)\[email protected](\n \"--inherit-from\",\n help=(\n \"Add a plugin inheriting from an existing plugin in the project\"\n \" or a discoverable plugin identified, by name.\"\n ),\n)\[email protected](\n \"--variant\",\n help=\"Add a specific (non-default) variant of the identified discoverable plugin.\",\n)\[email protected](\n \"--as\",\n \"as_name\",\n help=(\n \"Shorthand for '--inherit-from', that can be used to add a \"\n \"discoverable plugin to your project with a different name. \"\n \"Usage:\\b\\n\\nadd <type> <inherit-from> --as <name>\"\n ),\n)\[email protected](\n \"--from-ref\",\n \"plugin_yaml\",\n callback=_load_yaml_from_ref,\n help=\"Reference a plugin defintion to add from.\",\n)\[email protected](\n \"--python\",\n help=(\n \"The Python version to use for the plugin. Only applies to base plugins which \"\n \"have Python virtual environments, rather than inherited plugins which use the \"\n \"virtual environment of their base plugin, or plugins that run in a container.\"\n ),\n)\[email protected](\n \"--custom\",\n is_flag=True,\n help=(\n \"Add a custom plugin. The command will prompt you for the package's \"\n \"base plugin description metadata.\"\n ),\n)\[email protected](\n \"--no-install\",\n is_flag=True,\n help=\"Do not install the plugin after adding it to the project.\",\n)\n@pass_project()\[email protected]_context\ndef add( # noqa: WPS238\n ctx,\n project: Project,\n plugin_type: str,\n plugin_name: str,\n inherit_from: str | None = None,\n variant: str | None = None,\n as_name: str | None = None,\n plugin_yaml: dict | None = None,\n python: str | None = None,\n **flags,\n):\n \"\"\"\n Add a plugin to your project.\n\n \\b\\nRead more at https://docs.meltano.com/reference/command-line-interface#add\n \"\"\"\n tracker: Tracker = ctx.obj[\"tracker\"]\n\n plugin_type = PluginType.from_cli_argument(plugin_type)\n plugin_names = plugin_name # nargs=-1\n\n if as_name:\n # `add <type> <inherit-from> --as <name>``\n # is equivalent to:\n # `add <type> <name> --inherit-from <inherit-from>``\n inherit_from = plugin_names[0]\n plugin_names = [as_name]\n\n if flags[\"custom\"] and plugin_type in {\n PluginType.TRANSFORMS,\n PluginType.ORCHESTRATORS,\n }:\n tracker.track_command_event(CliEvent.aborted)\n raise CliError(f\"--custom is not supported for {plugin_type}\")\n\n plugin_refs = [\n PluginRef(plugin_type=plugin_type, name=name) for name in plugin_names\n ]\n dependencies_met, err = check_dependencies_met(\n plugin_refs=plugin_refs,\n plugins_service=project.plugins,\n )\n if not dependencies_met:\n tracker.track_command_event(CliEvent.aborted)\n raise CliError(f\"Failed to install plugin(s): {err}\")\n\n add_service = ProjectAddService(project)\n\n plugins: list[ProjectPlugin] = []\n for plugin in plugin_names:\n try:\n plugins.append(\n add_plugin(\n plugin_type,\n plugin,\n python=python,\n inherit_from=inherit_from,\n variant=variant,\n custom=flags[\"custom\"],\n add_service=add_service,\n plugin_yaml=plugin_yaml,\n ),\n )\n except Exception:\n # if the plugin is not known to meltano send what information we do have\n tracker.add_contexts(\n PluginsTrackingContext([(plugin, None) for plugin in plugins]),\n )\n tracker.track_command_event(CliEvent.aborted)\n raise\n\n required_plugins = add_required_plugins(\n plugins,\n add_service=add_service,\n )\n plugins.extend(required_plugins)\n tracker.add_contexts(\n PluginsTrackingContext([(candidate, None) for candidate in plugins]),\n )\n tracker.track_command_event(CliEvent.inflight)\n\n if not flags.get(\"no_install\"):\n success = install_plugins(project, plugins, reason=PluginInstallReason.ADD)\n\n if not success:\n tracker.track_command_event(CliEvent.failed)\n raise CliError(\"Failed to install plugin(s)\")\n\n _print_plugins(plugins)\n tracker.track_command_event(CliEvent.completed)\n\n\ndef _print_plugins(plugins):\n printed_empty_line = False\n for plugin in plugins:\n docs_url = plugin.docs or plugin.repo\n if not docs_url:\n continue\n\n if not printed_empty_line:\n click.echo()\n printed_empty_line = True\n\n click.echo(\n f\"To learn more about {plugin.type.descriptor} '{plugin.name}', \"\n f\"visit {docs_url}\",\n )\n", "path": "src/meltano/cli/add.py"}], "after_files": [{"content": "\"\"\"Plugin Add CLI.\"\"\"\n\nfrom __future__ import annotations\n\nimport typing as t\nfrom pathlib import Path\nfrom urllib.parse import urlparse\n\nimport click\nimport requests\n\nfrom meltano.cli.params import pass_project\nfrom meltano.cli.utils import (\n CliError,\n PartialInstrumentedCmd,\n add_plugin,\n add_required_plugins,\n check_dependencies_met,\n install_plugins,\n)\nfrom meltano.core.plugin import PluginRef, PluginType\nfrom meltano.core.plugin_install_service import PluginInstallReason\nfrom meltano.core.project_add_service import ProjectAddService\nfrom meltano.core.tracking.contexts import CliEvent, PluginsTrackingContext\nfrom meltano.core.yaml import yaml\n\nif t.TYPE_CHECKING:\n from meltano.core.plugin.project_plugin import ProjectPlugin\n from meltano.core.project import Project\n from meltano.core.tracking import Tracker\n\n\ndef _load_yaml_from_ref(_ctx, _param, value: str | None) -> dict:\n if not value:\n return\n\n try:\n url = urlparse(value)\n if url.scheme and url.netloc:\n response = requests.get(value, timeout=10)\n response.raise_for_status()\n content = response.text\n else:\n content = Path(value).read_text()\n\n except (ValueError, FileNotFoundError, IsADirectoryError) as e:\n raise click.BadParameter(e) from e\n\n return yaml.load(content) or {}\n\n\[email protected]( # noqa: WPS238\n cls=PartialInstrumentedCmd,\n short_help=\"Add a plugin to your project.\",\n)\[email protected](\"plugin_type\", type=click.Choice(PluginType.cli_arguments()))\[email protected](\"plugin_name\", nargs=-1, required=True)\[email protected](\n \"--inherit-from\",\n help=(\n \"Add a plugin inheriting from an existing plugin in the project\"\n \" or a discoverable plugin identified, by name.\"\n ),\n)\[email protected](\n \"--variant\",\n help=\"Add a specific (non-default) variant of the identified discoverable plugin.\",\n)\[email protected](\n \"--as\",\n \"as_name\",\n help=(\n \"Shorthand for '--inherit-from', that can be used to add a \"\n \"discoverable plugin to your project with a different name. \"\n \"Usage:\\b\\n\\nadd <type> <inherit-from> --as <name>\"\n ),\n)\[email protected](\n \"--from-ref\",\n \"plugin_yaml\",\n callback=_load_yaml_from_ref,\n help=\"Reference a plugin defintion to add from.\",\n)\[email protected](\n \"--python\",\n help=(\n \"The Python version to use for the plugin. Only applies to base plugins which \"\n \"have Python virtual environments, rather than inherited plugins which use the \"\n \"virtual environment of their base plugin, or plugins that run in a container.\"\n ),\n)\[email protected](\n \"--custom\",\n is_flag=True,\n help=(\n \"Add a custom plugin. The command will prompt you for the package's \"\n \"base plugin description metadata.\"\n ),\n)\[email protected](\n \"--no-install\",\n is_flag=True,\n help=\"Do not install the plugin after adding it to the project.\",\n)\[email protected](\n \"--force-install\",\n is_flag=True,\n help=\"Ignore the required Python version declared by the plugins.\",\n)\n@pass_project()\[email protected]_context\ndef add( # noqa: WPS238\n ctx,\n project: Project,\n plugin_type: str,\n plugin_name: str,\n inherit_from: str | None = None,\n variant: str | None = None,\n as_name: str | None = None,\n plugin_yaml: dict | None = None,\n python: str | None = None,\n **flags,\n):\n \"\"\"\n Add a plugin to your project.\n\n \\b\\nRead more at https://docs.meltano.com/reference/command-line-interface#add\n \"\"\"\n tracker: Tracker = ctx.obj[\"tracker\"]\n\n plugin_type = PluginType.from_cli_argument(plugin_type)\n plugin_names = plugin_name # nargs=-1\n\n if as_name:\n # `add <type> <inherit-from> --as <name>``\n # is equivalent to:\n # `add <type> <name> --inherit-from <inherit-from>``\n inherit_from = plugin_names[0]\n plugin_names = [as_name]\n\n if flags[\"custom\"] and plugin_type in {\n PluginType.TRANSFORMS,\n PluginType.ORCHESTRATORS,\n }:\n tracker.track_command_event(CliEvent.aborted)\n raise CliError(f\"--custom is not supported for {plugin_type}\")\n\n plugin_refs = [\n PluginRef(plugin_type=plugin_type, name=name) for name in plugin_names\n ]\n dependencies_met, err = check_dependencies_met(\n plugin_refs=plugin_refs,\n plugins_service=project.plugins,\n )\n if not dependencies_met:\n tracker.track_command_event(CliEvent.aborted)\n raise CliError(f\"Failed to install plugin(s): {err}\")\n\n add_service = ProjectAddService(project)\n\n plugins: list[ProjectPlugin] = []\n for plugin in plugin_names:\n try:\n plugins.append(\n add_plugin(\n plugin_type,\n plugin,\n python=python,\n inherit_from=inherit_from,\n variant=variant,\n custom=flags[\"custom\"],\n add_service=add_service,\n plugin_yaml=plugin_yaml,\n ),\n )\n except Exception:\n # if the plugin is not known to meltano send what information we do have\n tracker.add_contexts(\n PluginsTrackingContext([(plugin, None) for plugin in plugins]),\n )\n tracker.track_command_event(CliEvent.aborted)\n raise\n\n required_plugins = add_required_plugins(\n plugins,\n add_service=add_service,\n )\n plugins.extend(required_plugins)\n tracker.add_contexts(\n PluginsTrackingContext([(candidate, None) for candidate in plugins]),\n )\n tracker.track_command_event(CliEvent.inflight)\n\n if not flags.get(\"no_install\"):\n success = install_plugins(\n project,\n plugins,\n reason=PluginInstallReason.ADD,\n force=flags.get(\"force_install\", False),\n )\n\n if not success:\n tracker.track_command_event(CliEvent.failed)\n raise CliError(\"Failed to install plugin(s)\")\n\n _print_plugins(plugins)\n tracker.track_command_event(CliEvent.completed)\n\n\ndef _print_plugins(plugins):\n printed_empty_line = False\n for plugin in plugins:\n docs_url = plugin.docs or plugin.repo\n if not docs_url:\n continue\n\n if not printed_empty_line:\n click.echo()\n printed_empty_line = True\n\n click.echo(\n f\"To learn more about {plugin.type.descriptor} '{plugin.name}', \"\n f\"visit {docs_url}\",\n )\n", "path": "src/meltano/cli/add.py"}]}
| 2,388 | 239 |
gh_patches_debug_803
|
rasdani/github-patches
|
git_diff
|
googleapis__google-api-python-client-871
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
AttributeError: module 'googleapiclient' has no attribute '__version__'
When importing new version of google-api-python-client
`from apiclient import discovery`
i'm getting the error
`AttributeError: module 'googleapiclient' has no attribute '__version__'`
https://github.com/googleapis/google-api-python-client/blob/84d45619d753cb04d957651886231034194058b6/apiclient/__init__.py#L22
i guess this happens since you have removed `__version__` var from `googleapiclient/__init__.py`
https://github.com/googleapis/google-api-python-client/commit/f706cfd821ab7457e5db37abfc3619772657dd0e#diff-b926d296d4c856bcbf877809e4523562L15
can you please fix? @busunkim96 @mik-laj @crwilcox
Traceback:
``` from apiclient import discovery
File "/usr/local/lib/python3.7/site-packages/apiclient/__init__.py", line 22, in <module>
__version__ = googleapiclient.__version__
AttributeError: module 'googleapiclient' has no attribute '__version__'
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `apiclient/__init__.py`
Content:
```
1 """Retain apiclient as an alias for googleapiclient."""
2
3 from six import iteritems
4
5 import googleapiclient
6
7 from googleapiclient import channel
8 from googleapiclient import discovery
9 from googleapiclient import errors
10 from googleapiclient import http
11 from googleapiclient import mimeparse
12 from googleapiclient import model
13
14 try:
15 from googleapiclient import sample_tools
16 except ImportError:
17 # Silently ignore, because the vast majority of consumers won't use it and
18 # it has deep dependence on oauth2client, an optional dependency.
19 sample_tools = None
20 from googleapiclient import schema
21
22 __version__ = googleapiclient.__version__
23
24 _SUBMODULES = {
25 "channel": channel,
26 "discovery": discovery,
27 "errors": errors,
28 "http": http,
29 "mimeparse": mimeparse,
30 "model": model,
31 "sample_tools": sample_tools,
32 "schema": schema,
33 }
34
35 import sys
36
37 for module_name, module in iteritems(_SUBMODULES):
38 sys.modules["apiclient.%s" % module_name] = module
39
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/apiclient/__init__.py b/apiclient/__init__.py
--- a/apiclient/__init__.py
+++ b/apiclient/__init__.py
@@ -19,8 +19,6 @@
sample_tools = None
from googleapiclient import schema
-__version__ = googleapiclient.__version__
-
_SUBMODULES = {
"channel": channel,
"discovery": discovery,
|
{"golden_diff": "diff --git a/apiclient/__init__.py b/apiclient/__init__.py\n--- a/apiclient/__init__.py\n+++ b/apiclient/__init__.py\n@@ -19,8 +19,6 @@\n sample_tools = None\n from googleapiclient import schema\n \n-__version__ = googleapiclient.__version__\n-\n _SUBMODULES = {\n \"channel\": channel,\n \"discovery\": discovery,\n", "issue": "AttributeError: module 'googleapiclient' has no attribute '__version__'\nWhen importing new version of google-api-python-client\r\n`from apiclient import discovery`\r\ni'm getting the error\r\n`AttributeError: module 'googleapiclient' has no attribute '__version__'`\r\n\r\nhttps://github.com/googleapis/google-api-python-client/blob/84d45619d753cb04d957651886231034194058b6/apiclient/__init__.py#L22\r\n\r\ni guess this happens since you have removed `__version__` var from `googleapiclient/__init__.py`\r\n\r\nhttps://github.com/googleapis/google-api-python-client/commit/f706cfd821ab7457e5db37abfc3619772657dd0e#diff-b926d296d4c856bcbf877809e4523562L15\r\n\r\ncan you please fix? @busunkim96 @mik-laj @crwilcox \r\n\r\nTraceback:\r\n``` from apiclient import discovery\r\n File \"/usr/local/lib/python3.7/site-packages/apiclient/__init__.py\", line 22, in <module>\r\n __version__ = googleapiclient.__version__\r\nAttributeError: module 'googleapiclient' has no attribute '__version__'\n", "before_files": [{"content": "\"\"\"Retain apiclient as an alias for googleapiclient.\"\"\"\n\nfrom six import iteritems\n\nimport googleapiclient\n\nfrom googleapiclient import channel\nfrom googleapiclient import discovery\nfrom googleapiclient import errors\nfrom googleapiclient import http\nfrom googleapiclient import mimeparse\nfrom googleapiclient import model\n\ntry:\n from googleapiclient import sample_tools\nexcept ImportError:\n # Silently ignore, because the vast majority of consumers won't use it and\n # it has deep dependence on oauth2client, an optional dependency.\n sample_tools = None\nfrom googleapiclient import schema\n\n__version__ = googleapiclient.__version__\n\n_SUBMODULES = {\n \"channel\": channel,\n \"discovery\": discovery,\n \"errors\": errors,\n \"http\": http,\n \"mimeparse\": mimeparse,\n \"model\": model,\n \"sample_tools\": sample_tools,\n \"schema\": schema,\n}\n\nimport sys\n\nfor module_name, module in iteritems(_SUBMODULES):\n sys.modules[\"apiclient.%s\" % module_name] = module\n", "path": "apiclient/__init__.py"}], "after_files": [{"content": "\"\"\"Retain apiclient as an alias for googleapiclient.\"\"\"\n\nfrom six import iteritems\n\nimport googleapiclient\n\nfrom googleapiclient import channel\nfrom googleapiclient import discovery\nfrom googleapiclient import errors\nfrom googleapiclient import http\nfrom googleapiclient import mimeparse\nfrom googleapiclient import model\n\ntry:\n from googleapiclient import sample_tools\nexcept ImportError:\n # Silently ignore, because the vast majority of consumers won't use it and\n # it has deep dependence on oauth2client, an optional dependency.\n sample_tools = None\nfrom googleapiclient import schema\n\n_SUBMODULES = {\n \"channel\": channel,\n \"discovery\": discovery,\n \"errors\": errors,\n \"http\": http,\n \"mimeparse\": mimeparse,\n \"model\": model,\n \"sample_tools\": sample_tools,\n \"schema\": schema,\n}\n\nimport sys\n\nfor module_name, module in iteritems(_SUBMODULES):\n sys.modules[\"apiclient.%s\" % module_name] = module\n", "path": "apiclient/__init__.py"}]}
| 904 | 98 |
gh_patches_debug_6271
|
rasdani/github-patches
|
git_diff
|
saleor__saleor-4917
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Problem with filtering products
### What I'm trying to achieve
I want to filter out products from grocerries category but I receive results even when I should not.
### Steps to reproduce the problem
In groceries category there are two beers that have `abv` attribute with values: 51(5,1% alc) and 67(6,75 alc) but when filter with `abv` value set to 52 we receive `Carrot juice`
Correct query:
```
query{
products(first: 100, attributes: ["abv:51"]){
edges {
node {
id
name
}
}
}
}
```
and correct result:
```
{
"data": {
"products": {
"edges": [
{
"node": {
"id": "UHJvZHVjdDo4Mw==",
"name": "Seaman Lager"
}
}
]
}
}
}
```
Query with invalid value(abv set to 52):
```
query{
products(first: 100, attributes: ["abv:52"]){
edges {
node {
id
name
}
}
}
}
```
invalid result:
```
{
"data": {
"products": {
"edges": [
{
"node": {
"id": "UHJvZHVjdDo3Mw==",
"name": "Carrot Juice"
}
}
]
}
}
}
```
### What I expected to happen
We should receive empty product list if passing abv value 52.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `saleor/graphql/product/filters.py`
Content:
```
1 from collections import defaultdict
2
3 import django_filters
4 from django.db.models import Q, Sum
5 from graphene_django.filter import GlobalIDFilter, GlobalIDMultipleChoiceFilter
6
7 from ...product.filters import (
8 T_PRODUCT_FILTER_QUERIES,
9 filter_products_by_attributes_values,
10 )
11 from ...product.models import Attribute, Category, Collection, Product, ProductType
12 from ...search.backends import picker
13 from ..core.filters import EnumFilter, ListObjectTypeFilter, ObjectTypeFilter
14 from ..core.types import FilterInputObjectType
15 from ..core.types.common import PriceRangeInput
16 from ..core.utils import from_global_id_strict_type
17 from ..utils import filter_by_query_param, get_nodes
18 from . import types
19 from .enums import (
20 CollectionPublished,
21 ProductTypeConfigurable,
22 ProductTypeEnum,
23 StockAvailability,
24 )
25 from .types.attributes import AttributeInput
26
27
28 def filter_fields_containing_value(*search_fields: str):
29 """Create a icontains filters through given fields on a given query set object."""
30
31 def _filter_qs(qs, _, value):
32 if value:
33 qs = filter_by_query_param(qs, value, search_fields)
34 return qs
35
36 return _filter_qs
37
38
39 def _clean_product_attributes_filter_input(filter_value) -> T_PRODUCT_FILTER_QUERIES:
40 attributes = Attribute.objects.prefetch_related("values")
41 attributes_map = {attribute.slug: attribute.pk for attribute in attributes}
42 values_map = {
43 attr.slug: {value.slug: value.pk for value in attr.values.all()}
44 for attr in attributes
45 }
46 queries = defaultdict(list)
47
48 # Convert attribute:value pairs into a dictionary where
49 # attributes are keys and values are grouped in lists
50 for attr_name, val_slug in filter_value:
51 if attr_name not in attributes_map:
52 raise ValueError("Unknown attribute name: %r" % (attr_name,))
53 attr_pk = attributes_map[attr_name]
54 attr_val_pk = values_map[attr_name].get(val_slug, val_slug)
55 queries[attr_pk].append(attr_val_pk)
56
57 return queries
58
59
60 def filter_products_by_attributes(qs, filter_value):
61 queries = _clean_product_attributes_filter_input(filter_value)
62 return filter_products_by_attributes_values(qs, queries)
63
64
65 def filter_products_by_price(qs, price_lte=None, price_gte=None):
66 if price_lte:
67 qs = qs.filter(price_amount__lte=price_lte)
68 if price_gte:
69 qs = qs.filter(price_amount__gte=price_gte)
70 return qs
71
72
73 def filter_products_by_minimal_price(
74 qs, minimal_price_lte=None, minimal_price_gte=None
75 ):
76 if minimal_price_lte:
77 qs = qs.filter(minimal_variant_price_amount__lte=minimal_price_lte)
78 if minimal_price_gte:
79 qs = qs.filter(minimal_variant_price_amount__gte=minimal_price_gte)
80 return qs
81
82
83 def filter_products_by_categories(qs, categories):
84 categories = [
85 category.get_descendants(include_self=True) for category in categories
86 ]
87 ids = {category.id for tree in categories for category in tree}
88 return qs.filter(category__in=ids)
89
90
91 def filter_products_by_collections(qs, collections):
92 return qs.filter(collections__in=collections)
93
94
95 def sort_qs(qs, sort_by):
96 if sort_by:
97 qs = qs.order_by(sort_by["direction"] + sort_by["field"])
98 return qs
99
100
101 def filter_products_by_stock_availability(qs, stock_availability):
102 qs = qs.annotate(total_quantity=Sum("variants__quantity"))
103 if stock_availability == StockAvailability.IN_STOCK:
104 qs = qs.filter(total_quantity__gt=0)
105 elif stock_availability == StockAvailability.OUT_OF_STOCK:
106 qs = qs.filter(total_quantity=0)
107 return qs
108
109
110 def filter_attributes(qs, _, value):
111 if value:
112 value = [(v["slug"], v["value"]) for v in value]
113 qs = filter_products_by_attributes(qs, value)
114 return qs
115
116
117 def filter_categories(qs, _, value):
118 if value:
119 categories = get_nodes(value, types.Category)
120 qs = filter_products_by_categories(qs, categories)
121 return qs
122
123
124 def filter_collections(qs, _, value):
125 if value:
126 collections = get_nodes(value, types.Collection)
127 qs = filter_products_by_collections(qs, collections)
128 return qs
129
130
131 def filter_price(qs, _, value):
132 qs = filter_products_by_price(
133 qs, price_lte=value.get("lte"), price_gte=value.get("gte")
134 )
135 return qs
136
137
138 def filter_minimal_price(qs, _, value):
139 qs = filter_products_by_minimal_price(
140 qs, minimal_price_lte=value.get("lte"), minimal_price_gte=value.get("gte")
141 )
142 return qs
143
144
145 def filter_stock_availability(qs, _, value):
146 if value:
147 qs = filter_products_by_stock_availability(qs, value)
148 return qs
149
150
151 def filter_search(qs, _, value):
152 if value:
153 search = picker.pick_backend()
154 qs &= search(value).distinct()
155 return qs
156
157
158 def filter_collection_publish(qs, _, value):
159 if value == CollectionPublished.PUBLISHED:
160 qs = qs.filter(is_published=True)
161 elif value == CollectionPublished.HIDDEN:
162 qs = qs.filter(is_published=False)
163 return qs
164
165
166 def filter_product_type_configurable(qs, _, value):
167 if value == ProductTypeConfigurable.CONFIGURABLE:
168 qs = qs.filter(has_variants=True)
169 elif value == ProductTypeConfigurable.SIMPLE:
170 qs = qs.filter(has_variants=False)
171 return qs
172
173
174 def filter_product_type(qs, _, value):
175 if value == ProductTypeEnum.DIGITAL:
176 qs = qs.filter(is_digital=True)
177 elif value == ProductTypeEnum.SHIPPABLE:
178 qs = qs.filter(is_shipping_required=True)
179 return qs
180
181
182 def filter_attributes_by_product_types(qs, field, value):
183 if not value:
184 return qs
185
186 if field == "in_category":
187 category_id = from_global_id_strict_type(
188 value, only_type="Category", field=field
189 )
190 category = Category.objects.filter(pk=category_id).first()
191
192 if category is None:
193 return qs.none()
194
195 tree = category.get_descendants(include_self=True)
196 product_qs = Product.objects.filter(category__in=tree)
197
198 elif field == "in_collection":
199 collection_id = from_global_id_strict_type(
200 value, only_type="Collection", field=field
201 )
202 product_qs = Product.objects.filter(collections__id=collection_id)
203
204 else:
205 raise NotImplementedError(f"Filtering by {field} is unsupported")
206
207 product_types = set(product_qs.values_list("product_type_id", flat=True))
208 return qs.filter(
209 Q(product_types__in=product_types) | Q(product_variant_types__in=product_types)
210 )
211
212
213 class ProductFilter(django_filters.FilterSet):
214 is_published = django_filters.BooleanFilter()
215 collections = GlobalIDMultipleChoiceFilter(method=filter_collections)
216 categories = GlobalIDMultipleChoiceFilter(method=filter_categories)
217 price = ObjectTypeFilter(
218 input_class=PriceRangeInput, method=filter_price, field_name="price_amount"
219 )
220 minimal_price = ObjectTypeFilter(
221 input_class=PriceRangeInput,
222 method=filter_minimal_price,
223 field_name="minimal_price_amount",
224 )
225 attributes = ListObjectTypeFilter(
226 input_class=AttributeInput, method=filter_attributes
227 )
228 stock_availability = EnumFilter(
229 input_class=StockAvailability, method=filter_stock_availability
230 )
231 product_type = GlobalIDFilter()
232 search = django_filters.CharFilter(method=filter_search)
233
234 class Meta:
235 model = Product
236 fields = [
237 "is_published",
238 "collections",
239 "categories",
240 "price",
241 "attributes",
242 "stock_availability",
243 "product_type",
244 "search",
245 ]
246
247
248 class CollectionFilter(django_filters.FilterSet):
249 published = EnumFilter(
250 input_class=CollectionPublished, method=filter_collection_publish
251 )
252 search = django_filters.CharFilter(
253 method=filter_fields_containing_value("slug", "name")
254 )
255
256 class Meta:
257 model = Collection
258 fields = ["published", "search"]
259
260
261 class CategoryFilter(django_filters.FilterSet):
262 search = django_filters.CharFilter(
263 method=filter_fields_containing_value("slug", "name", "description")
264 )
265
266 class Meta:
267 model = Category
268 fields = ["search"]
269
270
271 class ProductTypeFilter(django_filters.FilterSet):
272 search = django_filters.CharFilter(method=filter_fields_containing_value("name"))
273
274 configurable = EnumFilter(
275 input_class=ProductTypeConfigurable, method=filter_product_type_configurable
276 )
277
278 product_type = EnumFilter(input_class=ProductTypeEnum, method=filter_product_type)
279
280 class Meta:
281 model = ProductType
282 fields = ["search", "configurable", "product_type"]
283
284
285 class AttributeFilter(django_filters.FilterSet):
286 # Search by attribute name and slug
287 search = django_filters.CharFilter(
288 method=filter_fields_containing_value("slug", "name")
289 )
290 ids = GlobalIDMultipleChoiceFilter(field_name="id")
291
292 in_collection = GlobalIDFilter(method=filter_attributes_by_product_types)
293 in_category = GlobalIDFilter(method=filter_attributes_by_product_types)
294
295 class Meta:
296 model = Attribute
297 fields = [
298 "value_required",
299 "is_variant_only",
300 "visible_in_storefront",
301 "filterable_in_storefront",
302 "filterable_in_dashboard",
303 "available_in_grid",
304 ]
305
306
307 class ProductFilterInput(FilterInputObjectType):
308 class Meta:
309 filterset_class = ProductFilter
310
311
312 class CollectionFilterInput(FilterInputObjectType):
313 class Meta:
314 filterset_class = CollectionFilter
315
316
317 class CategoryFilterInput(FilterInputObjectType):
318 class Meta:
319 filterset_class = CategoryFilter
320
321
322 class ProductTypeFilterInput(FilterInputObjectType):
323 class Meta:
324 filterset_class = ProductTypeFilter
325
326
327 class AttributeFilterInput(FilterInputObjectType):
328 class Meta:
329 filterset_class = AttributeFilter
330
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/saleor/graphql/product/filters.py b/saleor/graphql/product/filters.py
--- a/saleor/graphql/product/filters.py
+++ b/saleor/graphql/product/filters.py
@@ -51,7 +51,7 @@
if attr_name not in attributes_map:
raise ValueError("Unknown attribute name: %r" % (attr_name,))
attr_pk = attributes_map[attr_name]
- attr_val_pk = values_map[attr_name].get(val_slug, val_slug)
+ attr_val_pk = values_map[attr_name].get(val_slug)
queries[attr_pk].append(attr_val_pk)
return queries
|
{"golden_diff": "diff --git a/saleor/graphql/product/filters.py b/saleor/graphql/product/filters.py\n--- a/saleor/graphql/product/filters.py\n+++ b/saleor/graphql/product/filters.py\n@@ -51,7 +51,7 @@\n if attr_name not in attributes_map:\n raise ValueError(\"Unknown attribute name: %r\" % (attr_name,))\n attr_pk = attributes_map[attr_name]\n- attr_val_pk = values_map[attr_name].get(val_slug, val_slug)\n+ attr_val_pk = values_map[attr_name].get(val_slug)\n queries[attr_pk].append(attr_val_pk)\n \n return queries\n", "issue": "Problem with filtering products\n### What I'm trying to achieve\r\nI want to filter out products from grocerries category but I receive results even when I should not.\r\n\r\n### Steps to reproduce the problem\r\nIn groceries category there are two beers that have `abv` attribute with values: 51(5,1% alc) and 67(6,75 alc) but when filter with `abv` value set to 52 we receive `Carrot juice` \r\n\r\nCorrect query:\r\n```\r\nquery{\t\r\n products(first: 100, attributes: [\"abv:51\"]){\r\n \tedges {\r\n node {\r\n id\r\n name\r\n }\r\n }\r\n }\r\n}\r\n```\r\nand correct result:\r\n```\r\n{\r\n \"data\": {\r\n \"products\": {\r\n \"edges\": [\r\n {\r\n \"node\": {\r\n \"id\": \"UHJvZHVjdDo4Mw==\",\r\n \"name\": \"Seaman Lager\"\r\n }\r\n }\r\n ]\r\n }\r\n }\r\n}\r\n```\r\n\r\nQuery with invalid value(abv set to 52):\r\n```\r\nquery{\t\r\n products(first: 100, attributes: [\"abv:52\"]){\r\n \tedges {\r\n node {\r\n id\r\n name\r\n }\r\n }\r\n }\r\n}\r\n```\r\ninvalid result:\r\n```\r\n{\r\n \"data\": {\r\n \"products\": {\r\n \"edges\": [\r\n {\r\n \"node\": {\r\n \"id\": \"UHJvZHVjdDo3Mw==\",\r\n \"name\": \"Carrot Juice\"\r\n }\r\n }\r\n ]\r\n }\r\n }\r\n}\r\n```\r\n\r\n### What I expected to happen\r\nWe should receive empty product list if passing abv value 52.\n", "before_files": [{"content": "from collections import defaultdict\n\nimport django_filters\nfrom django.db.models import Q, Sum\nfrom graphene_django.filter import GlobalIDFilter, GlobalIDMultipleChoiceFilter\n\nfrom ...product.filters import (\n T_PRODUCT_FILTER_QUERIES,\n filter_products_by_attributes_values,\n)\nfrom ...product.models import Attribute, Category, Collection, Product, ProductType\nfrom ...search.backends import picker\nfrom ..core.filters import EnumFilter, ListObjectTypeFilter, ObjectTypeFilter\nfrom ..core.types import FilterInputObjectType\nfrom ..core.types.common import PriceRangeInput\nfrom ..core.utils import from_global_id_strict_type\nfrom ..utils import filter_by_query_param, get_nodes\nfrom . import types\nfrom .enums import (\n CollectionPublished,\n ProductTypeConfigurable,\n ProductTypeEnum,\n StockAvailability,\n)\nfrom .types.attributes import AttributeInput\n\n\ndef filter_fields_containing_value(*search_fields: str):\n \"\"\"Create a icontains filters through given fields on a given query set object.\"\"\"\n\n def _filter_qs(qs, _, value):\n if value:\n qs = filter_by_query_param(qs, value, search_fields)\n return qs\n\n return _filter_qs\n\n\ndef _clean_product_attributes_filter_input(filter_value) -> T_PRODUCT_FILTER_QUERIES:\n attributes = Attribute.objects.prefetch_related(\"values\")\n attributes_map = {attribute.slug: attribute.pk for attribute in attributes}\n values_map = {\n attr.slug: {value.slug: value.pk for value in attr.values.all()}\n for attr in attributes\n }\n queries = defaultdict(list)\n\n # Convert attribute:value pairs into a dictionary where\n # attributes are keys and values are grouped in lists\n for attr_name, val_slug in filter_value:\n if attr_name not in attributes_map:\n raise ValueError(\"Unknown attribute name: %r\" % (attr_name,))\n attr_pk = attributes_map[attr_name]\n attr_val_pk = values_map[attr_name].get(val_slug, val_slug)\n queries[attr_pk].append(attr_val_pk)\n\n return queries\n\n\ndef filter_products_by_attributes(qs, filter_value):\n queries = _clean_product_attributes_filter_input(filter_value)\n return filter_products_by_attributes_values(qs, queries)\n\n\ndef filter_products_by_price(qs, price_lte=None, price_gte=None):\n if price_lte:\n qs = qs.filter(price_amount__lte=price_lte)\n if price_gte:\n qs = qs.filter(price_amount__gte=price_gte)\n return qs\n\n\ndef filter_products_by_minimal_price(\n qs, minimal_price_lte=None, minimal_price_gte=None\n):\n if minimal_price_lte:\n qs = qs.filter(minimal_variant_price_amount__lte=minimal_price_lte)\n if minimal_price_gte:\n qs = qs.filter(minimal_variant_price_amount__gte=minimal_price_gte)\n return qs\n\n\ndef filter_products_by_categories(qs, categories):\n categories = [\n category.get_descendants(include_self=True) for category in categories\n ]\n ids = {category.id for tree in categories for category in tree}\n return qs.filter(category__in=ids)\n\n\ndef filter_products_by_collections(qs, collections):\n return qs.filter(collections__in=collections)\n\n\ndef sort_qs(qs, sort_by):\n if sort_by:\n qs = qs.order_by(sort_by[\"direction\"] + sort_by[\"field\"])\n return qs\n\n\ndef filter_products_by_stock_availability(qs, stock_availability):\n qs = qs.annotate(total_quantity=Sum(\"variants__quantity\"))\n if stock_availability == StockAvailability.IN_STOCK:\n qs = qs.filter(total_quantity__gt=0)\n elif stock_availability == StockAvailability.OUT_OF_STOCK:\n qs = qs.filter(total_quantity=0)\n return qs\n\n\ndef filter_attributes(qs, _, value):\n if value:\n value = [(v[\"slug\"], v[\"value\"]) for v in value]\n qs = filter_products_by_attributes(qs, value)\n return qs\n\n\ndef filter_categories(qs, _, value):\n if value:\n categories = get_nodes(value, types.Category)\n qs = filter_products_by_categories(qs, categories)\n return qs\n\n\ndef filter_collections(qs, _, value):\n if value:\n collections = get_nodes(value, types.Collection)\n qs = filter_products_by_collections(qs, collections)\n return qs\n\n\ndef filter_price(qs, _, value):\n qs = filter_products_by_price(\n qs, price_lte=value.get(\"lte\"), price_gte=value.get(\"gte\")\n )\n return qs\n\n\ndef filter_minimal_price(qs, _, value):\n qs = filter_products_by_minimal_price(\n qs, minimal_price_lte=value.get(\"lte\"), minimal_price_gte=value.get(\"gte\")\n )\n return qs\n\n\ndef filter_stock_availability(qs, _, value):\n if value:\n qs = filter_products_by_stock_availability(qs, value)\n return qs\n\n\ndef filter_search(qs, _, value):\n if value:\n search = picker.pick_backend()\n qs &= search(value).distinct()\n return qs\n\n\ndef filter_collection_publish(qs, _, value):\n if value == CollectionPublished.PUBLISHED:\n qs = qs.filter(is_published=True)\n elif value == CollectionPublished.HIDDEN:\n qs = qs.filter(is_published=False)\n return qs\n\n\ndef filter_product_type_configurable(qs, _, value):\n if value == ProductTypeConfigurable.CONFIGURABLE:\n qs = qs.filter(has_variants=True)\n elif value == ProductTypeConfigurable.SIMPLE:\n qs = qs.filter(has_variants=False)\n return qs\n\n\ndef filter_product_type(qs, _, value):\n if value == ProductTypeEnum.DIGITAL:\n qs = qs.filter(is_digital=True)\n elif value == ProductTypeEnum.SHIPPABLE:\n qs = qs.filter(is_shipping_required=True)\n return qs\n\n\ndef filter_attributes_by_product_types(qs, field, value):\n if not value:\n return qs\n\n if field == \"in_category\":\n category_id = from_global_id_strict_type(\n value, only_type=\"Category\", field=field\n )\n category = Category.objects.filter(pk=category_id).first()\n\n if category is None:\n return qs.none()\n\n tree = category.get_descendants(include_self=True)\n product_qs = Product.objects.filter(category__in=tree)\n\n elif field == \"in_collection\":\n collection_id = from_global_id_strict_type(\n value, only_type=\"Collection\", field=field\n )\n product_qs = Product.objects.filter(collections__id=collection_id)\n\n else:\n raise NotImplementedError(f\"Filtering by {field} is unsupported\")\n\n product_types = set(product_qs.values_list(\"product_type_id\", flat=True))\n return qs.filter(\n Q(product_types__in=product_types) | Q(product_variant_types__in=product_types)\n )\n\n\nclass ProductFilter(django_filters.FilterSet):\n is_published = django_filters.BooleanFilter()\n collections = GlobalIDMultipleChoiceFilter(method=filter_collections)\n categories = GlobalIDMultipleChoiceFilter(method=filter_categories)\n price = ObjectTypeFilter(\n input_class=PriceRangeInput, method=filter_price, field_name=\"price_amount\"\n )\n minimal_price = ObjectTypeFilter(\n input_class=PriceRangeInput,\n method=filter_minimal_price,\n field_name=\"minimal_price_amount\",\n )\n attributes = ListObjectTypeFilter(\n input_class=AttributeInput, method=filter_attributes\n )\n stock_availability = EnumFilter(\n input_class=StockAvailability, method=filter_stock_availability\n )\n product_type = GlobalIDFilter()\n search = django_filters.CharFilter(method=filter_search)\n\n class Meta:\n model = Product\n fields = [\n \"is_published\",\n \"collections\",\n \"categories\",\n \"price\",\n \"attributes\",\n \"stock_availability\",\n \"product_type\",\n \"search\",\n ]\n\n\nclass CollectionFilter(django_filters.FilterSet):\n published = EnumFilter(\n input_class=CollectionPublished, method=filter_collection_publish\n )\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\")\n )\n\n class Meta:\n model = Collection\n fields = [\"published\", \"search\"]\n\n\nclass CategoryFilter(django_filters.FilterSet):\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\", \"description\")\n )\n\n class Meta:\n model = Category\n fields = [\"search\"]\n\n\nclass ProductTypeFilter(django_filters.FilterSet):\n search = django_filters.CharFilter(method=filter_fields_containing_value(\"name\"))\n\n configurable = EnumFilter(\n input_class=ProductTypeConfigurable, method=filter_product_type_configurable\n )\n\n product_type = EnumFilter(input_class=ProductTypeEnum, method=filter_product_type)\n\n class Meta:\n model = ProductType\n fields = [\"search\", \"configurable\", \"product_type\"]\n\n\nclass AttributeFilter(django_filters.FilterSet):\n # Search by attribute name and slug\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\")\n )\n ids = GlobalIDMultipleChoiceFilter(field_name=\"id\")\n\n in_collection = GlobalIDFilter(method=filter_attributes_by_product_types)\n in_category = GlobalIDFilter(method=filter_attributes_by_product_types)\n\n class Meta:\n model = Attribute\n fields = [\n \"value_required\",\n \"is_variant_only\",\n \"visible_in_storefront\",\n \"filterable_in_storefront\",\n \"filterable_in_dashboard\",\n \"available_in_grid\",\n ]\n\n\nclass ProductFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = ProductFilter\n\n\nclass CollectionFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = CollectionFilter\n\n\nclass CategoryFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = CategoryFilter\n\n\nclass ProductTypeFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = ProductTypeFilter\n\n\nclass AttributeFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = AttributeFilter\n", "path": "saleor/graphql/product/filters.py"}], "after_files": [{"content": "from collections import defaultdict\n\nimport django_filters\nfrom django.db.models import Q, Sum\nfrom graphene_django.filter import GlobalIDFilter, GlobalIDMultipleChoiceFilter\n\nfrom ...product.filters import (\n T_PRODUCT_FILTER_QUERIES,\n filter_products_by_attributes_values,\n)\nfrom ...product.models import Attribute, Category, Collection, Product, ProductType\nfrom ...search.backends import picker\nfrom ..core.filters import EnumFilter, ListObjectTypeFilter, ObjectTypeFilter\nfrom ..core.types import FilterInputObjectType\nfrom ..core.types.common import PriceRangeInput\nfrom ..core.utils import from_global_id_strict_type\nfrom ..utils import filter_by_query_param, get_nodes\nfrom . import types\nfrom .enums import (\n CollectionPublished,\n ProductTypeConfigurable,\n ProductTypeEnum,\n StockAvailability,\n)\nfrom .types.attributes import AttributeInput\n\n\ndef filter_fields_containing_value(*search_fields: str):\n \"\"\"Create a icontains filters through given fields on a given query set object.\"\"\"\n\n def _filter_qs(qs, _, value):\n if value:\n qs = filter_by_query_param(qs, value, search_fields)\n return qs\n\n return _filter_qs\n\n\ndef _clean_product_attributes_filter_input(filter_value) -> T_PRODUCT_FILTER_QUERIES:\n attributes = Attribute.objects.prefetch_related(\"values\")\n attributes_map = {attribute.slug: attribute.pk for attribute in attributes}\n values_map = {\n attr.slug: {value.slug: value.pk for value in attr.values.all()}\n for attr in attributes\n }\n queries = defaultdict(list)\n\n # Convert attribute:value pairs into a dictionary where\n # attributes are keys and values are grouped in lists\n for attr_name, val_slug in filter_value:\n if attr_name not in attributes_map:\n raise ValueError(\"Unknown attribute name: %r\" % (attr_name,))\n attr_pk = attributes_map[attr_name]\n attr_val_pk = values_map[attr_name].get(val_slug)\n queries[attr_pk].append(attr_val_pk)\n\n return queries\n\n\ndef filter_products_by_attributes(qs, filter_value):\n queries = _clean_product_attributes_filter_input(filter_value)\n return filter_products_by_attributes_values(qs, queries)\n\n\ndef filter_products_by_price(qs, price_lte=None, price_gte=None):\n if price_lte:\n qs = qs.filter(price_amount__lte=price_lte)\n if price_gte:\n qs = qs.filter(price_amount__gte=price_gte)\n return qs\n\n\ndef filter_products_by_minimal_price(\n qs, minimal_price_lte=None, minimal_price_gte=None\n):\n if minimal_price_lte:\n qs = qs.filter(minimal_variant_price_amount__lte=minimal_price_lte)\n if minimal_price_gte:\n qs = qs.filter(minimal_variant_price_amount__gte=minimal_price_gte)\n return qs\n\n\ndef filter_products_by_categories(qs, categories):\n categories = [\n category.get_descendants(include_self=True) for category in categories\n ]\n ids = {category.id for tree in categories for category in tree}\n return qs.filter(category__in=ids)\n\n\ndef filter_products_by_collections(qs, collections):\n return qs.filter(collections__in=collections)\n\n\ndef sort_qs(qs, sort_by):\n if sort_by:\n qs = qs.order_by(sort_by[\"direction\"] + sort_by[\"field\"])\n return qs\n\n\ndef filter_products_by_stock_availability(qs, stock_availability):\n qs = qs.annotate(total_quantity=Sum(\"variants__quantity\"))\n if stock_availability == StockAvailability.IN_STOCK:\n qs = qs.filter(total_quantity__gt=0)\n elif stock_availability == StockAvailability.OUT_OF_STOCK:\n qs = qs.filter(total_quantity=0)\n return qs\n\n\ndef filter_attributes(qs, _, value):\n if value:\n value = [(v[\"slug\"], v[\"value\"]) for v in value]\n qs = filter_products_by_attributes(qs, value)\n return qs\n\n\ndef filter_categories(qs, _, value):\n if value:\n categories = get_nodes(value, types.Category)\n qs = filter_products_by_categories(qs, categories)\n return qs\n\n\ndef filter_collections(qs, _, value):\n if value:\n collections = get_nodes(value, types.Collection)\n qs = filter_products_by_collections(qs, collections)\n return qs\n\n\ndef filter_price(qs, _, value):\n qs = filter_products_by_price(\n qs, price_lte=value.get(\"lte\"), price_gte=value.get(\"gte\")\n )\n return qs\n\n\ndef filter_minimal_price(qs, _, value):\n qs = filter_products_by_minimal_price(\n qs, minimal_price_lte=value.get(\"lte\"), minimal_price_gte=value.get(\"gte\")\n )\n return qs\n\n\ndef filter_stock_availability(qs, _, value):\n if value:\n qs = filter_products_by_stock_availability(qs, value)\n return qs\n\n\ndef filter_search(qs, _, value):\n if value:\n search = picker.pick_backend()\n qs &= search(value).distinct()\n return qs\n\n\ndef filter_collection_publish(qs, _, value):\n if value == CollectionPublished.PUBLISHED:\n qs = qs.filter(is_published=True)\n elif value == CollectionPublished.HIDDEN:\n qs = qs.filter(is_published=False)\n return qs\n\n\ndef filter_product_type_configurable(qs, _, value):\n if value == ProductTypeConfigurable.CONFIGURABLE:\n qs = qs.filter(has_variants=True)\n elif value == ProductTypeConfigurable.SIMPLE:\n qs = qs.filter(has_variants=False)\n return qs\n\n\ndef filter_product_type(qs, _, value):\n if value == ProductTypeEnum.DIGITAL:\n qs = qs.filter(is_digital=True)\n elif value == ProductTypeEnum.SHIPPABLE:\n qs = qs.filter(is_shipping_required=True)\n return qs\n\n\ndef filter_attributes_by_product_types(qs, field, value):\n if not value:\n return qs\n\n if field == \"in_category\":\n category_id = from_global_id_strict_type(\n value, only_type=\"Category\", field=field\n )\n category = Category.objects.filter(pk=category_id).first()\n\n if category is None:\n return qs.none()\n\n tree = category.get_descendants(include_self=True)\n product_qs = Product.objects.filter(category__in=tree)\n\n elif field == \"in_collection\":\n collection_id = from_global_id_strict_type(\n value, only_type=\"Collection\", field=field\n )\n product_qs = Product.objects.filter(collections__id=collection_id)\n\n else:\n raise NotImplementedError(f\"Filtering by {field} is unsupported\")\n\n product_types = set(product_qs.values_list(\"product_type_id\", flat=True))\n return qs.filter(\n Q(product_types__in=product_types) | Q(product_variant_types__in=product_types)\n )\n\n\nclass ProductFilter(django_filters.FilterSet):\n is_published = django_filters.BooleanFilter()\n collections = GlobalIDMultipleChoiceFilter(method=filter_collections)\n categories = GlobalIDMultipleChoiceFilter(method=filter_categories)\n price = ObjectTypeFilter(\n input_class=PriceRangeInput, method=filter_price, field_name=\"price_amount\"\n )\n minimal_price = ObjectTypeFilter(\n input_class=PriceRangeInput,\n method=filter_minimal_price,\n field_name=\"minimal_price_amount\",\n )\n attributes = ListObjectTypeFilter(\n input_class=AttributeInput, method=filter_attributes\n )\n stock_availability = EnumFilter(\n input_class=StockAvailability, method=filter_stock_availability\n )\n product_type = GlobalIDFilter()\n search = django_filters.CharFilter(method=filter_search)\n\n class Meta:\n model = Product\n fields = [\n \"is_published\",\n \"collections\",\n \"categories\",\n \"price\",\n \"attributes\",\n \"stock_availability\",\n \"product_type\",\n \"search\",\n ]\n\n\nclass CollectionFilter(django_filters.FilterSet):\n published = EnumFilter(\n input_class=CollectionPublished, method=filter_collection_publish\n )\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\")\n )\n\n class Meta:\n model = Collection\n fields = [\"published\", \"search\"]\n\n\nclass CategoryFilter(django_filters.FilterSet):\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\", \"description\")\n )\n\n class Meta:\n model = Category\n fields = [\"search\"]\n\n\nclass ProductTypeFilter(django_filters.FilterSet):\n search = django_filters.CharFilter(method=filter_fields_containing_value(\"name\"))\n\n configurable = EnumFilter(\n input_class=ProductTypeConfigurable, method=filter_product_type_configurable\n )\n\n product_type = EnumFilter(input_class=ProductTypeEnum, method=filter_product_type)\n\n class Meta:\n model = ProductType\n fields = [\"search\", \"configurable\", \"product_type\"]\n\n\nclass AttributeFilter(django_filters.FilterSet):\n # Search by attribute name and slug\n search = django_filters.CharFilter(\n method=filter_fields_containing_value(\"slug\", \"name\")\n )\n ids = GlobalIDMultipleChoiceFilter(field_name=\"id\")\n\n in_collection = GlobalIDFilter(method=filter_attributes_by_product_types)\n in_category = GlobalIDFilter(method=filter_attributes_by_product_types)\n\n class Meta:\n model = Attribute\n fields = [\n \"value_required\",\n \"is_variant_only\",\n \"visible_in_storefront\",\n \"filterable_in_storefront\",\n \"filterable_in_dashboard\",\n \"available_in_grid\",\n ]\n\n\nclass ProductFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = ProductFilter\n\n\nclass CollectionFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = CollectionFilter\n\n\nclass CategoryFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = CategoryFilter\n\n\nclass ProductTypeFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = ProductTypeFilter\n\n\nclass AttributeFilterInput(FilterInputObjectType):\n class Meta:\n filterset_class = AttributeFilter\n", "path": "saleor/graphql/product/filters.py"}]}
| 3,711 | 138 |
gh_patches_debug_2263
|
rasdani/github-patches
|
git_diff
|
hpcaitech__ColossalAI-5321
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[tensor] fix some unittests
[tensor] fix some unittests
[tensor] fix some unittests
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `colossalai/accelerator/npu_accelerator.py`
Content:
```
1 #!/usr/bin/env python
2
3 from typing import Any, Callable, Dict, List, Optional, Tuple, Union
4
5 import torch
6 import torch.distributed as dist
7
8 from .base_accelerator import BaseAccelerator
9
10 try:
11 import torch_npu # noqa
12 except ImportError:
13 pass
14
15
16 __all__ = ["NpuAccelerator"]
17
18
19 class NpuAccelerator(BaseAccelerator):
20 """
21 Accelerator class for Huawei NPU devices.
22 """
23
24 def __init__(self):
25 super().__init__(name="npu", communication_backend="hccl", is_synchronous=False)
26
27 # =======================
28 # device APIs
29 # =======================
30 def get_version(self) -> str:
31 """
32 Return the version of the accelerator which torch is built against.
33 """
34 return torch.version.npu
35
36 def get_current_device(self) -> torch.device:
37 """
38 Return the current device.
39 """
40 return torch.device(f"npu:{torch.npu.current_device()}")
41
42 def current_device(self) -> int:
43 """
44 Return the current device index.
45 """
46 return torch.npu.current_device()
47
48 def set_device(self, device: Optional[Union[torch.device, int]] = None) -> None:
49 """
50 Bind the current process to a device.
51 """
52 if device is None:
53 if not dist.is_initialized():
54 raise RuntimeError("Cannot get current device when distributed is not initialized.")
55 device = dist.get_rank() % self.device_count()
56 torch.npu.set_device(device)
57
58 def get_device_name(self, device: Union[torch.device, int]) -> str:
59 """
60 Return the name of the device.
61 """
62 return torch.npu.get_device_name(device)
63
64 def synchronize(self, device: Union[torch.device, int] = None):
65 """
66 Synchronize the current process.
67 """
68 torch.npu.synchronize(device)
69
70 def is_available(self):
71 """
72 Check if the accelerator is available.
73 """
74 return torch.npu.is_available()
75
76 def device_count(self):
77 """
78 Return the number of devices on the machine.
79 """
80 return torch.npu.device_count()
81
82 def get_device_capability(self, device=None) -> Tuple[int, int]:
83 """
84 Gets the npu capability of a device.
85 """
86 return torch.npu.get_device_capability(device)
87
88 def get_device_name(self, device=None) -> str:
89 """
90 Gets the name of a device.
91 """
92 return torch.npu.get_device_name(device)
93
94 def get_device_properties(self, device):
95 """
96 Gets the properties of a device.
97 """
98 return torch.npu.get_device_properties(device)
99
100 def utilization(self, device=None) -> int:
101 """
102 Returns the percent of time over the past sample period during which one or more kernels was executing on the GPU as given by nvidia-smi
103 """
104 return torch.npu.utilization(device)
105
106 # =======================
107 # random number generator APIs
108 # =======================
109 def get_rng_state(self, device="npu") -> torch.Tensor:
110 """
111 Returns the random number generator state of the specified GPU as a ByteTensor.
112 """
113 return torch.npu.get_rng_state(device)
114
115 def get_rng_state_all(self) -> List[torch.Tensor]:
116 """
117 Returns a list of ByteTensor representing the random number states of all devices.
118 """
119 return torch.npu.get_rng_state_all()
120
121 def set_rng_state(self, new_state: torch.ByteTensor, device: str = "npu") -> None:
122 """
123 Sets the random number generator state of the specified GPU.
124 """
125 torch.npu.set_rng_state(new_state, device)
126
127 def set_rng_state_all(self, new_states: List[torch.ByteTensor]) -> None:
128 """
129 Sets the random number generator state of all devices.
130 """
131 torch.npu.set_rng_state_all(new_states)
132
133 def manual_seed(self, seed: int) -> None:
134 """
135 Sets the seed for generating random numbers for the current GPU.
136 """
137 torch.npu.manual_seed(seed)
138
139 def manual_seed_all(self, seed: int) -> None:
140 """
141 Set the random seed for the all processes.
142 """
143 torch.npu.manual_seed_all(seed)
144
145 def seed(self) -> None:
146 """
147 Sets the seed for generating random numbers to a random number for the current GPU.
148 """
149 torch.npu.seed()
150
151 def seed_all(self) -> None:
152 """
153 Sets the seed for generating random numbers to a random number on all GPUs.
154 """
155 torch.npu.seed_all()
156
157 def initial_seed(self) -> int:
158 """
159 Returns the current random seed of the current GPU.
160 """
161 return torch.npu.initial_seed()
162
163 # =======================
164 # memory management APIs
165 # =======================
166
167 def empty_cache(self) -> None:
168 """
169 Releases all unoccupied cached memory currently held by the caching allocator so that those can be used in other GPU application and visible in nvidia-smi.
170 """
171 torch.npu.empty_cache()
172
173 def memory_stats(self, device=None) -> Dict[str, Any]:
174 """
175 Returns a dictionary of npu memory allocator statistics for a given device.
176 """
177 return torch.npu.memory_stats(device=device)
178
179 def memory_summary(self, device=None, abbreviated=False) -> str:
180 """
181 Returns a human-readable printout of the current memory allocator statistics for a given device.
182 """
183 return torch.npu.memory_summary(device=device, abbreviated=abbreviated)
184
185 def memory_snapshot(self):
186 """
187 Returns a snapshot of the npu memory allocator state across all devices.
188 """
189 return torch.npu.memory_snapshot()
190
191 def memory_allocated(self, device=None) -> int:
192 """
193 Returns the current GPU memory occupied by tensors in bytes for a given device.
194 """
195 return torch.npu.memory_allocated(device=device)
196
197 def max_memory_allocated(self, device=None) -> int:
198 """
199 Returns the maximum GPU memory occupied by tensors in bytes for a given device.
200 """
201 return torch.npu.max_memory_allocated(device=device)
202
203 def reset_max_memory_allocated(self, device=None) -> None:
204 """
205 Resets the starting point in tracking maximum GPU memory occupied by tensors for a given device.
206 """
207 torch.npu.reset_max_memory_allocated(device=device)
208
209 def reset_max_memory_cached(self, device=None) -> None:
210 """
211 Resets the starting point in tracking maximum GPU memory managed by the caching allocator for a given device.
212 """
213 torch.npu.reset_max_memory_cached(device=device)
214
215 def memory_reserved(self, device=None) -> int:
216 """
217 Returns the current GPU memory managed by the caching allocator in bytes for a given device.
218 """
219 return torch.npu.memory_reserved(device=device)
220
221 def max_memory_reserved(self, device=None) -> int:
222 """
223 Returns the maximum GPU memory managed by the caching allocator in bytes for a given device.
224 """
225 return torch.npu.max_memory_reserved(device=device)
226
227 def set_per_process_memory_fraction(self, fraction: float, device=None) -> None:
228 """
229 Set memory fraction for a process.
230 """
231 torch.npu.set_per_process_memory_fraction(fraction, device=device)
232
233 def reset_peak_memory_stats(self, device=None) -> None:
234 """
235 Resets the "peak" stats tracked by the npu memory allocator.
236 """
237 torch.npu.reset_peak_memory_stats(device=device)
238
239 # =======================
240 # streams and events APIs
241 # =======================
242
243 def Stream(self, device=None, priority=0, **kwargs):
244 """
245 A npu stream is a linear sequence of execution that belongs to a specific device, independent from other streams. See npu-semantics for details.
246 """
247 return torch.npu.Stream(device, priority, **kwargs)
248
249 def Event(self, enable_timing: bool = False, blocking: bool = False, interprocess: bool = False):
250 """
251 npu events are synchronization markers that can be used to monitor the device's progress, to accurately measure timing, and to synchronize npu streams.
252 """
253 return torch.npu.Event(enable_timing, blocking, interprocess)
254
255 def current_stream(self, device=None):
256 """
257 Returns the currently selected Stream for a given device.
258 """
259 return torch.npu.current_stream(device)
260
261 def default_stream(self, device=None):
262 """
263 Returns the default Stream for a given device.
264 """
265 return torch.npu.default_stream(device)
266
267 def set_stream(self, stream_):
268 """
269 Sets the current stream.This is a wrapper API to set the stream.
270 """
271 torch.npu.set_stream(stream_)
272
273 def stream(self, stream_):
274 """
275 Wrapper around the Context-manager StreamContext that selects a given stream.
276 """
277 return torch.npu.stream(stream_)
278
279 # =======================
280 # amp APIs
281 # =======================
282 def autocast(
283 self, enabled: bool = True, dtype: torch.dtype = torch.float16, cache_enabled: bool = True
284 ) -> Callable:
285 """
286 Return autocast function
287 """
288 return torch.npu.amp.autocast(enabled=enabled, dtype=dtype, cache_enabled=cache_enabled)
289
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/colossalai/accelerator/npu_accelerator.py b/colossalai/accelerator/npu_accelerator.py
--- a/colossalai/accelerator/npu_accelerator.py
+++ b/colossalai/accelerator/npu_accelerator.py
@@ -31,7 +31,7 @@
"""
Return the version of the accelerator which torch is built against.
"""
- return torch.version.npu
+ return torch.version.cann
def get_current_device(self) -> torch.device:
"""
|
{"golden_diff": "diff --git a/colossalai/accelerator/npu_accelerator.py b/colossalai/accelerator/npu_accelerator.py\n--- a/colossalai/accelerator/npu_accelerator.py\n+++ b/colossalai/accelerator/npu_accelerator.py\n@@ -31,7 +31,7 @@\n \"\"\"\n Return the version of the accelerator which torch is built against.\n \"\"\"\n- return torch.version.npu\n+ return torch.version.cann\n \n def get_current_device(self) -> torch.device:\n \"\"\"\n", "issue": "[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\nfrom typing import Any, Callable, Dict, List, Optional, Tuple, Union\n\nimport torch\nimport torch.distributed as dist\n\nfrom .base_accelerator import BaseAccelerator\n\ntry:\n import torch_npu # noqa\nexcept ImportError:\n pass\n\n\n__all__ = [\"NpuAccelerator\"]\n\n\nclass NpuAccelerator(BaseAccelerator):\n \"\"\"\n Accelerator class for Huawei NPU devices.\n \"\"\"\n\n def __init__(self):\n super().__init__(name=\"npu\", communication_backend=\"hccl\", is_synchronous=False)\n\n # =======================\n # device APIs\n # =======================\n def get_version(self) -> str:\n \"\"\"\n Return the version of the accelerator which torch is built against.\n \"\"\"\n return torch.version.npu\n\n def get_current_device(self) -> torch.device:\n \"\"\"\n Return the current device.\n \"\"\"\n return torch.device(f\"npu:{torch.npu.current_device()}\")\n\n def current_device(self) -> int:\n \"\"\"\n Return the current device index.\n \"\"\"\n return torch.npu.current_device()\n\n def set_device(self, device: Optional[Union[torch.device, int]] = None) -> None:\n \"\"\"\n Bind the current process to a device.\n \"\"\"\n if device is None:\n if not dist.is_initialized():\n raise RuntimeError(\"Cannot get current device when distributed is not initialized.\")\n device = dist.get_rank() % self.device_count()\n torch.npu.set_device(device)\n\n def get_device_name(self, device: Union[torch.device, int]) -> str:\n \"\"\"\n Return the name of the device.\n \"\"\"\n return torch.npu.get_device_name(device)\n\n def synchronize(self, device: Union[torch.device, int] = None):\n \"\"\"\n Synchronize the current process.\n \"\"\"\n torch.npu.synchronize(device)\n\n def is_available(self):\n \"\"\"\n Check if the accelerator is available.\n \"\"\"\n return torch.npu.is_available()\n\n def device_count(self):\n \"\"\"\n Return the number of devices on the machine.\n \"\"\"\n return torch.npu.device_count()\n\n def get_device_capability(self, device=None) -> Tuple[int, int]:\n \"\"\"\n Gets the npu capability of a device.\n \"\"\"\n return torch.npu.get_device_capability(device)\n\n def get_device_name(self, device=None) -> str:\n \"\"\"\n Gets the name of a device.\n \"\"\"\n return torch.npu.get_device_name(device)\n\n def get_device_properties(self, device):\n \"\"\"\n Gets the properties of a device.\n \"\"\"\n return torch.npu.get_device_properties(device)\n\n def utilization(self, device=None) -> int:\n \"\"\"\n Returns the percent of time over the past sample period during which one or more kernels was executing on the GPU as given by nvidia-smi\n \"\"\"\n return torch.npu.utilization(device)\n\n # =======================\n # random number generator APIs\n # =======================\n def get_rng_state(self, device=\"npu\") -> torch.Tensor:\n \"\"\"\n Returns the random number generator state of the specified GPU as a ByteTensor.\n \"\"\"\n return torch.npu.get_rng_state(device)\n\n def get_rng_state_all(self) -> List[torch.Tensor]:\n \"\"\"\n Returns a list of ByteTensor representing the random number states of all devices.\n \"\"\"\n return torch.npu.get_rng_state_all()\n\n def set_rng_state(self, new_state: torch.ByteTensor, device: str = \"npu\") -> None:\n \"\"\"\n Sets the random number generator state of the specified GPU.\n \"\"\"\n torch.npu.set_rng_state(new_state, device)\n\n def set_rng_state_all(self, new_states: List[torch.ByteTensor]) -> None:\n \"\"\"\n Sets the random number generator state of all devices.\n \"\"\"\n torch.npu.set_rng_state_all(new_states)\n\n def manual_seed(self, seed: int) -> None:\n \"\"\"\n Sets the seed for generating random numbers for the current GPU.\n \"\"\"\n torch.npu.manual_seed(seed)\n\n def manual_seed_all(self, seed: int) -> None:\n \"\"\"\n Set the random seed for the all processes.\n \"\"\"\n torch.npu.manual_seed_all(seed)\n\n def seed(self) -> None:\n \"\"\"\n Sets the seed for generating random numbers to a random number for the current GPU.\n \"\"\"\n torch.npu.seed()\n\n def seed_all(self) -> None:\n \"\"\"\n Sets the seed for generating random numbers to a random number on all GPUs.\n \"\"\"\n torch.npu.seed_all()\n\n def initial_seed(self) -> int:\n \"\"\"\n Returns the current random seed of the current GPU.\n \"\"\"\n return torch.npu.initial_seed()\n\n # =======================\n # memory management APIs\n # =======================\n\n def empty_cache(self) -> None:\n \"\"\"\n Releases all unoccupied cached memory currently held by the caching allocator so that those can be used in other GPU application and visible in nvidia-smi.\n \"\"\"\n torch.npu.empty_cache()\n\n def memory_stats(self, device=None) -> Dict[str, Any]:\n \"\"\"\n Returns a dictionary of npu memory allocator statistics for a given device.\n \"\"\"\n return torch.npu.memory_stats(device=device)\n\n def memory_summary(self, device=None, abbreviated=False) -> str:\n \"\"\"\n Returns a human-readable printout of the current memory allocator statistics for a given device.\n \"\"\"\n return torch.npu.memory_summary(device=device, abbreviated=abbreviated)\n\n def memory_snapshot(self):\n \"\"\"\n Returns a snapshot of the npu memory allocator state across all devices.\n \"\"\"\n return torch.npu.memory_snapshot()\n\n def memory_allocated(self, device=None) -> int:\n \"\"\"\n Returns the current GPU memory occupied by tensors in bytes for a given device.\n \"\"\"\n return torch.npu.memory_allocated(device=device)\n\n def max_memory_allocated(self, device=None) -> int:\n \"\"\"\n Returns the maximum GPU memory occupied by tensors in bytes for a given device.\n \"\"\"\n return torch.npu.max_memory_allocated(device=device)\n\n def reset_max_memory_allocated(self, device=None) -> None:\n \"\"\"\n Resets the starting point in tracking maximum GPU memory occupied by tensors for a given device.\n \"\"\"\n torch.npu.reset_max_memory_allocated(device=device)\n\n def reset_max_memory_cached(self, device=None) -> None:\n \"\"\"\n Resets the starting point in tracking maximum GPU memory managed by the caching allocator for a given device.\n \"\"\"\n torch.npu.reset_max_memory_cached(device=device)\n\n def memory_reserved(self, device=None) -> int:\n \"\"\"\n Returns the current GPU memory managed by the caching allocator in bytes for a given device.\n \"\"\"\n return torch.npu.memory_reserved(device=device)\n\n def max_memory_reserved(self, device=None) -> int:\n \"\"\"\n Returns the maximum GPU memory managed by the caching allocator in bytes for a given device.\n \"\"\"\n return torch.npu.max_memory_reserved(device=device)\n\n def set_per_process_memory_fraction(self, fraction: float, device=None) -> None:\n \"\"\"\n Set memory fraction for a process.\n \"\"\"\n torch.npu.set_per_process_memory_fraction(fraction, device=device)\n\n def reset_peak_memory_stats(self, device=None) -> None:\n \"\"\"\n Resets the \"peak\" stats tracked by the npu memory allocator.\n \"\"\"\n torch.npu.reset_peak_memory_stats(device=device)\n\n # =======================\n # streams and events APIs\n # =======================\n\n def Stream(self, device=None, priority=0, **kwargs):\n \"\"\"\n A npu stream is a linear sequence of execution that belongs to a specific device, independent from other streams. See npu-semantics for details.\n \"\"\"\n return torch.npu.Stream(device, priority, **kwargs)\n\n def Event(self, enable_timing: bool = False, blocking: bool = False, interprocess: bool = False):\n \"\"\"\n npu events are synchronization markers that can be used to monitor the device's progress, to accurately measure timing, and to synchronize npu streams.\n \"\"\"\n return torch.npu.Event(enable_timing, blocking, interprocess)\n\n def current_stream(self, device=None):\n \"\"\"\n Returns the currently selected Stream for a given device.\n \"\"\"\n return torch.npu.current_stream(device)\n\n def default_stream(self, device=None):\n \"\"\"\n Returns the default Stream for a given device.\n \"\"\"\n return torch.npu.default_stream(device)\n\n def set_stream(self, stream_):\n \"\"\"\n Sets the current stream.This is a wrapper API to set the stream.\n \"\"\"\n torch.npu.set_stream(stream_)\n\n def stream(self, stream_):\n \"\"\"\n Wrapper around the Context-manager StreamContext that selects a given stream.\n \"\"\"\n return torch.npu.stream(stream_)\n\n # =======================\n # amp APIs\n # =======================\n def autocast(\n self, enabled: bool = True, dtype: torch.dtype = torch.float16, cache_enabled: bool = True\n ) -> Callable:\n \"\"\"\n Return autocast function\n \"\"\"\n return torch.npu.amp.autocast(enabled=enabled, dtype=dtype, cache_enabled=cache_enabled)\n", "path": "colossalai/accelerator/npu_accelerator.py"}], "after_files": [{"content": "#!/usr/bin/env python\n\nfrom typing import Any, Callable, Dict, List, Optional, Tuple, Union\n\nimport torch\nimport torch.distributed as dist\n\nfrom .base_accelerator import BaseAccelerator\n\ntry:\n import torch_npu # noqa\nexcept ImportError:\n pass\n\n\n__all__ = [\"NpuAccelerator\"]\n\n\nclass NpuAccelerator(BaseAccelerator):\n \"\"\"\n Accelerator class for Huawei NPU devices.\n \"\"\"\n\n def __init__(self):\n super().__init__(name=\"npu\", communication_backend=\"hccl\", is_synchronous=False)\n\n # =======================\n # device APIs\n # =======================\n def get_version(self) -> str:\n \"\"\"\n Return the version of the accelerator which torch is built against.\n \"\"\"\n return torch.version.cann\n\n def get_current_device(self) -> torch.device:\n \"\"\"\n Return the current device.\n \"\"\"\n return torch.device(f\"npu:{torch.npu.current_device()}\")\n\n def current_device(self) -> int:\n \"\"\"\n Return the current device index.\n \"\"\"\n return torch.npu.current_device()\n\n def set_device(self, device: Optional[Union[torch.device, int]] = None) -> None:\n \"\"\"\n Bind the current process to a device.\n \"\"\"\n if device is None:\n if not dist.is_initialized():\n raise RuntimeError(\"Cannot get current device when distributed is not initialized.\")\n device = dist.get_rank() % self.device_count()\n torch.npu.set_device(device)\n\n def get_device_name(self, device: Union[torch.device, int]) -> str:\n \"\"\"\n Return the name of the device.\n \"\"\"\n return torch.npu.get_device_name(device)\n\n def synchronize(self, device: Union[torch.device, int] = None):\n \"\"\"\n Synchronize the current process.\n \"\"\"\n torch.npu.synchronize(device)\n\n def is_available(self):\n \"\"\"\n Check if the accelerator is available.\n \"\"\"\n return torch.npu.is_available()\n\n def device_count(self):\n \"\"\"\n Return the number of devices on the machine.\n \"\"\"\n return torch.npu.device_count()\n\n def get_device_capability(self, device=None) -> Tuple[int, int]:\n \"\"\"\n Gets the npu capability of a device.\n \"\"\"\n return torch.npu.get_device_capability(device)\n\n def get_device_name(self, device=None) -> str:\n \"\"\"\n Gets the name of a device.\n \"\"\"\n return torch.npu.get_device_name(device)\n\n def get_device_properties(self, device):\n \"\"\"\n Gets the properties of a device.\n \"\"\"\n return torch.npu.get_device_properties(device)\n\n def utilization(self, device=None) -> int:\n \"\"\"\n Returns the percent of time over the past sample period during which one or more kernels was executing on the GPU as given by nvidia-smi\n \"\"\"\n return torch.npu.utilization(device)\n\n # =======================\n # random number generator APIs\n # =======================\n def get_rng_state(self, device=\"npu\") -> torch.Tensor:\n \"\"\"\n Returns the random number generator state of the specified GPU as a ByteTensor.\n \"\"\"\n return torch.npu.get_rng_state(device)\n\n def get_rng_state_all(self) -> List[torch.Tensor]:\n \"\"\"\n Returns a list of ByteTensor representing the random number states of all devices.\n \"\"\"\n return torch.npu.get_rng_state_all()\n\n def set_rng_state(self, new_state: torch.ByteTensor, device: str = \"npu\") -> None:\n \"\"\"\n Sets the random number generator state of the specified GPU.\n \"\"\"\n torch.npu.set_rng_state(new_state, device)\n\n def set_rng_state_all(self, new_states: List[torch.ByteTensor]) -> None:\n \"\"\"\n Sets the random number generator state of all devices.\n \"\"\"\n torch.npu.set_rng_state_all(new_states)\n\n def manual_seed(self, seed: int) -> None:\n \"\"\"\n Sets the seed for generating random numbers for the current GPU.\n \"\"\"\n torch.npu.manual_seed(seed)\n\n def manual_seed_all(self, seed: int) -> None:\n \"\"\"\n Set the random seed for the all processes.\n \"\"\"\n torch.npu.manual_seed_all(seed)\n\n def seed(self) -> None:\n \"\"\"\n Sets the seed for generating random numbers to a random number for the current GPU.\n \"\"\"\n torch.npu.seed()\n\n def seed_all(self) -> None:\n \"\"\"\n Sets the seed for generating random numbers to a random number on all GPUs.\n \"\"\"\n torch.npu.seed_all()\n\n def initial_seed(self) -> int:\n \"\"\"\n Returns the current random seed of the current GPU.\n \"\"\"\n return torch.npu.initial_seed()\n\n # =======================\n # memory management APIs\n # =======================\n\n def empty_cache(self) -> None:\n \"\"\"\n Releases all unoccupied cached memory currently held by the caching allocator so that those can be used in other GPU application and visible in nvidia-smi.\n \"\"\"\n torch.npu.empty_cache()\n\n def memory_stats(self, device=None) -> Dict[str, Any]:\n \"\"\"\n Returns a dictionary of npu memory allocator statistics for a given device.\n \"\"\"\n return torch.npu.memory_stats(device=device)\n\n def memory_summary(self, device=None, abbreviated=False) -> str:\n \"\"\"\n Returns a human-readable printout of the current memory allocator statistics for a given device.\n \"\"\"\n return torch.npu.memory_summary(device=device, abbreviated=abbreviated)\n\n def memory_snapshot(self):\n \"\"\"\n Returns a snapshot of the npu memory allocator state across all devices.\n \"\"\"\n return torch.npu.memory_snapshot()\n\n def memory_allocated(self, device=None) -> int:\n \"\"\"\n Returns the current GPU memory occupied by tensors in bytes for a given device.\n \"\"\"\n return torch.npu.memory_allocated(device=device)\n\n def max_memory_allocated(self, device=None) -> int:\n \"\"\"\n Returns the maximum GPU memory occupied by tensors in bytes for a given device.\n \"\"\"\n return torch.npu.max_memory_allocated(device=device)\n\n def reset_max_memory_allocated(self, device=None) -> None:\n \"\"\"\n Resets the starting point in tracking maximum GPU memory occupied by tensors for a given device.\n \"\"\"\n torch.npu.reset_max_memory_allocated(device=device)\n\n def reset_max_memory_cached(self, device=None) -> None:\n \"\"\"\n Resets the starting point in tracking maximum GPU memory managed by the caching allocator for a given device.\n \"\"\"\n torch.npu.reset_max_memory_cached(device=device)\n\n def memory_reserved(self, device=None) -> int:\n \"\"\"\n Returns the current GPU memory managed by the caching allocator in bytes for a given device.\n \"\"\"\n return torch.npu.memory_reserved(device=device)\n\n def max_memory_reserved(self, device=None) -> int:\n \"\"\"\n Returns the maximum GPU memory managed by the caching allocator in bytes for a given device.\n \"\"\"\n return torch.npu.max_memory_reserved(device=device)\n\n def set_per_process_memory_fraction(self, fraction: float, device=None) -> None:\n \"\"\"\n Set memory fraction for a process.\n \"\"\"\n torch.npu.set_per_process_memory_fraction(fraction, device=device)\n\n def reset_peak_memory_stats(self, device=None) -> None:\n \"\"\"\n Resets the \"peak\" stats tracked by the npu memory allocator.\n \"\"\"\n torch.npu.reset_peak_memory_stats(device=device)\n\n # =======================\n # streams and events APIs\n # =======================\n\n def Stream(self, device=None, priority=0, **kwargs):\n \"\"\"\n A npu stream is a linear sequence of execution that belongs to a specific device, independent from other streams. See npu-semantics for details.\n \"\"\"\n return torch.npu.Stream(device, priority, **kwargs)\n\n def Event(self, enable_timing: bool = False, blocking: bool = False, interprocess: bool = False):\n \"\"\"\n npu events are synchronization markers that can be used to monitor the device's progress, to accurately measure timing, and to synchronize npu streams.\n \"\"\"\n return torch.npu.Event(enable_timing, blocking, interprocess)\n\n def current_stream(self, device=None):\n \"\"\"\n Returns the currently selected Stream for a given device.\n \"\"\"\n return torch.npu.current_stream(device)\n\n def default_stream(self, device=None):\n \"\"\"\n Returns the default Stream for a given device.\n \"\"\"\n return torch.npu.default_stream(device)\n\n def set_stream(self, stream_):\n \"\"\"\n Sets the current stream.This is a wrapper API to set the stream.\n \"\"\"\n torch.npu.set_stream(stream_)\n\n def stream(self, stream_):\n \"\"\"\n Wrapper around the Context-manager StreamContext that selects a given stream.\n \"\"\"\n return torch.npu.stream(stream_)\n\n # =======================\n # amp APIs\n # =======================\n def autocast(\n self, enabled: bool = True, dtype: torch.dtype = torch.float16, cache_enabled: bool = True\n ) -> Callable:\n \"\"\"\n Return autocast function\n \"\"\"\n return torch.npu.amp.autocast(enabled=enabled, dtype=dtype, cache_enabled=cache_enabled)\n", "path": "colossalai/accelerator/npu_accelerator.py"}]}
| 3,013 | 118 |
gh_patches_debug_24929
|
rasdani/github-patches
|
git_diff
|
pytorch__TensorRT-2228
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
🐛 [Bug] Issue in `broadcastable` converter utility
## Bug Description
Encountered `IndexError` in `broadcastable` utility:
```python
File "~/TensorRT/py/torch_tensorrt/dynamo/conversion/converter_utils.py", line 118, in broadcastable
greater_tensor[i] != lesser_tensor[j]
IndexError: tuple index out of range
While executing %where_5 : [num_users=1] = call_function[target=torch.ops.aten.where.self](...)
```
## To Reproduce
Steps to reproduce the behavior:
Invoke `aten.where` converter with a 0D tensor.
## Expected behavior
Utility should not error.
## Environment
- Torch-TensorRT Version (e.g. 1.0.0): 8c62fca7649ac2e8e74ab7d9d66a20a164e2af51
- PyTorch Version (e.g. 1.0): `2.1.0.dev20230803+cu121`
## Additional context
Blocking compilation of #1697 with torch compile path.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `py/torch_tensorrt/dynamo/conversion/converter_utils.py`
Content:
```
1 import logging
2 import re
3 from typing import List, Optional
4
5 import tensorrt as trt
6 import torch
7 from torch.fx.node import Target
8 from torch_tensorrt.fx.converters.converter_utils import (
9 Frameworks,
10 unified_dtype_converter,
11 )
12 from torch_tensorrt.fx.types import TRTDataType, TRTNetwork, TRTTensor
13
14 from .._SourceIR import SourceIR
15 from .converter_registry import ConverterRegistry
16
17 _LOGGER: logging.Logger = logging.getLogger(__name__)
18
19
20 def get_node_name(node: torch.fx.Node) -> str:
21 # nn_module_stack preserves the call stack of pytorch nn.modules
22 # The call stack contains a detailed name of the module
23 # which shows exactly where the module is located in the
24 # network architecture.
25 stack_item = node.meta.get("nn_module_stack", None)
26 # The current node is the last item in the stack
27 mod_stack = stack_item.popitem() if stack_item else ""
28 node_name = str(node)
29 if mod_stack:
30 mod_name = str(mod_stack[0]).replace("___", "/")
31 # Clean up the module name
32 mod_name = re.sub("^.*__self", "", mod_name)
33 mod_name = re.sub(r"_(\d+)$", r"/\g<1>", mod_name)
34 node_name = mod_name + "/" + node_name
35 else:
36 # Try an alternative way to get the module info
37 # like the node.meta['source_fn'] attr
38 pass
39
40 _LOGGER.debug(f"Node meta name {node_name}")
41 return node_name
42
43
44 def dynamic_unsupported(node: torch.fx.Node) -> bool:
45 # Validate that none of the inputs to the node have Dynamic shapes
46 assert isinstance(
47 node, torch.fx.Node
48 ), "Inputs to validator functions must be FX Nodes"
49
50 # Check node value itself
51 if getattr(node.meta["val"], "_has_symbolic_sizes_strides", False):
52 return False
53
54 # Check node arguments individually
55 if any(
56 getattr(arg.meta["val"], "_has_symbolic_sizes_strides", False)
57 for arg in node.args
58 if isinstance(arg, torch.fx.Node)
59 ):
60 return False
61
62 # Check node keyword arguments individually
63 if any(
64 getattr(kwarg.meta["val"], "_has_symbolic_sizes_strides", False)
65 for kwarg in node.kwargs.values()
66 if isinstance(kwarg, torch.fx.Node)
67 ):
68 return False
69
70 return True
71
72
73 def cast_trt_tensor(
74 network: TRTNetwork,
75 input_val: TRTTensor,
76 dtype: TRTDataType,
77 name: str,
78 target: Target = "",
79 source_ir: Optional[SourceIR] = None,
80 ) -> TRTTensor:
81 """
82 Given a TRT Tensor, convert that Tensor to the specified dtype
83 Adds an Identity layer to the network which performs the conversion
84 Args:
85 network (TRTNetwork): A TensorRT network
86 input_val (TRTTensor): A TRT Tensor to cast to a new data type
87 dtype (TRTDataType, torch.dtype, np.dtype): The data type to cast the input Tensor to
88 name (str): Name of the calling layer
89 target (Target): Target of calling node
90 source_ir (SourceIR): SourceIR of calling converter
91 Returns:
92 A TensorRT ITensor which has been casted to the specified dtype
93 """
94 trt_dtype = unified_dtype_converter(dtype, Frameworks.TRT)
95
96 if input_val.dtype != trt_dtype:
97 source_ir = source_ir if source_ir is not None else SourceIR.UNKNOWN
98 target_str = ConverterRegistry.qualified_name_or_str(target)
99 target_name = f"{source_ir}_ops{('.' + target_str) if target_str else ''}"
100
101 identity_layer = network.add_identity(input_val)
102 identity_layer.set_output_type(0, trt_dtype)
103 identity_layer.name = f"Cast ITensor {input_val.name} from {input_val.dtype} to {trt_dtype} - [{target_name}]-[{name}]"
104 return identity_layer.get_output(0)
105 else:
106 return input_val
107
108
109 def cast_int_int_div_trt_tensor(
110 network: TRTNetwork,
111 lhs_val: TRTTensor,
112 rhs_val: TRTTensor,
113 name: str,
114 ) -> List[TRTTensor]:
115 """
116 Given two `int` data type TRT Tensor to div operation, cast the TRT Tensor to float type
117 Args:
118 network (TRTNetwork): A TensorRT network
119 lhs_val (TRTTensor): A TRT Tensor numerator
120 rhs_val (TRTTensor): A TRT Tensor numerator
121 name (str): Name of calling layer
122 Returns:
123 A list of lhs_val and rhs_val casted to the approriate datatype
124 """
125 if (lhs_val.dtype == trt.int8 or lhs_val.dtype == trt.int32) and (
126 rhs_val.dtype == trt.int8 or rhs_val.dtype == trt.int32
127 ):
128 lhs_val = cast_trt_tensor(network, lhs_val, trt.float32, name)
129 rhs_val = cast_trt_tensor(network, rhs_val, trt.float32, name)
130 return [lhs_val, rhs_val]
131
132
133 def broadcastable(
134 a: TRTTensor,
135 b: TRTTensor,
136 ) -> bool:
137 "Check if two tensors are broadcastable according to torch rules"
138 a_shape = tuple(a.shape)
139 b_shape = tuple(b.shape)
140 # check from the trailing
141 diff = len(a_shape) - len(b_shape)
142 if diff == 0:
143 return True
144 if diff > 0:
145 max = len(a_shape)
146 min = len(b_shape)
147 greater_tensor = a_shape
148 lesser_tensor = b_shape
149 elif diff < 0:
150 max = len(b_shape)
151 min = len(a_shape)
152 greater_tensor = b_shape
153 lesser_tensor = a_shape
154 j = min - 1
155 for i in range(max - 1, diff - 1, -1):
156 if not (
157 greater_tensor[i] != lesser_tensor[j]
158 and (greater_tensor[i] == 1 or lesser_tensor[i] == 1)
159 ):
160 return False
161 return True
162
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/py/torch_tensorrt/dynamo/conversion/converter_utils.py b/py/torch_tensorrt/dynamo/conversion/converter_utils.py
--- a/py/torch_tensorrt/dynamo/conversion/converter_utils.py
+++ b/py/torch_tensorrt/dynamo/conversion/converter_utils.py
@@ -137,25 +137,23 @@
"Check if two tensors are broadcastable according to torch rules"
a_shape = tuple(a.shape)
b_shape = tuple(b.shape)
+
# check from the trailing
diff = len(a_shape) - len(b_shape)
- if diff == 0:
+
+ # Validate tensors have same rank and shape
+ if diff == 0 and all(a_shape[i] == b_shape[i] for i in range(len(a_shape))):
return True
+
+ # Left-pad the shorter dimension with ones
if diff > 0:
- max = len(a_shape)
- min = len(b_shape)
- greater_tensor = a_shape
- lesser_tensor = b_shape
- elif diff < 0:
- max = len(b_shape)
- min = len(a_shape)
- greater_tensor = b_shape
- lesser_tensor = a_shape
- j = min - 1
- for i in range(max - 1, diff - 1, -1):
- if not (
- greater_tensor[i] != lesser_tensor[j]
- and (greater_tensor[i] == 1 or lesser_tensor[i] == 1)
- ):
+ b_shape = (1,) * abs(diff) + b_shape
+ else:
+ a_shape = (1,) * abs(diff) + a_shape
+
+ # Validate one of the following conditions for broadcastability per-dimension
+ # 1. Equal number of dimensions or 2. Dimension has shape 1
+ for i in range(len(a_shape)):
+ if not (a_shape[i] == b_shape[i] or a_shape[i] == 1 or b_shape[i] == 1):
return False
return True
|
{"golden_diff": "diff --git a/py/torch_tensorrt/dynamo/conversion/converter_utils.py b/py/torch_tensorrt/dynamo/conversion/converter_utils.py\n--- a/py/torch_tensorrt/dynamo/conversion/converter_utils.py\n+++ b/py/torch_tensorrt/dynamo/conversion/converter_utils.py\n@@ -137,25 +137,23 @@\n \"Check if two tensors are broadcastable according to torch rules\"\n a_shape = tuple(a.shape)\n b_shape = tuple(b.shape)\n+\n # check from the trailing\n diff = len(a_shape) - len(b_shape)\n- if diff == 0:\n+\n+ # Validate tensors have same rank and shape\n+ if diff == 0 and all(a_shape[i] == b_shape[i] for i in range(len(a_shape))):\n return True\n+\n+ # Left-pad the shorter dimension with ones\n if diff > 0:\n- max = len(a_shape)\n- min = len(b_shape)\n- greater_tensor = a_shape\n- lesser_tensor = b_shape\n- elif diff < 0:\n- max = len(b_shape)\n- min = len(a_shape)\n- greater_tensor = b_shape\n- lesser_tensor = a_shape\n- j = min - 1\n- for i in range(max - 1, diff - 1, -1):\n- if not (\n- greater_tensor[i] != lesser_tensor[j]\n- and (greater_tensor[i] == 1 or lesser_tensor[i] == 1)\n- ):\n+ b_shape = (1,) * abs(diff) + b_shape\n+ else:\n+ a_shape = (1,) * abs(diff) + a_shape\n+\n+ # Validate one of the following conditions for broadcastability per-dimension\n+ # 1. Equal number of dimensions or 2. Dimension has shape 1\n+ for i in range(len(a_shape)):\n+ if not (a_shape[i] == b_shape[i] or a_shape[i] == 1 or b_shape[i] == 1):\n return False\n return True\n", "issue": "\ud83d\udc1b [Bug] Issue in `broadcastable` converter utility\n## Bug Description\r\n\r\nEncountered `IndexError` in `broadcastable` utility:\r\n\r\n```python\r\nFile \"~/TensorRT/py/torch_tensorrt/dynamo/conversion/converter_utils.py\", line 118, in broadcastable\r\n greater_tensor[i] != lesser_tensor[j]\r\n\r\nIndexError: tuple index out of range\r\n\r\nWhile executing %where_5 : [num_users=1] = call_function[target=torch.ops.aten.where.self](...)\r\n```\r\n\r\n## To Reproduce\r\n\r\nSteps to reproduce the behavior:\r\n\r\nInvoke `aten.where` converter with a 0D tensor.\r\n\r\n## Expected behavior\r\n\r\nUtility should not error.\r\n\r\n## Environment\r\n\r\n - Torch-TensorRT Version (e.g. 1.0.0): 8c62fca7649ac2e8e74ab7d9d66a20a164e2af51\r\n - PyTorch Version (e.g. 1.0): `2.1.0.dev20230803+cu121`\r\n\r\n## Additional context\r\n\r\nBlocking compilation of #1697 with torch compile path.\r\n\n", "before_files": [{"content": "import logging\nimport re\nfrom typing import List, Optional\n\nimport tensorrt as trt\nimport torch\nfrom torch.fx.node import Target\nfrom torch_tensorrt.fx.converters.converter_utils import (\n Frameworks,\n unified_dtype_converter,\n)\nfrom torch_tensorrt.fx.types import TRTDataType, TRTNetwork, TRTTensor\n\nfrom .._SourceIR import SourceIR\nfrom .converter_registry import ConverterRegistry\n\n_LOGGER: logging.Logger = logging.getLogger(__name__)\n\n\ndef get_node_name(node: torch.fx.Node) -> str:\n # nn_module_stack preserves the call stack of pytorch nn.modules\n # The call stack contains a detailed name of the module\n # which shows exactly where the module is located in the\n # network architecture.\n stack_item = node.meta.get(\"nn_module_stack\", None)\n # The current node is the last item in the stack\n mod_stack = stack_item.popitem() if stack_item else \"\"\n node_name = str(node)\n if mod_stack:\n mod_name = str(mod_stack[0]).replace(\"___\", \"/\")\n # Clean up the module name\n mod_name = re.sub(\"^.*__self\", \"\", mod_name)\n mod_name = re.sub(r\"_(\\d+)$\", r\"/\\g<1>\", mod_name)\n node_name = mod_name + \"/\" + node_name\n else:\n # Try an alternative way to get the module info\n # like the node.meta['source_fn'] attr\n pass\n\n _LOGGER.debug(f\"Node meta name {node_name}\")\n return node_name\n\n\ndef dynamic_unsupported(node: torch.fx.Node) -> bool:\n # Validate that none of the inputs to the node have Dynamic shapes\n assert isinstance(\n node, torch.fx.Node\n ), \"Inputs to validator functions must be FX Nodes\"\n\n # Check node value itself\n if getattr(node.meta[\"val\"], \"_has_symbolic_sizes_strides\", False):\n return False\n\n # Check node arguments individually\n if any(\n getattr(arg.meta[\"val\"], \"_has_symbolic_sizes_strides\", False)\n for arg in node.args\n if isinstance(arg, torch.fx.Node)\n ):\n return False\n\n # Check node keyword arguments individually\n if any(\n getattr(kwarg.meta[\"val\"], \"_has_symbolic_sizes_strides\", False)\n for kwarg in node.kwargs.values()\n if isinstance(kwarg, torch.fx.Node)\n ):\n return False\n\n return True\n\n\ndef cast_trt_tensor(\n network: TRTNetwork,\n input_val: TRTTensor,\n dtype: TRTDataType,\n name: str,\n target: Target = \"\",\n source_ir: Optional[SourceIR] = None,\n) -> TRTTensor:\n \"\"\"\n Given a TRT Tensor, convert that Tensor to the specified dtype\n Adds an Identity layer to the network which performs the conversion\n Args:\n network (TRTNetwork): A TensorRT network\n input_val (TRTTensor): A TRT Tensor to cast to a new data type\n dtype (TRTDataType, torch.dtype, np.dtype): The data type to cast the input Tensor to\n name (str): Name of the calling layer\n target (Target): Target of calling node\n source_ir (SourceIR): SourceIR of calling converter\n Returns:\n A TensorRT ITensor which has been casted to the specified dtype\n \"\"\"\n trt_dtype = unified_dtype_converter(dtype, Frameworks.TRT)\n\n if input_val.dtype != trt_dtype:\n source_ir = source_ir if source_ir is not None else SourceIR.UNKNOWN\n target_str = ConverterRegistry.qualified_name_or_str(target)\n target_name = f\"{source_ir}_ops{('.' + target_str) if target_str else ''}\"\n\n identity_layer = network.add_identity(input_val)\n identity_layer.set_output_type(0, trt_dtype)\n identity_layer.name = f\"Cast ITensor {input_val.name} from {input_val.dtype} to {trt_dtype} - [{target_name}]-[{name}]\"\n return identity_layer.get_output(0)\n else:\n return input_val\n\n\ndef cast_int_int_div_trt_tensor(\n network: TRTNetwork,\n lhs_val: TRTTensor,\n rhs_val: TRTTensor,\n name: str,\n) -> List[TRTTensor]:\n \"\"\"\n Given two `int` data type TRT Tensor to div operation, cast the TRT Tensor to float type\n Args:\n network (TRTNetwork): A TensorRT network\n lhs_val (TRTTensor): A TRT Tensor numerator\n rhs_val (TRTTensor): A TRT Tensor numerator\n name (str): Name of calling layer\n Returns:\n A list of lhs_val and rhs_val casted to the approriate datatype\n \"\"\"\n if (lhs_val.dtype == trt.int8 or lhs_val.dtype == trt.int32) and (\n rhs_val.dtype == trt.int8 or rhs_val.dtype == trt.int32\n ):\n lhs_val = cast_trt_tensor(network, lhs_val, trt.float32, name)\n rhs_val = cast_trt_tensor(network, rhs_val, trt.float32, name)\n return [lhs_val, rhs_val]\n\n\ndef broadcastable(\n a: TRTTensor,\n b: TRTTensor,\n) -> bool:\n \"Check if two tensors are broadcastable according to torch rules\"\n a_shape = tuple(a.shape)\n b_shape = tuple(b.shape)\n # check from the trailing\n diff = len(a_shape) - len(b_shape)\n if diff == 0:\n return True\n if diff > 0:\n max = len(a_shape)\n min = len(b_shape)\n greater_tensor = a_shape\n lesser_tensor = b_shape\n elif diff < 0:\n max = len(b_shape)\n min = len(a_shape)\n greater_tensor = b_shape\n lesser_tensor = a_shape\n j = min - 1\n for i in range(max - 1, diff - 1, -1):\n if not (\n greater_tensor[i] != lesser_tensor[j]\n and (greater_tensor[i] == 1 or lesser_tensor[i] == 1)\n ):\n return False\n return True\n", "path": "py/torch_tensorrt/dynamo/conversion/converter_utils.py"}], "after_files": [{"content": "import logging\nimport re\nfrom typing import List, Optional\n\nimport tensorrt as trt\nimport torch\nfrom torch.fx.node import Target\nfrom torch_tensorrt.fx.converters.converter_utils import (\n Frameworks,\n unified_dtype_converter,\n)\nfrom torch_tensorrt.fx.types import TRTDataType, TRTNetwork, TRTTensor\n\nfrom .._SourceIR import SourceIR\nfrom .converter_registry import ConverterRegistry\n\n_LOGGER: logging.Logger = logging.getLogger(__name__)\n\n\ndef get_node_name(node: torch.fx.Node) -> str:\n # nn_module_stack preserves the call stack of pytorch nn.modules\n # The call stack contains a detailed name of the module\n # which shows exactly where the module is located in the\n # network architecture.\n stack_item = node.meta.get(\"nn_module_stack\", None)\n # The current node is the last item in the stack\n mod_stack = stack_item.popitem() if stack_item else \"\"\n node_name = str(node)\n if mod_stack:\n mod_name = str(mod_stack[0]).replace(\"___\", \"/\")\n # Clean up the module name\n mod_name = re.sub(\"^.*__self\", \"\", mod_name)\n mod_name = re.sub(r\"_(\\d+)$\", r\"/\\g<1>\", mod_name)\n node_name = mod_name + \"/\" + node_name\n else:\n # Try an alternative way to get the module info\n # like the node.meta['source_fn'] attr\n pass\n\n _LOGGER.debug(f\"Node meta name {node_name}\")\n return node_name\n\n\ndef dynamic_unsupported(node: torch.fx.Node) -> bool:\n # Validate that none of the inputs to the node have Dynamic shapes\n assert isinstance(\n node, torch.fx.Node\n ), \"Inputs to validator functions must be FX Nodes\"\n\n # Check node value itself\n if getattr(node.meta[\"val\"], \"_has_symbolic_sizes_strides\", False):\n return False\n\n # Check node arguments individually\n if any(\n getattr(arg.meta[\"val\"], \"_has_symbolic_sizes_strides\", False)\n for arg in node.args\n if isinstance(arg, torch.fx.Node)\n ):\n return False\n\n # Check node keyword arguments individually\n if any(\n getattr(kwarg.meta[\"val\"], \"_has_symbolic_sizes_strides\", False)\n for kwarg in node.kwargs.values()\n if isinstance(kwarg, torch.fx.Node)\n ):\n return False\n\n return True\n\n\ndef cast_trt_tensor(\n network: TRTNetwork,\n input_val: TRTTensor,\n dtype: TRTDataType,\n name: str,\n target: Target = \"\",\n source_ir: Optional[SourceIR] = None,\n) -> TRTTensor:\n \"\"\"\n Given a TRT Tensor, convert that Tensor to the specified dtype\n Adds an Identity layer to the network which performs the conversion\n Args:\n network (TRTNetwork): A TensorRT network\n input_val (TRTTensor): A TRT Tensor to cast to a new data type\n dtype (TRTDataType, torch.dtype, np.dtype): The data type to cast the input Tensor to\n name (str): Name of the calling layer\n target (Target): Target of calling node\n source_ir (SourceIR): SourceIR of calling converter\n Returns:\n A TensorRT ITensor which has been casted to the specified dtype\n \"\"\"\n trt_dtype = unified_dtype_converter(dtype, Frameworks.TRT)\n\n if input_val.dtype != trt_dtype:\n source_ir = source_ir if source_ir is not None else SourceIR.UNKNOWN\n target_str = ConverterRegistry.qualified_name_or_str(target)\n target_name = f\"{source_ir}_ops{('.' + target_str) if target_str else ''}\"\n\n identity_layer = network.add_identity(input_val)\n identity_layer.set_output_type(0, trt_dtype)\n identity_layer.name = f\"Cast ITensor {input_val.name} from {input_val.dtype} to {trt_dtype} - [{target_name}]-[{name}]\"\n return identity_layer.get_output(0)\n else:\n return input_val\n\n\ndef cast_int_int_div_trt_tensor(\n network: TRTNetwork,\n lhs_val: TRTTensor,\n rhs_val: TRTTensor,\n name: str,\n) -> List[TRTTensor]:\n \"\"\"\n Given two `int` data type TRT Tensor to div operation, cast the TRT Tensor to float type\n Args:\n network (TRTNetwork): A TensorRT network\n lhs_val (TRTTensor): A TRT Tensor numerator\n rhs_val (TRTTensor): A TRT Tensor numerator\n name (str): Name of calling layer\n Returns:\n A list of lhs_val and rhs_val casted to the approriate datatype\n \"\"\"\n if (lhs_val.dtype == trt.int8 or lhs_val.dtype == trt.int32) and (\n rhs_val.dtype == trt.int8 or rhs_val.dtype == trt.int32\n ):\n lhs_val = cast_trt_tensor(network, lhs_val, trt.float32, name)\n rhs_val = cast_trt_tensor(network, rhs_val, trt.float32, name)\n return [lhs_val, rhs_val]\n\n\ndef broadcastable(\n a: TRTTensor,\n b: TRTTensor,\n) -> bool:\n \"Check if two tensors are broadcastable according to torch rules\"\n a_shape = tuple(a.shape)\n b_shape = tuple(b.shape)\n\n # check from the trailing\n diff = len(a_shape) - len(b_shape)\n\n # Validate tensors have same rank and shape\n if diff == 0 and all(a_shape[i] == b_shape[i] for i in range(len(a_shape))):\n return True\n\n # Left-pad the shorter dimension with ones\n if diff > 0:\n b_shape = (1,) * abs(diff) + b_shape\n else:\n a_shape = (1,) * abs(diff) + a_shape\n\n # Validate one of the following conditions for broadcastability per-dimension\n # 1. Equal number of dimensions or 2. Dimension has shape 1\n for i in range(len(a_shape)):\n if not (a_shape[i] == b_shape[i] or a_shape[i] == 1 or b_shape[i] == 1):\n return False\n return True\n", "path": "py/torch_tensorrt/dynamo/conversion/converter_utils.py"}]}
| 2,264 | 460 |
gh_patches_debug_19691
|
rasdani/github-patches
|
git_diff
|
aws-cloudformation__cfn-lint-2967
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Invalid Aliases when using SSM dynamic references
### CloudFormation Lint Version
0.83.0
### What operating system are you using?
Mac
### Describe the bug
When using a dynamic reference to resolve the Alias domain, cfn-lint fails indicating it's an invalid alias. Shouldn't the code check if this is a `REGEX_DYN_REF` in https://github.com/aws-cloudformation/cfn-lint/blob/main/src/cfnlint/rules/resources/cloudfront/Aliases.py and ignore if so?
A workaround would be to use "!Sub" which apparently is ignored already (`FUNCTIONS`). Shouldn't we also ignore when `REGEX_DYN_REF`?
### Expected behavior
E3013 shouldn't be informed, since there's no way to validate the dynamic-reference value from cfn-lint perspective (?)
### Reproduction template
```
CloudFront:
Type: AWS::CloudFront::Distribution
Properties:
DistributionConfig:
Enabled: true
Aliases:
- "{{resolve:ssm:/env/fqdns/certifier}}"
DefaultRootObject: index.html
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/cfnlint/rules/resources/cloudfront/Aliases.py`
Content:
```
1 """
2 Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
3 SPDX-License-Identifier: MIT-0
4 """
5 import regex as re
6
7 from cfnlint.helpers import FUNCTIONS
8 from cfnlint.rules import CloudFormationLintRule, RuleMatch
9
10
11 class Aliases(CloudFormationLintRule):
12 """Check if CloudFront Aliases are valid domain names"""
13
14 id = "E3013"
15 shortdesc = "CloudFront Aliases"
16 description = "CloudFront aliases should contain valid domain names"
17 source_url = "https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-cloudfront-distribution-distributionconfig.html#cfn-cloudfront-distribution-distributionconfig-aliases"
18 tags = ["properties", "cloudfront"]
19
20 def match(self, cfn):
21 """Check cloudfront Resource Parameters"""
22
23 matches = []
24
25 valid_domain = re.compile(
26 r"^(?:[a-z0-9\*](?:[a-z0-9-]{0,61}[a-z0-9])?\.)+[a-z0-9][a-z0-9-]{0,61}[a-z0-9]$"
27 )
28
29 results = cfn.get_resource_properties(
30 ["AWS::CloudFront::Distribution", "DistributionConfig"]
31 )
32 for result in results:
33 aliases = result["Value"].get("Aliases")
34 if aliases:
35 for alias in aliases:
36 if isinstance(alias, str) and alias not in FUNCTIONS:
37 wildcard = alias.split(".")
38 if "*" in wildcard[1:]:
39 path = result["Path"] + ["Aliases"]
40 message = f'Invalid use of wildcards: {alias} at {"/".join(result["Path"])}'
41 matches.append(RuleMatch(path, message))
42 if not re.match(valid_domain, alias):
43 path = result["Path"] + ["Aliases"]
44 message = f'Invalid alias found: {alias} at {"/".join(result["Path"])}'
45 matches.append(RuleMatch(path, message))
46
47 return matches
48
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/cfnlint/rules/resources/cloudfront/Aliases.py b/src/cfnlint/rules/resources/cloudfront/Aliases.py
--- a/src/cfnlint/rules/resources/cloudfront/Aliases.py
+++ b/src/cfnlint/rules/resources/cloudfront/Aliases.py
@@ -4,7 +4,7 @@
"""
import regex as re
-from cfnlint.helpers import FUNCTIONS
+from cfnlint.helpers import FUNCTIONS, REGEX_DYN_REF
from cfnlint.rules import CloudFormationLintRule, RuleMatch
@@ -35,6 +35,8 @@
for alias in aliases:
if isinstance(alias, str) and alias not in FUNCTIONS:
wildcard = alias.split(".")
+ if re.match(REGEX_DYN_REF, alias):
+ continue
if "*" in wildcard[1:]:
path = result["Path"] + ["Aliases"]
message = f'Invalid use of wildcards: {alias} at {"/".join(result["Path"])}'
|
{"golden_diff": "diff --git a/src/cfnlint/rules/resources/cloudfront/Aliases.py b/src/cfnlint/rules/resources/cloudfront/Aliases.py\n--- a/src/cfnlint/rules/resources/cloudfront/Aliases.py\n+++ b/src/cfnlint/rules/resources/cloudfront/Aliases.py\n@@ -4,7 +4,7 @@\n \"\"\"\n import regex as re\n \n-from cfnlint.helpers import FUNCTIONS\n+from cfnlint.helpers import FUNCTIONS, REGEX_DYN_REF\n from cfnlint.rules import CloudFormationLintRule, RuleMatch\n \n \n@@ -35,6 +35,8 @@\n for alias in aliases:\n if isinstance(alias, str) and alias not in FUNCTIONS:\n wildcard = alias.split(\".\")\n+ if re.match(REGEX_DYN_REF, alias):\n+ continue\n if \"*\" in wildcard[1:]:\n path = result[\"Path\"] + [\"Aliases\"]\n message = f'Invalid use of wildcards: {alias} at {\"/\".join(result[\"Path\"])}'\n", "issue": "Invalid Aliases when using SSM dynamic references\n### CloudFormation Lint Version\r\n\r\n0.83.0\r\n\r\n### What operating system are you using?\r\n\r\nMac\r\n\r\n### Describe the bug\r\n\r\nWhen using a dynamic reference to resolve the Alias domain, cfn-lint fails indicating it's an invalid alias. Shouldn't the code check if this is a `REGEX_DYN_REF` in https://github.com/aws-cloudformation/cfn-lint/blob/main/src/cfnlint/rules/resources/cloudfront/Aliases.py and ignore if so?\r\n\r\nA workaround would be to use \"!Sub\" which apparently is ignored already (`FUNCTIONS`). Shouldn't we also ignore when `REGEX_DYN_REF`?\r\n\r\n### Expected behavior\r\n\r\nE3013 shouldn't be informed, since there's no way to validate the dynamic-reference value from cfn-lint perspective (?)\r\n\r\n### Reproduction template\r\n\r\n```\r\nCloudFront:\r\n Type: AWS::CloudFront::Distribution\r\n Properties:\r\n DistributionConfig:\r\n Enabled: true\r\n Aliases:\r\n - \"{{resolve:ssm:/env/fqdns/certifier}}\"\r\n DefaultRootObject: index.html\r\n``` \n", "before_files": [{"content": "\"\"\"\nCopyright Amazon.com, Inc. or its affiliates. All Rights Reserved.\nSPDX-License-Identifier: MIT-0\n\"\"\"\nimport regex as re\n\nfrom cfnlint.helpers import FUNCTIONS\nfrom cfnlint.rules import CloudFormationLintRule, RuleMatch\n\n\nclass Aliases(CloudFormationLintRule):\n \"\"\"Check if CloudFront Aliases are valid domain names\"\"\"\n\n id = \"E3013\"\n shortdesc = \"CloudFront Aliases\"\n description = \"CloudFront aliases should contain valid domain names\"\n source_url = \"https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-cloudfront-distribution-distributionconfig.html#cfn-cloudfront-distribution-distributionconfig-aliases\"\n tags = [\"properties\", \"cloudfront\"]\n\n def match(self, cfn):\n \"\"\"Check cloudfront Resource Parameters\"\"\"\n\n matches = []\n\n valid_domain = re.compile(\n r\"^(?:[a-z0-9\\*](?:[a-z0-9-]{0,61}[a-z0-9])?\\.)+[a-z0-9][a-z0-9-]{0,61}[a-z0-9]$\"\n )\n\n results = cfn.get_resource_properties(\n [\"AWS::CloudFront::Distribution\", \"DistributionConfig\"]\n )\n for result in results:\n aliases = result[\"Value\"].get(\"Aliases\")\n if aliases:\n for alias in aliases:\n if isinstance(alias, str) and alias not in FUNCTIONS:\n wildcard = alias.split(\".\")\n if \"*\" in wildcard[1:]:\n path = result[\"Path\"] + [\"Aliases\"]\n message = f'Invalid use of wildcards: {alias} at {\"/\".join(result[\"Path\"])}'\n matches.append(RuleMatch(path, message))\n if not re.match(valid_domain, alias):\n path = result[\"Path\"] + [\"Aliases\"]\n message = f'Invalid alias found: {alias} at {\"/\".join(result[\"Path\"])}'\n matches.append(RuleMatch(path, message))\n\n return matches\n", "path": "src/cfnlint/rules/resources/cloudfront/Aliases.py"}], "after_files": [{"content": "\"\"\"\nCopyright Amazon.com, Inc. or its affiliates. All Rights Reserved.\nSPDX-License-Identifier: MIT-0\n\"\"\"\nimport regex as re\n\nfrom cfnlint.helpers import FUNCTIONS, REGEX_DYN_REF\nfrom cfnlint.rules import CloudFormationLintRule, RuleMatch\n\n\nclass Aliases(CloudFormationLintRule):\n \"\"\"Check if CloudFront Aliases are valid domain names\"\"\"\n\n id = \"E3013\"\n shortdesc = \"CloudFront Aliases\"\n description = \"CloudFront aliases should contain valid domain names\"\n source_url = \"https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-cloudfront-distribution-distributionconfig.html#cfn-cloudfront-distribution-distributionconfig-aliases\"\n tags = [\"properties\", \"cloudfront\"]\n\n def match(self, cfn):\n \"\"\"Check cloudfront Resource Parameters\"\"\"\n\n matches = []\n\n valid_domain = re.compile(\n r\"^(?:[a-z0-9\\*](?:[a-z0-9-]{0,61}[a-z0-9])?\\.)+[a-z0-9][a-z0-9-]{0,61}[a-z0-9]$\"\n )\n\n results = cfn.get_resource_properties(\n [\"AWS::CloudFront::Distribution\", \"DistributionConfig\"]\n )\n for result in results:\n aliases = result[\"Value\"].get(\"Aliases\")\n if aliases:\n for alias in aliases:\n if isinstance(alias, str) and alias not in FUNCTIONS:\n wildcard = alias.split(\".\")\n if re.match(REGEX_DYN_REF, alias):\n continue\n if \"*\" in wildcard[1:]:\n path = result[\"Path\"] + [\"Aliases\"]\n message = f'Invalid use of wildcards: {alias} at {\"/\".join(result[\"Path\"])}'\n matches.append(RuleMatch(path, message))\n if not re.match(valid_domain, alias):\n path = result[\"Path\"] + [\"Aliases\"]\n message = f'Invalid alias found: {alias} at {\"/\".join(result[\"Path\"])}'\n matches.append(RuleMatch(path, message))\n\n return matches\n", "path": "src/cfnlint/rules/resources/cloudfront/Aliases.py"}]}
| 1,033 | 213 |
gh_patches_debug_25570
|
rasdani/github-patches
|
git_diff
|
Zeroto521__my-data-toolkit-649
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
EHN: New geoaccessor `toposimiplify`
<!--
Thanks for contributing a pull request!
Please follow these standard acronyms to start the commit message:
- ENH: enhancement
- BUG: bug fix
- DOC: documentation
- TYP: type annotations
- TST: addition or modification of tests
- MAINT: maintenance commit (refactoring, typos, etc.)
- BLD: change related to building
- REL: related to releasing
- API: an (incompatible) API change
- DEP: deprecate something, or remove a deprecated object
- DEV: development tool or utility
- REV: revert an earlier commit
- PERF: performance improvement
- BOT: always commit via a bot
- CI: related to CI or CD
- CLN: Code cleanup
-->
- [ ] closes #xxxx
- [x] whatsnew entry
simplify multi-geometries method but could keep the shared edges
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `dtoolkit/geoaccessor/geoseries/toposimplify.py`
Content:
```
1 from __future__ import annotations
2
3 from typing import Literal
4
5 import geopandas as gpd
6 from pandas.util._decorators import doc
7
8 from dtoolkit.accessor.dataframe import to_series # noqa: F401
9 from dtoolkit.geoaccessor.register import register_geoseries_method
10
11
12 @register_geoseries_method
13 @doc(klass=":class:`~geopandas.GeoSeries`")
14 def toposimplify(
15 s: gpd.GeoSeries,
16 /,
17 tolerance: float,
18 simplify_algorithm: Literal["dp", "vw"] = "dp",
19 simplify_with: Literal["shapely", "simplification"] = "shapely",
20 prevent_oversimplify: bool = True,
21 ) -> gpd.GeoSeries:
22 """
23 Apply :func:`topojson.Topology.toposimplify` to {klass} to keep **shared edges**.
24
25 Parameters
26 ----------
27 tolerance : float
28 All parts of a simplified geometry will be no more than tolerance distance from
29 the original.
30
31 simplify_algorithm : {{'dp', 'vw'}}, default 'dp'
32 ``vw`` will only be selected if ``simplify_with`` is set to ``simplification``.
33 - ``dp`` : Douglas-Peucker
34 - ``vw`` : Visvalingam-Whyatt
35
36 simplify_with : {{'shapely', 'simplification'}}, default 'shapely'
37 Sets the package to use for simplifying. Shapely adopts solely Douglas-Peucker
38 and simplification both Douglas-Peucker and Visvalingam-Whyatt. The package
39 simplification is known to be quicker than shapely.
40
41 prevent_oversimplify : bool, default True
42 If `True`, the simplification is slower, but the likelihood of producing
43 valid geometries is higher as it prevents oversimplification. Simplification
44 happens on paths separately, so this setting is especially relevant for rings
45 with no partial shared paths. This is also known as a topology-preserving
46 variant of simplification.
47
48 Returns
49 -------
50 {klass}
51
52 Raises
53 ------
54 ModuleNotFoundError
55 - If don't have module named 'topojson'.
56 - If don't have module named 'simplification'.
57
58 See Also
59 --------
60 geopandas.GeoSeries.simplify
61 dtoolkit.geoaccessor.geoseries.toposimplify
62 dtoolkit.geoaccessor.geodataframe.toposimplify
63 topojson.Topology.toposimplify
64 https://mattijn.github.io/topojson/api/topojson.core.topology.html#toposimplify
65
66 Examples
67 --------
68 >>> import dtoolkit.geoaccessor
69 >>> import geopandas as gpd
70 >>> df = (
71 ... gpd.read_file(gpd.datasets.get_path("naturalearth_lowres"))
72 ... .query('continent == "Africa"')
73 ... )
74 >>> df.head()
75 pop_est ... geometry
76 1 58005463.0 ... POLYGON ((33.90371 -0.95000, 34.07262 -1.05982...
77 2 603253.0 ... POLYGON ((-8.66559 27.65643, -8.66512 27.58948...
78 11 86790567.0 ... POLYGON ((29.34000 -4.49998, 29.51999 -5.41998...
79 12 10192317.3 ... POLYGON ((41.58513 -1.68325, 40.99300 -0.85829...
80 13 52573973.0 ... POLYGON ((39.20222 -4.67677, 37.76690 -3.67712...
81 <BLANKLINE>
82 [5 rows x 6 columns]
83 >>> df.toposimplify(0.1).head()
84 pop_est ... geometry
85 1 58005463.0 ... POLYGON ((33.90367 -0.95000, 30.76984 -1.01452...
86 2 603253.0 ... POLYGON ((-8.66561 27.65644, -8.81786 27.65644...
87 11 86790567.0 ... POLYGON ((29.33999 -4.50001, 29.27634 -3.29391...
88 12 10192317.3 ... POLYGON ((41.58509 -1.68324, 42.04156 -0.91918...
89 13 52573973.0 ... POLYGON ((39.20220 -4.67675, 39.60493 -4.34650...
90 <BLANKLINE>
91 [5 rows x 6 columns]
92
93 .. plot::
94
95 import dtoolkit.geoaccessor
96 import geopandas as gpd
97 import matplotlib.pyplot as plt
98
99
100 df = (
101 gpd.read_file(gpd.datasets.get_path("naturalearth_lowres"))
102 .query('continent == "Africa"')
103 )
104
105 fig, (ax1, ax2) = plt.subplots(ncols=2, sharex=True, sharey=True)
106
107 df.simplify(1).plot(ax=ax1, alpha=0.7)
108 df.toposimplify(1).plot(ax=ax2, alpha=0.7)
109
110 ax1.set_title("simplify")
111 ax1.set_axis_off()
112 ax2.set_title("toposimplify")
113 ax2.set_axis_off()
114 fig.tight_layout()
115 plt.show()
116 """
117 from topojson import Topology
118
119 return (
120 Topology(
121 s,
122 toposimplify=tolerance,
123 simplify_algorithm=simplify_algorithm,
124 simplify_with=simplify_with,
125 prevent_oversimplify=prevent_oversimplify,
126 )
127 # `to_gdf` return is a GeoDataFrame, require GeoSeries
128 .to_gdf(crs=s.crs).to_series()
129 # To fix https://github.com/mattijn/topojson/issues/164
130 .set_axis(s.index)
131 )
132
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/dtoolkit/geoaccessor/geoseries/toposimplify.py b/dtoolkit/geoaccessor/geoseries/toposimplify.py
--- a/dtoolkit/geoaccessor/geoseries/toposimplify.py
+++ b/dtoolkit/geoaccessor/geoseries/toposimplify.py
@@ -20,7 +20,12 @@
prevent_oversimplify: bool = True,
) -> gpd.GeoSeries:
"""
- Apply :func:`topojson.Topology.toposimplify` to {klass} to keep **shared edges**.
+ Returns a {klass} containing a simplified representation of each geometry.
+ Similar to :meth:`~geopandas.GeoSeries.simplify`, but keeps shared edges.
+
+ .. image:: ../../../../_static/simplify-vs-toposimplify.png
+ :width: 80%
+ :align: center
Parameters
----------
@@ -89,30 +94,6 @@
13 52573973.0 ... POLYGON ((39.20220 -4.67675, 39.60493 -4.34650...
<BLANKLINE>
[5 rows x 6 columns]
-
- .. plot::
-
- import dtoolkit.geoaccessor
- import geopandas as gpd
- import matplotlib.pyplot as plt
-
-
- df = (
- gpd.read_file(gpd.datasets.get_path("naturalearth_lowres"))
- .query('continent == "Africa"')
- )
-
- fig, (ax1, ax2) = plt.subplots(ncols=2, sharex=True, sharey=True)
-
- df.simplify(1).plot(ax=ax1, alpha=0.7)
- df.toposimplify(1).plot(ax=ax2, alpha=0.7)
-
- ax1.set_title("simplify")
- ax1.set_axis_off()
- ax2.set_title("toposimplify")
- ax2.set_axis_off()
- fig.tight_layout()
- plt.show()
"""
from topojson import Topology
|
{"golden_diff": "diff --git a/dtoolkit/geoaccessor/geoseries/toposimplify.py b/dtoolkit/geoaccessor/geoseries/toposimplify.py\n--- a/dtoolkit/geoaccessor/geoseries/toposimplify.py\n+++ b/dtoolkit/geoaccessor/geoseries/toposimplify.py\n@@ -20,7 +20,12 @@\n prevent_oversimplify: bool = True,\n ) -> gpd.GeoSeries:\n \"\"\"\n- Apply :func:`topojson.Topology.toposimplify` to {klass} to keep **shared edges**.\n+ Returns a {klass} containing a simplified representation of each geometry.\n+ Similar to :meth:`~geopandas.GeoSeries.simplify`, but keeps shared edges.\n+\n+ .. image:: ../../../../_static/simplify-vs-toposimplify.png\n+ :width: 80%\n+ :align: center\n \n Parameters\n ----------\n@@ -89,30 +94,6 @@\n 13 52573973.0 ... POLYGON ((39.20220 -4.67675, 39.60493 -4.34650...\n <BLANKLINE>\n [5 rows x 6 columns]\n-\n- .. plot::\n-\n- import dtoolkit.geoaccessor\n- import geopandas as gpd\n- import matplotlib.pyplot as plt\n-\n-\n- df = (\n- gpd.read_file(gpd.datasets.get_path(\"naturalearth_lowres\"))\n- .query('continent == \"Africa\"')\n- )\n-\n- fig, (ax1, ax2) = plt.subplots(ncols=2, sharex=True, sharey=True)\n-\n- df.simplify(1).plot(ax=ax1, alpha=0.7)\n- df.toposimplify(1).plot(ax=ax2, alpha=0.7)\n-\n- ax1.set_title(\"simplify\")\n- ax1.set_axis_off()\n- ax2.set_title(\"toposimplify\")\n- ax2.set_axis_off()\n- fig.tight_layout()\n- plt.show()\n \"\"\"\n from topojson import Topology\n", "issue": "EHN: New geoaccessor `toposimiplify`\n<!--\r\nThanks for contributing a pull request!\r\n\r\nPlease follow these standard acronyms to start the commit message:\r\n\r\n- ENH: enhancement\r\n- BUG: bug fix\r\n- DOC: documentation\r\n- TYP: type annotations\r\n- TST: addition or modification of tests\r\n- MAINT: maintenance commit (refactoring, typos, etc.)\r\n- BLD: change related to building\r\n- REL: related to releasing\r\n- API: an (incompatible) API change\r\n- DEP: deprecate something, or remove a deprecated object\r\n- DEV: development tool or utility\r\n- REV: revert an earlier commit\r\n- PERF: performance improvement\r\n- BOT: always commit via a bot\r\n- CI: related to CI or CD\r\n- CLN: Code cleanup\r\n-->\r\n\r\n- [ ] closes #xxxx\r\n- [x] whatsnew entry\r\n\r\nsimplify multi-geometries method but could keep the shared edges\n", "before_files": [{"content": "from __future__ import annotations\n\nfrom typing import Literal\n\nimport geopandas as gpd\nfrom pandas.util._decorators import doc\n\nfrom dtoolkit.accessor.dataframe import to_series # noqa: F401\nfrom dtoolkit.geoaccessor.register import register_geoseries_method\n\n\n@register_geoseries_method\n@doc(klass=\":class:`~geopandas.GeoSeries`\")\ndef toposimplify(\n s: gpd.GeoSeries,\n /,\n tolerance: float,\n simplify_algorithm: Literal[\"dp\", \"vw\"] = \"dp\",\n simplify_with: Literal[\"shapely\", \"simplification\"] = \"shapely\",\n prevent_oversimplify: bool = True,\n) -> gpd.GeoSeries:\n \"\"\"\n Apply :func:`topojson.Topology.toposimplify` to {klass} to keep **shared edges**.\n\n Parameters\n ----------\n tolerance : float\n All parts of a simplified geometry will be no more than tolerance distance from\n the original.\n\n simplify_algorithm : {{'dp', 'vw'}}, default 'dp'\n ``vw`` will only be selected if ``simplify_with`` is set to ``simplification``.\n - ``dp`` : Douglas-Peucker\n - ``vw`` : Visvalingam-Whyatt\n\n simplify_with : {{'shapely', 'simplification'}}, default 'shapely'\n Sets the package to use for simplifying. Shapely adopts solely Douglas-Peucker\n and simplification both Douglas-Peucker and Visvalingam-Whyatt. The package\n simplification is known to be quicker than shapely.\n\n prevent_oversimplify : bool, default True\n If `True`, the simplification is slower, but the likelihood of producing\n valid geometries is higher as it prevents oversimplification. Simplification\n happens on paths separately, so this setting is especially relevant for rings\n with no partial shared paths. This is also known as a topology-preserving\n variant of simplification.\n\n Returns\n -------\n {klass}\n\n Raises\n ------\n ModuleNotFoundError\n - If don't have module named 'topojson'.\n - If don't have module named 'simplification'.\n\n See Also\n --------\n geopandas.GeoSeries.simplify\n dtoolkit.geoaccessor.geoseries.toposimplify\n dtoolkit.geoaccessor.geodataframe.toposimplify\n topojson.Topology.toposimplify\n https://mattijn.github.io/topojson/api/topojson.core.topology.html#toposimplify\n\n Examples\n --------\n >>> import dtoolkit.geoaccessor\n >>> import geopandas as gpd\n >>> df = (\n ... gpd.read_file(gpd.datasets.get_path(\"naturalearth_lowres\"))\n ... .query('continent == \"Africa\"')\n ... )\n >>> df.head()\n pop_est ... geometry\n 1 58005463.0 ... POLYGON ((33.90371 -0.95000, 34.07262 -1.05982...\n 2 603253.0 ... POLYGON ((-8.66559 27.65643, -8.66512 27.58948...\n 11 86790567.0 ... POLYGON ((29.34000 -4.49998, 29.51999 -5.41998...\n 12 10192317.3 ... POLYGON ((41.58513 -1.68325, 40.99300 -0.85829...\n 13 52573973.0 ... POLYGON ((39.20222 -4.67677, 37.76690 -3.67712...\n <BLANKLINE>\n [5 rows x 6 columns]\n >>> df.toposimplify(0.1).head()\n pop_est ... geometry\n 1 58005463.0 ... POLYGON ((33.90367 -0.95000, 30.76984 -1.01452...\n 2 603253.0 ... POLYGON ((-8.66561 27.65644, -8.81786 27.65644...\n 11 86790567.0 ... POLYGON ((29.33999 -4.50001, 29.27634 -3.29391...\n 12 10192317.3 ... POLYGON ((41.58509 -1.68324, 42.04156 -0.91918...\n 13 52573973.0 ... POLYGON ((39.20220 -4.67675, 39.60493 -4.34650...\n <BLANKLINE>\n [5 rows x 6 columns]\n\n .. plot::\n\n import dtoolkit.geoaccessor\n import geopandas as gpd\n import matplotlib.pyplot as plt\n\n\n df = (\n gpd.read_file(gpd.datasets.get_path(\"naturalearth_lowres\"))\n .query('continent == \"Africa\"')\n )\n\n fig, (ax1, ax2) = plt.subplots(ncols=2, sharex=True, sharey=True)\n\n df.simplify(1).plot(ax=ax1, alpha=0.7)\n df.toposimplify(1).plot(ax=ax2, alpha=0.7)\n\n ax1.set_title(\"simplify\")\n ax1.set_axis_off()\n ax2.set_title(\"toposimplify\")\n ax2.set_axis_off()\n fig.tight_layout()\n plt.show()\n \"\"\"\n from topojson import Topology\n\n return (\n Topology(\n s,\n toposimplify=tolerance,\n simplify_algorithm=simplify_algorithm,\n simplify_with=simplify_with,\n prevent_oversimplify=prevent_oversimplify,\n )\n # `to_gdf` return is a GeoDataFrame, require GeoSeries\n .to_gdf(crs=s.crs).to_series()\n # To fix https://github.com/mattijn/topojson/issues/164\n .set_axis(s.index)\n )\n", "path": "dtoolkit/geoaccessor/geoseries/toposimplify.py"}], "after_files": [{"content": "from __future__ import annotations\n\nfrom typing import Literal\n\nimport geopandas as gpd\nfrom pandas.util._decorators import doc\n\nfrom dtoolkit.accessor.dataframe import to_series # noqa: F401\nfrom dtoolkit.geoaccessor.register import register_geoseries_method\n\n\n@register_geoseries_method\n@doc(klass=\":class:`~geopandas.GeoSeries`\")\ndef toposimplify(\n s: gpd.GeoSeries,\n /,\n tolerance: float,\n simplify_algorithm: Literal[\"dp\", \"vw\"] = \"dp\",\n simplify_with: Literal[\"shapely\", \"simplification\"] = \"shapely\",\n prevent_oversimplify: bool = True,\n) -> gpd.GeoSeries:\n \"\"\"\n Returns a {klass} containing a simplified representation of each geometry.\n Similar to :meth:`~geopandas.GeoSeries.simplify`, but keeps shared edges.\n\n .. image:: ../../../../_static/simplify-vs-toposimplify.png\n :width: 80%\n :align: center\n\n Parameters\n ----------\n tolerance : float\n All parts of a simplified geometry will be no more than tolerance distance from\n the original.\n\n simplify_algorithm : {{'dp', 'vw'}}, default 'dp'\n ``vw`` will only be selected if ``simplify_with`` is set to ``simplification``.\n - ``dp`` : Douglas-Peucker\n - ``vw`` : Visvalingam-Whyatt\n\n simplify_with : {{'shapely', 'simplification'}}, default 'shapely'\n Sets the package to use for simplifying. Shapely adopts solely Douglas-Peucker\n and simplification both Douglas-Peucker and Visvalingam-Whyatt. The package\n simplification is known to be quicker than shapely.\n\n prevent_oversimplify : bool, default True\n If `True`, the simplification is slower, but the likelihood of producing\n valid geometries is higher as it prevents oversimplification. Simplification\n happens on paths separately, so this setting is especially relevant for rings\n with no partial shared paths. This is also known as a topology-preserving\n variant of simplification.\n\n Returns\n -------\n {klass}\n\n Raises\n ------\n ModuleNotFoundError\n - If don't have module named 'topojson'.\n - If don't have module named 'simplification'.\n\n See Also\n --------\n geopandas.GeoSeries.simplify\n dtoolkit.geoaccessor.geoseries.toposimplify\n dtoolkit.geoaccessor.geodataframe.toposimplify\n topojson.Topology.toposimplify\n https://mattijn.github.io/topojson/api/topojson.core.topology.html#toposimplify\n\n Examples\n --------\n >>> import dtoolkit.geoaccessor\n >>> import geopandas as gpd\n >>> df = (\n ... gpd.read_file(gpd.datasets.get_path(\"naturalearth_lowres\"))\n ... .query('continent == \"Africa\"')\n ... )\n >>> df.head()\n pop_est ... geometry\n 1 58005463.0 ... POLYGON ((33.90371 -0.95000, 34.07262 -1.05982...\n 2 603253.0 ... POLYGON ((-8.66559 27.65643, -8.66512 27.58948...\n 11 86790567.0 ... POLYGON ((29.34000 -4.49998, 29.51999 -5.41998...\n 12 10192317.3 ... POLYGON ((41.58513 -1.68325, 40.99300 -0.85829...\n 13 52573973.0 ... POLYGON ((39.20222 -4.67677, 37.76690 -3.67712...\n <BLANKLINE>\n [5 rows x 6 columns]\n >>> df.toposimplify(0.1).head()\n pop_est ... geometry\n 1 58005463.0 ... POLYGON ((33.90367 -0.95000, 30.76984 -1.01452...\n 2 603253.0 ... POLYGON ((-8.66561 27.65644, -8.81786 27.65644...\n 11 86790567.0 ... POLYGON ((29.33999 -4.50001, 29.27634 -3.29391...\n 12 10192317.3 ... POLYGON ((41.58509 -1.68324, 42.04156 -0.91918...\n 13 52573973.0 ... POLYGON ((39.20220 -4.67675, 39.60493 -4.34650...\n <BLANKLINE>\n [5 rows x 6 columns]\n \"\"\"\n from topojson import Topology\n\n return (\n Topology(\n s,\n toposimplify=tolerance,\n simplify_algorithm=simplify_algorithm,\n simplify_with=simplify_with,\n prevent_oversimplify=prevent_oversimplify,\n )\n # `to_gdf` return is a GeoDataFrame, require GeoSeries\n .to_gdf(crs=s.crs).to_series()\n # To fix https://github.com/mattijn/topojson/issues/164\n .set_axis(s.index)\n )\n", "path": "dtoolkit/geoaccessor/geoseries/toposimplify.py"}]}
| 2,324 | 489 |
gh_patches_debug_13562
|
rasdani/github-patches
|
git_diff
|
projectmesa__mesa-1984
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
JupyterViz space view limits
**What's the problem this feature will solve?**
At the moment, when drawing spaces with JupyterViz, the axes limits are left to be automatically determined by Matplotlib. Would it not make more sense (in particular of continuous spaces, to derive xlim and xmax from the space itself?
**Describe the solution you'd like**
The simplest fix would be to adapt `_draw_grid` to use `space.width` and `space.height` while `_draw_continuos_space` uses `space.x_min`, `space.x_max`, `space.y_min`, and `space.y_max`. Are there reasons where the user might want to rely on matpltolib's automatic determination of the bounds? If so, then a slightly more sophisticated solution would be needed where kwargs are passed to from jupyterviz to the underlying space_drawer function. So you would add a `space_drawer_kwargs` keyword argument.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mesa/experimental/components/matplotlib.py`
Content:
```
1 from typing import Optional
2
3 import networkx as nx
4 import solara
5 from matplotlib.figure import Figure
6 from matplotlib.ticker import MaxNLocator
7
8 import mesa
9
10
11 @solara.component
12 def SpaceMatplotlib(model, agent_portrayal, dependencies: Optional[list[any]] = None):
13 space_fig = Figure()
14 space_ax = space_fig.subplots()
15 space = getattr(model, "grid", None)
16 if space is None:
17 # Sometimes the space is defined as model.space instead of model.grid
18 space = model.space
19 if isinstance(space, mesa.space.NetworkGrid):
20 _draw_network_grid(space, space_ax, agent_portrayal)
21 elif isinstance(space, mesa.space.ContinuousSpace):
22 _draw_continuous_space(space, space_ax, agent_portrayal)
23 else:
24 _draw_grid(space, space_ax, agent_portrayal)
25 space_ax.set_axis_off()
26 solara.FigureMatplotlib(space_fig, format="png", dependencies=dependencies)
27
28
29 def _draw_grid(space, space_ax, agent_portrayal):
30 def portray(g):
31 x = []
32 y = []
33 s = [] # size
34 c = [] # color
35 for i in range(g.width):
36 for j in range(g.height):
37 content = g._grid[i][j]
38 if not content:
39 continue
40 if not hasattr(content, "__iter__"):
41 # Is a single grid
42 content = [content]
43 for agent in content:
44 data = agent_portrayal(agent)
45 x.append(i)
46 y.append(j)
47 if "size" in data:
48 s.append(data["size"])
49 if "color" in data:
50 c.append(data["color"])
51 out = {"x": x, "y": y}
52 if len(s) > 0:
53 out["s"] = s
54 if len(c) > 0:
55 out["c"] = c
56 return out
57
58 space_ax.scatter(**portray(space))
59
60
61 def _draw_network_grid(space, space_ax, agent_portrayal):
62 graph = space.G
63 pos = nx.spring_layout(graph, seed=0)
64 nx.draw(
65 graph,
66 ax=space_ax,
67 pos=pos,
68 **agent_portrayal(graph),
69 )
70
71
72 def _draw_continuous_space(space, space_ax, agent_portrayal):
73 def portray(space):
74 x = []
75 y = []
76 s = [] # size
77 c = [] # color
78 for agent in space._agent_to_index:
79 data = agent_portrayal(agent)
80 _x, _y = agent.pos
81 x.append(_x)
82 y.append(_y)
83 if "size" in data:
84 s.append(data["size"])
85 if "color" in data:
86 c.append(data["color"])
87 out = {"x": x, "y": y}
88 if len(s) > 0:
89 out["s"] = s
90 if len(c) > 0:
91 out["c"] = c
92 return out
93
94 space_ax.scatter(**portray(space))
95
96
97 def make_plot(model, measure):
98 fig = Figure()
99 ax = fig.subplots()
100 df = model.datacollector.get_model_vars_dataframe()
101 if isinstance(measure, str):
102 ax.plot(df.loc[:, measure])
103 ax.set_ylabel(measure)
104 elif isinstance(measure, dict):
105 for m, color in measure.items():
106 ax.plot(df.loc[:, m], label=m, color=color)
107 fig.legend()
108 elif isinstance(measure, (list, tuple)):
109 for m in measure:
110 ax.plot(df.loc[:, m], label=m)
111 fig.legend()
112 # Set integer x axis
113 ax.xaxis.set_major_locator(MaxNLocator(integer=True))
114 solara.FigureMatplotlib(fig)
115
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/mesa/experimental/components/matplotlib.py b/mesa/experimental/components/matplotlib.py
--- a/mesa/experimental/components/matplotlib.py
+++ b/mesa/experimental/components/matplotlib.py
@@ -55,6 +55,8 @@
out["c"] = c
return out
+ space_ax.set_xlim(-1, space.width)
+ space_ax.set_ylim(-1, space.height)
space_ax.scatter(**portray(space))
@@ -91,6 +93,14 @@
out["c"] = c
return out
+ width = space.x_max - space.x_min
+ x_padding = width / 20
+ height = space.y_max - space.y_min
+ y_padding = height / 20
+ space_ax.set_xlim(space.x_min - x_padding, space.x_max + x_padding)
+ space_ax.set_ylim(space.y_min - y_padding, space.y_max + y_padding)
+ space_ax.scatter(**portray(space))
+
space_ax.scatter(**portray(space))
|
{"golden_diff": "diff --git a/mesa/experimental/components/matplotlib.py b/mesa/experimental/components/matplotlib.py\n--- a/mesa/experimental/components/matplotlib.py\n+++ b/mesa/experimental/components/matplotlib.py\n@@ -55,6 +55,8 @@\n out[\"c\"] = c\n return out\n \n+ space_ax.set_xlim(-1, space.width)\n+ space_ax.set_ylim(-1, space.height)\n space_ax.scatter(**portray(space))\n \n \n@@ -91,6 +93,14 @@\n out[\"c\"] = c\n return out\n \n+ width = space.x_max - space.x_min\n+ x_padding = width / 20\n+ height = space.y_max - space.y_min\n+ y_padding = height / 20\n+ space_ax.set_xlim(space.x_min - x_padding, space.x_max + x_padding)\n+ space_ax.set_ylim(space.y_min - y_padding, space.y_max + y_padding)\n+ space_ax.scatter(**portray(space))\n+\n space_ax.scatter(**portray(space))\n", "issue": "JupyterViz space view limits\n**What's the problem this feature will solve?**\r\nAt the moment, when drawing spaces with JupyterViz, the axes limits are left to be automatically determined by Matplotlib. Would it not make more sense (in particular of continuous spaces, to derive xlim and xmax from the space itself?\r\n\r\n**Describe the solution you'd like**\r\nThe simplest fix would be to adapt `_draw_grid` to use `space.width` and `space.height` while `_draw_continuos_space` uses `space.x_min`, `space.x_max`, `space.y_min`, and `space.y_max`. Are there reasons where the user might want to rely on matpltolib's automatic determination of the bounds? If so, then a slightly more sophisticated solution would be needed where kwargs are passed to from jupyterviz to the underlying space_drawer function. So you would add a `space_drawer_kwargs` keyword argument.\r\n\r\n\n", "before_files": [{"content": "from typing import Optional\n\nimport networkx as nx\nimport solara\nfrom matplotlib.figure import Figure\nfrom matplotlib.ticker import MaxNLocator\n\nimport mesa\n\n\[email protected]\ndef SpaceMatplotlib(model, agent_portrayal, dependencies: Optional[list[any]] = None):\n space_fig = Figure()\n space_ax = space_fig.subplots()\n space = getattr(model, \"grid\", None)\n if space is None:\n # Sometimes the space is defined as model.space instead of model.grid\n space = model.space\n if isinstance(space, mesa.space.NetworkGrid):\n _draw_network_grid(space, space_ax, agent_portrayal)\n elif isinstance(space, mesa.space.ContinuousSpace):\n _draw_continuous_space(space, space_ax, agent_portrayal)\n else:\n _draw_grid(space, space_ax, agent_portrayal)\n space_ax.set_axis_off()\n solara.FigureMatplotlib(space_fig, format=\"png\", dependencies=dependencies)\n\n\ndef _draw_grid(space, space_ax, agent_portrayal):\n def portray(g):\n x = []\n y = []\n s = [] # size\n c = [] # color\n for i in range(g.width):\n for j in range(g.height):\n content = g._grid[i][j]\n if not content:\n continue\n if not hasattr(content, \"__iter__\"):\n # Is a single grid\n content = [content]\n for agent in content:\n data = agent_portrayal(agent)\n x.append(i)\n y.append(j)\n if \"size\" in data:\n s.append(data[\"size\"])\n if \"color\" in data:\n c.append(data[\"color\"])\n out = {\"x\": x, \"y\": y}\n if len(s) > 0:\n out[\"s\"] = s\n if len(c) > 0:\n out[\"c\"] = c\n return out\n\n space_ax.scatter(**portray(space))\n\n\ndef _draw_network_grid(space, space_ax, agent_portrayal):\n graph = space.G\n pos = nx.spring_layout(graph, seed=0)\n nx.draw(\n graph,\n ax=space_ax,\n pos=pos,\n **agent_portrayal(graph),\n )\n\n\ndef _draw_continuous_space(space, space_ax, agent_portrayal):\n def portray(space):\n x = []\n y = []\n s = [] # size\n c = [] # color\n for agent in space._agent_to_index:\n data = agent_portrayal(agent)\n _x, _y = agent.pos\n x.append(_x)\n y.append(_y)\n if \"size\" in data:\n s.append(data[\"size\"])\n if \"color\" in data:\n c.append(data[\"color\"])\n out = {\"x\": x, \"y\": y}\n if len(s) > 0:\n out[\"s\"] = s\n if len(c) > 0:\n out[\"c\"] = c\n return out\n\n space_ax.scatter(**portray(space))\n\n\ndef make_plot(model, measure):\n fig = Figure()\n ax = fig.subplots()\n df = model.datacollector.get_model_vars_dataframe()\n if isinstance(measure, str):\n ax.plot(df.loc[:, measure])\n ax.set_ylabel(measure)\n elif isinstance(measure, dict):\n for m, color in measure.items():\n ax.plot(df.loc[:, m], label=m, color=color)\n fig.legend()\n elif isinstance(measure, (list, tuple)):\n for m in measure:\n ax.plot(df.loc[:, m], label=m)\n fig.legend()\n # Set integer x axis\n ax.xaxis.set_major_locator(MaxNLocator(integer=True))\n solara.FigureMatplotlib(fig)\n", "path": "mesa/experimental/components/matplotlib.py"}], "after_files": [{"content": "from typing import Optional\n\nimport networkx as nx\nimport solara\nfrom matplotlib.figure import Figure\nfrom matplotlib.ticker import MaxNLocator\n\nimport mesa\n\n\[email protected]\ndef SpaceMatplotlib(model, agent_portrayal, dependencies: Optional[list[any]] = None):\n space_fig = Figure()\n space_ax = space_fig.subplots()\n space = getattr(model, \"grid\", None)\n if space is None:\n # Sometimes the space is defined as model.space instead of model.grid\n space = model.space\n if isinstance(space, mesa.space.NetworkGrid):\n _draw_network_grid(space, space_ax, agent_portrayal)\n elif isinstance(space, mesa.space.ContinuousSpace):\n _draw_continuous_space(space, space_ax, agent_portrayal)\n else:\n _draw_grid(space, space_ax, agent_portrayal)\n space_ax.set_axis_off()\n solara.FigureMatplotlib(space_fig, format=\"png\", dependencies=dependencies)\n\n\ndef _draw_grid(space, space_ax, agent_portrayal):\n def portray(g):\n x = []\n y = []\n s = [] # size\n c = [] # color\n for i in range(g.width):\n for j in range(g.height):\n content = g._grid[i][j]\n if not content:\n continue\n if not hasattr(content, \"__iter__\"):\n # Is a single grid\n content = [content]\n for agent in content:\n data = agent_portrayal(agent)\n x.append(i)\n y.append(j)\n if \"size\" in data:\n s.append(data[\"size\"])\n if \"color\" in data:\n c.append(data[\"color\"])\n out = {\"x\": x, \"y\": y}\n if len(s) > 0:\n out[\"s\"] = s\n if len(c) > 0:\n out[\"c\"] = c\n return out\n\n space_ax.set_xlim(-1, space.width)\n space_ax.set_ylim(-1, space.height)\n space_ax.scatter(**portray(space))\n\n\ndef _draw_network_grid(space, space_ax, agent_portrayal):\n graph = space.G\n pos = nx.spring_layout(graph, seed=0)\n nx.draw(\n graph,\n ax=space_ax,\n pos=pos,\n **agent_portrayal(graph),\n )\n\n\ndef _draw_continuous_space(space, space_ax, agent_portrayal):\n def portray(space):\n x = []\n y = []\n s = [] # size\n c = [] # color\n for agent in space._agent_to_index:\n data = agent_portrayal(agent)\n _x, _y = agent.pos\n x.append(_x)\n y.append(_y)\n if \"size\" in data:\n s.append(data[\"size\"])\n if \"color\" in data:\n c.append(data[\"color\"])\n out = {\"x\": x, \"y\": y}\n if len(s) > 0:\n out[\"s\"] = s\n if len(c) > 0:\n out[\"c\"] = c\n return out\n\n width = space.x_max - space.x_min\n x_padding = width / 20\n height = space.y_max - space.y_min\n y_padding = height / 20\n space_ax.set_xlim(space.x_min - x_padding, space.x_max + x_padding)\n space_ax.set_ylim(space.y_min - y_padding, space.y_max + y_padding)\n space_ax.scatter(**portray(space))\n\n space_ax.scatter(**portray(space))\n\n\ndef make_plot(model, measure):\n fig = Figure()\n ax = fig.subplots()\n df = model.datacollector.get_model_vars_dataframe()\n if isinstance(measure, str):\n ax.plot(df.loc[:, measure])\n ax.set_ylabel(measure)\n elif isinstance(measure, dict):\n for m, color in measure.items():\n ax.plot(df.loc[:, m], label=m, color=color)\n fig.legend()\n elif isinstance(measure, (list, tuple)):\n for m in measure:\n ax.plot(df.loc[:, m], label=m)\n fig.legend()\n # Set integer x axis\n ax.xaxis.set_major_locator(MaxNLocator(integer=True))\n solara.FigureMatplotlib(fig)\n", "path": "mesa/experimental/components/matplotlib.py"}]}
| 1,507 | 233 |
gh_patches_debug_39815
|
rasdani/github-patches
|
git_diff
|
nilearn__nilearn-773
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Misleading (wrong?) code in Oasis VBM example
As pointed out on http://neurostars.org/p/3388/, the Oasis VBM code applies the NiftiMasker twice, and thus ends up smoothing the data twice.
It does that to do screening of low variance features. I think that the right way to do this would be to use a sklearn.feature_selection.VarianceThreshold in the pipeline.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `examples/decoding/plot_oasis_vbm.py`
Content:
```
1 """
2 Voxel-Based Morphometry on Oasis dataset
3 ========================================
4
5 This example uses Voxel-Based Morphometry (VBM) to study the relationship
6 between aging and gray matter density.
7
8 The data come from the `OASIS <http://www.oasis-brains.org/>`_ project.
9 If you use it, you need to agree with the data usage agreement available
10 on the website.
11
12 It has been run through a standard VBM pipeline (using SPM8 and
13 NewSegment) to create VBM maps, which we study here.
14
15 Predictive modeling analysis: VBM bio-markers of aging?
16 --------------------------------------------------------
17
18 We run a standard SVM-ANOVA nilearn pipeline to predict age from the VBM
19 data. We use only 100 subjects from the OASIS dataset to limit the memory
20 usage.
21
22 Note that for an actual predictive modeling study of aging, the study
23 should be ran on the full set of subjects. Also, parameters such as the
24 smoothing applied to the data and the number of features selected by the
25 Anova step should be set by nested cross-validation, as they impact
26 significantly the prediction score.
27
28 Brain mapping with mass univariate
29 -----------------------------------
30
31 SVM weights are very noisy, partly because heavy smoothing is detrimental
32 for the prediction here. A standard analysis using mass-univariate GLM
33 (here permuted to have exact correction for multiple comparisons) gives a
34 much clearer view of the important regions.
35
36 ____
37
38 """
39 # Authors: Elvis Dhomatob, <[email protected]>, Apr. 2014
40 # Virgile Fritsch, <[email protected]>, Apr 2014
41 # Gael Varoquaux, Apr 2014
42 import numpy as np
43 from scipy import linalg
44 import matplotlib.pyplot as plt
45 from nilearn import datasets
46 from nilearn.input_data import NiftiMasker
47
48 n_subjects = 100 # more subjects requires more memory
49
50 ### Load Oasis dataset ########################################################
51 oasis_dataset = datasets.fetch_oasis_vbm(n_subjects=n_subjects)
52 gray_matter_map_filenames = oasis_dataset.gray_matter_maps
53 age = oasis_dataset.ext_vars['age'].astype(float)
54
55 # print basic information on the dataset
56 print('First gray-matter anatomy image (3D) is located at: %s' %
57 oasis_dataset.gray_matter_maps[0]) # 3D data
58 print('First white-matter anatomy image (3D) is located at: %s' %
59 oasis_dataset.white_matter_maps[0]) # 3D data
60
61 ### Preprocess data ###########################################################
62 nifti_masker = NiftiMasker(
63 standardize=False,
64 smoothing_fwhm=2,
65 memory='nilearn_cache') # cache options
66 # remove features with too low between-subject variance
67 gm_maps_masked = nifti_masker.fit_transform(gray_matter_map_filenames)
68 gm_maps_masked[:, gm_maps_masked.var(0) < 0.01] = 0.
69 # final masking
70 new_images = nifti_masker.inverse_transform(gm_maps_masked)
71 gm_maps_masked = nifti_masker.fit_transform(new_images)
72 n_samples, n_features = gm_maps_masked.shape
73 print("%d samples, %d features" % (n_subjects, n_features))
74
75 ### Prediction with SVR #######################################################
76 print("ANOVA + SVR")
77 # Define the prediction function to be used.
78 # Here we use a Support Vector Classification, with a linear kernel
79 from sklearn.svm import SVR
80 svr = SVR(kernel='linear')
81
82 # Dimension reduction
83 from sklearn.feature_selection import SelectKBest, f_regression
84
85 # Here we use a classical univariate feature selection based on F-test,
86 # namely Anova.
87 feature_selection = SelectKBest(f_regression, k=2000)
88
89 # We have our predictor (SVR), our feature selection (SelectKBest), and now,
90 # we can plug them together in a *pipeline* that performs the two operations
91 # successively:
92 from sklearn.pipeline import Pipeline
93 anova_svr = Pipeline([('anova', feature_selection), ('svr', svr)])
94
95 ### Fit and predict
96 anova_svr.fit(gm_maps_masked, age)
97 age_pred = anova_svr.predict(gm_maps_masked)
98
99 # Visualization
100 # Look at the SVR's discriminating weights
101 coef = svr.coef_
102 # reverse feature selection
103 coef = feature_selection.inverse_transform(coef)
104 # reverse masking
105 weight_img = nifti_masker.inverse_transform(coef)
106
107 # Create the figure
108 from nilearn.plotting import plot_stat_map, show
109 bg_filename = gray_matter_map_filenames[0]
110 z_slice = 0
111 from nilearn.image.resampling import coord_transform
112 affine = weight_img.get_affine()
113 _, _, k_slice = coord_transform(0, 0, z_slice,
114 linalg.inv(affine))
115 k_slice = np.round(k_slice)
116
117 fig = plt.figure(figsize=(5.5, 7.5), facecolor='k')
118 weight_slice_data = weight_img.get_data()[..., k_slice, 0]
119 vmax = max(-np.min(weight_slice_data), np.max(weight_slice_data)) * 0.5
120 display = plot_stat_map(weight_img, bg_img=bg_filename,
121 display_mode='z', cut_coords=[z_slice],
122 figure=fig, vmax=vmax)
123 display.title('SVM weights', y=1.2)
124
125 # Measure accuracy with cross validation
126 from sklearn.cross_validation import cross_val_score
127 cv_scores = cross_val_score(anova_svr, gm_maps_masked, age)
128
129 # Return the corresponding mean prediction accuracy
130 prediction_accuracy = np.mean(cv_scores)
131 print("=== ANOVA ===")
132 print("Prediction accuracy: %f" % prediction_accuracy)
133 print("")
134
135 ### Inference with massively univariate model #################################
136 print("Massively univariate model")
137
138 # Statistical inference
139 from nilearn.mass_univariate import permuted_ols
140 neg_log_pvals, t_scores_original_data, _ = permuted_ols(
141 age, gm_maps_masked, # + intercept as a covariate by default
142 n_perm=1000, # 1,000 in the interest of time; 10000 would be better
143 n_jobs=1) # can be changed to use more CPUs
144 signed_neg_log_pvals = neg_log_pvals * np.sign(t_scores_original_data)
145 signed_neg_log_pvals_unmasked = nifti_masker.inverse_transform(
146 signed_neg_log_pvals)
147
148 # Show results
149 threshold = -np.log10(0.1) # 10% corrected
150
151 fig = plt.figure(figsize=(5.5, 7.5), facecolor='k')
152
153 display = plot_stat_map(signed_neg_log_pvals_unmasked, bg_img=bg_filename,
154 threshold=threshold, cmap=plt.cm.RdBu_r,
155 display_mode='z', cut_coords=[z_slice],
156 figure=fig)
157 title = ('Negative $\log_{10}$ p-values'
158 '\n(Non-parametric + max-type correction)')
159 display.title(title, y=1.2)
160
161 signed_neg_log_pvals_slice_data = \
162 signed_neg_log_pvals_unmasked.get_data()[..., k_slice, 0]
163 n_detections = (np.abs(signed_neg_log_pvals_slice_data) > threshold).sum()
164 print('\n%d detections' % n_detections)
165
166 show()
167
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/examples/decoding/plot_oasis_vbm.py b/examples/decoding/plot_oasis_vbm.py
--- a/examples/decoding/plot_oasis_vbm.py
+++ b/examples/decoding/plot_oasis_vbm.py
@@ -63,12 +63,7 @@
standardize=False,
smoothing_fwhm=2,
memory='nilearn_cache') # cache options
-# remove features with too low between-subject variance
gm_maps_masked = nifti_masker.fit_transform(gray_matter_map_filenames)
-gm_maps_masked[:, gm_maps_masked.var(0) < 0.01] = 0.
-# final masking
-new_images = nifti_masker.inverse_transform(gm_maps_masked)
-gm_maps_masked = nifti_masker.fit_transform(new_images)
n_samples, n_features = gm_maps_masked.shape
print("%d samples, %d features" % (n_subjects, n_features))
@@ -80,7 +75,11 @@
svr = SVR(kernel='linear')
# Dimension reduction
-from sklearn.feature_selection import SelectKBest, f_regression
+from sklearn.feature_selection import VarianceThreshold, SelectKBest, \
+ f_regression
+
+# Remove features with too low between-subject variance
+variance_threshold = VarianceThreshold(threshold=.01)
# Here we use a classical univariate feature selection based on F-test,
# namely Anova.
@@ -90,7 +89,10 @@
# we can plug them together in a *pipeline* that performs the two operations
# successively:
from sklearn.pipeline import Pipeline
-anova_svr = Pipeline([('anova', feature_selection), ('svr', svr)])
+anova_svr = Pipeline([
+ ('variance_threshold', variance_threshold),
+ ('anova', feature_selection),
+ ('svr', svr)])
### Fit and predict
anova_svr.fit(gm_maps_masked, age)
@@ -101,6 +103,8 @@
coef = svr.coef_
# reverse feature selection
coef = feature_selection.inverse_transform(coef)
+# reverse variance threshold
+coef = variance_threshold.inverse_transform(coef)
# reverse masking
weight_img = nifti_masker.inverse_transform(coef)
@@ -137,13 +141,14 @@
# Statistical inference
from nilearn.mass_univariate import permuted_ols
+data = variance_threshold.fit_transform(gm_maps_masked)
neg_log_pvals, t_scores_original_data, _ = permuted_ols(
- age, gm_maps_masked, # + intercept as a covariate by default
- n_perm=1000, # 1,000 in the interest of time; 10000 would be better
+ age, data, # + intercept as a covariate by default
+ n_perm=2000, # 1,000 in the interest of time; 10000 would be better
n_jobs=1) # can be changed to use more CPUs
signed_neg_log_pvals = neg_log_pvals * np.sign(t_scores_original_data)
signed_neg_log_pvals_unmasked = nifti_masker.inverse_transform(
- signed_neg_log_pvals)
+ variance_threshold.inverse_transform(signed_neg_log_pvals))
# Show results
threshold = -np.log10(0.1) # 10% corrected
|
{"golden_diff": "diff --git a/examples/decoding/plot_oasis_vbm.py b/examples/decoding/plot_oasis_vbm.py\n--- a/examples/decoding/plot_oasis_vbm.py\n+++ b/examples/decoding/plot_oasis_vbm.py\n@@ -63,12 +63,7 @@\n standardize=False,\n smoothing_fwhm=2,\n memory='nilearn_cache') # cache options\n-# remove features with too low between-subject variance\n gm_maps_masked = nifti_masker.fit_transform(gray_matter_map_filenames)\n-gm_maps_masked[:, gm_maps_masked.var(0) < 0.01] = 0.\n-# final masking\n-new_images = nifti_masker.inverse_transform(gm_maps_masked)\n-gm_maps_masked = nifti_masker.fit_transform(new_images)\n n_samples, n_features = gm_maps_masked.shape\n print(\"%d samples, %d features\" % (n_subjects, n_features))\n \n@@ -80,7 +75,11 @@\n svr = SVR(kernel='linear')\n \n # Dimension reduction\n-from sklearn.feature_selection import SelectKBest, f_regression\n+from sklearn.feature_selection import VarianceThreshold, SelectKBest, \\\n+ f_regression\n+\n+# Remove features with too low between-subject variance\n+variance_threshold = VarianceThreshold(threshold=.01)\n \n # Here we use a classical univariate feature selection based on F-test,\n # namely Anova.\n@@ -90,7 +89,10 @@\n # we can plug them together in a *pipeline* that performs the two operations\n # successively:\n from sklearn.pipeline import Pipeline\n-anova_svr = Pipeline([('anova', feature_selection), ('svr', svr)])\n+anova_svr = Pipeline([\n+ ('variance_threshold', variance_threshold),\n+ ('anova', feature_selection),\n+ ('svr', svr)])\n \n ### Fit and predict\n anova_svr.fit(gm_maps_masked, age)\n@@ -101,6 +103,8 @@\n coef = svr.coef_\n # reverse feature selection\n coef = feature_selection.inverse_transform(coef)\n+# reverse variance threshold\n+coef = variance_threshold.inverse_transform(coef)\n # reverse masking\n weight_img = nifti_masker.inverse_transform(coef)\n \n@@ -137,13 +141,14 @@\n \n # Statistical inference\n from nilearn.mass_univariate import permuted_ols\n+data = variance_threshold.fit_transform(gm_maps_masked)\n neg_log_pvals, t_scores_original_data, _ = permuted_ols(\n- age, gm_maps_masked, # + intercept as a covariate by default\n- n_perm=1000, # 1,000 in the interest of time; 10000 would be better\n+ age, data, # + intercept as a covariate by default\n+ n_perm=2000, # 1,000 in the interest of time; 10000 would be better\n n_jobs=1) # can be changed to use more CPUs\n signed_neg_log_pvals = neg_log_pvals * np.sign(t_scores_original_data)\n signed_neg_log_pvals_unmasked = nifti_masker.inverse_transform(\n- signed_neg_log_pvals)\n+ variance_threshold.inverse_transform(signed_neg_log_pvals))\n \n # Show results\n threshold = -np.log10(0.1) # 10% corrected\n", "issue": "Misleading (wrong?) code in Oasis VBM example\nAs pointed out on http://neurostars.org/p/3388/, the Oasis VBM code applies the NiftiMasker twice, and thus ends up smoothing the data twice.\n\nIt does that to do screening of low variance features. I think that the right way to do this would be to use a sklearn.feature_selection.VarianceThreshold in the pipeline.\n\n", "before_files": [{"content": "\"\"\"\nVoxel-Based Morphometry on Oasis dataset\n========================================\n\nThis example uses Voxel-Based Morphometry (VBM) to study the relationship\nbetween aging and gray matter density.\n\nThe data come from the `OASIS <http://www.oasis-brains.org/>`_ project.\nIf you use it, you need to agree with the data usage agreement available\non the website.\n\nIt has been run through a standard VBM pipeline (using SPM8 and\nNewSegment) to create VBM maps, which we study here.\n\nPredictive modeling analysis: VBM bio-markers of aging?\n--------------------------------------------------------\n\nWe run a standard SVM-ANOVA nilearn pipeline to predict age from the VBM\ndata. We use only 100 subjects from the OASIS dataset to limit the memory\nusage.\n\nNote that for an actual predictive modeling study of aging, the study\nshould be ran on the full set of subjects. Also, parameters such as the\nsmoothing applied to the data and the number of features selected by the\nAnova step should be set by nested cross-validation, as they impact\nsignificantly the prediction score.\n\nBrain mapping with mass univariate\n-----------------------------------\n\nSVM weights are very noisy, partly because heavy smoothing is detrimental\nfor the prediction here. A standard analysis using mass-univariate GLM\n(here permuted to have exact correction for multiple comparisons) gives a\nmuch clearer view of the important regions.\n\n____\n\n\"\"\"\n# Authors: Elvis Dhomatob, <[email protected]>, Apr. 2014\n# Virgile Fritsch, <[email protected]>, Apr 2014\n# Gael Varoquaux, Apr 2014\nimport numpy as np\nfrom scipy import linalg\nimport matplotlib.pyplot as plt\nfrom nilearn import datasets\nfrom nilearn.input_data import NiftiMasker\n\nn_subjects = 100 # more subjects requires more memory\n\n### Load Oasis dataset ########################################################\noasis_dataset = datasets.fetch_oasis_vbm(n_subjects=n_subjects)\ngray_matter_map_filenames = oasis_dataset.gray_matter_maps\nage = oasis_dataset.ext_vars['age'].astype(float)\n\n# print basic information on the dataset\nprint('First gray-matter anatomy image (3D) is located at: %s' %\n oasis_dataset.gray_matter_maps[0]) # 3D data\nprint('First white-matter anatomy image (3D) is located at: %s' %\n oasis_dataset.white_matter_maps[0]) # 3D data\n\n### Preprocess data ###########################################################\nnifti_masker = NiftiMasker(\n standardize=False,\n smoothing_fwhm=2,\n memory='nilearn_cache') # cache options\n# remove features with too low between-subject variance\ngm_maps_masked = nifti_masker.fit_transform(gray_matter_map_filenames)\ngm_maps_masked[:, gm_maps_masked.var(0) < 0.01] = 0.\n# final masking\nnew_images = nifti_masker.inverse_transform(gm_maps_masked)\ngm_maps_masked = nifti_masker.fit_transform(new_images)\nn_samples, n_features = gm_maps_masked.shape\nprint(\"%d samples, %d features\" % (n_subjects, n_features))\n\n### Prediction with SVR #######################################################\nprint(\"ANOVA + SVR\")\n# Define the prediction function to be used.\n# Here we use a Support Vector Classification, with a linear kernel\nfrom sklearn.svm import SVR\nsvr = SVR(kernel='linear')\n\n# Dimension reduction\nfrom sklearn.feature_selection import SelectKBest, f_regression\n\n# Here we use a classical univariate feature selection based on F-test,\n# namely Anova.\nfeature_selection = SelectKBest(f_regression, k=2000)\n\n# We have our predictor (SVR), our feature selection (SelectKBest), and now,\n# we can plug them together in a *pipeline* that performs the two operations\n# successively:\nfrom sklearn.pipeline import Pipeline\nanova_svr = Pipeline([('anova', feature_selection), ('svr', svr)])\n\n### Fit and predict\nanova_svr.fit(gm_maps_masked, age)\nage_pred = anova_svr.predict(gm_maps_masked)\n\n# Visualization\n# Look at the SVR's discriminating weights\ncoef = svr.coef_\n# reverse feature selection\ncoef = feature_selection.inverse_transform(coef)\n# reverse masking\nweight_img = nifti_masker.inverse_transform(coef)\n\n# Create the figure\nfrom nilearn.plotting import plot_stat_map, show\nbg_filename = gray_matter_map_filenames[0]\nz_slice = 0\nfrom nilearn.image.resampling import coord_transform\naffine = weight_img.get_affine()\n_, _, k_slice = coord_transform(0, 0, z_slice,\n linalg.inv(affine))\nk_slice = np.round(k_slice)\n\nfig = plt.figure(figsize=(5.5, 7.5), facecolor='k')\nweight_slice_data = weight_img.get_data()[..., k_slice, 0]\nvmax = max(-np.min(weight_slice_data), np.max(weight_slice_data)) * 0.5\ndisplay = plot_stat_map(weight_img, bg_img=bg_filename,\n display_mode='z', cut_coords=[z_slice],\n figure=fig, vmax=vmax)\ndisplay.title('SVM weights', y=1.2)\n\n# Measure accuracy with cross validation\nfrom sklearn.cross_validation import cross_val_score\ncv_scores = cross_val_score(anova_svr, gm_maps_masked, age)\n\n# Return the corresponding mean prediction accuracy\nprediction_accuracy = np.mean(cv_scores)\nprint(\"=== ANOVA ===\")\nprint(\"Prediction accuracy: %f\" % prediction_accuracy)\nprint(\"\")\n\n### Inference with massively univariate model #################################\nprint(\"Massively univariate model\")\n\n# Statistical inference\nfrom nilearn.mass_univariate import permuted_ols\nneg_log_pvals, t_scores_original_data, _ = permuted_ols(\n age, gm_maps_masked, # + intercept as a covariate by default\n n_perm=1000, # 1,000 in the interest of time; 10000 would be better\n n_jobs=1) # can be changed to use more CPUs\nsigned_neg_log_pvals = neg_log_pvals * np.sign(t_scores_original_data)\nsigned_neg_log_pvals_unmasked = nifti_masker.inverse_transform(\n signed_neg_log_pvals)\n\n# Show results\nthreshold = -np.log10(0.1) # 10% corrected\n\nfig = plt.figure(figsize=(5.5, 7.5), facecolor='k')\n\ndisplay = plot_stat_map(signed_neg_log_pvals_unmasked, bg_img=bg_filename,\n threshold=threshold, cmap=plt.cm.RdBu_r,\n display_mode='z', cut_coords=[z_slice],\n figure=fig)\ntitle = ('Negative $\\log_{10}$ p-values'\n '\\n(Non-parametric + max-type correction)')\ndisplay.title(title, y=1.2)\n\nsigned_neg_log_pvals_slice_data = \\\n signed_neg_log_pvals_unmasked.get_data()[..., k_slice, 0]\nn_detections = (np.abs(signed_neg_log_pvals_slice_data) > threshold).sum()\nprint('\\n%d detections' % n_detections)\n\nshow()\n", "path": "examples/decoding/plot_oasis_vbm.py"}], "after_files": [{"content": "\"\"\"\nVoxel-Based Morphometry on Oasis dataset\n========================================\n\nThis example uses Voxel-Based Morphometry (VBM) to study the relationship\nbetween aging and gray matter density.\n\nThe data come from the `OASIS <http://www.oasis-brains.org/>`_ project.\nIf you use it, you need to agree with the data usage agreement available\non the website.\n\nIt has been run through a standard VBM pipeline (using SPM8 and\nNewSegment) to create VBM maps, which we study here.\n\nPredictive modeling analysis: VBM bio-markers of aging?\n--------------------------------------------------------\n\nWe run a standard SVM-ANOVA nilearn pipeline to predict age from the VBM\ndata. We use only 100 subjects from the OASIS dataset to limit the memory\nusage.\n\nNote that for an actual predictive modeling study of aging, the study\nshould be ran on the full set of subjects. Also, parameters such as the\nsmoothing applied to the data and the number of features selected by the\nAnova step should be set by nested cross-validation, as they impact\nsignificantly the prediction score.\n\nBrain mapping with mass univariate\n-----------------------------------\n\nSVM weights are very noisy, partly because heavy smoothing is detrimental\nfor the prediction here. A standard analysis using mass-univariate GLM\n(here permuted to have exact correction for multiple comparisons) gives a\nmuch clearer view of the important regions.\n\n____\n\n\"\"\"\n# Authors: Elvis Dhomatob, <[email protected]>, Apr. 2014\n# Virgile Fritsch, <[email protected]>, Apr 2014\n# Gael Varoquaux, Apr 2014\nimport numpy as np\nfrom scipy import linalg\nimport matplotlib.pyplot as plt\nfrom nilearn import datasets\nfrom nilearn.input_data import NiftiMasker\n\nn_subjects = 100 # more subjects requires more memory\n\n### Load Oasis dataset ########################################################\noasis_dataset = datasets.fetch_oasis_vbm(n_subjects=n_subjects)\ngray_matter_map_filenames = oasis_dataset.gray_matter_maps\nage = oasis_dataset.ext_vars['age'].astype(float)\n\n# print basic information on the dataset\nprint('First gray-matter anatomy image (3D) is located at: %s' %\n oasis_dataset.gray_matter_maps[0]) # 3D data\nprint('First white-matter anatomy image (3D) is located at: %s' %\n oasis_dataset.white_matter_maps[0]) # 3D data\n\n### Preprocess data ###########################################################\nnifti_masker = NiftiMasker(\n standardize=False,\n smoothing_fwhm=2,\n memory='nilearn_cache') # cache options\ngm_maps_masked = nifti_masker.fit_transform(gray_matter_map_filenames)\nn_samples, n_features = gm_maps_masked.shape\nprint(\"%d samples, %d features\" % (n_subjects, n_features))\n\n### Prediction with SVR #######################################################\nprint(\"ANOVA + SVR\")\n# Define the prediction function to be used.\n# Here we use a Support Vector Classification, with a linear kernel\nfrom sklearn.svm import SVR\nsvr = SVR(kernel='linear')\n\n# Dimension reduction\nfrom sklearn.feature_selection import VarianceThreshold, SelectKBest, \\\n f_regression\n\n# Remove features with too low between-subject variance\nvariance_threshold = VarianceThreshold(threshold=.01)\n\n# Here we use a classical univariate feature selection based on F-test,\n# namely Anova.\nfeature_selection = SelectKBest(f_regression, k=2000)\n\n# We have our predictor (SVR), our feature selection (SelectKBest), and now,\n# we can plug them together in a *pipeline* that performs the two operations\n# successively:\nfrom sklearn.pipeline import Pipeline\nanova_svr = Pipeline([\n ('variance_threshold', variance_threshold),\n ('anova', feature_selection),\n ('svr', svr)])\n\n### Fit and predict\nanova_svr.fit(gm_maps_masked, age)\nage_pred = anova_svr.predict(gm_maps_masked)\n\n# Visualization\n# Look at the SVR's discriminating weights\ncoef = svr.coef_\n# reverse feature selection\ncoef = feature_selection.inverse_transform(coef)\n# reverse variance threshold\ncoef = variance_threshold.inverse_transform(coef)\n# reverse masking\nweight_img = nifti_masker.inverse_transform(coef)\n\n# Create the figure\nfrom nilearn.plotting import plot_stat_map, show\nbg_filename = gray_matter_map_filenames[0]\nz_slice = 0\nfrom nilearn.image.resampling import coord_transform\naffine = weight_img.get_affine()\n_, _, k_slice = coord_transform(0, 0, z_slice,\n linalg.inv(affine))\nk_slice = np.round(k_slice)\n\nfig = plt.figure(figsize=(5.5, 7.5), facecolor='k')\nweight_slice_data = weight_img.get_data()[..., k_slice, 0]\nvmax = max(-np.min(weight_slice_data), np.max(weight_slice_data)) * 0.5\ndisplay = plot_stat_map(weight_img, bg_img=bg_filename,\n display_mode='z', cut_coords=[z_slice],\n figure=fig, vmax=vmax)\ndisplay.title('SVM weights', y=1.2)\n\n# Measure accuracy with cross validation\nfrom sklearn.cross_validation import cross_val_score\ncv_scores = cross_val_score(anova_svr, gm_maps_masked, age)\n\n# Return the corresponding mean prediction accuracy\nprediction_accuracy = np.mean(cv_scores)\nprint(\"=== ANOVA ===\")\nprint(\"Prediction accuracy: %f\" % prediction_accuracy)\nprint(\"\")\n\n### Inference with massively univariate model #################################\nprint(\"Massively univariate model\")\n\n# Statistical inference\nfrom nilearn.mass_univariate import permuted_ols\ndata = variance_threshold.fit_transform(gm_maps_masked)\nneg_log_pvals, t_scores_original_data, _ = permuted_ols(\n age, data, # + intercept as a covariate by default\n n_perm=2000, # 1,000 in the interest of time; 10000 would be better\n n_jobs=1) # can be changed to use more CPUs\nsigned_neg_log_pvals = neg_log_pvals * np.sign(t_scores_original_data)\nsigned_neg_log_pvals_unmasked = nifti_masker.inverse_transform(\n variance_threshold.inverse_transform(signed_neg_log_pvals))\n\n# Show results\nthreshold = -np.log10(0.1) # 10% corrected\n\nfig = plt.figure(figsize=(5.5, 7.5), facecolor='k')\n\ndisplay = plot_stat_map(signed_neg_log_pvals_unmasked, bg_img=bg_filename,\n threshold=threshold, cmap=plt.cm.RdBu_r,\n display_mode='z', cut_coords=[z_slice],\n figure=fig)\ntitle = ('Negative $\\log_{10}$ p-values'\n '\\n(Non-parametric + max-type correction)')\ndisplay.title(title, y=1.2)\n\nsigned_neg_log_pvals_slice_data = \\\n signed_neg_log_pvals_unmasked.get_data()[..., k_slice, 0]\nn_detections = (np.abs(signed_neg_log_pvals_slice_data) > threshold).sum()\nprint('\\n%d detections' % n_detections)\n\nshow()\n", "path": "examples/decoding/plot_oasis_vbm.py"}]}
| 2,362 | 769 |
gh_patches_debug_27391
|
rasdani/github-patches
|
git_diff
|
jupyterhub__jupyterhub-250
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Admin panel - Add ability to import multiple users at once
It would be nice to add a list of users as a bulk list of newline delimited users.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `jupyterhub/apihandlers/users.py`
Content:
```
1 """User handlers"""
2
3 # Copyright (c) Jupyter Development Team.
4 # Distributed under the terms of the Modified BSD License.
5
6 import json
7
8 from tornado import gen, web
9
10 from .. import orm
11 from ..utils import admin_only
12 from .base import APIHandler
13
14
15 class UserListAPIHandler(APIHandler):
16 @admin_only
17 def get(self):
18 users = self.db.query(orm.User)
19 data = [ self.user_model(u) for u in users ]
20 self.write(json.dumps(data))
21
22
23 def admin_or_self(method):
24 """Decorator for restricting access to either the target user or admin"""
25 def m(self, name):
26 current = self.get_current_user()
27 if current is None:
28 raise web.HTTPError(403)
29 if not (current.name == name or current.admin):
30 raise web.HTTPError(403)
31
32 # raise 404 if not found
33 if not self.find_user(name):
34 raise web.HTTPError(404)
35 return method(self, name)
36 return m
37
38 class UserAPIHandler(APIHandler):
39
40 @admin_or_self
41 def get(self, name):
42 user = self.find_user(name)
43 self.write(json.dumps(self.user_model(user)))
44
45 @admin_only
46 @gen.coroutine
47 def post(self, name):
48 data = self.get_json_body()
49 user = self.find_user(name)
50 if user is not None:
51 raise web.HTTPError(400, "User %s already exists" % name)
52
53 user = self.user_from_username(name)
54 if data:
55 self._check_user_model(data)
56 if 'admin' in data:
57 user.admin = data['admin']
58 self.db.commit()
59
60 try:
61 yield gen.maybe_future(self.authenticator.add_user(user))
62 except Exception:
63 self.log.error("Failed to create user: %s" % name, exc_info=True)
64 self.db.delete(user)
65 self.db.commit()
66 raise web.HTTPError(400, "Failed to create user: %s" % name)
67
68 self.write(json.dumps(self.user_model(user)))
69 self.set_status(201)
70
71 @admin_only
72 @gen.coroutine
73 def delete(self, name):
74 user = self.find_user(name)
75 if user is None:
76 raise web.HTTPError(404)
77 if user.name == self.get_current_user().name:
78 raise web.HTTPError(400, "Cannot delete yourself!")
79 if user.stop_pending:
80 raise web.HTTPError(400, "%s's server is in the process of stopping, please wait." % name)
81 if user.running:
82 yield self.stop_single_user(user)
83 if user.stop_pending:
84 raise web.HTTPError(400, "%s's server is in the process of stopping, please wait." % name)
85
86 yield gen.maybe_future(self.authenticator.delete_user(user))
87
88 # remove from the db
89 self.db.delete(user)
90 self.db.commit()
91
92 self.set_status(204)
93
94 @admin_only
95 def patch(self, name):
96 user = self.find_user(name)
97 if user is None:
98 raise web.HTTPError(404)
99 data = self.get_json_body()
100 self._check_user_model(data)
101 for key, value in data.items():
102 setattr(user, key, value)
103 self.db.commit()
104 self.write(json.dumps(self.user_model(user)))
105
106
107 class UserServerAPIHandler(APIHandler):
108 @gen.coroutine
109 @admin_or_self
110 def post(self, name):
111 user = self.find_user(name)
112 if user.spawner:
113 state = yield user.spawner.poll()
114 if state is None:
115 raise web.HTTPError(400, "%s's server is already running" % name)
116
117 yield self.spawn_single_user(user)
118 status = 202 if user.spawn_pending else 201
119 self.set_status(status)
120
121 @gen.coroutine
122 @admin_or_self
123 def delete(self, name):
124 user = self.find_user(name)
125 if user.stop_pending:
126 self.set_status(202)
127 return
128 if not user.running:
129 raise web.HTTPError(400, "%s's server is not running" % name)
130 status = yield user.spawner.poll()
131 if status is not None:
132 raise web.HTTPError(400, "%s's server is not running" % name)
133 yield self.stop_single_user(user)
134 status = 202 if user.stop_pending else 204
135 self.set_status(status)
136
137 class UserAdminAccessAPIHandler(APIHandler):
138 """Grant admins access to single-user servers
139
140 This handler sets the necessary cookie for an admin to login to a single-user server.
141 """
142 @admin_only
143 def post(self, name):
144 current = self.get_current_user()
145 self.log.warn("Admin user %s has requested access to %s's server",
146 current.name, name,
147 )
148 if not self.settings.get('admin_access', False):
149 raise web.HTTPError(403, "admin access to user servers disabled")
150 user = self.find_user(name)
151 if user is None:
152 raise web.HTTPError(404)
153 if not user.running:
154 raise web.HTTPError(400, "%s's server is not running" % name)
155 self.set_server_cookie(user)
156
157
158 default_handlers = [
159 (r"/api/users", UserListAPIHandler),
160 (r"/api/users/([^/]+)", UserAPIHandler),
161 (r"/api/users/([^/]+)/server", UserServerAPIHandler),
162 (r"/api/users/([^/]+)/admin-access", UserAdminAccessAPIHandler),
163 ]
164
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/jupyterhub/apihandlers/users.py b/jupyterhub/apihandlers/users.py
--- a/jupyterhub/apihandlers/users.py
+++ b/jupyterhub/apihandlers/users.py
@@ -18,6 +18,49 @@
users = self.db.query(orm.User)
data = [ self.user_model(u) for u in users ]
self.write(json.dumps(data))
+
+ @admin_only
+ @gen.coroutine
+ def post(self):
+ data = self.get_json_body()
+ if not data or not isinstance(data, dict) or not data.get('usernames'):
+ raise web.HTTPError(400, "Must specify at least one user to create")
+
+ usernames = data.pop('usernames')
+ self._check_user_model(data)
+ # admin is set for all users
+ # to create admin and non-admin users requires at least two API requests
+ admin = data.get('admin', False)
+
+ to_create = []
+ for name in usernames:
+ user = self.find_user(name)
+ if user is not None:
+ self.log.warn("User %s already exists" % name)
+ else:
+ to_create.append(name)
+
+ if not to_create:
+ raise web.HTTPError(400, "All %i users already exist" % len(usernames))
+
+ created = []
+ for name in to_create:
+ user = self.user_from_username(name)
+ if admin:
+ user.admin = True
+ self.db.commit()
+ try:
+ yield gen.maybe_future(self.authenticator.add_user(user))
+ except Exception:
+ self.log.error("Failed to create user: %s" % name, exc_info=True)
+ self.db.delete(user)
+ self.db.commit()
+ raise web.HTTPError(400, "Failed to create user: %s" % name)
+ else:
+ created.append(user)
+
+ self.write(json.dumps([ self.user_model(u) for u in created ]))
+ self.set_status(201)
def admin_or_self(method):
|
{"golden_diff": "diff --git a/jupyterhub/apihandlers/users.py b/jupyterhub/apihandlers/users.py\n--- a/jupyterhub/apihandlers/users.py\n+++ b/jupyterhub/apihandlers/users.py\n@@ -18,6 +18,49 @@\n users = self.db.query(orm.User)\n data = [ self.user_model(u) for u in users ]\n self.write(json.dumps(data))\n+ \n+ @admin_only\n+ @gen.coroutine\n+ def post(self):\n+ data = self.get_json_body()\n+ if not data or not isinstance(data, dict) or not data.get('usernames'):\n+ raise web.HTTPError(400, \"Must specify at least one user to create\")\n+ \n+ usernames = data.pop('usernames')\n+ self._check_user_model(data)\n+ # admin is set for all users\n+ # to create admin and non-admin users requires at least two API requests\n+ admin = data.get('admin', False)\n+ \n+ to_create = []\n+ for name in usernames:\n+ user = self.find_user(name)\n+ if user is not None:\n+ self.log.warn(\"User %s already exists\" % name)\n+ else:\n+ to_create.append(name)\n+ \n+ if not to_create:\n+ raise web.HTTPError(400, \"All %i users already exist\" % len(usernames))\n+ \n+ created = []\n+ for name in to_create:\n+ user = self.user_from_username(name)\n+ if admin:\n+ user.admin = True\n+ self.db.commit()\n+ try:\n+ yield gen.maybe_future(self.authenticator.add_user(user))\n+ except Exception:\n+ self.log.error(\"Failed to create user: %s\" % name, exc_info=True)\n+ self.db.delete(user)\n+ self.db.commit()\n+ raise web.HTTPError(400, \"Failed to create user: %s\" % name)\n+ else:\n+ created.append(user)\n+ \n+ self.write(json.dumps([ self.user_model(u) for u in created ]))\n+ self.set_status(201)\n \n \n def admin_or_self(method):\n", "issue": "Admin panel - Add ability to import multiple users at once\nIt would be nice to add a list of users as a bulk list of newline delimited users.\n\n", "before_files": [{"content": "\"\"\"User handlers\"\"\"\n\n# Copyright (c) Jupyter Development Team.\n# Distributed under the terms of the Modified BSD License.\n\nimport json\n\nfrom tornado import gen, web\n\nfrom .. import orm\nfrom ..utils import admin_only\nfrom .base import APIHandler\n\n\nclass UserListAPIHandler(APIHandler):\n @admin_only\n def get(self):\n users = self.db.query(orm.User)\n data = [ self.user_model(u) for u in users ]\n self.write(json.dumps(data))\n\n\ndef admin_or_self(method):\n \"\"\"Decorator for restricting access to either the target user or admin\"\"\"\n def m(self, name):\n current = self.get_current_user()\n if current is None:\n raise web.HTTPError(403)\n if not (current.name == name or current.admin):\n raise web.HTTPError(403)\n \n # raise 404 if not found\n if not self.find_user(name):\n raise web.HTTPError(404)\n return method(self, name)\n return m\n\nclass UserAPIHandler(APIHandler):\n \n @admin_or_self\n def get(self, name):\n user = self.find_user(name)\n self.write(json.dumps(self.user_model(user)))\n \n @admin_only\n @gen.coroutine\n def post(self, name):\n data = self.get_json_body()\n user = self.find_user(name)\n if user is not None:\n raise web.HTTPError(400, \"User %s already exists\" % name)\n \n user = self.user_from_username(name)\n if data:\n self._check_user_model(data)\n if 'admin' in data:\n user.admin = data['admin']\n self.db.commit()\n \n try:\n yield gen.maybe_future(self.authenticator.add_user(user))\n except Exception:\n self.log.error(\"Failed to create user: %s\" % name, exc_info=True)\n self.db.delete(user)\n self.db.commit()\n raise web.HTTPError(400, \"Failed to create user: %s\" % name)\n \n self.write(json.dumps(self.user_model(user)))\n self.set_status(201)\n \n @admin_only\n @gen.coroutine\n def delete(self, name):\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n if user.name == self.get_current_user().name:\n raise web.HTTPError(400, \"Cannot delete yourself!\")\n if user.stop_pending:\n raise web.HTTPError(400, \"%s's server is in the process of stopping, please wait.\" % name)\n if user.running:\n yield self.stop_single_user(user)\n if user.stop_pending:\n raise web.HTTPError(400, \"%s's server is in the process of stopping, please wait.\" % name)\n \n yield gen.maybe_future(self.authenticator.delete_user(user))\n \n # remove from the db\n self.db.delete(user)\n self.db.commit()\n \n self.set_status(204)\n \n @admin_only\n def patch(self, name):\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n data = self.get_json_body()\n self._check_user_model(data)\n for key, value in data.items():\n setattr(user, key, value)\n self.db.commit()\n self.write(json.dumps(self.user_model(user)))\n\n\nclass UserServerAPIHandler(APIHandler):\n @gen.coroutine\n @admin_or_self\n def post(self, name):\n user = self.find_user(name)\n if user.spawner:\n state = yield user.spawner.poll()\n if state is None:\n raise web.HTTPError(400, \"%s's server is already running\" % name)\n\n yield self.spawn_single_user(user)\n status = 202 if user.spawn_pending else 201\n self.set_status(status)\n\n @gen.coroutine\n @admin_or_self\n def delete(self, name):\n user = self.find_user(name)\n if user.stop_pending:\n self.set_status(202)\n return\n if not user.running:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n status = yield user.spawner.poll()\n if status is not None:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n yield self.stop_single_user(user)\n status = 202 if user.stop_pending else 204\n self.set_status(status)\n\nclass UserAdminAccessAPIHandler(APIHandler):\n \"\"\"Grant admins access to single-user servers\n \n This handler sets the necessary cookie for an admin to login to a single-user server.\n \"\"\"\n @admin_only\n def post(self, name):\n current = self.get_current_user()\n self.log.warn(\"Admin user %s has requested access to %s's server\",\n current.name, name,\n )\n if not self.settings.get('admin_access', False):\n raise web.HTTPError(403, \"admin access to user servers disabled\")\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n if not user.running:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n self.set_server_cookie(user)\n\n\ndefault_handlers = [\n (r\"/api/users\", UserListAPIHandler),\n (r\"/api/users/([^/]+)\", UserAPIHandler),\n (r\"/api/users/([^/]+)/server\", UserServerAPIHandler),\n (r\"/api/users/([^/]+)/admin-access\", UserAdminAccessAPIHandler),\n]\n", "path": "jupyterhub/apihandlers/users.py"}], "after_files": [{"content": "\"\"\"User handlers\"\"\"\n\n# Copyright (c) Jupyter Development Team.\n# Distributed under the terms of the Modified BSD License.\n\nimport json\n\nfrom tornado import gen, web\n\nfrom .. import orm\nfrom ..utils import admin_only\nfrom .base import APIHandler\n\n\nclass UserListAPIHandler(APIHandler):\n @admin_only\n def get(self):\n users = self.db.query(orm.User)\n data = [ self.user_model(u) for u in users ]\n self.write(json.dumps(data))\n \n @admin_only\n @gen.coroutine\n def post(self):\n data = self.get_json_body()\n if not data or not isinstance(data, dict) or not data.get('usernames'):\n raise web.HTTPError(400, \"Must specify at least one user to create\")\n \n usernames = data.pop('usernames')\n self._check_user_model(data)\n # admin is set for all users\n # to create admin and non-admin users requires at least two API requests\n admin = data.get('admin', False)\n \n to_create = []\n for name in usernames:\n user = self.find_user(name)\n if user is not None:\n self.log.warn(\"User %s already exists\" % name)\n else:\n to_create.append(name)\n \n if not to_create:\n raise web.HTTPError(400, \"All %i users already exist\" % len(usernames))\n \n created = []\n for name in to_create:\n user = self.user_from_username(name)\n if admin:\n user.admin = True\n self.db.commit()\n try:\n yield gen.maybe_future(self.authenticator.add_user(user))\n except Exception:\n self.log.error(\"Failed to create user: %s\" % name, exc_info=True)\n self.db.delete(user)\n self.db.commit()\n raise web.HTTPError(400, \"Failed to create user: %s\" % name)\n else:\n created.append(user)\n \n self.write(json.dumps([ self.user_model(u) for u in created ]))\n self.set_status(201)\n\n\ndef admin_or_self(method):\n \"\"\"Decorator for restricting access to either the target user or admin\"\"\"\n def m(self, name):\n current = self.get_current_user()\n if current is None:\n raise web.HTTPError(403)\n if not (current.name == name or current.admin):\n raise web.HTTPError(403)\n \n # raise 404 if not found\n if not self.find_user(name):\n raise web.HTTPError(404)\n return method(self, name)\n return m\n\nclass UserAPIHandler(APIHandler):\n \n @admin_or_self\n def get(self, name):\n user = self.find_user(name)\n self.write(json.dumps(self.user_model(user)))\n \n @admin_only\n @gen.coroutine\n def post(self, name):\n data = self.get_json_body()\n user = self.find_user(name)\n if user is not None:\n raise web.HTTPError(400, \"User %s already exists\" % name)\n \n user = self.user_from_username(name)\n if data:\n self._check_user_model(data)\n if 'admin' in data:\n user.admin = data['admin']\n self.db.commit()\n \n try:\n yield gen.maybe_future(self.authenticator.add_user(user))\n except Exception:\n self.log.error(\"Failed to create user: %s\" % name, exc_info=True)\n self.db.delete(user)\n self.db.commit()\n raise web.HTTPError(400, \"Failed to create user: %s\" % name)\n \n self.write(json.dumps(self.user_model(user)))\n self.set_status(201)\n \n @admin_only\n @gen.coroutine\n def delete(self, name):\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n if user.name == self.get_current_user().name:\n raise web.HTTPError(400, \"Cannot delete yourself!\")\n if user.stop_pending:\n raise web.HTTPError(400, \"%s's server is in the process of stopping, please wait.\" % name)\n if user.running:\n yield self.stop_single_user(user)\n if user.stop_pending:\n raise web.HTTPError(400, \"%s's server is in the process of stopping, please wait.\" % name)\n \n yield gen.maybe_future(self.authenticator.delete_user(user))\n \n # remove from the db\n self.db.delete(user)\n self.db.commit()\n \n self.set_status(204)\n \n @admin_only\n def patch(self, name):\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n data = self.get_json_body()\n self._check_user_model(data)\n for key, value in data.items():\n setattr(user, key, value)\n self.db.commit()\n self.write(json.dumps(self.user_model(user)))\n\n\nclass UserServerAPIHandler(APIHandler):\n @gen.coroutine\n @admin_or_self\n def post(self, name):\n user = self.find_user(name)\n if user.spawner:\n state = yield user.spawner.poll()\n if state is None:\n raise web.HTTPError(400, \"%s's server is already running\" % name)\n\n yield self.spawn_single_user(user)\n status = 202 if user.spawn_pending else 201\n self.set_status(status)\n\n @gen.coroutine\n @admin_or_self\n def delete(self, name):\n user = self.find_user(name)\n if user.stop_pending:\n self.set_status(202)\n return\n if not user.running:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n status = yield user.spawner.poll()\n if status is not None:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n yield self.stop_single_user(user)\n status = 202 if user.stop_pending else 204\n self.set_status(status)\n\nclass UserAdminAccessAPIHandler(APIHandler):\n \"\"\"Grant admins access to single-user servers\n \n This handler sets the necessary cookie for an admin to login to a single-user server.\n \"\"\"\n @admin_only\n def post(self, name):\n current = self.get_current_user()\n self.log.warn(\"Admin user %s has requested access to %s's server\",\n current.name, name,\n )\n if not self.settings.get('admin_access', False):\n raise web.HTTPError(403, \"admin access to user servers disabled\")\n user = self.find_user(name)\n if user is None:\n raise web.HTTPError(404)\n if not user.running:\n raise web.HTTPError(400, \"%s's server is not running\" % name)\n self.set_server_cookie(user)\n\n\ndefault_handlers = [\n (r\"/api/users\", UserListAPIHandler),\n (r\"/api/users/([^/]+)\", UserAPIHandler),\n (r\"/api/users/([^/]+)/server\", UserServerAPIHandler),\n (r\"/api/users/([^/]+)/admin-access\", UserAdminAccessAPIHandler),\n]\n", "path": "jupyterhub/apihandlers/users.py"}]}
| 1,914 | 475 |
gh_patches_debug_2900
|
rasdani/github-patches
|
git_diff
|
Chia-Network__chia-blockchain-15508
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[Bug] Module `chia.wallet.puzzles.clawback` not found
### What happened?
When installing `1.8.2-rc3` or `master` via `pip`, the module `chia.wallet.puzzles.clawback` is missing. The files are not included because the packages are not listed in `setup.py`. This is also true of the `prefarm` sibling package.
### Version
1.8.2-rc3
### What platform are you using?
Linux
### What ui mode are you using?
CLI
### Relevant log output
```shell
$ pip install git+https://github.com/chia-network/chia-blockchain
Collecting git+https://github.com/chia-network/chia-blockchain
Cloning https://github.com/chia-network/chia-blockchain to /tmp/pip-req-build-m26feywu
Running command git clone --filter=blob:none --quiet https://github.com/chia-network/chia-blockchain /tmp/pip-req-build-m26feywu
Resolved https://github.com/chia-network/chia-blockchain to commit 49140b2b3c0c128f2464c0b4e50c496e7029939d
Running command git submodule update --init --recursive -q
[snip]
$ python3
>>> import chia.wallet.wallet
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/local/lib/python3.11/site-packages/chia/wallet/wallet.py", line 23, in <module>
from chia.wallet.coin_selection import select_coins
File "/usr/local/lib/python3.11/site-packages/chia/wallet/coin_selection.py", line 10, in <module>
from chia.wallet.wallet_coin_record import WalletCoinRecord
File "/usr/local/lib/python3.11/site-packages/chia/wallet/wallet_coin_record.py", line 11, in <module>
from chia.wallet.puzzles.clawback.metadata import ClawbackMetadata, ClawbackVersion
ModuleNotFoundError: No module named 'chia.wallet.puzzles.clawback'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 from __future__ import annotations
2
3 import os
4 import sys
5
6 from setuptools import setup
7
8 dependencies = [
9 "aiofiles==23.1.0", # Async IO for files
10 "anyio==3.6.2",
11 "boto3==1.26.148", # AWS S3 for DL s3 plugin
12 "blspy==1.0.16", # Signature library
13 "chiavdf==1.0.8", # timelord and vdf verification
14 "chiabip158==1.2", # bip158-style wallet filters
15 "chiapos==1.0.11", # proof of space
16 "clvm==0.9.7",
17 "clvm_tools==0.4.6", # Currying, Program.to, other conveniences
18 "chia_rs==0.2.7",
19 "clvm-tools-rs==0.1.34", # Rust implementation of clvm_tools' compiler
20 "aiohttp==3.8.4", # HTTP server for full node rpc
21 "aiosqlite==0.19.0", # asyncio wrapper for sqlite, to store blocks
22 "bitstring==4.0.2", # Binary data management library
23 "colorama==0.4.6", # Colorizes terminal output
24 "colorlog==6.7.0", # Adds color to logs
25 "concurrent-log-handler==0.9.24", # Concurrently log and rotate logs
26 "cryptography==41.0.0", # Python cryptography library for TLS - keyring conflict
27 "filelock==3.12.0", # For reading and writing config multiprocess and multithread safely (non-reentrant locks)
28 "keyring==23.13.1", # Store keys in MacOS Keychain, Windows Credential Locker
29 "PyYAML==6.0", # Used for config file format
30 "setproctitle==1.3.2", # Gives the chia processes readable names
31 "sortedcontainers==2.4.0", # For maintaining sorted mempools
32 "click==8.1.3", # For the CLI
33 "dnspython==2.3.0", # Query DNS seeds
34 "watchdog==2.2.0", # Filesystem event watching - watches keyring.yaml
35 "dnslib==0.9.23", # dns lib
36 "typing-extensions==4.6.0", # typing backports like Protocol and TypedDict
37 "zstd==1.5.5.1",
38 "packaging==23.1",
39 "psutil==5.9.4",
40 ]
41
42 upnp_dependencies = [
43 "miniupnpc==2.2.2", # Allows users to open ports on their router
44 ]
45
46 dev_dependencies = [
47 "build",
48 # >=7.2.4 for https://github.com/nedbat/coveragepy/issues/1604
49 "coverage>=7.2.4",
50 "diff-cover",
51 "pre-commit",
52 "py3createtorrent",
53 "pylint",
54 "pytest",
55 "pytest-asyncio>=0.18.1", # require attribute 'fixture'
56 "pytest-cov",
57 "pytest-monitor; sys_platform == 'linux'",
58 "pytest-xdist",
59 "twine",
60 "isort",
61 "flake8",
62 "mypy",
63 "black==23.3.0",
64 "aiohttp_cors", # For blackd
65 "ipython", # For asyncio debugging
66 "pyinstaller==5.11.0",
67 "types-aiofiles",
68 "types-cryptography",
69 "types-pkg_resources",
70 "types-pyyaml",
71 "types-setuptools",
72 ]
73
74 legacy_keyring_dependencies = [
75 "keyrings.cryptfile==1.3.9",
76 ]
77
78 kwargs = dict(
79 name="chia-blockchain",
80 author="Mariano Sorgente",
81 author_email="[email protected]",
82 description="Chia blockchain full node, farmer, timelord, and wallet.",
83 url="https://chia.net/",
84 license="Apache License",
85 python_requires=">=3.7, <4",
86 keywords="chia blockchain node",
87 install_requires=dependencies,
88 extras_require=dict(
89 dev=dev_dependencies,
90 upnp=upnp_dependencies,
91 legacy_keyring=legacy_keyring_dependencies,
92 ),
93 packages=[
94 "build_scripts",
95 "chia",
96 "chia.cmds",
97 "chia.clvm",
98 "chia.consensus",
99 "chia.daemon",
100 "chia.data_layer",
101 "chia.full_node",
102 "chia.timelord",
103 "chia.farmer",
104 "chia.harvester",
105 "chia.introducer",
106 "chia.plot_sync",
107 "chia.plotters",
108 "chia.plotting",
109 "chia.pools",
110 "chia.protocols",
111 "chia.rpc",
112 "chia.seeder",
113 "chia.server",
114 "chia.simulator",
115 "chia.types.blockchain_format",
116 "chia.types",
117 "chia.util",
118 "chia.wallet",
119 "chia.wallet.db_wallet",
120 "chia.wallet.puzzles",
121 "chia.wallet.cat_wallet",
122 "chia.wallet.did_wallet",
123 "chia.wallet.nft_wallet",
124 "chia.wallet.trading",
125 "chia.wallet.util",
126 "chia.wallet.vc_wallet",
127 "chia.wallet.vc_wallet.vc_puzzles",
128 "chia.wallet.vc_wallet.cr_puzzles",
129 "chia.ssl",
130 "mozilla-ca",
131 ],
132 entry_points={
133 "console_scripts": [
134 "chia = chia.cmds.chia:main",
135 "chia_daemon = chia.daemon.server:main",
136 "chia_wallet = chia.server.start_wallet:main",
137 "chia_full_node = chia.server.start_full_node:main",
138 "chia_harvester = chia.server.start_harvester:main",
139 "chia_farmer = chia.server.start_farmer:main",
140 "chia_introducer = chia.server.start_introducer:main",
141 "chia_crawler = chia.seeder.start_crawler:main",
142 "chia_seeder = chia.seeder.dns_server:main",
143 "chia_timelord = chia.server.start_timelord:main",
144 "chia_timelord_launcher = chia.timelord.timelord_launcher:main",
145 "chia_full_node_simulator = chia.simulator.start_simulator:main",
146 "chia_data_layer = chia.server.start_data_layer:main",
147 "chia_data_layer_http = chia.data_layer.data_layer_server:main",
148 "chia_data_layer_s3_plugin = chia.data_layer.s3_plugin_service:run_server",
149 ]
150 },
151 package_data={
152 "chia": ["pyinstaller.spec"],
153 "": ["*.clsp", "*.clsp.hex", "*.clvm", "*.clib", "py.typed"],
154 "chia.util": ["initial-*.yaml", "english.txt"],
155 "chia.ssl": ["chia_ca.crt", "chia_ca.key", "dst_root_ca.pem"],
156 "mozilla-ca": ["cacert.pem"],
157 },
158 long_description=open("README.md").read(),
159 long_description_content_type="text/markdown",
160 zip_safe=False,
161 project_urls={
162 "Source": "https://github.com/Chia-Network/chia-blockchain/",
163 "Changelog": "https://github.com/Chia-Network/chia-blockchain/blob/main/CHANGELOG.md",
164 },
165 )
166
167 if "setup_file" in sys.modules:
168 # include dev deps in regular deps when run in snyk
169 dependencies.extend(dev_dependencies)
170
171 if len(os.environ.get("CHIA_SKIP_SETUP", "")) < 1:
172 setup(**kwargs) # type: ignore
173
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -118,6 +118,8 @@
"chia.wallet",
"chia.wallet.db_wallet",
"chia.wallet.puzzles",
+ "chia.wallet.puzzles.clawback",
+ "chia.wallet.puzzles.prefarm",
"chia.wallet.cat_wallet",
"chia.wallet.did_wallet",
"chia.wallet.nft_wallet",
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -118,6 +118,8 @@\n \"chia.wallet\",\n \"chia.wallet.db_wallet\",\n \"chia.wallet.puzzles\",\n+ \"chia.wallet.puzzles.clawback\",\n+ \"chia.wallet.puzzles.prefarm\",\n \"chia.wallet.cat_wallet\",\n \"chia.wallet.did_wallet\",\n \"chia.wallet.nft_wallet\",\n", "issue": "[Bug] Module `chia.wallet.puzzles.clawback` not found\n### What happened?\r\n\r\nWhen installing `1.8.2-rc3` or `master` via `pip`, the module `chia.wallet.puzzles.clawback` is missing. The files are not included because the packages are not listed in `setup.py`. This is also true of the `prefarm` sibling package.\r\n\r\n### Version\r\n\r\n1.8.2-rc3\r\n\r\n### What platform are you using?\r\n\r\nLinux\r\n\r\n### What ui mode are you using?\r\n\r\nCLI\r\n\r\n### Relevant log output\r\n\r\n```shell\r\n$ pip install git+https://github.com/chia-network/chia-blockchain\r\nCollecting git+https://github.com/chia-network/chia-blockchain\r\n Cloning https://github.com/chia-network/chia-blockchain to /tmp/pip-req-build-m26feywu\r\n Running command git clone --filter=blob:none --quiet https://github.com/chia-network/chia-blockchain /tmp/pip-req-build-m26feywu\r\n Resolved https://github.com/chia-network/chia-blockchain to commit 49140b2b3c0c128f2464c0b4e50c496e7029939d\r\n Running command git submodule update --init --recursive -q\r\n[snip]\r\n$ python3\r\n>>> import chia.wallet.wallet\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/usr/local/lib/python3.11/site-packages/chia/wallet/wallet.py\", line 23, in <module>\r\n from chia.wallet.coin_selection import select_coins\r\n File \"/usr/local/lib/python3.11/site-packages/chia/wallet/coin_selection.py\", line 10, in <module>\r\n from chia.wallet.wallet_coin_record import WalletCoinRecord\r\n File \"/usr/local/lib/python3.11/site-packages/chia/wallet/wallet_coin_record.py\", line 11, in <module>\r\n from chia.wallet.puzzles.clawback.metadata import ClawbackMetadata, ClawbackVersion\r\nModuleNotFoundError: No module named 'chia.wallet.puzzles.clawback'\r\n```\r\n\n", "before_files": [{"content": "from __future__ import annotations\n\nimport os\nimport sys\n\nfrom setuptools import setup\n\ndependencies = [\n \"aiofiles==23.1.0\", # Async IO for files\n \"anyio==3.6.2\",\n \"boto3==1.26.148\", # AWS S3 for DL s3 plugin\n \"blspy==1.0.16\", # Signature library\n \"chiavdf==1.0.8\", # timelord and vdf verification\n \"chiabip158==1.2\", # bip158-style wallet filters\n \"chiapos==1.0.11\", # proof of space\n \"clvm==0.9.7\",\n \"clvm_tools==0.4.6\", # Currying, Program.to, other conveniences\n \"chia_rs==0.2.7\",\n \"clvm-tools-rs==0.1.34\", # Rust implementation of clvm_tools' compiler\n \"aiohttp==3.8.4\", # HTTP server for full node rpc\n \"aiosqlite==0.19.0\", # asyncio wrapper for sqlite, to store blocks\n \"bitstring==4.0.2\", # Binary data management library\n \"colorama==0.4.6\", # Colorizes terminal output\n \"colorlog==6.7.0\", # Adds color to logs\n \"concurrent-log-handler==0.9.24\", # Concurrently log and rotate logs\n \"cryptography==41.0.0\", # Python cryptography library for TLS - keyring conflict\n \"filelock==3.12.0\", # For reading and writing config multiprocess and multithread safely (non-reentrant locks)\n \"keyring==23.13.1\", # Store keys in MacOS Keychain, Windows Credential Locker\n \"PyYAML==6.0\", # Used for config file format\n \"setproctitle==1.3.2\", # Gives the chia processes readable names\n \"sortedcontainers==2.4.0\", # For maintaining sorted mempools\n \"click==8.1.3\", # For the CLI\n \"dnspython==2.3.0\", # Query DNS seeds\n \"watchdog==2.2.0\", # Filesystem event watching - watches keyring.yaml\n \"dnslib==0.9.23\", # dns lib\n \"typing-extensions==4.6.0\", # typing backports like Protocol and TypedDict\n \"zstd==1.5.5.1\",\n \"packaging==23.1\",\n \"psutil==5.9.4\",\n]\n\nupnp_dependencies = [\n \"miniupnpc==2.2.2\", # Allows users to open ports on their router\n]\n\ndev_dependencies = [\n \"build\",\n # >=7.2.4 for https://github.com/nedbat/coveragepy/issues/1604\n \"coverage>=7.2.4\",\n \"diff-cover\",\n \"pre-commit\",\n \"py3createtorrent\",\n \"pylint\",\n \"pytest\",\n \"pytest-asyncio>=0.18.1\", # require attribute 'fixture'\n \"pytest-cov\",\n \"pytest-monitor; sys_platform == 'linux'\",\n \"pytest-xdist\",\n \"twine\",\n \"isort\",\n \"flake8\",\n \"mypy\",\n \"black==23.3.0\",\n \"aiohttp_cors\", # For blackd\n \"ipython\", # For asyncio debugging\n \"pyinstaller==5.11.0\",\n \"types-aiofiles\",\n \"types-cryptography\",\n \"types-pkg_resources\",\n \"types-pyyaml\",\n \"types-setuptools\",\n]\n\nlegacy_keyring_dependencies = [\n \"keyrings.cryptfile==1.3.9\",\n]\n\nkwargs = dict(\n name=\"chia-blockchain\",\n author=\"Mariano Sorgente\",\n author_email=\"[email protected]\",\n description=\"Chia blockchain full node, farmer, timelord, and wallet.\",\n url=\"https://chia.net/\",\n license=\"Apache License\",\n python_requires=\">=3.7, <4\",\n keywords=\"chia blockchain node\",\n install_requires=dependencies,\n extras_require=dict(\n dev=dev_dependencies,\n upnp=upnp_dependencies,\n legacy_keyring=legacy_keyring_dependencies,\n ),\n packages=[\n \"build_scripts\",\n \"chia\",\n \"chia.cmds\",\n \"chia.clvm\",\n \"chia.consensus\",\n \"chia.daemon\",\n \"chia.data_layer\",\n \"chia.full_node\",\n \"chia.timelord\",\n \"chia.farmer\",\n \"chia.harvester\",\n \"chia.introducer\",\n \"chia.plot_sync\",\n \"chia.plotters\",\n \"chia.plotting\",\n \"chia.pools\",\n \"chia.protocols\",\n \"chia.rpc\",\n \"chia.seeder\",\n \"chia.server\",\n \"chia.simulator\",\n \"chia.types.blockchain_format\",\n \"chia.types\",\n \"chia.util\",\n \"chia.wallet\",\n \"chia.wallet.db_wallet\",\n \"chia.wallet.puzzles\",\n \"chia.wallet.cat_wallet\",\n \"chia.wallet.did_wallet\",\n \"chia.wallet.nft_wallet\",\n \"chia.wallet.trading\",\n \"chia.wallet.util\",\n \"chia.wallet.vc_wallet\",\n \"chia.wallet.vc_wallet.vc_puzzles\",\n \"chia.wallet.vc_wallet.cr_puzzles\",\n \"chia.ssl\",\n \"mozilla-ca\",\n ],\n entry_points={\n \"console_scripts\": [\n \"chia = chia.cmds.chia:main\",\n \"chia_daemon = chia.daemon.server:main\",\n \"chia_wallet = chia.server.start_wallet:main\",\n \"chia_full_node = chia.server.start_full_node:main\",\n \"chia_harvester = chia.server.start_harvester:main\",\n \"chia_farmer = chia.server.start_farmer:main\",\n \"chia_introducer = chia.server.start_introducer:main\",\n \"chia_crawler = chia.seeder.start_crawler:main\",\n \"chia_seeder = chia.seeder.dns_server:main\",\n \"chia_timelord = chia.server.start_timelord:main\",\n \"chia_timelord_launcher = chia.timelord.timelord_launcher:main\",\n \"chia_full_node_simulator = chia.simulator.start_simulator:main\",\n \"chia_data_layer = chia.server.start_data_layer:main\",\n \"chia_data_layer_http = chia.data_layer.data_layer_server:main\",\n \"chia_data_layer_s3_plugin = chia.data_layer.s3_plugin_service:run_server\",\n ]\n },\n package_data={\n \"chia\": [\"pyinstaller.spec\"],\n \"\": [\"*.clsp\", \"*.clsp.hex\", \"*.clvm\", \"*.clib\", \"py.typed\"],\n \"chia.util\": [\"initial-*.yaml\", \"english.txt\"],\n \"chia.ssl\": [\"chia_ca.crt\", \"chia_ca.key\", \"dst_root_ca.pem\"],\n \"mozilla-ca\": [\"cacert.pem\"],\n },\n long_description=open(\"README.md\").read(),\n long_description_content_type=\"text/markdown\",\n zip_safe=False,\n project_urls={\n \"Source\": \"https://github.com/Chia-Network/chia-blockchain/\",\n \"Changelog\": \"https://github.com/Chia-Network/chia-blockchain/blob/main/CHANGELOG.md\",\n },\n)\n\nif \"setup_file\" in sys.modules:\n # include dev deps in regular deps when run in snyk\n dependencies.extend(dev_dependencies)\n\nif len(os.environ.get(\"CHIA_SKIP_SETUP\", \"\")) < 1:\n setup(**kwargs) # type: ignore\n", "path": "setup.py"}], "after_files": [{"content": "from __future__ import annotations\n\nimport os\nimport sys\n\nfrom setuptools import setup\n\ndependencies = [\n \"aiofiles==23.1.0\", # Async IO for files\n \"anyio==3.6.2\",\n \"boto3==1.26.148\", # AWS S3 for DL s3 plugin\n \"blspy==1.0.16\", # Signature library\n \"chiavdf==1.0.8\", # timelord and vdf verification\n \"chiabip158==1.2\", # bip158-style wallet filters\n \"chiapos==1.0.11\", # proof of space\n \"clvm==0.9.7\",\n \"clvm_tools==0.4.6\", # Currying, Program.to, other conveniences\n \"chia_rs==0.2.7\",\n \"clvm-tools-rs==0.1.34\", # Rust implementation of clvm_tools' compiler\n \"aiohttp==3.8.4\", # HTTP server for full node rpc\n \"aiosqlite==0.19.0\", # asyncio wrapper for sqlite, to store blocks\n \"bitstring==4.0.2\", # Binary data management library\n \"colorama==0.4.6\", # Colorizes terminal output\n \"colorlog==6.7.0\", # Adds color to logs\n \"concurrent-log-handler==0.9.24\", # Concurrently log and rotate logs\n \"cryptography==41.0.0\", # Python cryptography library for TLS - keyring conflict\n \"filelock==3.12.0\", # For reading and writing config multiprocess and multithread safely (non-reentrant locks)\n \"keyring==23.13.1\", # Store keys in MacOS Keychain, Windows Credential Locker\n \"PyYAML==6.0\", # Used for config file format\n \"setproctitle==1.3.2\", # Gives the chia processes readable names\n \"sortedcontainers==2.4.0\", # For maintaining sorted mempools\n \"click==8.1.3\", # For the CLI\n \"dnspython==2.3.0\", # Query DNS seeds\n \"watchdog==2.2.0\", # Filesystem event watching - watches keyring.yaml\n \"dnslib==0.9.23\", # dns lib\n \"typing-extensions==4.6.0\", # typing backports like Protocol and TypedDict\n \"zstd==1.5.5.1\",\n \"packaging==23.1\",\n \"psutil==5.9.4\",\n]\n\nupnp_dependencies = [\n \"miniupnpc==2.2.2\", # Allows users to open ports on their router\n]\n\ndev_dependencies = [\n \"build\",\n # >=7.2.4 for https://github.com/nedbat/coveragepy/issues/1604\n \"coverage>=7.2.4\",\n \"diff-cover\",\n \"pre-commit\",\n \"py3createtorrent\",\n \"pylint\",\n \"pytest\",\n \"pytest-asyncio>=0.18.1\", # require attribute 'fixture'\n \"pytest-cov\",\n \"pytest-monitor; sys_platform == 'linux'\",\n \"pytest-xdist\",\n \"twine\",\n \"isort\",\n \"flake8\",\n \"mypy\",\n \"black==23.3.0\",\n \"aiohttp_cors\", # For blackd\n \"ipython\", # For asyncio debugging\n \"pyinstaller==5.11.0\",\n \"types-aiofiles\",\n \"types-cryptography\",\n \"types-pkg_resources\",\n \"types-pyyaml\",\n \"types-setuptools\",\n]\n\nlegacy_keyring_dependencies = [\n \"keyrings.cryptfile==1.3.9\",\n]\n\nkwargs = dict(\n name=\"chia-blockchain\",\n author=\"Mariano Sorgente\",\n author_email=\"[email protected]\",\n description=\"Chia blockchain full node, farmer, timelord, and wallet.\",\n url=\"https://chia.net/\",\n license=\"Apache License\",\n python_requires=\">=3.7, <4\",\n keywords=\"chia blockchain node\",\n install_requires=dependencies,\n extras_require=dict(\n dev=dev_dependencies,\n upnp=upnp_dependencies,\n legacy_keyring=legacy_keyring_dependencies,\n ),\n packages=[\n \"build_scripts\",\n \"chia\",\n \"chia.cmds\",\n \"chia.clvm\",\n \"chia.consensus\",\n \"chia.daemon\",\n \"chia.data_layer\",\n \"chia.full_node\",\n \"chia.timelord\",\n \"chia.farmer\",\n \"chia.harvester\",\n \"chia.introducer\",\n \"chia.plot_sync\",\n \"chia.plotters\",\n \"chia.plotting\",\n \"chia.pools\",\n \"chia.protocols\",\n \"chia.rpc\",\n \"chia.seeder\",\n \"chia.server\",\n \"chia.simulator\",\n \"chia.types.blockchain_format\",\n \"chia.types\",\n \"chia.util\",\n \"chia.wallet\",\n \"chia.wallet.db_wallet\",\n \"chia.wallet.puzzles\",\n \"chia.wallet.puzzles.clawback\",\n \"chia.wallet.puzzles.prefarm\",\n \"chia.wallet.cat_wallet\",\n \"chia.wallet.did_wallet\",\n \"chia.wallet.nft_wallet\",\n \"chia.wallet.trading\",\n \"chia.wallet.util\",\n \"chia.wallet.vc_wallet\",\n \"chia.wallet.vc_wallet.vc_puzzles\",\n \"chia.wallet.vc_wallet.cr_puzzles\",\n \"chia.ssl\",\n \"mozilla-ca\",\n ],\n entry_points={\n \"console_scripts\": [\n \"chia = chia.cmds.chia:main\",\n \"chia_daemon = chia.daemon.server:main\",\n \"chia_wallet = chia.server.start_wallet:main\",\n \"chia_full_node = chia.server.start_full_node:main\",\n \"chia_harvester = chia.server.start_harvester:main\",\n \"chia_farmer = chia.server.start_farmer:main\",\n \"chia_introducer = chia.server.start_introducer:main\",\n \"chia_crawler = chia.seeder.start_crawler:main\",\n \"chia_seeder = chia.seeder.dns_server:main\",\n \"chia_timelord = chia.server.start_timelord:main\",\n \"chia_timelord_launcher = chia.timelord.timelord_launcher:main\",\n \"chia_full_node_simulator = chia.simulator.start_simulator:main\",\n \"chia_data_layer = chia.server.start_data_layer:main\",\n \"chia_data_layer_http = chia.data_layer.data_layer_server:main\",\n \"chia_data_layer_s3_plugin = chia.data_layer.s3_plugin_service:run_server\",\n ]\n },\n package_data={\n \"chia\": [\"pyinstaller.spec\"],\n \"\": [\"*.clsp\", \"*.clsp.hex\", \"*.clvm\", \"*.clib\", \"py.typed\"],\n \"chia.util\": [\"initial-*.yaml\", \"english.txt\"],\n \"chia.ssl\": [\"chia_ca.crt\", \"chia_ca.key\", \"dst_root_ca.pem\"],\n \"mozilla-ca\": [\"cacert.pem\"],\n },\n long_description=open(\"README.md\").read(),\n long_description_content_type=\"text/markdown\",\n zip_safe=False,\n project_urls={\n \"Source\": \"https://github.com/Chia-Network/chia-blockchain/\",\n \"Changelog\": \"https://github.com/Chia-Network/chia-blockchain/blob/main/CHANGELOG.md\",\n },\n)\n\nif \"setup_file\" in sys.modules:\n # include dev deps in regular deps when run in snyk\n dependencies.extend(dev_dependencies)\n\nif len(os.environ.get(\"CHIA_SKIP_SETUP\", \"\")) < 1:\n setup(**kwargs) # type: ignore\n", "path": "setup.py"}]}
| 2,909 | 99 |
gh_patches_debug_26616
|
rasdani/github-patches
|
git_diff
|
kivy__kivy-3859
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Button behavior has a property MIN_STATE_TIME
This property can't be used in kv cause of the uppercase.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `kivy/uix/behaviors/button.py`
Content:
```
1 '''See :class:`ButtonBehavior` for details.
2 '''
3
4 __all__ = ('ButtonBehavior', )
5
6 from kivy.clock import Clock
7 from kivy.properties import OptionProperty, ObjectProperty, BooleanProperty
8 from time import time
9
10
11 class ButtonBehavior(object):
12 '''
13 This `mixin <https://en.wikipedia.org/wiki/Mixin>`_ class provides
14 :class:`~kivy.uix.button.Button` behavior.
15
16 :Events:
17 `on_press`
18 Fired when the button is pressed.
19 `on_release`
20 Fired when the button is released (i.e. the touch/click that
21 pressed the button goes away).
22 '''
23
24 state = OptionProperty('normal', options=('normal', 'down'))
25 '''The state of the button, must be one of 'normal' or 'down'.
26 The state is 'down' only when the button is currently touched/clicked,
27 otherwise its 'normal'.
28
29 :attr:`state` is an :class:`~kivy.properties.OptionProperty` and defaults
30 to 'normal'.
31 '''
32
33 last_touch = ObjectProperty(None)
34 '''Contains the last relevant touch received by the Button. This can
35 be used in `on_press` or `on_release` in order to know which touch
36 dispatched the event.
37
38 .. versionadded:: 1.8.0
39
40 :attr:`last_touch` is a :class:`~kivy.properties.ObjectProperty` and
41 defaults to `None`.
42 '''
43
44 MIN_STATE_TIME = 0.035
45 '''The minimum period of time which the widget must remain in the
46 `'down'` state.
47
48 :attr:`MIN_STATE_TIME` is a float and defaults to 0.035.
49 '''
50
51 always_release = BooleanProperty(True)
52 '''This determines whether or not the widget fires an `on_release` event if
53 the touch_up is outside the widget.
54
55 .. versionadded:: 1.9.0
56
57 :attr:`always_release` is a :class:`~kivy.properties.BooleanProperty` and
58 defaults to `True`.
59 '''
60
61 def __init__(self, **kwargs):
62 self.register_event_type('on_press')
63 self.register_event_type('on_release')
64 super(ButtonBehavior, self).__init__(**kwargs)
65 self.__state_event = None
66 self.__touch_time = None
67 self.fbind('state', self.cancel_event)
68
69 def _do_press(self):
70 self.state = 'down'
71
72 def _do_release(self, *args):
73 self.state = 'normal'
74
75 def cancel_event(self, *args):
76 if self.__state_event:
77 self.__state_event.cancel()
78 self.__state_event = None
79
80 def on_touch_down(self, touch):
81 if super(ButtonBehavior, self).on_touch_down(touch):
82 return True
83 if touch.is_mouse_scrolling:
84 return False
85 if not self.collide_point(touch.x, touch.y):
86 return False
87 if self in touch.ud:
88 return False
89 touch.grab(self)
90 touch.ud[self] = True
91 self.last_touch = touch
92 self.__touch_time = time()
93 self._do_press()
94 self.dispatch('on_press')
95 return True
96
97 def on_touch_move(self, touch):
98 if touch.grab_current is self:
99 return True
100 if super(ButtonBehavior, self).on_touch_move(touch):
101 return True
102 return self in touch.ud
103
104 def on_touch_up(self, touch):
105 if touch.grab_current is not self:
106 return super(ButtonBehavior, self).on_touch_up(touch)
107 assert(self in touch.ud)
108 touch.ungrab(self)
109 self.last_touch = touch
110
111 if (not self.always_release
112 and not self.collide_point(*touch.pos)):
113 self.state = 'normal'
114 return
115
116 touchtime = time() - self.__touch_time
117 if touchtime < self.MIN_STATE_TIME:
118 self.__state_event = Clock.schedule_once(
119 self._do_release, self.MIN_STATE_TIME - touchtime)
120 else:
121 self._do_release()
122 self.dispatch('on_release')
123 return True
124
125 def on_press(self):
126 pass
127
128 def on_release(self):
129 pass
130
131 def trigger_action(self, duration=0.1):
132 '''Trigger whatever action(s) have been bound to the button by calling
133 both the on_press and on_release callbacks.
134
135 This simulates a quick button press without using any touch events.
136
137 Duration is the length of the press in seconds. Pass 0 if you want
138 the action to happen instantly.
139
140 .. versionadded:: 1.8.0
141 '''
142 self._do_press()
143 self.dispatch('on_press')
144
145 def trigger_release(dt):
146 self._do_release()
147 self.dispatch('on_release')
148 if not duration:
149 trigger_release(0)
150 else:
151 Clock.schedule_once(trigger_release, duration)
152
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/kivy/uix/behaviors/button.py b/kivy/uix/behaviors/button.py
--- a/kivy/uix/behaviors/button.py
+++ b/kivy/uix/behaviors/button.py
@@ -4,8 +4,10 @@
__all__ = ('ButtonBehavior', )
from kivy.clock import Clock
-from kivy.properties import OptionProperty, ObjectProperty, BooleanProperty
+from kivy.properties import OptionProperty, ObjectProperty, \
+ BooleanProperty, NumericProperty, AliasProperty
from time import time
+from kivy.logger import Logger
class ButtonBehavior(object):
@@ -41,13 +43,26 @@
defaults to `None`.
'''
- MIN_STATE_TIME = 0.035
+ min_state_time = NumericProperty(0.035)
'''The minimum period of time which the widget must remain in the
`'down'` state.
- :attr:`MIN_STATE_TIME` is a float and defaults to 0.035.
+ :attr:`min_state_time` is a float and defaults to 0.035.
'''
+ def _set_min_state_time(self, value):
+ Logger.warning(
+ 'MIN_STATE_TIME is deprecated, use min_state_time instead')
+ self.min_state_time = value
+
+ def _get_min_state_time(self):
+ Logger.warning(
+ 'MIN_STATE_TIME is deprecated, use min_state_time instead')
+ return self.min_state_time
+
+ MIN_STATE_TIME = AliasProperty(
+ _get_min_state_time, _set_min_state_time, bind=('min_state_time', ))
+
always_release = BooleanProperty(True)
'''This determines whether or not the widget fires an `on_release` event if
the touch_up is outside the widget.
|
{"golden_diff": "diff --git a/kivy/uix/behaviors/button.py b/kivy/uix/behaviors/button.py\n--- a/kivy/uix/behaviors/button.py\n+++ b/kivy/uix/behaviors/button.py\n@@ -4,8 +4,10 @@\n __all__ = ('ButtonBehavior', )\n \n from kivy.clock import Clock\n-from kivy.properties import OptionProperty, ObjectProperty, BooleanProperty\n+from kivy.properties import OptionProperty, ObjectProperty, \\\n+ BooleanProperty, NumericProperty, AliasProperty\n from time import time\n+from kivy.logger import Logger\n \n \n class ButtonBehavior(object):\n@@ -41,13 +43,26 @@\n defaults to `None`.\n '''\n \n- MIN_STATE_TIME = 0.035\n+ min_state_time = NumericProperty(0.035)\n '''The minimum period of time which the widget must remain in the\n `'down'` state.\n \n- :attr:`MIN_STATE_TIME` is a float and defaults to 0.035.\n+ :attr:`min_state_time` is a float and defaults to 0.035.\n '''\n \n+ def _set_min_state_time(self, value):\n+ Logger.warning(\n+ 'MIN_STATE_TIME is deprecated, use min_state_time instead')\n+ self.min_state_time = value\n+\n+ def _get_min_state_time(self):\n+ Logger.warning(\n+ 'MIN_STATE_TIME is deprecated, use min_state_time instead')\n+ return self.min_state_time\n+\n+ MIN_STATE_TIME = AliasProperty(\n+ _get_min_state_time, _set_min_state_time, bind=('min_state_time', ))\n+\n always_release = BooleanProperty(True)\n '''This determines whether or not the widget fires an `on_release` event if\n the touch_up is outside the widget.\n", "issue": "Button behavior has a property MIN_STATE_TIME\nThis property can't be used in kv cause of the uppercase.\n\n", "before_files": [{"content": "'''See :class:`ButtonBehavior` for details.\n'''\n\n__all__ = ('ButtonBehavior', )\n\nfrom kivy.clock import Clock\nfrom kivy.properties import OptionProperty, ObjectProperty, BooleanProperty\nfrom time import time\n\n\nclass ButtonBehavior(object):\n '''\n This `mixin <https://en.wikipedia.org/wiki/Mixin>`_ class provides\n :class:`~kivy.uix.button.Button` behavior.\n\n :Events:\n `on_press`\n Fired when the button is pressed.\n `on_release`\n Fired when the button is released (i.e. the touch/click that\n pressed the button goes away).\n '''\n\n state = OptionProperty('normal', options=('normal', 'down'))\n '''The state of the button, must be one of 'normal' or 'down'.\n The state is 'down' only when the button is currently touched/clicked,\n otherwise its 'normal'.\n\n :attr:`state` is an :class:`~kivy.properties.OptionProperty` and defaults\n to 'normal'.\n '''\n\n last_touch = ObjectProperty(None)\n '''Contains the last relevant touch received by the Button. This can\n be used in `on_press` or `on_release` in order to know which touch\n dispatched the event.\n\n .. versionadded:: 1.8.0\n\n :attr:`last_touch` is a :class:`~kivy.properties.ObjectProperty` and\n defaults to `None`.\n '''\n\n MIN_STATE_TIME = 0.035\n '''The minimum period of time which the widget must remain in the\n `'down'` state.\n\n :attr:`MIN_STATE_TIME` is a float and defaults to 0.035.\n '''\n\n always_release = BooleanProperty(True)\n '''This determines whether or not the widget fires an `on_release` event if\n the touch_up is outside the widget.\n\n .. versionadded:: 1.9.0\n\n :attr:`always_release` is a :class:`~kivy.properties.BooleanProperty` and\n defaults to `True`.\n '''\n\n def __init__(self, **kwargs):\n self.register_event_type('on_press')\n self.register_event_type('on_release')\n super(ButtonBehavior, self).__init__(**kwargs)\n self.__state_event = None\n self.__touch_time = None\n self.fbind('state', self.cancel_event)\n\n def _do_press(self):\n self.state = 'down'\n\n def _do_release(self, *args):\n self.state = 'normal'\n\n def cancel_event(self, *args):\n if self.__state_event:\n self.__state_event.cancel()\n self.__state_event = None\n\n def on_touch_down(self, touch):\n if super(ButtonBehavior, self).on_touch_down(touch):\n return True\n if touch.is_mouse_scrolling:\n return False\n if not self.collide_point(touch.x, touch.y):\n return False\n if self in touch.ud:\n return False\n touch.grab(self)\n touch.ud[self] = True\n self.last_touch = touch\n self.__touch_time = time()\n self._do_press()\n self.dispatch('on_press')\n return True\n\n def on_touch_move(self, touch):\n if touch.grab_current is self:\n return True\n if super(ButtonBehavior, self).on_touch_move(touch):\n return True\n return self in touch.ud\n\n def on_touch_up(self, touch):\n if touch.grab_current is not self:\n return super(ButtonBehavior, self).on_touch_up(touch)\n assert(self in touch.ud)\n touch.ungrab(self)\n self.last_touch = touch\n\n if (not self.always_release\n and not self.collide_point(*touch.pos)):\n self.state = 'normal'\n return\n\n touchtime = time() - self.__touch_time\n if touchtime < self.MIN_STATE_TIME:\n self.__state_event = Clock.schedule_once(\n self._do_release, self.MIN_STATE_TIME - touchtime)\n else:\n self._do_release()\n self.dispatch('on_release')\n return True\n\n def on_press(self):\n pass\n\n def on_release(self):\n pass\n\n def trigger_action(self, duration=0.1):\n '''Trigger whatever action(s) have been bound to the button by calling\n both the on_press and on_release callbacks.\n\n This simulates a quick button press without using any touch events.\n\n Duration is the length of the press in seconds. Pass 0 if you want\n the action to happen instantly.\n\n .. versionadded:: 1.8.0\n '''\n self._do_press()\n self.dispatch('on_press')\n\n def trigger_release(dt):\n self._do_release()\n self.dispatch('on_release')\n if not duration:\n trigger_release(0)\n else:\n Clock.schedule_once(trigger_release, duration)\n", "path": "kivy/uix/behaviors/button.py"}], "after_files": [{"content": "'''See :class:`ButtonBehavior` for details.\n'''\n\n__all__ = ('ButtonBehavior', )\n\nfrom kivy.clock import Clock\nfrom kivy.properties import OptionProperty, ObjectProperty, \\\n BooleanProperty, NumericProperty, AliasProperty\nfrom time import time\nfrom kivy.logger import Logger\n\n\nclass ButtonBehavior(object):\n '''\n This `mixin <https://en.wikipedia.org/wiki/Mixin>`_ class provides\n :class:`~kivy.uix.button.Button` behavior.\n\n :Events:\n `on_press`\n Fired when the button is pressed.\n `on_release`\n Fired when the button is released (i.e. the touch/click that\n pressed the button goes away).\n '''\n\n state = OptionProperty('normal', options=('normal', 'down'))\n '''The state of the button, must be one of 'normal' or 'down'.\n The state is 'down' only when the button is currently touched/clicked,\n otherwise its 'normal'.\n\n :attr:`state` is an :class:`~kivy.properties.OptionProperty` and defaults\n to 'normal'.\n '''\n\n last_touch = ObjectProperty(None)\n '''Contains the last relevant touch received by the Button. This can\n be used in `on_press` or `on_release` in order to know which touch\n dispatched the event.\n\n .. versionadded:: 1.8.0\n\n :attr:`last_touch` is a :class:`~kivy.properties.ObjectProperty` and\n defaults to `None`.\n '''\n\n min_state_time = NumericProperty(0.035)\n '''The minimum period of time which the widget must remain in the\n `'down'` state.\n\n :attr:`min_state_time` is a float and defaults to 0.035.\n '''\n\n def _set_min_state_time(self, value):\n Logger.warning(\n 'MIN_STATE_TIME is deprecated, use min_state_time instead')\n self.min_state_time = value\n\n def _get_min_state_time(self):\n Logger.warning(\n 'MIN_STATE_TIME is deprecated, use min_state_time instead')\n return self.min_state_time\n\n MIN_STATE_TIME = AliasProperty(\n _get_min_state_time, _set_min_state_time, bind=('min_state_time', ))\n\n always_release = BooleanProperty(True)\n '''This determines whether or not the widget fires an `on_release` event if\n the touch_up is outside the widget.\n\n .. versionadded:: 1.9.0\n\n :attr:`always_release` is a :class:`~kivy.properties.BooleanProperty` and\n defaults to `True`.\n '''\n\n def __init__(self, **kwargs):\n self.register_event_type('on_press')\n self.register_event_type('on_release')\n super(ButtonBehavior, self).__init__(**kwargs)\n self.__state_event = None\n self.__touch_time = None\n self.fbind('state', self.cancel_event)\n\n def _do_press(self):\n self.state = 'down'\n\n def _do_release(self, *args):\n self.state = 'normal'\n\n def cancel_event(self, *args):\n if self.__state_event:\n self.__state_event.cancel()\n self.__state_event = None\n\n def on_touch_down(self, touch):\n if super(ButtonBehavior, self).on_touch_down(touch):\n return True\n if touch.is_mouse_scrolling:\n return False\n if not self.collide_point(touch.x, touch.y):\n return False\n if self in touch.ud:\n return False\n touch.grab(self)\n touch.ud[self] = True\n self.last_touch = touch\n self.__touch_time = time()\n self._do_press()\n self.dispatch('on_press')\n return True\n\n def on_touch_move(self, touch):\n if touch.grab_current is self:\n return True\n if super(ButtonBehavior, self).on_touch_move(touch):\n return True\n return self in touch.ud\n\n def on_touch_up(self, touch):\n if touch.grab_current is not self:\n return super(ButtonBehavior, self).on_touch_up(touch)\n assert(self in touch.ud)\n touch.ungrab(self)\n self.last_touch = touch\n\n if (not self.always_release\n and not self.collide_point(*touch.pos)):\n self.state = 'normal'\n return\n\n touchtime = time() - self.__touch_time\n if touchtime < self.MIN_STATE_TIME:\n self.__state_event = Clock.schedule_once(\n self._do_release, self.MIN_STATE_TIME - touchtime)\n else:\n self._do_release()\n self.dispatch('on_release')\n return True\n\n def on_press(self):\n pass\n\n def on_release(self):\n pass\n\n def trigger_action(self, duration=0.1):\n '''Trigger whatever action(s) have been bound to the button by calling\n both the on_press and on_release callbacks.\n\n This simulates a quick button press without using any touch events.\n\n Duration is the length of the press in seconds. Pass 0 if you want\n the action to happen instantly.\n\n .. versionadded:: 1.8.0\n '''\n self._do_press()\n self.dispatch('on_press')\n\n def trigger_release(dt):\n self._do_release()\n self.dispatch('on_release')\n if not duration:\n trigger_release(0)\n else:\n Clock.schedule_once(trigger_release, duration)\n", "path": "kivy/uix/behaviors/button.py"}]}
| 1,702 | 401 |
gh_patches_debug_33958
|
rasdani/github-patches
|
git_diff
|
googleapis__google-auth-library-python-97
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add GAE Signer
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `google/auth/app_engine.py`
Content:
```
1 # Copyright 2016 Google Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Google App Engine standard environment credentials.
16
17 This module provides authentication for application running on App Engine in
18 the standard environment using the `App Identity API`_.
19
20
21 .. _App Identity API:
22 https://cloud.google.com/appengine/docs/python/appidentity/
23 """
24
25 import datetime
26
27 from google.auth import _helpers
28 from google.auth import credentials
29
30 try:
31 from google.appengine.api import app_identity
32 except ImportError:
33 app_identity = None
34
35
36 def get_project_id():
37 """Gets the project ID for the current App Engine application.
38
39 Returns:
40 str: The project ID
41
42 Raises:
43 EnvironmentError: If the App Engine APIs are unavailable.
44 """
45 if app_identity is None:
46 raise EnvironmentError(
47 'The App Engine APIs are not available.')
48 return app_identity.get_application_id()
49
50
51 class Credentials(credentials.Scoped, credentials.Signing,
52 credentials.Credentials):
53 """App Engine standard environment credentials.
54
55 These credentials use the App Engine App Identity API to obtain access
56 tokens.
57 """
58
59 def __init__(self, scopes=None, service_account_id=None):
60 """
61 Args:
62 scopes (Sequence[str]): Scopes to request from the App Identity
63 API.
64 service_account_id (str): The service account ID passed into
65 :func:`google.appengine.api.app_identity.get_access_token`.
66 If not specified, the default application service account
67 ID will be used.
68
69 Raises:
70 EnvironmentError: If the App Engine APIs are unavailable.
71 """
72 if app_identity is None:
73 raise EnvironmentError(
74 'The App Engine APIs are not available.')
75
76 super(Credentials, self).__init__()
77 self._scopes = scopes
78 self._service_account_id = service_account_id
79
80 @_helpers.copy_docstring(credentials.Credentials)
81 def refresh(self, request):
82 # pylint: disable=unused-argument
83 token, ttl = app_identity.get_access_token(
84 self._scopes, self._service_account_id)
85 expiry = _helpers.utcnow() + datetime.timedelta(seconds=ttl)
86
87 self.token, self.expiry = token, expiry
88
89 @property
90 def service_account_email(self):
91 """The service account email."""
92 if self._service_account_id is None:
93 self._service_account_id = app_identity.get_service_account_name()
94 return self._service_account_id
95
96 @property
97 def requires_scopes(self):
98 """Checks if the credentials requires scopes.
99
100 Returns:
101 bool: True if there are no scopes set otherwise False.
102 """
103 return not self._scopes
104
105 @_helpers.copy_docstring(credentials.Scoped)
106 def with_scopes(self, scopes):
107 return Credentials(
108 scopes=scopes, service_account_id=self._service_account_id)
109
110 @_helpers.copy_docstring(credentials.Signing)
111 def sign_bytes(self, message):
112 return app_identity.sign_blob(message)
113
114 @property
115 @_helpers.copy_docstring(credentials.Signing)
116 def signer_email(self):
117 return self.service_account_email
118
```
Path: `google/auth/crypt.py`
Content:
```
1 # Copyright 2016 Google Inc.
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """Cryptography helpers for verifying and signing messages.
16
17 Uses the ``rsa``, ``pyasn1`` and ``pyasn1_modules`` packages
18 to parse PEM files storing PKCS#1 or PKCS#8 keys as well as
19 certificates. There is no support for p12 files.
20
21 The simplest way to verify signatures is using :func:`verify_signature`::
22
23 cert = open('certs.pem').read()
24 valid = crypt.verify_signature(message, signature, cert)
25
26 If you're going to verify many messages with the same certificate, you can use
27 :class:`Verifier`::
28
29 cert = open('certs.pem').read()
30 verifier = crypt.Verifier.from_string(cert)
31 valid = verifier.verify(message, signature)
32
33
34 To sign messages use :class:`Signer` with a private key::
35
36 private_key = open('private_key.pem').read()
37 signer = crypt.Signer(private_key)
38 signature = signer.sign(message)
39
40 """
41
42 from pyasn1.codec.der import decoder
43 from pyasn1_modules import pem
44 from pyasn1_modules.rfc2459 import Certificate
45 from pyasn1_modules.rfc5208 import PrivateKeyInfo
46 import rsa
47 import six
48
49 from google.auth import _helpers
50
51 _POW2 = (128, 64, 32, 16, 8, 4, 2, 1)
52 _CERTIFICATE_MARKER = b'-----BEGIN CERTIFICATE-----'
53 _PKCS1_MARKER = ('-----BEGIN RSA PRIVATE KEY-----',
54 '-----END RSA PRIVATE KEY-----')
55 _PKCS8_MARKER = ('-----BEGIN PRIVATE KEY-----',
56 '-----END PRIVATE KEY-----')
57 _PKCS8_SPEC = PrivateKeyInfo()
58
59
60 def _bit_list_to_bytes(bit_list):
61 """Converts an iterable of 1s and 0s to bytes.
62
63 Combines the list 8 at a time, treating each group of 8 bits
64 as a single byte.
65
66 Args:
67 bit_list (Sequence): Sequence of 1s and 0s.
68
69 Returns:
70 bytes: The decoded bytes.
71 """
72 num_bits = len(bit_list)
73 byte_vals = bytearray()
74 for start in six.moves.xrange(0, num_bits, 8):
75 curr_bits = bit_list[start:start + 8]
76 char_val = sum(val * digit
77 for val, digit in six.moves.zip(_POW2, curr_bits))
78 byte_vals.append(char_val)
79 return bytes(byte_vals)
80
81
82 class Verifier(object):
83 """This object is used to verify cryptographic signatures.
84
85 Args:
86 public_key (rsa.key.PublicKey): The public key used to verify
87 signatures.
88 """
89
90 def __init__(self, public_key):
91 self._pubkey = public_key
92
93 def verify(self, message, signature):
94 """Verifies a message against a cryptographic signature.
95
96 Args:
97 message (Union[str, bytes]): The message to verify.
98 signature (Union[str, bytes]): The cryptography signature to check.
99
100 Returns:
101 bool: True if message was signed by the private key associated
102 with the public key that this object was constructed with.
103 """
104 message = _helpers.to_bytes(message)
105 try:
106 return rsa.pkcs1.verify(message, signature, self._pubkey)
107 except (ValueError, rsa.pkcs1.VerificationError):
108 return False
109
110 @classmethod
111 def from_string(cls, public_key):
112 """Construct an Verifier instance from a public key or public
113 certificate string.
114
115 Args:
116 public_key (Union[str, bytes]): The public key in PEM format or the
117 x509 public key certificate.
118
119 Returns:
120 Verifier: The constructed verifier.
121
122 Raises:
123 ValueError: If the public_key can't be parsed.
124 """
125 public_key = _helpers.to_bytes(public_key)
126 is_x509_cert = _CERTIFICATE_MARKER in public_key
127
128 # If this is a certificate, extract the public key info.
129 if is_x509_cert:
130 der = rsa.pem.load_pem(public_key, 'CERTIFICATE')
131 asn1_cert, remaining = decoder.decode(der, asn1Spec=Certificate())
132 if remaining != b'':
133 raise ValueError('Unused bytes', remaining)
134
135 cert_info = asn1_cert['tbsCertificate']['subjectPublicKeyInfo']
136 key_bytes = _bit_list_to_bytes(cert_info['subjectPublicKey'])
137 pubkey = rsa.PublicKey.load_pkcs1(key_bytes, 'DER')
138 else:
139 pubkey = rsa.PublicKey.load_pkcs1(public_key, 'PEM')
140 return cls(pubkey)
141
142
143 def verify_signature(message, signature, certs):
144 """Verify a cryptographic signature.
145
146 Checks that the provided ``signature`` was generated from ``bytes`` using
147 the private key associated with the ``cert``.
148
149 Args:
150 message (Union[str, bytes]): The plaintext message.
151 signature (Union[str, bytes]): The cryptographic signature to check.
152 certs (Union[Sequence, str, bytes]): The certificate or certificates
153 to use to check the signature.
154
155 Returns:
156 bool: True if the signature is valid, otherwise False.
157 """
158 if isinstance(certs, (six.text_type, six.binary_type)):
159 certs = [certs]
160
161 for cert in certs:
162 verifier = Verifier.from_string(cert)
163 if verifier.verify(message, signature):
164 return True
165 return False
166
167
168 class Signer(object):
169 """Signs messages with a private key.
170
171 Args:
172 private_key (rsa.key.PrivateKey): The private key to sign with.
173 key_id (str): Optional key ID used to identify this private key. This
174 can be useful to associate the private key with its associated
175 public key or certificate.
176 """
177
178 def __init__(self, private_key, key_id=None):
179 self._key = private_key
180 self.key_id = key_id
181
182 def sign(self, message):
183 """Signs a message.
184
185 Args:
186 message (Union[str, bytes]): The message to be signed.
187
188 Returns:
189 bytes: The signature of the message for the given key.
190 """
191 message = _helpers.to_bytes(message)
192 return rsa.pkcs1.sign(message, self._key, 'SHA-256')
193
194 @classmethod
195 def from_string(cls, key, key_id=None):
196 """Construct an Signer instance from a private key in PEM format.
197
198 Args:
199 key (str): Private key in PEM format.
200 key_id (str): An optional key id used to identify the private key.
201
202 Returns:
203 Signer: The constructed signer.
204
205 Raises:
206 ValueError: If the key cannot be parsed as PKCS#1 or PKCS#8 in
207 PEM format.
208 """
209 key = _helpers.from_bytes(key) # PEM expects str in Python 3
210 marker_id, key_bytes = pem.readPemBlocksFromFile(
211 six.StringIO(key), _PKCS1_MARKER, _PKCS8_MARKER)
212
213 # Key is in pkcs1 format.
214 if marker_id == 0:
215 private_key = rsa.key.PrivateKey.load_pkcs1(
216 key_bytes, format='DER')
217 # Key is in pkcs8.
218 elif marker_id == 1:
219 key_info, remaining = decoder.decode(
220 key_bytes, asn1Spec=_PKCS8_SPEC)
221 if remaining != b'':
222 raise ValueError('Unused bytes', remaining)
223 private_key_info = key_info.getComponentByName('privateKey')
224 private_key = rsa.key.PrivateKey.load_pkcs1(
225 private_key_info.asOctets(), format='DER')
226 else:
227 raise ValueError('No key could be detected.')
228
229 return cls(private_key, key_id=key_id)
230
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/google/auth/app_engine.py b/google/auth/app_engine.py
--- a/google/auth/app_engine.py
+++ b/google/auth/app_engine.py
@@ -12,10 +12,10 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-"""Google App Engine standard environment credentials.
+"""Google App Engine standard environment support.
-This module provides authentication for application running on App Engine in
-the standard environment using the `App Identity API`_.
+This module provides authentication and signing for applications running on App
+Engine in the standard environment using the `App Identity API`_.
.. _App Identity API:
@@ -33,6 +33,29 @@
app_identity = None
+class Signer(object):
+ """Signs messages using the App Engine app identity service.
+
+ This can be used in place of :class:`google.auth.crypt.Signer` when
+ running in the App Engine standard environment.
+ """
+ def __init__(self):
+ self.key_id = None
+
+ @staticmethod
+ def sign(message):
+ """Signs a message.
+
+ Args:
+ message (Union[str, bytes]): The message to be signed.
+
+ Returns:
+ bytes: The signature of the message.
+ """
+ message = _helpers.to_bytes(message)
+ return app_identity.sign_blob(message)
+
+
def get_project_id():
"""Gets the project ID for the current App Engine application.
@@ -109,7 +132,7 @@
@_helpers.copy_docstring(credentials.Signing)
def sign_bytes(self, message):
- return app_identity.sign_blob(message)
+ return Signer().sign(message)
@property
@_helpers.copy_docstring(credentials.Signing)
diff --git a/google/auth/crypt.py b/google/auth/crypt.py
--- a/google/auth/crypt.py
+++ b/google/auth/crypt.py
@@ -186,7 +186,7 @@
message (Union[str, bytes]): The message to be signed.
Returns:
- bytes: The signature of the message for the given key.
+ bytes: The signature of the message.
"""
message = _helpers.to_bytes(message)
return rsa.pkcs1.sign(message, self._key, 'SHA-256')
|
{"golden_diff": "diff --git a/google/auth/app_engine.py b/google/auth/app_engine.py\n--- a/google/auth/app_engine.py\n+++ b/google/auth/app_engine.py\n@@ -12,10 +12,10 @@\n # See the License for the specific language governing permissions and\n # limitations under the License.\n \n-\"\"\"Google App Engine standard environment credentials.\n+\"\"\"Google App Engine standard environment support.\n \n-This module provides authentication for application running on App Engine in\n-the standard environment using the `App Identity API`_.\n+This module provides authentication and signing for applications running on App\n+Engine in the standard environment using the `App Identity API`_.\n \n \n .. _App Identity API:\n@@ -33,6 +33,29 @@\n app_identity = None\n \n \n+class Signer(object):\n+ \"\"\"Signs messages using the App Engine app identity service.\n+\n+ This can be used in place of :class:`google.auth.crypt.Signer` when\n+ running in the App Engine standard environment.\n+ \"\"\"\n+ def __init__(self):\n+ self.key_id = None\n+\n+ @staticmethod\n+ def sign(message):\n+ \"\"\"Signs a message.\n+\n+ Args:\n+ message (Union[str, bytes]): The message to be signed.\n+\n+ Returns:\n+ bytes: The signature of the message.\n+ \"\"\"\n+ message = _helpers.to_bytes(message)\n+ return app_identity.sign_blob(message)\n+\n+\n def get_project_id():\n \"\"\"Gets the project ID for the current App Engine application.\n \n@@ -109,7 +132,7 @@\n \n @_helpers.copy_docstring(credentials.Signing)\n def sign_bytes(self, message):\n- return app_identity.sign_blob(message)\n+ return Signer().sign(message)\n \n @property\n @_helpers.copy_docstring(credentials.Signing)\ndiff --git a/google/auth/crypt.py b/google/auth/crypt.py\n--- a/google/auth/crypt.py\n+++ b/google/auth/crypt.py\n@@ -186,7 +186,7 @@\n message (Union[str, bytes]): The message to be signed.\n \n Returns:\n- bytes: The signature of the message for the given key.\n+ bytes: The signature of the message.\n \"\"\"\n message = _helpers.to_bytes(message)\n return rsa.pkcs1.sign(message, self._key, 'SHA-256')\n", "issue": "Add GAE Signer\n\n", "before_files": [{"content": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Google App Engine standard environment credentials.\n\nThis module provides authentication for application running on App Engine in\nthe standard environment using the `App Identity API`_.\n\n\n.. _App Identity API:\n https://cloud.google.com/appengine/docs/python/appidentity/\n\"\"\"\n\nimport datetime\n\nfrom google.auth import _helpers\nfrom google.auth import credentials\n\ntry:\n from google.appengine.api import app_identity\nexcept ImportError:\n app_identity = None\n\n\ndef get_project_id():\n \"\"\"Gets the project ID for the current App Engine application.\n\n Returns:\n str: The project ID\n\n Raises:\n EnvironmentError: If the App Engine APIs are unavailable.\n \"\"\"\n if app_identity is None:\n raise EnvironmentError(\n 'The App Engine APIs are not available.')\n return app_identity.get_application_id()\n\n\nclass Credentials(credentials.Scoped, credentials.Signing,\n credentials.Credentials):\n \"\"\"App Engine standard environment credentials.\n\n These credentials use the App Engine App Identity API to obtain access\n tokens.\n \"\"\"\n\n def __init__(self, scopes=None, service_account_id=None):\n \"\"\"\n Args:\n scopes (Sequence[str]): Scopes to request from the App Identity\n API.\n service_account_id (str): The service account ID passed into\n :func:`google.appengine.api.app_identity.get_access_token`.\n If not specified, the default application service account\n ID will be used.\n\n Raises:\n EnvironmentError: If the App Engine APIs are unavailable.\n \"\"\"\n if app_identity is None:\n raise EnvironmentError(\n 'The App Engine APIs are not available.')\n\n super(Credentials, self).__init__()\n self._scopes = scopes\n self._service_account_id = service_account_id\n\n @_helpers.copy_docstring(credentials.Credentials)\n def refresh(self, request):\n # pylint: disable=unused-argument\n token, ttl = app_identity.get_access_token(\n self._scopes, self._service_account_id)\n expiry = _helpers.utcnow() + datetime.timedelta(seconds=ttl)\n\n self.token, self.expiry = token, expiry\n\n @property\n def service_account_email(self):\n \"\"\"The service account email.\"\"\"\n if self._service_account_id is None:\n self._service_account_id = app_identity.get_service_account_name()\n return self._service_account_id\n\n @property\n def requires_scopes(self):\n \"\"\"Checks if the credentials requires scopes.\n\n Returns:\n bool: True if there are no scopes set otherwise False.\n \"\"\"\n return not self._scopes\n\n @_helpers.copy_docstring(credentials.Scoped)\n def with_scopes(self, scopes):\n return Credentials(\n scopes=scopes, service_account_id=self._service_account_id)\n\n @_helpers.copy_docstring(credentials.Signing)\n def sign_bytes(self, message):\n return app_identity.sign_blob(message)\n\n @property\n @_helpers.copy_docstring(credentials.Signing)\n def signer_email(self):\n return self.service_account_email\n", "path": "google/auth/app_engine.py"}, {"content": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Cryptography helpers for verifying and signing messages.\n\nUses the ``rsa``, ``pyasn1`` and ``pyasn1_modules`` packages\nto parse PEM files storing PKCS#1 or PKCS#8 keys as well as\ncertificates. There is no support for p12 files.\n\nThe simplest way to verify signatures is using :func:`verify_signature`::\n\n cert = open('certs.pem').read()\n valid = crypt.verify_signature(message, signature, cert)\n\nIf you're going to verify many messages with the same certificate, you can use\n:class:`Verifier`::\n\n cert = open('certs.pem').read()\n verifier = crypt.Verifier.from_string(cert)\n valid = verifier.verify(message, signature)\n\n\nTo sign messages use :class:`Signer` with a private key::\n\n private_key = open('private_key.pem').read()\n signer = crypt.Signer(private_key)\n signature = signer.sign(message)\n\n\"\"\"\n\nfrom pyasn1.codec.der import decoder\nfrom pyasn1_modules import pem\nfrom pyasn1_modules.rfc2459 import Certificate\nfrom pyasn1_modules.rfc5208 import PrivateKeyInfo\nimport rsa\nimport six\n\nfrom google.auth import _helpers\n\n_POW2 = (128, 64, 32, 16, 8, 4, 2, 1)\n_CERTIFICATE_MARKER = b'-----BEGIN CERTIFICATE-----'\n_PKCS1_MARKER = ('-----BEGIN RSA PRIVATE KEY-----',\n '-----END RSA PRIVATE KEY-----')\n_PKCS8_MARKER = ('-----BEGIN PRIVATE KEY-----',\n '-----END PRIVATE KEY-----')\n_PKCS8_SPEC = PrivateKeyInfo()\n\n\ndef _bit_list_to_bytes(bit_list):\n \"\"\"Converts an iterable of 1s and 0s to bytes.\n\n Combines the list 8 at a time, treating each group of 8 bits\n as a single byte.\n\n Args:\n bit_list (Sequence): Sequence of 1s and 0s.\n\n Returns:\n bytes: The decoded bytes.\n \"\"\"\n num_bits = len(bit_list)\n byte_vals = bytearray()\n for start in six.moves.xrange(0, num_bits, 8):\n curr_bits = bit_list[start:start + 8]\n char_val = sum(val * digit\n for val, digit in six.moves.zip(_POW2, curr_bits))\n byte_vals.append(char_val)\n return bytes(byte_vals)\n\n\nclass Verifier(object):\n \"\"\"This object is used to verify cryptographic signatures.\n\n Args:\n public_key (rsa.key.PublicKey): The public key used to verify\n signatures.\n \"\"\"\n\n def __init__(self, public_key):\n self._pubkey = public_key\n\n def verify(self, message, signature):\n \"\"\"Verifies a message against a cryptographic signature.\n\n Args:\n message (Union[str, bytes]): The message to verify.\n signature (Union[str, bytes]): The cryptography signature to check.\n\n Returns:\n bool: True if message was signed by the private key associated\n with the public key that this object was constructed with.\n \"\"\"\n message = _helpers.to_bytes(message)\n try:\n return rsa.pkcs1.verify(message, signature, self._pubkey)\n except (ValueError, rsa.pkcs1.VerificationError):\n return False\n\n @classmethod\n def from_string(cls, public_key):\n \"\"\"Construct an Verifier instance from a public key or public\n certificate string.\n\n Args:\n public_key (Union[str, bytes]): The public key in PEM format or the\n x509 public key certificate.\n\n Returns:\n Verifier: The constructed verifier.\n\n Raises:\n ValueError: If the public_key can't be parsed.\n \"\"\"\n public_key = _helpers.to_bytes(public_key)\n is_x509_cert = _CERTIFICATE_MARKER in public_key\n\n # If this is a certificate, extract the public key info.\n if is_x509_cert:\n der = rsa.pem.load_pem(public_key, 'CERTIFICATE')\n asn1_cert, remaining = decoder.decode(der, asn1Spec=Certificate())\n if remaining != b'':\n raise ValueError('Unused bytes', remaining)\n\n cert_info = asn1_cert['tbsCertificate']['subjectPublicKeyInfo']\n key_bytes = _bit_list_to_bytes(cert_info['subjectPublicKey'])\n pubkey = rsa.PublicKey.load_pkcs1(key_bytes, 'DER')\n else:\n pubkey = rsa.PublicKey.load_pkcs1(public_key, 'PEM')\n return cls(pubkey)\n\n\ndef verify_signature(message, signature, certs):\n \"\"\"Verify a cryptographic signature.\n\n Checks that the provided ``signature`` was generated from ``bytes`` using\n the private key associated with the ``cert``.\n\n Args:\n message (Union[str, bytes]): The plaintext message.\n signature (Union[str, bytes]): The cryptographic signature to check.\n certs (Union[Sequence, str, bytes]): The certificate or certificates\n to use to check the signature.\n\n Returns:\n bool: True if the signature is valid, otherwise False.\n \"\"\"\n if isinstance(certs, (six.text_type, six.binary_type)):\n certs = [certs]\n\n for cert in certs:\n verifier = Verifier.from_string(cert)\n if verifier.verify(message, signature):\n return True\n return False\n\n\nclass Signer(object):\n \"\"\"Signs messages with a private key.\n\n Args:\n private_key (rsa.key.PrivateKey): The private key to sign with.\n key_id (str): Optional key ID used to identify this private key. This\n can be useful to associate the private key with its associated\n public key or certificate.\n \"\"\"\n\n def __init__(self, private_key, key_id=None):\n self._key = private_key\n self.key_id = key_id\n\n def sign(self, message):\n \"\"\"Signs a message.\n\n Args:\n message (Union[str, bytes]): The message to be signed.\n\n Returns:\n bytes: The signature of the message for the given key.\n \"\"\"\n message = _helpers.to_bytes(message)\n return rsa.pkcs1.sign(message, self._key, 'SHA-256')\n\n @classmethod\n def from_string(cls, key, key_id=None):\n \"\"\"Construct an Signer instance from a private key in PEM format.\n\n Args:\n key (str): Private key in PEM format.\n key_id (str): An optional key id used to identify the private key.\n\n Returns:\n Signer: The constructed signer.\n\n Raises:\n ValueError: If the key cannot be parsed as PKCS#1 or PKCS#8 in\n PEM format.\n \"\"\"\n key = _helpers.from_bytes(key) # PEM expects str in Python 3\n marker_id, key_bytes = pem.readPemBlocksFromFile(\n six.StringIO(key), _PKCS1_MARKER, _PKCS8_MARKER)\n\n # Key is in pkcs1 format.\n if marker_id == 0:\n private_key = rsa.key.PrivateKey.load_pkcs1(\n key_bytes, format='DER')\n # Key is in pkcs8.\n elif marker_id == 1:\n key_info, remaining = decoder.decode(\n key_bytes, asn1Spec=_PKCS8_SPEC)\n if remaining != b'':\n raise ValueError('Unused bytes', remaining)\n private_key_info = key_info.getComponentByName('privateKey')\n private_key = rsa.key.PrivateKey.load_pkcs1(\n private_key_info.asOctets(), format='DER')\n else:\n raise ValueError('No key could be detected.')\n\n return cls(private_key, key_id=key_id)\n", "path": "google/auth/crypt.py"}], "after_files": [{"content": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Google App Engine standard environment support.\n\nThis module provides authentication and signing for applications running on App\nEngine in the standard environment using the `App Identity API`_.\n\n\n.. _App Identity API:\n https://cloud.google.com/appengine/docs/python/appidentity/\n\"\"\"\n\nimport datetime\n\nfrom google.auth import _helpers\nfrom google.auth import credentials\n\ntry:\n from google.appengine.api import app_identity\nexcept ImportError:\n app_identity = None\n\n\nclass Signer(object):\n \"\"\"Signs messages using the App Engine app identity service.\n\n This can be used in place of :class:`google.auth.crypt.Signer` when\n running in the App Engine standard environment.\n \"\"\"\n def __init__(self):\n self.key_id = None\n\n @staticmethod\n def sign(message):\n \"\"\"Signs a message.\n\n Args:\n message (Union[str, bytes]): The message to be signed.\n\n Returns:\n bytes: The signature of the message.\n \"\"\"\n message = _helpers.to_bytes(message)\n return app_identity.sign_blob(message)\n\n\ndef get_project_id():\n \"\"\"Gets the project ID for the current App Engine application.\n\n Returns:\n str: The project ID\n\n Raises:\n EnvironmentError: If the App Engine APIs are unavailable.\n \"\"\"\n if app_identity is None:\n raise EnvironmentError(\n 'The App Engine APIs are not available.')\n return app_identity.get_application_id()\n\n\nclass Credentials(credentials.Scoped, credentials.Signing,\n credentials.Credentials):\n \"\"\"App Engine standard environment credentials.\n\n These credentials use the App Engine App Identity API to obtain access\n tokens.\n \"\"\"\n\n def __init__(self, scopes=None, service_account_id=None):\n \"\"\"\n Args:\n scopes (Sequence[str]): Scopes to request from the App Identity\n API.\n service_account_id (str): The service account ID passed into\n :func:`google.appengine.api.app_identity.get_access_token`.\n If not specified, the default application service account\n ID will be used.\n\n Raises:\n EnvironmentError: If the App Engine APIs are unavailable.\n \"\"\"\n if app_identity is None:\n raise EnvironmentError(\n 'The App Engine APIs are not available.')\n\n super(Credentials, self).__init__()\n self._scopes = scopes\n self._service_account_id = service_account_id\n\n @_helpers.copy_docstring(credentials.Credentials)\n def refresh(self, request):\n # pylint: disable=unused-argument\n token, ttl = app_identity.get_access_token(\n self._scopes, self._service_account_id)\n expiry = _helpers.utcnow() + datetime.timedelta(seconds=ttl)\n\n self.token, self.expiry = token, expiry\n\n @property\n def service_account_email(self):\n \"\"\"The service account email.\"\"\"\n if self._service_account_id is None:\n self._service_account_id = app_identity.get_service_account_name()\n return self._service_account_id\n\n @property\n def requires_scopes(self):\n \"\"\"Checks if the credentials requires scopes.\n\n Returns:\n bool: True if there are no scopes set otherwise False.\n \"\"\"\n return not self._scopes\n\n @_helpers.copy_docstring(credentials.Scoped)\n def with_scopes(self, scopes):\n return Credentials(\n scopes=scopes, service_account_id=self._service_account_id)\n\n @_helpers.copy_docstring(credentials.Signing)\n def sign_bytes(self, message):\n return Signer().sign(message)\n\n @property\n @_helpers.copy_docstring(credentials.Signing)\n def signer_email(self):\n return self.service_account_email\n", "path": "google/auth/app_engine.py"}, {"content": "# Copyright 2016 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Cryptography helpers for verifying and signing messages.\n\nUses the ``rsa``, ``pyasn1`` and ``pyasn1_modules`` packages\nto parse PEM files storing PKCS#1 or PKCS#8 keys as well as\ncertificates. There is no support for p12 files.\n\nThe simplest way to verify signatures is using :func:`verify_signature`::\n\n cert = open('certs.pem').read()\n valid = crypt.verify_signature(message, signature, cert)\n\nIf you're going to verify many messages with the same certificate, you can use\n:class:`Verifier`::\n\n cert = open('certs.pem').read()\n verifier = crypt.Verifier.from_string(cert)\n valid = verifier.verify(message, signature)\n\n\nTo sign messages use :class:`Signer` with a private key::\n\n private_key = open('private_key.pem').read()\n signer = crypt.Signer(private_key)\n signature = signer.sign(message)\n\n\"\"\"\n\nfrom pyasn1.codec.der import decoder\nfrom pyasn1_modules import pem\nfrom pyasn1_modules.rfc2459 import Certificate\nfrom pyasn1_modules.rfc5208 import PrivateKeyInfo\nimport rsa\nimport six\n\nfrom google.auth import _helpers\n\n_POW2 = (128, 64, 32, 16, 8, 4, 2, 1)\n_CERTIFICATE_MARKER = b'-----BEGIN CERTIFICATE-----'\n_PKCS1_MARKER = ('-----BEGIN RSA PRIVATE KEY-----',\n '-----END RSA PRIVATE KEY-----')\n_PKCS8_MARKER = ('-----BEGIN PRIVATE KEY-----',\n '-----END PRIVATE KEY-----')\n_PKCS8_SPEC = PrivateKeyInfo()\n\n\ndef _bit_list_to_bytes(bit_list):\n \"\"\"Converts an iterable of 1s and 0s to bytes.\n\n Combines the list 8 at a time, treating each group of 8 bits\n as a single byte.\n\n Args:\n bit_list (Sequence): Sequence of 1s and 0s.\n\n Returns:\n bytes: The decoded bytes.\n \"\"\"\n num_bits = len(bit_list)\n byte_vals = bytearray()\n for start in six.moves.xrange(0, num_bits, 8):\n curr_bits = bit_list[start:start + 8]\n char_val = sum(val * digit\n for val, digit in six.moves.zip(_POW2, curr_bits))\n byte_vals.append(char_val)\n return bytes(byte_vals)\n\n\nclass Verifier(object):\n \"\"\"This object is used to verify cryptographic signatures.\n\n Args:\n public_key (rsa.key.PublicKey): The public key used to verify\n signatures.\n \"\"\"\n\n def __init__(self, public_key):\n self._pubkey = public_key\n\n def verify(self, message, signature):\n \"\"\"Verifies a message against a cryptographic signature.\n\n Args:\n message (Union[str, bytes]): The message to verify.\n signature (Union[str, bytes]): The cryptography signature to check.\n\n Returns:\n bool: True if message was signed by the private key associated\n with the public key that this object was constructed with.\n \"\"\"\n message = _helpers.to_bytes(message)\n try:\n return rsa.pkcs1.verify(message, signature, self._pubkey)\n except (ValueError, rsa.pkcs1.VerificationError):\n return False\n\n @classmethod\n def from_string(cls, public_key):\n \"\"\"Construct an Verifier instance from a public key or public\n certificate string.\n\n Args:\n public_key (Union[str, bytes]): The public key in PEM format or the\n x509 public key certificate.\n\n Returns:\n Verifier: The constructed verifier.\n\n Raises:\n ValueError: If the public_key can't be parsed.\n \"\"\"\n public_key = _helpers.to_bytes(public_key)\n is_x509_cert = _CERTIFICATE_MARKER in public_key\n\n # If this is a certificate, extract the public key info.\n if is_x509_cert:\n der = rsa.pem.load_pem(public_key, 'CERTIFICATE')\n asn1_cert, remaining = decoder.decode(der, asn1Spec=Certificate())\n if remaining != b'':\n raise ValueError('Unused bytes', remaining)\n\n cert_info = asn1_cert['tbsCertificate']['subjectPublicKeyInfo']\n key_bytes = _bit_list_to_bytes(cert_info['subjectPublicKey'])\n pubkey = rsa.PublicKey.load_pkcs1(key_bytes, 'DER')\n else:\n pubkey = rsa.PublicKey.load_pkcs1(public_key, 'PEM')\n return cls(pubkey)\n\n\ndef verify_signature(message, signature, certs):\n \"\"\"Verify a cryptographic signature.\n\n Checks that the provided ``signature`` was generated from ``bytes`` using\n the private key associated with the ``cert``.\n\n Args:\n message (Union[str, bytes]): The plaintext message.\n signature (Union[str, bytes]): The cryptographic signature to check.\n certs (Union[Sequence, str, bytes]): The certificate or certificates\n to use to check the signature.\n\n Returns:\n bool: True if the signature is valid, otherwise False.\n \"\"\"\n if isinstance(certs, (six.text_type, six.binary_type)):\n certs = [certs]\n\n for cert in certs:\n verifier = Verifier.from_string(cert)\n if verifier.verify(message, signature):\n return True\n return False\n\n\nclass Signer(object):\n \"\"\"Signs messages with a private key.\n\n Args:\n private_key (rsa.key.PrivateKey): The private key to sign with.\n key_id (str): Optional key ID used to identify this private key. This\n can be useful to associate the private key with its associated\n public key or certificate.\n \"\"\"\n\n def __init__(self, private_key, key_id=None):\n self._key = private_key\n self.key_id = key_id\n\n def sign(self, message):\n \"\"\"Signs a message.\n\n Args:\n message (Union[str, bytes]): The message to be signed.\n\n Returns:\n bytes: The signature of the message.\n \"\"\"\n message = _helpers.to_bytes(message)\n return rsa.pkcs1.sign(message, self._key, 'SHA-256')\n\n @classmethod\n def from_string(cls, key, key_id=None):\n \"\"\"Construct an Signer instance from a private key in PEM format.\n\n Args:\n key (str): Private key in PEM format.\n key_id (str): An optional key id used to identify the private key.\n\n Returns:\n Signer: The constructed signer.\n\n Raises:\n ValueError: If the key cannot be parsed as PKCS#1 or PKCS#8 in\n PEM format.\n \"\"\"\n key = _helpers.from_bytes(key) # PEM expects str in Python 3\n marker_id, key_bytes = pem.readPemBlocksFromFile(\n six.StringIO(key), _PKCS1_MARKER, _PKCS8_MARKER)\n\n # Key is in pkcs1 format.\n if marker_id == 0:\n private_key = rsa.key.PrivateKey.load_pkcs1(\n key_bytes, format='DER')\n # Key is in pkcs8.\n elif marker_id == 1:\n key_info, remaining = decoder.decode(\n key_bytes, asn1Spec=_PKCS8_SPEC)\n if remaining != b'':\n raise ValueError('Unused bytes', remaining)\n private_key_info = key_info.getComponentByName('privateKey')\n private_key = rsa.key.PrivateKey.load_pkcs1(\n private_key_info.asOctets(), format='DER')\n else:\n raise ValueError('No key could be detected.')\n\n return cls(private_key, key_id=key_id)\n", "path": "google/auth/crypt.py"}]}
| 3,684 | 518 |
gh_patches_debug_40206
|
rasdani/github-patches
|
git_diff
|
networkx__networkx-3072
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Behavior of `nx.union_all` when `graphs=[]`
I'm running into a case where I'm passing in an empty list to `nx.union_all`, and it returns a None type.
While this is not necessarily the wrong thing to do, it is not documented.
Intuitively, I would expect the result of union with no inputs to be an empty graph, but the issue here is that you don't know what type the graph should be. Therefore I think the best behavior would be to raise a ValueError indicating that the input cannot be empty. This would make it more clear where the code is failing.
Current behavior:
```python
>>> nx.union_all([nx.path_graph([1, 2])])
<networkx.classes.graph.Graph at 0x7f6fb15d1ac8>
>>> nx.union_all([nx.path_graph([1, 2]), nx.path_graph([3, 4])])
<networkx.classes.graph.Graph at 0x7f6fb1477ac8>
>>> print(nx.union_all([]))
None
```
Proposed Behavior:
```python
>>> print(nx.union_all([]))
ValueError: Cannot union_all an empty list
```
Behavior of `nx.union_all` when `graphs=[]`
I'm running into a case where I'm passing in an empty list to `nx.union_all`, and it returns a None type.
While this is not necessarily the wrong thing to do, it is not documented.
Intuitively, I would expect the result of union with no inputs to be an empty graph, but the issue here is that you don't know what type the graph should be. Therefore I think the best behavior would be to raise a ValueError indicating that the input cannot be empty. This would make it more clear where the code is failing.
Current behavior:
```python
>>> nx.union_all([nx.path_graph([1, 2])])
<networkx.classes.graph.Graph at 0x7f6fb15d1ac8>
>>> nx.union_all([nx.path_graph([1, 2]), nx.path_graph([3, 4])])
<networkx.classes.graph.Graph at 0x7f6fb1477ac8>
>>> print(nx.union_all([]))
None
```
Proposed Behavior:
```python
>>> print(nx.union_all([]))
ValueError: Cannot union_all an empty list
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `networkx/algorithms/operators/all.py`
Content:
```
1 """Operations on many graphs.
2 """
3 # Copyright (C) 2013 by
4 # Aric Hagberg <[email protected]>
5 # Dan Schult <[email protected]>
6 # Pieter Swart <[email protected]>
7 # All rights reserved.
8 # BSD license.
9 try:
10 from itertools import izip_longest as zip_longest
11 except ImportError: # Python3 has zip_longest
12 from itertools import zip_longest
13 import networkx as nx
14
15 __author__ = """\n""".join(['Robert King <[email protected]>',
16 'Aric Hagberg <[email protected]>'])
17
18 __all__ = ['union_all', 'compose_all', 'disjoint_union_all',
19 'intersection_all']
20
21
22 def union_all(graphs, rename=(None,)):
23 """Return the union of all graphs.
24
25 The graphs must be disjoint, otherwise an exception is raised.
26
27 Parameters
28 ----------
29 graphs : list of graphs
30 List of NetworkX graphs
31
32 rename : bool , default=(None, None)
33 Node names of G and H can be changed by specifying the tuple
34 rename=('G-','H-') (for example). Node "u" in G is then renamed
35 "G-u" and "v" in H is renamed "H-v".
36
37 Returns
38 -------
39 U : a graph with the same type as the first graph in list
40
41 Notes
42 -----
43 To force a disjoint union with node relabeling, use
44 disjoint_union_all(G,H) or convert_node_labels_to integers().
45
46 Graph, edge, and node attributes are propagated to the union graph.
47 If a graph attribute is present in multiple graphs, then the value
48 from the last graph in the list with that attribute is used.
49
50 See Also
51 --------
52 union
53 disjoint_union_all
54 """
55 graphs_names = zip_longest(graphs, rename)
56 U, gname = next(graphs_names)
57 for H, hname in graphs_names:
58 U = nx.union(U, H, (gname, hname))
59 gname = None
60 return U
61
62
63 def disjoint_union_all(graphs):
64 """Return the disjoint union of all graphs.
65
66 This operation forces distinct integer node labels starting with 0
67 for the first graph in the list and numbering consecutively.
68
69 Parameters
70 ----------
71 graphs : list
72 List of NetworkX graphs
73
74 Returns
75 -------
76 U : A graph with the same type as the first graph in list
77
78 Notes
79 -----
80 It is recommended that the graphs be either all directed or all undirected.
81
82 Graph, edge, and node attributes are propagated to the union graph.
83 If a graph attribute is present in multiple graphs, then the value
84 from the last graph in the list with that attribute is used.
85 """
86 graphs = iter(graphs)
87 U = next(graphs)
88 for H in graphs:
89 U = nx.disjoint_union(U, H)
90 return U
91
92
93 def compose_all(graphs):
94 """Return the composition of all graphs.
95
96 Composition is the simple union of the node sets and edge sets.
97 The node sets of the supplied graphs need not be disjoint.
98
99 Parameters
100 ----------
101 graphs : list
102 List of NetworkX graphs
103
104 Returns
105 -------
106 C : A graph with the same type as the first graph in list
107
108 Notes
109 -----
110 It is recommended that the supplied graphs be either all directed or all
111 undirected.
112
113 Graph, edge, and node attributes are propagated to the union graph.
114 If a graph attribute is present in multiple graphs, then the value
115 from the last graph in the list with that attribute is used.
116 """
117 graphs = iter(graphs)
118 C = next(graphs)
119 for H in graphs:
120 C = nx.compose(C, H)
121 return C
122
123
124 def intersection_all(graphs):
125 """Return a new graph that contains only the edges that exist in
126 all graphs.
127
128 All supplied graphs must have the same node set.
129
130 Parameters
131 ----------
132 graphs : list
133 List of NetworkX graphs
134
135 Returns
136 -------
137 R : A new graph with the same type as the first graph in list
138
139 Notes
140 -----
141 Attributes from the graph, nodes, and edges are not copied to the new
142 graph.
143 """
144 graphs = iter(graphs)
145 R = next(graphs)
146 for H in graphs:
147 R = nx.intersection(R, H)
148 return R
149
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/networkx/algorithms/operators/all.py b/networkx/algorithms/operators/all.py
--- a/networkx/algorithms/operators/all.py
+++ b/networkx/algorithms/operators/all.py
@@ -38,6 +38,11 @@
-------
U : a graph with the same type as the first graph in list
+ Raises
+ ------
+ ValueError
+ If `graphs` is an empty list.
+
Notes
-----
To force a disjoint union with node relabeling, use
@@ -52,6 +57,8 @@
union
disjoint_union_all
"""
+ if not graphs:
+ raise ValueError('cannot apply union_all to an empty list')
graphs_names = zip_longest(graphs, rename)
U, gname = next(graphs_names)
for H, hname in graphs_names:
@@ -75,6 +82,11 @@
-------
U : A graph with the same type as the first graph in list
+ Raises
+ ------
+ ValueError
+ If `graphs` is an empty list.
+
Notes
-----
It is recommended that the graphs be either all directed or all undirected.
@@ -83,6 +95,8 @@
If a graph attribute is present in multiple graphs, then the value
from the last graph in the list with that attribute is used.
"""
+ if not graphs:
+ raise ValueError('cannot apply disjoint_union_all to an empty list')
graphs = iter(graphs)
U = next(graphs)
for H in graphs:
@@ -105,6 +119,11 @@
-------
C : A graph with the same type as the first graph in list
+ Raises
+ ------
+ ValueError
+ If `graphs` is an empty list.
+
Notes
-----
It is recommended that the supplied graphs be either all directed or all
@@ -114,6 +133,8 @@
If a graph attribute is present in multiple graphs, then the value
from the last graph in the list with that attribute is used.
"""
+ if not graphs:
+ raise ValueError('cannot apply compose_all to an empty list')
graphs = iter(graphs)
C = next(graphs)
for H in graphs:
@@ -136,11 +157,18 @@
-------
R : A new graph with the same type as the first graph in list
+ Raises
+ ------
+ ValueError
+ If `graphs` is an empty list.
+
Notes
-----
Attributes from the graph, nodes, and edges are not copied to the new
graph.
"""
+ if not graphs:
+ raise ValueError('cannot apply intersection_all to an empty list')
graphs = iter(graphs)
R = next(graphs)
for H in graphs:
|
{"golden_diff": "diff --git a/networkx/algorithms/operators/all.py b/networkx/algorithms/operators/all.py\n--- a/networkx/algorithms/operators/all.py\n+++ b/networkx/algorithms/operators/all.py\n@@ -38,6 +38,11 @@\n -------\n U : a graph with the same type as the first graph in list\n \n+ Raises\n+ ------\n+ ValueError\n+ If `graphs` is an empty list.\n+\n Notes\n -----\n To force a disjoint union with node relabeling, use\n@@ -52,6 +57,8 @@\n union\n disjoint_union_all\n \"\"\"\n+ if not graphs:\n+ raise ValueError('cannot apply union_all to an empty list')\n graphs_names = zip_longest(graphs, rename)\n U, gname = next(graphs_names)\n for H, hname in graphs_names:\n@@ -75,6 +82,11 @@\n -------\n U : A graph with the same type as the first graph in list\n \n+ Raises\n+ ------\n+ ValueError\n+ If `graphs` is an empty list.\n+\n Notes\n -----\n It is recommended that the graphs be either all directed or all undirected.\n@@ -83,6 +95,8 @@\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n+ if not graphs:\n+ raise ValueError('cannot apply disjoint_union_all to an empty list')\n graphs = iter(graphs)\n U = next(graphs)\n for H in graphs:\n@@ -105,6 +119,11 @@\n -------\n C : A graph with the same type as the first graph in list\n \n+ Raises\n+ ------\n+ ValueError\n+ If `graphs` is an empty list.\n+\n Notes\n -----\n It is recommended that the supplied graphs be either all directed or all\n@@ -114,6 +133,8 @@\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n+ if not graphs:\n+ raise ValueError('cannot apply compose_all to an empty list')\n graphs = iter(graphs)\n C = next(graphs)\n for H in graphs:\n@@ -136,11 +157,18 @@\n -------\n R : A new graph with the same type as the first graph in list\n \n+ Raises\n+ ------\n+ ValueError\n+ If `graphs` is an empty list.\n+\n Notes\n -----\n Attributes from the graph, nodes, and edges are not copied to the new\n graph.\n \"\"\"\n+ if not graphs:\n+ raise ValueError('cannot apply intersection_all to an empty list')\n graphs = iter(graphs)\n R = next(graphs)\n for H in graphs:\n", "issue": "Behavior of `nx.union_all` when `graphs=[]`\nI'm running into a case where I'm passing in an empty list to `nx.union_all`, and it returns a None type. \r\n\r\nWhile this is not necessarily the wrong thing to do, it is not documented.\r\n\r\nIntuitively, I would expect the result of union with no inputs to be an empty graph, but the issue here is that you don't know what type the graph should be. Therefore I think the best behavior would be to raise a ValueError indicating that the input cannot be empty. This would make it more clear where the code is failing. \r\n\r\nCurrent behavior:\r\n\r\n```python\r\n\r\n>>> nx.union_all([nx.path_graph([1, 2])])\r\n<networkx.classes.graph.Graph at 0x7f6fb15d1ac8>\r\n\r\n>>> nx.union_all([nx.path_graph([1, 2]), nx.path_graph([3, 4])])\r\n<networkx.classes.graph.Graph at 0x7f6fb1477ac8>\r\n\r\n>>> print(nx.union_all([]))\r\nNone\r\n\r\n```\r\n\r\nProposed Behavior:\r\n\r\n```python\r\n>>> print(nx.union_all([]))\r\nValueError: Cannot union_all an empty list\r\n```\nBehavior of `nx.union_all` when `graphs=[]`\nI'm running into a case where I'm passing in an empty list to `nx.union_all`, and it returns a None type. \r\n\r\nWhile this is not necessarily the wrong thing to do, it is not documented.\r\n\r\nIntuitively, I would expect the result of union with no inputs to be an empty graph, but the issue here is that you don't know what type the graph should be. Therefore I think the best behavior would be to raise a ValueError indicating that the input cannot be empty. This would make it more clear where the code is failing. \r\n\r\nCurrent behavior:\r\n\r\n```python\r\n\r\n>>> nx.union_all([nx.path_graph([1, 2])])\r\n<networkx.classes.graph.Graph at 0x7f6fb15d1ac8>\r\n\r\n>>> nx.union_all([nx.path_graph([1, 2]), nx.path_graph([3, 4])])\r\n<networkx.classes.graph.Graph at 0x7f6fb1477ac8>\r\n\r\n>>> print(nx.union_all([]))\r\nNone\r\n\r\n```\r\n\r\nProposed Behavior:\r\n\r\n```python\r\n>>> print(nx.union_all([]))\r\nValueError: Cannot union_all an empty list\r\n```\n", "before_files": [{"content": "\"\"\"Operations on many graphs.\n\"\"\"\n# Copyright (C) 2013 by\n# Aric Hagberg <[email protected]>\n# Dan Schult <[email protected]>\n# Pieter Swart <[email protected]>\n# All rights reserved.\n# BSD license.\ntry:\n from itertools import izip_longest as zip_longest\nexcept ImportError: # Python3 has zip_longest\n from itertools import zip_longest\nimport networkx as nx\n\n__author__ = \"\"\"\\n\"\"\".join(['Robert King <[email protected]>',\n 'Aric Hagberg <[email protected]>'])\n\n__all__ = ['union_all', 'compose_all', 'disjoint_union_all',\n 'intersection_all']\n\n\ndef union_all(graphs, rename=(None,)):\n \"\"\"Return the union of all graphs.\n\n The graphs must be disjoint, otherwise an exception is raised.\n\n Parameters\n ----------\n graphs : list of graphs\n List of NetworkX graphs\n\n rename : bool , default=(None, None)\n Node names of G and H can be changed by specifying the tuple\n rename=('G-','H-') (for example). Node \"u\" in G is then renamed\n \"G-u\" and \"v\" in H is renamed \"H-v\".\n\n Returns\n -------\n U : a graph with the same type as the first graph in list\n\n Notes\n -----\n To force a disjoint union with node relabeling, use\n disjoint_union_all(G,H) or convert_node_labels_to integers().\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n\n See Also\n --------\n union\n disjoint_union_all\n \"\"\"\n graphs_names = zip_longest(graphs, rename)\n U, gname = next(graphs_names)\n for H, hname in graphs_names:\n U = nx.union(U, H, (gname, hname))\n gname = None\n return U\n\n\ndef disjoint_union_all(graphs):\n \"\"\"Return the disjoint union of all graphs.\n\n This operation forces distinct integer node labels starting with 0\n for the first graph in the list and numbering consecutively.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n U : A graph with the same type as the first graph in list\n\n Notes\n -----\n It is recommended that the graphs be either all directed or all undirected.\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n graphs = iter(graphs)\n U = next(graphs)\n for H in graphs:\n U = nx.disjoint_union(U, H)\n return U\n\n\ndef compose_all(graphs):\n \"\"\"Return the composition of all graphs.\n\n Composition is the simple union of the node sets and edge sets.\n The node sets of the supplied graphs need not be disjoint.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n C : A graph with the same type as the first graph in list\n\n Notes\n -----\n It is recommended that the supplied graphs be either all directed or all\n undirected.\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n graphs = iter(graphs)\n C = next(graphs)\n for H in graphs:\n C = nx.compose(C, H)\n return C\n\n\ndef intersection_all(graphs):\n \"\"\"Return a new graph that contains only the edges that exist in\n all graphs.\n\n All supplied graphs must have the same node set.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n R : A new graph with the same type as the first graph in list\n\n Notes\n -----\n Attributes from the graph, nodes, and edges are not copied to the new\n graph.\n \"\"\"\n graphs = iter(graphs)\n R = next(graphs)\n for H in graphs:\n R = nx.intersection(R, H)\n return R\n", "path": "networkx/algorithms/operators/all.py"}], "after_files": [{"content": "\"\"\"Operations on many graphs.\n\"\"\"\n# Copyright (C) 2013 by\n# Aric Hagberg <[email protected]>\n# Dan Schult <[email protected]>\n# Pieter Swart <[email protected]>\n# All rights reserved.\n# BSD license.\ntry:\n from itertools import izip_longest as zip_longest\nexcept ImportError: # Python3 has zip_longest\n from itertools import zip_longest\nimport networkx as nx\n\n__author__ = \"\"\"\\n\"\"\".join(['Robert King <[email protected]>',\n 'Aric Hagberg <[email protected]>'])\n\n__all__ = ['union_all', 'compose_all', 'disjoint_union_all',\n 'intersection_all']\n\n\ndef union_all(graphs, rename=(None,)):\n \"\"\"Return the union of all graphs.\n\n The graphs must be disjoint, otherwise an exception is raised.\n\n Parameters\n ----------\n graphs : list of graphs\n List of NetworkX graphs\n\n rename : bool , default=(None, None)\n Node names of G and H can be changed by specifying the tuple\n rename=('G-','H-') (for example). Node \"u\" in G is then renamed\n \"G-u\" and \"v\" in H is renamed \"H-v\".\n\n Returns\n -------\n U : a graph with the same type as the first graph in list\n\n Raises\n ------\n ValueError\n If `graphs` is an empty list.\n\n Notes\n -----\n To force a disjoint union with node relabeling, use\n disjoint_union_all(G,H) or convert_node_labels_to integers().\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n\n See Also\n --------\n union\n disjoint_union_all\n \"\"\"\n if not graphs:\n raise ValueError('cannot apply union_all to an empty list')\n graphs_names = zip_longest(graphs, rename)\n U, gname = next(graphs_names)\n for H, hname in graphs_names:\n U = nx.union(U, H, (gname, hname))\n gname = None\n return U\n\n\ndef disjoint_union_all(graphs):\n \"\"\"Return the disjoint union of all graphs.\n\n This operation forces distinct integer node labels starting with 0\n for the first graph in the list and numbering consecutively.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n U : A graph with the same type as the first graph in list\n\n Raises\n ------\n ValueError\n If `graphs` is an empty list.\n\n Notes\n -----\n It is recommended that the graphs be either all directed or all undirected.\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n if not graphs:\n raise ValueError('cannot apply disjoint_union_all to an empty list')\n graphs = iter(graphs)\n U = next(graphs)\n for H in graphs:\n U = nx.disjoint_union(U, H)\n return U\n\n\ndef compose_all(graphs):\n \"\"\"Return the composition of all graphs.\n\n Composition is the simple union of the node sets and edge sets.\n The node sets of the supplied graphs need not be disjoint.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n C : A graph with the same type as the first graph in list\n\n Raises\n ------\n ValueError\n If `graphs` is an empty list.\n\n Notes\n -----\n It is recommended that the supplied graphs be either all directed or all\n undirected.\n\n Graph, edge, and node attributes are propagated to the union graph.\n If a graph attribute is present in multiple graphs, then the value\n from the last graph in the list with that attribute is used.\n \"\"\"\n if not graphs:\n raise ValueError('cannot apply compose_all to an empty list')\n graphs = iter(graphs)\n C = next(graphs)\n for H in graphs:\n C = nx.compose(C, H)\n return C\n\n\ndef intersection_all(graphs):\n \"\"\"Return a new graph that contains only the edges that exist in\n all graphs.\n\n All supplied graphs must have the same node set.\n\n Parameters\n ----------\n graphs : list\n List of NetworkX graphs\n\n Returns\n -------\n R : A new graph with the same type as the first graph in list\n\n Raises\n ------\n ValueError\n If `graphs` is an empty list.\n\n Notes\n -----\n Attributes from the graph, nodes, and edges are not copied to the new\n graph.\n \"\"\"\n if not graphs:\n raise ValueError('cannot apply intersection_all to an empty list')\n graphs = iter(graphs)\n R = next(graphs)\n for H in graphs:\n R = nx.intersection(R, H)\n return R\n", "path": "networkx/algorithms/operators/all.py"}]}
| 2,099 | 637 |
gh_patches_debug_14403
|
rasdani/github-patches
|
git_diff
|
dbt-labs__dbt-core-4359
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Ensure that all dbt invocations have an invocation_id
This is a sub-issue of #4260.
In order to make sure we can distinguish logs from different runs we should always have an invocation_id. If tracking is off, it isn't created today so we would need to create it. (to add it to logging, add it as a class property on the superclass) (previously considered: preserve process id like it is today, global run_id reported at start, or allow some sort of user markers).
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `core/dbt/lib.py`
Content:
```
1 # TODO: this file is one big TODO
2 import os
3 from dbt.exceptions import RuntimeException
4 from dbt import flags
5 from collections import namedtuple
6
7 RuntimeArgs = namedtuple(
8 'RuntimeArgs', 'project_dir profiles_dir single_threaded profile_name'
9 )
10
11
12 def get_dbt_config(project_dir, single_threaded=False):
13 from dbt.config.runtime import RuntimeConfig
14 import dbt.adapters.factory
15
16 if os.getenv('DBT_PROFILES_DIR'):
17 profiles_dir = os.getenv('DBT_PROFILES_DIR')
18 else:
19 profiles_dir = os.path.expanduser("~/.dbt")
20
21 # Construct a phony config
22 config = RuntimeConfig.from_args(RuntimeArgs(
23 project_dir, profiles_dir, single_threaded, 'user'
24 ))
25 # Clear previously registered adapters--
26 # this fixes cacheing behavior on the dbt-server
27 flags.set_from_args('', config)
28 dbt.adapters.factory.reset_adapters()
29 # Load the relevant adapter
30 dbt.adapters.factory.register_adapter(config)
31
32 return config
33
34
35 def get_task_by_type(type):
36 # TODO: we need to tell dbt-server what tasks are available
37 from dbt.task.run import RunTask
38 from dbt.task.list import ListTask
39 from dbt.task.seed import SeedTask
40 from dbt.task.test import TestTask
41 from dbt.task.build import BuildTask
42 from dbt.task.snapshot import SnapshotTask
43 from dbt.task.run_operation import RunOperationTask
44
45 if type == 'run':
46 return RunTask
47 elif type == 'test':
48 return TestTask
49 elif type == 'list':
50 return ListTask
51 elif type == 'seed':
52 return SeedTask
53 elif type == 'build':
54 return BuildTask
55 elif type == 'snapshot':
56 return SnapshotTask
57 elif type == 'run_operation':
58 return RunOperationTask
59
60 raise RuntimeException('not a valid task')
61
62
63 def create_task(type, args, manifest, config):
64 task = get_task_by_type(type)
65
66 def no_op(*args, **kwargs):
67 pass
68
69 # TODO: yuck, let's rethink tasks a little
70 task = task(args, config)
71
72 # Wow! We can monkeypatch taskCls.load_manifest to return _our_ manifest
73 task.load_manifest = no_op
74 task.manifest = manifest
75 return task
76
77
78 def _get_operation_node(manifest, project_path, sql):
79 from dbt.parser.manifest import process_node
80 from dbt.parser.sql import SqlBlockParser
81 import dbt.adapters.factory
82
83 config = get_dbt_config(project_path)
84 block_parser = SqlBlockParser(
85 project=config,
86 manifest=manifest,
87 root_project=config,
88 )
89
90 adapter = dbt.adapters.factory.get_adapter(config)
91 # TODO : This needs a real name?
92 sql_node = block_parser.parse_remote(sql, 'name')
93 process_node(config, manifest, sql_node)
94 return config, sql_node, adapter
95
96
97 def compile_sql(manifest, project_path, sql):
98 from dbt.task.sql import SqlCompileRunner
99
100 config, node, adapter = _get_operation_node(manifest, project_path, sql)
101 runner = SqlCompileRunner(config, adapter, node, 1, 1)
102 return runner.safe_run(manifest)
103
104
105 def execute_sql(manifest, project_path, sql):
106 from dbt.task.sql import SqlExecuteRunner
107
108 config, node, adapter = _get_operation_node(manifest, project_path, sql)
109 runner = SqlExecuteRunner(config, adapter, node, 1, 1)
110 # TODO: use same interface for runner
111 return runner.safe_run(manifest)
112
113
114 def parse_to_manifest(config):
115 from dbt.parser.manifest import ManifestLoader
116
117 return ManifestLoader.get_full_manifest(config)
118
119
120 def deserialize_manifest(manifest_msgpack):
121 from dbt.contracts.graph.manifest import Manifest
122
123 return Manifest.from_msgpack(manifest_msgpack)
124
125
126 def serialize_manifest(manifest):
127 # TODO: what should this take as an arg?
128 return manifest.to_msgpack()
129
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/core/dbt/lib.py b/core/dbt/lib.py
--- a/core/dbt/lib.py
+++ b/core/dbt/lib.py
@@ -12,6 +12,7 @@
def get_dbt_config(project_dir, single_threaded=False):
from dbt.config.runtime import RuntimeConfig
import dbt.adapters.factory
+ import dbt.events.functions
if os.getenv('DBT_PROFILES_DIR'):
profiles_dir = os.getenv('DBT_PROFILES_DIR')
@@ -28,6 +29,8 @@
dbt.adapters.factory.reset_adapters()
# Load the relevant adapter
dbt.adapters.factory.register_adapter(config)
+ # Set invocation id
+ dbt.events.functions.set_invocation_id()
return config
|
{"golden_diff": "diff --git a/core/dbt/lib.py b/core/dbt/lib.py\n--- a/core/dbt/lib.py\n+++ b/core/dbt/lib.py\n@@ -12,6 +12,7 @@\n def get_dbt_config(project_dir, single_threaded=False):\n from dbt.config.runtime import RuntimeConfig\n import dbt.adapters.factory\n+ import dbt.events.functions\n \n if os.getenv('DBT_PROFILES_DIR'):\n profiles_dir = os.getenv('DBT_PROFILES_DIR')\n@@ -28,6 +29,8 @@\n dbt.adapters.factory.reset_adapters()\n # Load the relevant adapter\n dbt.adapters.factory.register_adapter(config)\n+ # Set invocation id\n+ dbt.events.functions.set_invocation_id()\n \n return config\n", "issue": "Ensure that all dbt invocations have an invocation_id\nThis is a sub-issue of #4260. \n\nIn order to make sure we can distinguish logs from different runs we should always have an invocation_id. If tracking is off, it isn't created today so we would need to create it. (to add it to logging, add it as a class property on the superclass) (previously considered: preserve process id like it is today, global run_id reported at start, or allow some sort of user markers).\n", "before_files": [{"content": "# TODO: this file is one big TODO\nimport os\nfrom dbt.exceptions import RuntimeException\nfrom dbt import flags\nfrom collections import namedtuple\n\nRuntimeArgs = namedtuple(\n 'RuntimeArgs', 'project_dir profiles_dir single_threaded profile_name'\n)\n\n\ndef get_dbt_config(project_dir, single_threaded=False):\n from dbt.config.runtime import RuntimeConfig\n import dbt.adapters.factory\n\n if os.getenv('DBT_PROFILES_DIR'):\n profiles_dir = os.getenv('DBT_PROFILES_DIR')\n else:\n profiles_dir = os.path.expanduser(\"~/.dbt\")\n\n # Construct a phony config\n config = RuntimeConfig.from_args(RuntimeArgs(\n project_dir, profiles_dir, single_threaded, 'user'\n ))\n # Clear previously registered adapters--\n # this fixes cacheing behavior on the dbt-server\n flags.set_from_args('', config)\n dbt.adapters.factory.reset_adapters()\n # Load the relevant adapter\n dbt.adapters.factory.register_adapter(config)\n\n return config\n\n\ndef get_task_by_type(type):\n # TODO: we need to tell dbt-server what tasks are available\n from dbt.task.run import RunTask\n from dbt.task.list import ListTask\n from dbt.task.seed import SeedTask\n from dbt.task.test import TestTask\n from dbt.task.build import BuildTask\n from dbt.task.snapshot import SnapshotTask\n from dbt.task.run_operation import RunOperationTask\n\n if type == 'run':\n return RunTask\n elif type == 'test':\n return TestTask\n elif type == 'list':\n return ListTask\n elif type == 'seed':\n return SeedTask\n elif type == 'build':\n return BuildTask\n elif type == 'snapshot':\n return SnapshotTask\n elif type == 'run_operation':\n return RunOperationTask\n\n raise RuntimeException('not a valid task')\n\n\ndef create_task(type, args, manifest, config):\n task = get_task_by_type(type)\n\n def no_op(*args, **kwargs):\n pass\n\n # TODO: yuck, let's rethink tasks a little\n task = task(args, config)\n\n # Wow! We can monkeypatch taskCls.load_manifest to return _our_ manifest\n task.load_manifest = no_op\n task.manifest = manifest\n return task\n\n\ndef _get_operation_node(manifest, project_path, sql):\n from dbt.parser.manifest import process_node\n from dbt.parser.sql import SqlBlockParser\n import dbt.adapters.factory\n\n config = get_dbt_config(project_path)\n block_parser = SqlBlockParser(\n project=config,\n manifest=manifest,\n root_project=config,\n )\n\n adapter = dbt.adapters.factory.get_adapter(config)\n # TODO : This needs a real name?\n sql_node = block_parser.parse_remote(sql, 'name')\n process_node(config, manifest, sql_node)\n return config, sql_node, adapter\n\n\ndef compile_sql(manifest, project_path, sql):\n from dbt.task.sql import SqlCompileRunner\n\n config, node, adapter = _get_operation_node(manifest, project_path, sql)\n runner = SqlCompileRunner(config, adapter, node, 1, 1)\n return runner.safe_run(manifest)\n\n\ndef execute_sql(manifest, project_path, sql):\n from dbt.task.sql import SqlExecuteRunner\n\n config, node, adapter = _get_operation_node(manifest, project_path, sql)\n runner = SqlExecuteRunner(config, adapter, node, 1, 1)\n # TODO: use same interface for runner\n return runner.safe_run(manifest)\n\n\ndef parse_to_manifest(config):\n from dbt.parser.manifest import ManifestLoader\n\n return ManifestLoader.get_full_manifest(config)\n\n\ndef deserialize_manifest(manifest_msgpack):\n from dbt.contracts.graph.manifest import Manifest\n\n return Manifest.from_msgpack(manifest_msgpack)\n\n\ndef serialize_manifest(manifest):\n # TODO: what should this take as an arg?\n return manifest.to_msgpack()\n", "path": "core/dbt/lib.py"}], "after_files": [{"content": "# TODO: this file is one big TODO\nimport os\nfrom dbt.exceptions import RuntimeException\nfrom dbt import flags\nfrom collections import namedtuple\n\nRuntimeArgs = namedtuple(\n 'RuntimeArgs', 'project_dir profiles_dir single_threaded profile_name'\n)\n\n\ndef get_dbt_config(project_dir, single_threaded=False):\n from dbt.config.runtime import RuntimeConfig\n import dbt.adapters.factory\n import dbt.events.functions\n\n if os.getenv('DBT_PROFILES_DIR'):\n profiles_dir = os.getenv('DBT_PROFILES_DIR')\n else:\n profiles_dir = os.path.expanduser(\"~/.dbt\")\n\n # Construct a phony config\n config = RuntimeConfig.from_args(RuntimeArgs(\n project_dir, profiles_dir, single_threaded, 'user'\n ))\n # Clear previously registered adapters--\n # this fixes cacheing behavior on the dbt-server\n flags.set_from_args('', config)\n dbt.adapters.factory.reset_adapters()\n # Load the relevant adapter\n dbt.adapters.factory.register_adapter(config)\n # Set invocation id\n dbt.events.functions.set_invocation_id()\n\n return config\n\n\ndef get_task_by_type(type):\n # TODO: we need to tell dbt-server what tasks are available\n from dbt.task.run import RunTask\n from dbt.task.list import ListTask\n from dbt.task.seed import SeedTask\n from dbt.task.test import TestTask\n from dbt.task.build import BuildTask\n from dbt.task.snapshot import SnapshotTask\n from dbt.task.run_operation import RunOperationTask\n\n if type == 'run':\n return RunTask\n elif type == 'test':\n return TestTask\n elif type == 'list':\n return ListTask\n elif type == 'seed':\n return SeedTask\n elif type == 'build':\n return BuildTask\n elif type == 'snapshot':\n return SnapshotTask\n elif type == 'run_operation':\n return RunOperationTask\n\n raise RuntimeException('not a valid task')\n\n\ndef create_task(type, args, manifest, config):\n task = get_task_by_type(type)\n\n def no_op(*args, **kwargs):\n pass\n\n # TODO: yuck, let's rethink tasks a little\n task = task(args, config)\n\n # Wow! We can monkeypatch taskCls.load_manifest to return _our_ manifest\n task.load_manifest = no_op\n task.manifest = manifest\n return task\n\n\ndef _get_operation_node(manifest, project_path, sql):\n from dbt.parser.manifest import process_node\n from dbt.parser.sql import SqlBlockParser\n import dbt.adapters.factory\n\n config = get_dbt_config(project_path)\n block_parser = SqlBlockParser(\n project=config,\n manifest=manifest,\n root_project=config,\n )\n\n adapter = dbt.adapters.factory.get_adapter(config)\n # TODO : This needs a real name?\n sql_node = block_parser.parse_remote(sql, 'name')\n process_node(config, manifest, sql_node)\n return config, sql_node, adapter\n\n\ndef compile_sql(manifest, project_path, sql):\n from dbt.task.sql import SqlCompileRunner\n\n config, node, adapter = _get_operation_node(manifest, project_path, sql)\n runner = SqlCompileRunner(config, adapter, node, 1, 1)\n return runner.safe_run(manifest)\n\n\ndef execute_sql(manifest, project_path, sql):\n from dbt.task.sql import SqlExecuteRunner\n\n config, node, adapter = _get_operation_node(manifest, project_path, sql)\n runner = SqlExecuteRunner(config, adapter, node, 1, 1)\n # TODO: use same interface for runner\n return runner.safe_run(manifest)\n\n\ndef parse_to_manifest(config):\n from dbt.parser.manifest import ManifestLoader\n\n return ManifestLoader.get_full_manifest(config)\n\n\ndef deserialize_manifest(manifest_msgpack):\n from dbt.contracts.graph.manifest import Manifest\n\n return Manifest.from_msgpack(manifest_msgpack)\n\n\ndef serialize_manifest(manifest):\n # TODO: what should this take as an arg?\n return manifest.to_msgpack()\n", "path": "core/dbt/lib.py"}]}
| 1,531 | 166 |
gh_patches_debug_57288
|
rasdani/github-patches
|
git_diff
|
microsoft__DeepSpeed-5134
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[BUG] TypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods
**Describe the bug**
TypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods handles_memory_backpressure, resolves_data_dependency, use_host_timers
**To Reproduce**
Steps to reproduce the behavior:
1. git clone https://github.com/OpenBMB/MiniCPM.git
2. follow setup step.
3. run `!bash lora_finetune.sh` via `lora_finetune.ipynb`
**Expected behavior**
runnable
**ds_report output**
Please run `ds_report` to give us details about your setup.
**Screenshots**
```
class CudaEventTimer(object):
File "/Users/katopz/git/MiniCPM/finetune/.venv/lib/python3.11/site-packages/deepspeed/utils/timer.py", line 33, in CudaEventTimer
def __init__(self, start_event: get_accelerator().Event, end_event: get_accelerator().Event):
^^^^^^^^^^^^^^^^^
File "/Users/katopz/git/MiniCPM/finetune/.venv/lib/python3.11/site-packages/deepspeed/accelerator/real_accelerator.py", line 184, in get_accelerator
ds_accelerator = MPS_Accelerator()
^^^^^^^^^^^^^^^^^
TypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods handles_memory_backpressure, resolves_data_dependency, use_host_timers
```
**System info (please complete the following information):**
- OS: macOS 14.2.1 (23C71)
- metal
- Python 3.11.7
**Launcher context**
deepspeed
**Docker context**
no
**Additional context**
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `accelerator/mps_accelerator.py`
Content:
```
1 # Copyright (c) Microsoft Corporation.
2 # SPDX-License-Identifier: Apache-2.0
3
4 # DeepSpeed Team
5
6 import torch
7
8 from .abstract_accelerator import DeepSpeedAccelerator
9
10 # During setup stage torch may not be installed, pass on no torch will
11 # allow op builder related API to be executed.
12 try:
13 import torch.mps
14 except ImportError:
15 pass
16
17
18 class MPS_Accelerator(DeepSpeedAccelerator):
19
20 def __init__(self):
21 self._name = "mps"
22 self._communication_backend_name = None
23
24 def is_synchronized_device(self):
25 return False
26
27 # Device APIs
28 def device_name(self, device_index=None):
29 if device_index is None:
30 return "mps"
31 return "mps:{}".format(device_index)
32
33 def device(self, device_index):
34 return torch.device("mps", index=0)
35
36 def set_device(self, device_index):
37 return
38
39 def current_device(self):
40 return torch.device("mps", index=0)
41
42 def current_device_name(self):
43 return "mps:0"
44
45 def device_count(self):
46 return 1
47
48 def synchronize(self, device_index=None):
49 return torch.mps.synchronize()
50
51 # RNG APIs
52 def random(self):
53 return torch.random
54
55 def set_rng_state(self, new_state, device_index=None):
56 return torch.mps.set_rng_state(new_state)
57
58 def get_rng_state(self, device_index=None):
59 return torch.mps.get_rng_state()
60
61 def manual_seed(self, seed):
62 return torch.mps.manual_seed(seed)
63
64 def manual_seed_all(self, seed):
65 return torch.mps.manual_seed(seed)
66
67 def seed(self):
68 return torch.mps.seed()
69
70 def initial_seed(self, seed):
71 return
72
73 def default_generator(self, device_index):
74 return
75
76 # Streams/Events
77 @property
78 def Stream(self):
79 return None
80
81 def stream(self, stream):
82 return None
83
84 def current_stream(self, device_index=None):
85 return None
86
87 def default_stream(self, device_index=None):
88 return None
89
90 @property
91 def Event(self):
92 return None
93
94 # Memory management
95 def empty_cache(self):
96 return torch.mps.empty_cache()
97
98 def memory_allocated(self, device_index=None):
99 return torch.mps.current_allocated_memory()
100
101 def max_memory_allocated(self, device_index=None):
102 return torch.mps.driver_allocated_memory()
103
104 def set_per_process_memory_fraction(self, fraction):
105 return torch.mps.set_per_process_memory_fraction(fraction)
106
107 def reset_max_memory_allocated(self, device_index=None):
108 return
109
110 def memory_cached(self, device_index=None):
111 return
112
113 def max_memory_cached(self, device_index=None):
114 return
115
116 def reset_max_memory_cached(self, device_index=None):
117 return
118
119 def memory_stats(self, device_index=None):
120 return
121
122 def reset_peak_memory_stats(self, device_index=None):
123 return
124
125 def memory_reserved(self, device_index=None):
126 return
127
128 def max_memory_reserved(self, device_index=None):
129 return
130
131 def total_memory(self, device_index=None):
132 return
133
134 def available_memory(self, device_index=None):
135 return
136
137 # Data types
138 def is_bf16_supported(self):
139 return False
140
141 def is_fp16_supported(self):
142 return False
143
144 def supported_dtypes(self):
145 return [torch.float]
146
147 # Misc
148 def amp(self):
149 return
150
151 def is_available(self):
152 return hasattr(torch.backends, "mps") and torch.backends.mps.is_available()
153
154 def range_push(self, msg):
155 return
156
157 def range_pop(self):
158 return
159
160 def lazy_call(self, callback):
161 return
162
163 def communication_backend_name(self):
164 return self._communication_backend_name
165
166 def is_triton_supported(self):
167 return False
168
169 # Graph operations
170 def create_graph(self):
171 return None
172
173 def capture_to_graph(self, graph, pool=None, stream=None):
174 from deepspeed.runtime.utils import noop_context
175 return noop_context()
176
177 def replay_graph(self, graph):
178 return
179
180 # Tensor operations
181 @property
182 def BFloat16Tensor(self):
183 return
184
185 @property
186 def ByteTensor(self):
187 return
188
189 @property
190 def DoubleTensor(self):
191 return
192
193 @property
194 def FloatTensor(self):
195 return
196
197 @property
198 def HalfTensor(self):
199 return
200
201 @property
202 def IntTensor(self):
203 return
204
205 @property
206 def LongTensor(self):
207 return
208
209 def pin_memory(self, tensor, align_bytes=1):
210 return tensor.pin_memory()
211
212 def is_pinned(self, tensor):
213 return tensor.is_pinned()
214
215 def on_accelerator(self, tensor):
216 device_str = str(tensor.device)
217 if device_str.startswith("mps"):
218 return True
219 else:
220 return False
221
222 def op_builder_dir(self):
223 try:
224 # is op_builder from deepspeed or a 3p version? this should only succeed if it's deepspeed
225 # if successful this also means we're doing a local install and not JIT compile path
226 from op_builder import __deepspeed__ # noqa: F401 # type: ignore
227
228 return "op_builder"
229 except ImportError:
230 return "deepspeed.ops.op_builder"
231
232 # create an instance of op builder, specified by class_name
233 def create_op_builder(self, op_name):
234 builder_class = self.get_op_builder(op_name)
235 if builder_class is not None:
236 return builder_class()
237 return None
238
239 # return an op builder class, specified by class_name
240 def get_op_builder(self, class_name):
241 from deepspeed.ops.op_builder.cpu import NotImplementedBuilder
242
243 return NotImplementedBuilder
244
245 def build_extension(self):
246 from torch.utils.cpp_extension import BuildExtension
247
248 return BuildExtension
249
250 def export_envs(self):
251 return []
252
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/accelerator/mps_accelerator.py b/accelerator/mps_accelerator.py
--- a/accelerator/mps_accelerator.py
+++ b/accelerator/mps_accelerator.py
@@ -24,6 +24,15 @@
def is_synchronized_device(self):
return False
+ def use_host_timers(self):
+ return self.is_synchronized_device()
+
+ def resolves_data_dependency(self):
+ return self.is_synchronized_device()
+
+ def handles_memory_backpressure(self):
+ return self.is_synchronized_device()
+
# Device APIs
def device_name(self, device_index=None):
if device_index is None:
|
{"golden_diff": "diff --git a/accelerator/mps_accelerator.py b/accelerator/mps_accelerator.py\n--- a/accelerator/mps_accelerator.py\n+++ b/accelerator/mps_accelerator.py\n@@ -24,6 +24,15 @@\n def is_synchronized_device(self):\n return False\n \n+ def use_host_timers(self):\n+ return self.is_synchronized_device()\n+\n+ def resolves_data_dependency(self):\n+ return self.is_synchronized_device()\n+\n+ def handles_memory_backpressure(self):\n+ return self.is_synchronized_device()\n+\n # Device APIs\n def device_name(self, device_index=None):\n if device_index is None:\n", "issue": "[BUG] TypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods\n**Describe the bug**\r\nTypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods handles_memory_backpressure, resolves_data_dependency, use_host_timers\r\n\r\n**To Reproduce**\r\nSteps to reproduce the behavior:\r\n1. git clone https://github.com/OpenBMB/MiniCPM.git\r\n2. follow setup step.\r\n3. run `!bash lora_finetune.sh` via `lora_finetune.ipynb`\r\n\r\n**Expected behavior**\r\nrunnable\r\n\r\n**ds_report output**\r\nPlease run `ds_report` to give us details about your setup.\r\n\r\n**Screenshots**\r\n```\r\n class CudaEventTimer(object):\r\n File \"/Users/katopz/git/MiniCPM/finetune/.venv/lib/python3.11/site-packages/deepspeed/utils/timer.py\", line 33, in CudaEventTimer\r\n def __init__(self, start_event: get_accelerator().Event, end_event: get_accelerator().Event):\r\n ^^^^^^^^^^^^^^^^^\r\n File \"/Users/katopz/git/MiniCPM/finetune/.venv/lib/python3.11/site-packages/deepspeed/accelerator/real_accelerator.py\", line 184, in get_accelerator\r\n ds_accelerator = MPS_Accelerator()\r\n ^^^^^^^^^^^^^^^^^\r\nTypeError: Can't instantiate abstract class MPS_Accelerator with abstract methods handles_memory_backpressure, resolves_data_dependency, use_host_timers\r\n```\r\n\r\n**System info (please complete the following information):**\r\n - OS: macOS 14.2.1 (23C71)\r\n - metal\r\n - Python 3.11.7\r\n\r\n**Launcher context**\r\ndeepspeed\r\n\r\n**Docker context**\r\nno\r\n\r\n**Additional context**\r\n\r\n\n", "before_files": [{"content": "# Copyright (c) Microsoft Corporation.\n# SPDX-License-Identifier: Apache-2.0\n\n# DeepSpeed Team\n\nimport torch\n\nfrom .abstract_accelerator import DeepSpeedAccelerator\n\n# During setup stage torch may not be installed, pass on no torch will\n# allow op builder related API to be executed.\ntry:\n import torch.mps\nexcept ImportError:\n pass\n\n\nclass MPS_Accelerator(DeepSpeedAccelerator):\n\n def __init__(self):\n self._name = \"mps\"\n self._communication_backend_name = None\n\n def is_synchronized_device(self):\n return False\n\n # Device APIs\n def device_name(self, device_index=None):\n if device_index is None:\n return \"mps\"\n return \"mps:{}\".format(device_index)\n\n def device(self, device_index):\n return torch.device(\"mps\", index=0)\n\n def set_device(self, device_index):\n return\n\n def current_device(self):\n return torch.device(\"mps\", index=0)\n\n def current_device_name(self):\n return \"mps:0\"\n\n def device_count(self):\n return 1\n\n def synchronize(self, device_index=None):\n return torch.mps.synchronize()\n\n # RNG APIs\n def random(self):\n return torch.random\n\n def set_rng_state(self, new_state, device_index=None):\n return torch.mps.set_rng_state(new_state)\n\n def get_rng_state(self, device_index=None):\n return torch.mps.get_rng_state()\n\n def manual_seed(self, seed):\n return torch.mps.manual_seed(seed)\n\n def manual_seed_all(self, seed):\n return torch.mps.manual_seed(seed)\n\n def seed(self):\n return torch.mps.seed()\n\n def initial_seed(self, seed):\n return\n\n def default_generator(self, device_index):\n return\n\n # Streams/Events\n @property\n def Stream(self):\n return None\n\n def stream(self, stream):\n return None\n\n def current_stream(self, device_index=None):\n return None\n\n def default_stream(self, device_index=None):\n return None\n\n @property\n def Event(self):\n return None\n\n # Memory management\n def empty_cache(self):\n return torch.mps.empty_cache()\n\n def memory_allocated(self, device_index=None):\n return torch.mps.current_allocated_memory()\n\n def max_memory_allocated(self, device_index=None):\n return torch.mps.driver_allocated_memory()\n\n def set_per_process_memory_fraction(self, fraction):\n return torch.mps.set_per_process_memory_fraction(fraction)\n\n def reset_max_memory_allocated(self, device_index=None):\n return\n\n def memory_cached(self, device_index=None):\n return\n\n def max_memory_cached(self, device_index=None):\n return\n\n def reset_max_memory_cached(self, device_index=None):\n return\n\n def memory_stats(self, device_index=None):\n return\n\n def reset_peak_memory_stats(self, device_index=None):\n return\n\n def memory_reserved(self, device_index=None):\n return\n\n def max_memory_reserved(self, device_index=None):\n return\n\n def total_memory(self, device_index=None):\n return\n\n def available_memory(self, device_index=None):\n return\n\n # Data types\n def is_bf16_supported(self):\n return False\n\n def is_fp16_supported(self):\n return False\n\n def supported_dtypes(self):\n return [torch.float]\n\n # Misc\n def amp(self):\n return\n\n def is_available(self):\n return hasattr(torch.backends, \"mps\") and torch.backends.mps.is_available()\n\n def range_push(self, msg):\n return\n\n def range_pop(self):\n return\n\n def lazy_call(self, callback):\n return\n\n def communication_backend_name(self):\n return self._communication_backend_name\n\n def is_triton_supported(self):\n return False\n\n # Graph operations\n def create_graph(self):\n return None\n\n def capture_to_graph(self, graph, pool=None, stream=None):\n from deepspeed.runtime.utils import noop_context\n return noop_context()\n\n def replay_graph(self, graph):\n return\n\n # Tensor operations\n @property\n def BFloat16Tensor(self):\n return\n\n @property\n def ByteTensor(self):\n return\n\n @property\n def DoubleTensor(self):\n return\n\n @property\n def FloatTensor(self):\n return\n\n @property\n def HalfTensor(self):\n return\n\n @property\n def IntTensor(self):\n return\n\n @property\n def LongTensor(self):\n return\n\n def pin_memory(self, tensor, align_bytes=1):\n return tensor.pin_memory()\n\n def is_pinned(self, tensor):\n return tensor.is_pinned()\n\n def on_accelerator(self, tensor):\n device_str = str(tensor.device)\n if device_str.startswith(\"mps\"):\n return True\n else:\n return False\n\n def op_builder_dir(self):\n try:\n # is op_builder from deepspeed or a 3p version? this should only succeed if it's deepspeed\n # if successful this also means we're doing a local install and not JIT compile path\n from op_builder import __deepspeed__ # noqa: F401 # type: ignore\n\n return \"op_builder\"\n except ImportError:\n return \"deepspeed.ops.op_builder\"\n\n # create an instance of op builder, specified by class_name\n def create_op_builder(self, op_name):\n builder_class = self.get_op_builder(op_name)\n if builder_class is not None:\n return builder_class()\n return None\n\n # return an op builder class, specified by class_name\n def get_op_builder(self, class_name):\n from deepspeed.ops.op_builder.cpu import NotImplementedBuilder\n\n return NotImplementedBuilder\n\n def build_extension(self):\n from torch.utils.cpp_extension import BuildExtension\n\n return BuildExtension\n\n def export_envs(self):\n return []\n", "path": "accelerator/mps_accelerator.py"}], "after_files": [{"content": "# Copyright (c) Microsoft Corporation.\n# SPDX-License-Identifier: Apache-2.0\n\n# DeepSpeed Team\n\nimport torch\n\nfrom .abstract_accelerator import DeepSpeedAccelerator\n\n# During setup stage torch may not be installed, pass on no torch will\n# allow op builder related API to be executed.\ntry:\n import torch.mps\nexcept ImportError:\n pass\n\n\nclass MPS_Accelerator(DeepSpeedAccelerator):\n\n def __init__(self):\n self._name = \"mps\"\n self._communication_backend_name = None\n\n def is_synchronized_device(self):\n return False\n\n def use_host_timers(self):\n return self.is_synchronized_device()\n\n def resolves_data_dependency(self):\n return self.is_synchronized_device()\n\n def handles_memory_backpressure(self):\n return self.is_synchronized_device()\n\n # Device APIs\n def device_name(self, device_index=None):\n if device_index is None:\n return \"mps\"\n return \"mps:{}\".format(device_index)\n\n def device(self, device_index):\n return torch.device(\"mps\", index=0)\n\n def set_device(self, device_index):\n return\n\n def current_device(self):\n return torch.device(\"mps\", index=0)\n\n def current_device_name(self):\n return \"mps:0\"\n\n def device_count(self):\n return 1\n\n def synchronize(self, device_index=None):\n return torch.mps.synchronize()\n\n # RNG APIs\n def random(self):\n return torch.random\n\n def set_rng_state(self, new_state, device_index=None):\n return torch.mps.set_rng_state(new_state)\n\n def get_rng_state(self, device_index=None):\n return torch.mps.get_rng_state()\n\n def manual_seed(self, seed):\n return torch.mps.manual_seed(seed)\n\n def manual_seed_all(self, seed):\n return torch.mps.manual_seed(seed)\n\n def seed(self):\n return torch.mps.seed()\n\n def initial_seed(self, seed):\n return\n\n def default_generator(self, device_index):\n return\n\n # Streams/Events\n @property\n def Stream(self):\n return None\n\n def stream(self, stream):\n return None\n\n def current_stream(self, device_index=None):\n return None\n\n def default_stream(self, device_index=None):\n return None\n\n @property\n def Event(self):\n return None\n\n # Memory management\n def empty_cache(self):\n return torch.mps.empty_cache()\n\n def memory_allocated(self, device_index=None):\n return torch.mps.current_allocated_memory()\n\n def max_memory_allocated(self, device_index=None):\n return torch.mps.driver_allocated_memory()\n\n def set_per_process_memory_fraction(self, fraction):\n return torch.mps.set_per_process_memory_fraction(fraction)\n\n def reset_max_memory_allocated(self, device_index=None):\n return\n\n def memory_cached(self, device_index=None):\n return\n\n def max_memory_cached(self, device_index=None):\n return\n\n def reset_max_memory_cached(self, device_index=None):\n return\n\n def memory_stats(self, device_index=None):\n return\n\n def reset_peak_memory_stats(self, device_index=None):\n return\n\n def memory_reserved(self, device_index=None):\n return\n\n def max_memory_reserved(self, device_index=None):\n return\n\n def total_memory(self, device_index=None):\n return\n\n def available_memory(self, device_index=None):\n return\n\n # Data types\n def is_bf16_supported(self):\n return False\n\n def is_fp16_supported(self):\n return False\n\n def supported_dtypes(self):\n return [torch.float]\n\n # Misc\n def amp(self):\n return\n\n def is_available(self):\n return hasattr(torch.backends, \"mps\") and torch.backends.mps.is_available()\n\n def range_push(self, msg):\n return\n\n def range_pop(self):\n return\n\n def lazy_call(self, callback):\n return\n\n def communication_backend_name(self):\n return self._communication_backend_name\n\n def is_triton_supported(self):\n return False\n\n # Graph operations\n def create_graph(self):\n return None\n\n def capture_to_graph(self, graph, pool=None, stream=None):\n from deepspeed.runtime.utils import noop_context\n return noop_context()\n\n def replay_graph(self, graph):\n return\n\n # Tensor operations\n @property\n def BFloat16Tensor(self):\n return\n\n @property\n def ByteTensor(self):\n return\n\n @property\n def DoubleTensor(self):\n return\n\n @property\n def FloatTensor(self):\n return\n\n @property\n def HalfTensor(self):\n return\n\n @property\n def IntTensor(self):\n return\n\n @property\n def LongTensor(self):\n return\n\n def pin_memory(self, tensor, align_bytes=1):\n return tensor.pin_memory()\n\n def is_pinned(self, tensor):\n return tensor.is_pinned()\n\n def on_accelerator(self, tensor):\n device_str = str(tensor.device)\n if device_str.startswith(\"mps\"):\n return True\n else:\n return False\n\n def op_builder_dir(self):\n try:\n # is op_builder from deepspeed or a 3p version? this should only succeed if it's deepspeed\n # if successful this also means we're doing a local install and not JIT compile path\n from op_builder import __deepspeed__ # noqa: F401 # type: ignore\n\n return \"op_builder\"\n except ImportError:\n return \"deepspeed.ops.op_builder\"\n\n # create an instance of op builder, specified by class_name\n def create_op_builder(self, op_name):\n builder_class = self.get_op_builder(op_name)\n if builder_class is not None:\n return builder_class()\n return None\n\n # return an op builder class, specified by class_name\n def get_op_builder(self, class_name):\n from deepspeed.ops.op_builder.cpu import NotImplementedBuilder\n\n return NotImplementedBuilder\n\n def build_extension(self):\n from torch.utils.cpp_extension import BuildExtension\n\n return BuildExtension\n\n def export_envs(self):\n return []\n", "path": "accelerator/mps_accelerator.py"}]}
| 2,634 | 147 |
gh_patches_debug_15241
|
rasdani/github-patches
|
git_diff
|
numpy__numpy-14912
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Numpy FFTs return data types don't match input data types
<!-- Please describe the issue in detail here, and fill in the fields below -->
Unlike other functions, the NumPy FFT functions do not return arrays whose data types match the input arrays. Instead, `complex64` is promoted to `complex128`.
### Reproducing code example:
<!-- A short code example that reproduces the problem/missing feature. It should be
self-contained, i.e., possible to run as-is via 'python myproblem.py' -->
```python
import sys
import numpy as np
print(np.__version__, sys.version)
input = np.ones((5, 5), dtype='complex64')
output = np.square(input)
assert output.dtype == 'complex64', f"{output.dtype} square" # assert passes
output = np.fft.fft2(input)
assert output.dtype == 'complex64', f"{output.dtype} fft" # assert fails
```
### Numpy/Python version information:
<!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' -->
1.17.3 3.8.0 | packaged by conda-forge | (default, Nov 6 2019, 23:20:36)
[GCC 7.3.0]
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `numpy/fft/__init__.py`
Content:
```
1 """
2 Discrete Fourier Transform (:mod:`numpy.fft`)
3 =============================================
4
5 .. currentmodule:: numpy.fft
6
7 Standard FFTs
8 -------------
9
10 .. autosummary::
11 :toctree: generated/
12
13 fft Discrete Fourier transform.
14 ifft Inverse discrete Fourier transform.
15 fft2 Discrete Fourier transform in two dimensions.
16 ifft2 Inverse discrete Fourier transform in two dimensions.
17 fftn Discrete Fourier transform in N-dimensions.
18 ifftn Inverse discrete Fourier transform in N dimensions.
19
20 Real FFTs
21 ---------
22
23 .. autosummary::
24 :toctree: generated/
25
26 rfft Real discrete Fourier transform.
27 irfft Inverse real discrete Fourier transform.
28 rfft2 Real discrete Fourier transform in two dimensions.
29 irfft2 Inverse real discrete Fourier transform in two dimensions.
30 rfftn Real discrete Fourier transform in N dimensions.
31 irfftn Inverse real discrete Fourier transform in N dimensions.
32
33 Hermitian FFTs
34 --------------
35
36 .. autosummary::
37 :toctree: generated/
38
39 hfft Hermitian discrete Fourier transform.
40 ihfft Inverse Hermitian discrete Fourier transform.
41
42 Helper routines
43 ---------------
44
45 .. autosummary::
46 :toctree: generated/
47
48 fftfreq Discrete Fourier Transform sample frequencies.
49 rfftfreq DFT sample frequencies (for usage with rfft, irfft).
50 fftshift Shift zero-frequency component to center of spectrum.
51 ifftshift Inverse of fftshift.
52
53
54 Background information
55 ----------------------
56
57 Fourier analysis is fundamentally a method for expressing a function as a
58 sum of periodic components, and for recovering the function from those
59 components. When both the function and its Fourier transform are
60 replaced with discretized counterparts, it is called the discrete Fourier
61 transform (DFT). The DFT has become a mainstay of numerical computing in
62 part because of a very fast algorithm for computing it, called the Fast
63 Fourier Transform (FFT), which was known to Gauss (1805) and was brought
64 to light in its current form by Cooley and Tukey [CT]_. Press et al. [NR]_
65 provide an accessible introduction to Fourier analysis and its
66 applications.
67
68 Because the discrete Fourier transform separates its input into
69 components that contribute at discrete frequencies, it has a great number
70 of applications in digital signal processing, e.g., for filtering, and in
71 this context the discretized input to the transform is customarily
72 referred to as a *signal*, which exists in the *time domain*. The output
73 is called a *spectrum* or *transform* and exists in the *frequency
74 domain*.
75
76 Implementation details
77 ----------------------
78
79 There are many ways to define the DFT, varying in the sign of the
80 exponent, normalization, etc. In this implementation, the DFT is defined
81 as
82
83 .. math::
84 A_k = \\sum_{m=0}^{n-1} a_m \\exp\\left\\{-2\\pi i{mk \\over n}\\right\\}
85 \\qquad k = 0,\\ldots,n-1.
86
87 The DFT is in general defined for complex inputs and outputs, and a
88 single-frequency component at linear frequency :math:`f` is
89 represented by a complex exponential
90 :math:`a_m = \\exp\\{2\\pi i\\,f m\\Delta t\\}`, where :math:`\\Delta t`
91 is the sampling interval.
92
93 The values in the result follow so-called "standard" order: If ``A =
94 fft(a, n)``, then ``A[0]`` contains the zero-frequency term (the sum of
95 the signal), which is always purely real for real inputs. Then ``A[1:n/2]``
96 contains the positive-frequency terms, and ``A[n/2+1:]`` contains the
97 negative-frequency terms, in order of decreasingly negative frequency.
98 For an even number of input points, ``A[n/2]`` represents both positive and
99 negative Nyquist frequency, and is also purely real for real input. For
100 an odd number of input points, ``A[(n-1)/2]`` contains the largest positive
101 frequency, while ``A[(n+1)/2]`` contains the largest negative frequency.
102 The routine ``np.fft.fftfreq(n)`` returns an array giving the frequencies
103 of corresponding elements in the output. The routine
104 ``np.fft.fftshift(A)`` shifts transforms and their frequencies to put the
105 zero-frequency components in the middle, and ``np.fft.ifftshift(A)`` undoes
106 that shift.
107
108 When the input `a` is a time-domain signal and ``A = fft(a)``, ``np.abs(A)``
109 is its amplitude spectrum and ``np.abs(A)**2`` is its power spectrum.
110 The phase spectrum is obtained by ``np.angle(A)``.
111
112 The inverse DFT is defined as
113
114 .. math::
115 a_m = \\frac{1}{n}\\sum_{k=0}^{n-1}A_k\\exp\\left\\{2\\pi i{mk\\over n}\\right\\}
116 \\qquad m = 0,\\ldots,n-1.
117
118 It differs from the forward transform by the sign of the exponential
119 argument and the default normalization by :math:`1/n`.
120
121 Normalization
122 -------------
123 The default normalization has the direct transforms unscaled and the inverse
124 transforms are scaled by :math:`1/n`. It is possible to obtain unitary
125 transforms by setting the keyword argument ``norm`` to ``"ortho"`` (default is
126 `None`) so that both direct and inverse transforms will be scaled by
127 :math:`1/\\sqrt{n}`.
128
129 Real and Hermitian transforms
130 -----------------------------
131
132 When the input is purely real, its transform is Hermitian, i.e., the
133 component at frequency :math:`f_k` is the complex conjugate of the
134 component at frequency :math:`-f_k`, which means that for real
135 inputs there is no information in the negative frequency components that
136 is not already available from the positive frequency components.
137 The family of `rfft` functions is
138 designed to operate on real inputs, and exploits this symmetry by
139 computing only the positive frequency components, up to and including the
140 Nyquist frequency. Thus, ``n`` input points produce ``n/2+1`` complex
141 output points. The inverses of this family assumes the same symmetry of
142 its input, and for an output of ``n`` points uses ``n/2+1`` input points.
143
144 Correspondingly, when the spectrum is purely real, the signal is
145 Hermitian. The `hfft` family of functions exploits this symmetry by
146 using ``n/2+1`` complex points in the input (time) domain for ``n`` real
147 points in the frequency domain.
148
149 In higher dimensions, FFTs are used, e.g., for image analysis and
150 filtering. The computational efficiency of the FFT means that it can
151 also be a faster way to compute large convolutions, using the property
152 that a convolution in the time domain is equivalent to a point-by-point
153 multiplication in the frequency domain.
154
155 Higher dimensions
156 -----------------
157
158 In two dimensions, the DFT is defined as
159
160 .. math::
161 A_{kl} = \\sum_{m=0}^{M-1} \\sum_{n=0}^{N-1}
162 a_{mn}\\exp\\left\\{-2\\pi i \\left({mk\\over M}+{nl\\over N}\\right)\\right\\}
163 \\qquad k = 0, \\ldots, M-1;\\quad l = 0, \\ldots, N-1,
164
165 which extends in the obvious way to higher dimensions, and the inverses
166 in higher dimensions also extend in the same way.
167
168 References
169 ----------
170
171 .. [CT] Cooley, James W., and John W. Tukey, 1965, "An algorithm for the
172 machine calculation of complex Fourier series," *Math. Comput.*
173 19: 297-301.
174
175 .. [NR] Press, W., Teukolsky, S., Vetterline, W.T., and Flannery, B.P.,
176 2007, *Numerical Recipes: The Art of Scientific Computing*, ch.
177 12-13. Cambridge Univ. Press, Cambridge, UK.
178
179 Examples
180 --------
181
182 For examples, see the various functions.
183
184 """
185
186 from __future__ import division, absolute_import, print_function
187
188 from ._pocketfft import *
189 from .helper import *
190
191 from numpy._pytesttester import PytestTester
192 test = PytestTester(__name__)
193 del PytestTester
194
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/numpy/fft/__init__.py b/numpy/fft/__init__.py
--- a/numpy/fft/__init__.py
+++ b/numpy/fft/__init__.py
@@ -118,8 +118,16 @@
It differs from the forward transform by the sign of the exponential
argument and the default normalization by :math:`1/n`.
+Type Promotion
+--------------
+
+`numpy.fft` promotes ``float32`` and ``complex64`` arrays to ``float64`` and
+``complex128`` arrays respectively. For an FFT implementation that does not
+promote input arrays, see `scipy.fftpack`.
+
Normalization
-------------
+
The default normalization has the direct transforms unscaled and the inverse
transforms are scaled by :math:`1/n`. It is possible to obtain unitary
transforms by setting the keyword argument ``norm`` to ``"ortho"`` (default is
|
{"golden_diff": "diff --git a/numpy/fft/__init__.py b/numpy/fft/__init__.py\n--- a/numpy/fft/__init__.py\n+++ b/numpy/fft/__init__.py\n@@ -118,8 +118,16 @@\n It differs from the forward transform by the sign of the exponential\n argument and the default normalization by :math:`1/n`.\n \n+Type Promotion\n+--------------\n+\n+`numpy.fft` promotes ``float32`` and ``complex64`` arrays to ``float64`` and\n+``complex128`` arrays respectively. For an FFT implementation that does not\n+promote input arrays, see `scipy.fftpack`.\n+\n Normalization\n -------------\n+\n The default normalization has the direct transforms unscaled and the inverse\n transforms are scaled by :math:`1/n`. It is possible to obtain unitary\n transforms by setting the keyword argument ``norm`` to ``\"ortho\"`` (default is\n", "issue": "Numpy FFTs return data types don't match input data types\n<!-- Please describe the issue in detail here, and fill in the fields below -->\r\n\r\nUnlike other functions, the NumPy FFT functions do not return arrays whose data types match the input arrays. Instead, `complex64` is promoted to `complex128`.\r\n\r\n### Reproducing code example:\r\n\r\n<!-- A short code example that reproduces the problem/missing feature. It should be\r\nself-contained, i.e., possible to run as-is via 'python myproblem.py' -->\r\n\r\n```python\r\nimport sys\r\nimport numpy as np\r\n\r\nprint(np.__version__, sys.version)\r\n\r\ninput = np.ones((5, 5), dtype='complex64')\r\n\r\noutput = np.square(input)\r\nassert output.dtype == 'complex64', f\"{output.dtype} square\" # assert passes\r\n\r\noutput = np.fft.fft2(input)\r\nassert output.dtype == 'complex64', f\"{output.dtype} fft\" # assert fails\r\n```\r\n\r\n### Numpy/Python version information:\r\n\r\n<!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' -->\r\n1.17.3 3.8.0 | packaged by conda-forge | (default, Nov 6 2019, 23:20:36) \r\n[GCC 7.3.0]\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nDiscrete Fourier Transform (:mod:`numpy.fft`)\n=============================================\n\n.. currentmodule:: numpy.fft\n\nStandard FFTs\n-------------\n\n.. autosummary::\n :toctree: generated/\n\n fft Discrete Fourier transform.\n ifft Inverse discrete Fourier transform.\n fft2 Discrete Fourier transform in two dimensions.\n ifft2 Inverse discrete Fourier transform in two dimensions.\n fftn Discrete Fourier transform in N-dimensions.\n ifftn Inverse discrete Fourier transform in N dimensions.\n\nReal FFTs\n---------\n\n.. autosummary::\n :toctree: generated/\n\n rfft Real discrete Fourier transform.\n irfft Inverse real discrete Fourier transform.\n rfft2 Real discrete Fourier transform in two dimensions.\n irfft2 Inverse real discrete Fourier transform in two dimensions.\n rfftn Real discrete Fourier transform in N dimensions.\n irfftn Inverse real discrete Fourier transform in N dimensions.\n\nHermitian FFTs\n--------------\n\n.. autosummary::\n :toctree: generated/\n\n hfft Hermitian discrete Fourier transform.\n ihfft Inverse Hermitian discrete Fourier transform.\n\nHelper routines\n---------------\n\n.. autosummary::\n :toctree: generated/\n\n fftfreq Discrete Fourier Transform sample frequencies.\n rfftfreq DFT sample frequencies (for usage with rfft, irfft).\n fftshift Shift zero-frequency component to center of spectrum.\n ifftshift Inverse of fftshift.\n\n\nBackground information\n----------------------\n\nFourier analysis is fundamentally a method for expressing a function as a\nsum of periodic components, and for recovering the function from those\ncomponents. When both the function and its Fourier transform are\nreplaced with discretized counterparts, it is called the discrete Fourier\ntransform (DFT). The DFT has become a mainstay of numerical computing in\npart because of a very fast algorithm for computing it, called the Fast\nFourier Transform (FFT), which was known to Gauss (1805) and was brought\nto light in its current form by Cooley and Tukey [CT]_. Press et al. [NR]_\nprovide an accessible introduction to Fourier analysis and its\napplications.\n\nBecause the discrete Fourier transform separates its input into\ncomponents that contribute at discrete frequencies, it has a great number\nof applications in digital signal processing, e.g., for filtering, and in\nthis context the discretized input to the transform is customarily\nreferred to as a *signal*, which exists in the *time domain*. The output\nis called a *spectrum* or *transform* and exists in the *frequency\ndomain*.\n\nImplementation details\n----------------------\n\nThere are many ways to define the DFT, varying in the sign of the\nexponent, normalization, etc. In this implementation, the DFT is defined\nas\n\n.. math::\n A_k = \\\\sum_{m=0}^{n-1} a_m \\\\exp\\\\left\\\\{-2\\\\pi i{mk \\\\over n}\\\\right\\\\}\n \\\\qquad k = 0,\\\\ldots,n-1.\n\nThe DFT is in general defined for complex inputs and outputs, and a\nsingle-frequency component at linear frequency :math:`f` is\nrepresented by a complex exponential\n:math:`a_m = \\\\exp\\\\{2\\\\pi i\\\\,f m\\\\Delta t\\\\}`, where :math:`\\\\Delta t`\nis the sampling interval.\n\nThe values in the result follow so-called \"standard\" order: If ``A =\nfft(a, n)``, then ``A[0]`` contains the zero-frequency term (the sum of\nthe signal), which is always purely real for real inputs. Then ``A[1:n/2]``\ncontains the positive-frequency terms, and ``A[n/2+1:]`` contains the\nnegative-frequency terms, in order of decreasingly negative frequency.\nFor an even number of input points, ``A[n/2]`` represents both positive and\nnegative Nyquist frequency, and is also purely real for real input. For\nan odd number of input points, ``A[(n-1)/2]`` contains the largest positive\nfrequency, while ``A[(n+1)/2]`` contains the largest negative frequency.\nThe routine ``np.fft.fftfreq(n)`` returns an array giving the frequencies\nof corresponding elements in the output. The routine\n``np.fft.fftshift(A)`` shifts transforms and their frequencies to put the\nzero-frequency components in the middle, and ``np.fft.ifftshift(A)`` undoes\nthat shift.\n\nWhen the input `a` is a time-domain signal and ``A = fft(a)``, ``np.abs(A)``\nis its amplitude spectrum and ``np.abs(A)**2`` is its power spectrum.\nThe phase spectrum is obtained by ``np.angle(A)``.\n\nThe inverse DFT is defined as\n\n.. math::\n a_m = \\\\frac{1}{n}\\\\sum_{k=0}^{n-1}A_k\\\\exp\\\\left\\\\{2\\\\pi i{mk\\\\over n}\\\\right\\\\}\n \\\\qquad m = 0,\\\\ldots,n-1.\n\nIt differs from the forward transform by the sign of the exponential\nargument and the default normalization by :math:`1/n`.\n\nNormalization\n-------------\nThe default normalization has the direct transforms unscaled and the inverse\ntransforms are scaled by :math:`1/n`. It is possible to obtain unitary\ntransforms by setting the keyword argument ``norm`` to ``\"ortho\"`` (default is\n`None`) so that both direct and inverse transforms will be scaled by\n:math:`1/\\\\sqrt{n}`.\n\nReal and Hermitian transforms\n-----------------------------\n\nWhen the input is purely real, its transform is Hermitian, i.e., the\ncomponent at frequency :math:`f_k` is the complex conjugate of the\ncomponent at frequency :math:`-f_k`, which means that for real\ninputs there is no information in the negative frequency components that\nis not already available from the positive frequency components.\nThe family of `rfft` functions is\ndesigned to operate on real inputs, and exploits this symmetry by\ncomputing only the positive frequency components, up to and including the\nNyquist frequency. Thus, ``n`` input points produce ``n/2+1`` complex\noutput points. The inverses of this family assumes the same symmetry of\nits input, and for an output of ``n`` points uses ``n/2+1`` input points.\n\nCorrespondingly, when the spectrum is purely real, the signal is\nHermitian. The `hfft` family of functions exploits this symmetry by\nusing ``n/2+1`` complex points in the input (time) domain for ``n`` real\npoints in the frequency domain.\n\nIn higher dimensions, FFTs are used, e.g., for image analysis and\nfiltering. The computational efficiency of the FFT means that it can\nalso be a faster way to compute large convolutions, using the property\nthat a convolution in the time domain is equivalent to a point-by-point\nmultiplication in the frequency domain.\n\nHigher dimensions\n-----------------\n\nIn two dimensions, the DFT is defined as\n\n.. math::\n A_{kl} = \\\\sum_{m=0}^{M-1} \\\\sum_{n=0}^{N-1}\n a_{mn}\\\\exp\\\\left\\\\{-2\\\\pi i \\\\left({mk\\\\over M}+{nl\\\\over N}\\\\right)\\\\right\\\\}\n \\\\qquad k = 0, \\\\ldots, M-1;\\\\quad l = 0, \\\\ldots, N-1,\n\nwhich extends in the obvious way to higher dimensions, and the inverses\nin higher dimensions also extend in the same way.\n\nReferences\n----------\n\n.. [CT] Cooley, James W., and John W. Tukey, 1965, \"An algorithm for the\n machine calculation of complex Fourier series,\" *Math. Comput.*\n 19: 297-301.\n\n.. [NR] Press, W., Teukolsky, S., Vetterline, W.T., and Flannery, B.P.,\n 2007, *Numerical Recipes: The Art of Scientific Computing*, ch.\n 12-13. Cambridge Univ. Press, Cambridge, UK.\n\nExamples\n--------\n\nFor examples, see the various functions.\n\n\"\"\"\n\nfrom __future__ import division, absolute_import, print_function\n\nfrom ._pocketfft import *\nfrom .helper import *\n\nfrom numpy._pytesttester import PytestTester\ntest = PytestTester(__name__)\ndel PytestTester\n", "path": "numpy/fft/__init__.py"}], "after_files": [{"content": "\"\"\"\nDiscrete Fourier Transform (:mod:`numpy.fft`)\n=============================================\n\n.. currentmodule:: numpy.fft\n\nStandard FFTs\n-------------\n\n.. autosummary::\n :toctree: generated/\n\n fft Discrete Fourier transform.\n ifft Inverse discrete Fourier transform.\n fft2 Discrete Fourier transform in two dimensions.\n ifft2 Inverse discrete Fourier transform in two dimensions.\n fftn Discrete Fourier transform in N-dimensions.\n ifftn Inverse discrete Fourier transform in N dimensions.\n\nReal FFTs\n---------\n\n.. autosummary::\n :toctree: generated/\n\n rfft Real discrete Fourier transform.\n irfft Inverse real discrete Fourier transform.\n rfft2 Real discrete Fourier transform in two dimensions.\n irfft2 Inverse real discrete Fourier transform in two dimensions.\n rfftn Real discrete Fourier transform in N dimensions.\n irfftn Inverse real discrete Fourier transform in N dimensions.\n\nHermitian FFTs\n--------------\n\n.. autosummary::\n :toctree: generated/\n\n hfft Hermitian discrete Fourier transform.\n ihfft Inverse Hermitian discrete Fourier transform.\n\nHelper routines\n---------------\n\n.. autosummary::\n :toctree: generated/\n\n fftfreq Discrete Fourier Transform sample frequencies.\n rfftfreq DFT sample frequencies (for usage with rfft, irfft).\n fftshift Shift zero-frequency component to center of spectrum.\n ifftshift Inverse of fftshift.\n\n\nBackground information\n----------------------\n\nFourier analysis is fundamentally a method for expressing a function as a\nsum of periodic components, and for recovering the function from those\ncomponents. When both the function and its Fourier transform are\nreplaced with discretized counterparts, it is called the discrete Fourier\ntransform (DFT). The DFT has become a mainstay of numerical computing in\npart because of a very fast algorithm for computing it, called the Fast\nFourier Transform (FFT), which was known to Gauss (1805) and was brought\nto light in its current form by Cooley and Tukey [CT]_. Press et al. [NR]_\nprovide an accessible introduction to Fourier analysis and its\napplications.\n\nBecause the discrete Fourier transform separates its input into\ncomponents that contribute at discrete frequencies, it has a great number\nof applications in digital signal processing, e.g., for filtering, and in\nthis context the discretized input to the transform is customarily\nreferred to as a *signal*, which exists in the *time domain*. The output\nis called a *spectrum* or *transform* and exists in the *frequency\ndomain*.\n\nImplementation details\n----------------------\n\nThere are many ways to define the DFT, varying in the sign of the\nexponent, normalization, etc. In this implementation, the DFT is defined\nas\n\n.. math::\n A_k = \\\\sum_{m=0}^{n-1} a_m \\\\exp\\\\left\\\\{-2\\\\pi i{mk \\\\over n}\\\\right\\\\}\n \\\\qquad k = 0,\\\\ldots,n-1.\n\nThe DFT is in general defined for complex inputs and outputs, and a\nsingle-frequency component at linear frequency :math:`f` is\nrepresented by a complex exponential\n:math:`a_m = \\\\exp\\\\{2\\\\pi i\\\\,f m\\\\Delta t\\\\}`, where :math:`\\\\Delta t`\nis the sampling interval.\n\nThe values in the result follow so-called \"standard\" order: If ``A =\nfft(a, n)``, then ``A[0]`` contains the zero-frequency term (the sum of\nthe signal), which is always purely real for real inputs. Then ``A[1:n/2]``\ncontains the positive-frequency terms, and ``A[n/2+1:]`` contains the\nnegative-frequency terms, in order of decreasingly negative frequency.\nFor an even number of input points, ``A[n/2]`` represents both positive and\nnegative Nyquist frequency, and is also purely real for real input. For\nan odd number of input points, ``A[(n-1)/2]`` contains the largest positive\nfrequency, while ``A[(n+1)/2]`` contains the largest negative frequency.\nThe routine ``np.fft.fftfreq(n)`` returns an array giving the frequencies\nof corresponding elements in the output. The routine\n``np.fft.fftshift(A)`` shifts transforms and their frequencies to put the\nzero-frequency components in the middle, and ``np.fft.ifftshift(A)`` undoes\nthat shift.\n\nWhen the input `a` is a time-domain signal and ``A = fft(a)``, ``np.abs(A)``\nis its amplitude spectrum and ``np.abs(A)**2`` is its power spectrum.\nThe phase spectrum is obtained by ``np.angle(A)``.\n\nThe inverse DFT is defined as\n\n.. math::\n a_m = \\\\frac{1}{n}\\\\sum_{k=0}^{n-1}A_k\\\\exp\\\\left\\\\{2\\\\pi i{mk\\\\over n}\\\\right\\\\}\n \\\\qquad m = 0,\\\\ldots,n-1.\n\nIt differs from the forward transform by the sign of the exponential\nargument and the default normalization by :math:`1/n`.\n\nType Promotion\n--------------\n\n`numpy.fft` promotes ``float32`` and ``complex64`` arrays to ``float64`` and\n``complex128`` arrays respectively. For an FFT implementation that does not\npromote input arrays, see `scipy.fftpack`.\n\nNormalization\n-------------\n\nThe default normalization has the direct transforms unscaled and the inverse\ntransforms are scaled by :math:`1/n`. It is possible to obtain unitary\ntransforms by setting the keyword argument ``norm`` to ``\"ortho\"`` (default is\n`None`) so that both direct and inverse transforms will be scaled by\n:math:`1/\\\\sqrt{n}`.\n\nReal and Hermitian transforms\n-----------------------------\n\nWhen the input is purely real, its transform is Hermitian, i.e., the\ncomponent at frequency :math:`f_k` is the complex conjugate of the\ncomponent at frequency :math:`-f_k`, which means that for real\ninputs there is no information in the negative frequency components that\nis not already available from the positive frequency components.\nThe family of `rfft` functions is\ndesigned to operate on real inputs, and exploits this symmetry by\ncomputing only the positive frequency components, up to and including the\nNyquist frequency. Thus, ``n`` input points produce ``n/2+1`` complex\noutput points. The inverses of this family assumes the same symmetry of\nits input, and for an output of ``n`` points uses ``n/2+1`` input points.\n\nCorrespondingly, when the spectrum is purely real, the signal is\nHermitian. The `hfft` family of functions exploits this symmetry by\nusing ``n/2+1`` complex points in the input (time) domain for ``n`` real\npoints in the frequency domain.\n\nIn higher dimensions, FFTs are used, e.g., for image analysis and\nfiltering. The computational efficiency of the FFT means that it can\nalso be a faster way to compute large convolutions, using the property\nthat a convolution in the time domain is equivalent to a point-by-point\nmultiplication in the frequency domain.\n\nHigher dimensions\n-----------------\n\nIn two dimensions, the DFT is defined as\n\n.. math::\n A_{kl} = \\\\sum_{m=0}^{M-1} \\\\sum_{n=0}^{N-1}\n a_{mn}\\\\exp\\\\left\\\\{-2\\\\pi i \\\\left({mk\\\\over M}+{nl\\\\over N}\\\\right)\\\\right\\\\}\n \\\\qquad k = 0, \\\\ldots, M-1;\\\\quad l = 0, \\\\ldots, N-1,\n\nwhich extends in the obvious way to higher dimensions, and the inverses\nin higher dimensions also extend in the same way.\n\nReferences\n----------\n\n.. [CT] Cooley, James W., and John W. Tukey, 1965, \"An algorithm for the\n machine calculation of complex Fourier series,\" *Math. Comput.*\n 19: 297-301.\n\n.. [NR] Press, W., Teukolsky, S., Vetterline, W.T., and Flannery, B.P.,\n 2007, *Numerical Recipes: The Art of Scientific Computing*, ch.\n 12-13. Cambridge Univ. Press, Cambridge, UK.\n\nExamples\n--------\n\nFor examples, see the various functions.\n\n\"\"\"\n\nfrom __future__ import division, absolute_import, print_function\n\nfrom ._pocketfft import *\nfrom .helper import *\n\nfrom numpy._pytesttester import PytestTester\ntest = PytestTester(__name__)\ndel PytestTester\n", "path": "numpy/fft/__init__.py"}]}
| 2,903 | 207 |
gh_patches_debug_35417
|
rasdani/github-patches
|
git_diff
|
holoviz__panel-2418
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
unexpected Interactivity of "layered" plotly objects in Tabs
My first one of these so apologies if it's missing info / already flagged.
#### Software info:
python: 3.6.9
Panel = 0.6.2
plotly = 4.2.1
bokeh = 1.3.4
jupyter notebook server = 6.0.0
browser: Google chrome (and same behavior in embedded html output)
OS: Windows 8.1
#### Description of expected behavior and the observed behavior
Below applies to within jupyter notebook, in browser window (.show()) and in html output (.save(embed = True))
__Expected behavior:__ plotly objects within panel tabs to have same interactivity as when not in tabs.
__observed behavior:__ Only plotly object in "bottom level" (last in list of tabs, "fig2" tab in example) retains full interactivity (pan, zoom, select, legend trace select etc.)). All other tab "levels" (tabs other than the last one in tab list, "fig1" tab inexample) retain only legend select interactivity. Interactions with the area bound by the axis (where a crosshair is seen) in "fig1" results in changes to the "fig2" plotly object.
#### Complete, minimal, self-contained example code that reproduces the issue
```
import plotly.graph_objs as go
import panel as pn
import numpy as np
pn.extension('plotly')
x = np.linspace(0,10,100)
y = np.sin(x)
y2 = np.cos(x)
data1 = [go.Scatter(x = x, y = y, name = 'scatter sin(x)', mode="markers+lines"),
go.Scatter(x = x, y = y2, name = 'scatter cos(x)', mode="markers+lines")]
data2 = [go.Bar(x = x, y = y, name = 'bar sin(x)'),
go.Bar(x = x, y = y2, name = 'bar cos(x)')]
fig1 = go.Figure(data = data1)
fig2 = go.Figure(data = data2)
pn.Column('## App with plotly objects in tabs:',
'Loss of interactivity on "fig1" tab plotly object',
pn.Tabs(('fig1', pn.Pane(fig1)),
('fig2', pn.Pane(fig2)))
)
```
#### Screenshots of issue

--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `panel/pane/plotly.py`
Content:
```
1 """
2 Defines a PlotlyPane which renders a plotly plot using PlotlyPlot
3 bokeh model.
4 """
5 import numpy as np
6 import param
7
8 from bokeh.models import ColumnDataSource
9 from pyviz_comms import JupyterComm
10
11 from .base import PaneBase
12 from ..util import isdatetime, lazy_load
13 from ..viewable import Layoutable
14
15
16
17 class Plotly(PaneBase):
18 """
19 Plotly panes allow rendering plotly Figures and traces.
20
21 For efficiency any array objects found inside a Figure are added
22 to a ColumnDataSource which allows using binary transport to sync
23 the figure on bokeh server and via Comms.
24 """
25
26 click_data = param.Dict(doc="Click callback data")
27
28 clickannotation_data = param.Dict(doc="Clickannotation callback data")
29
30 config = param.Dict(doc="Config data")
31
32 hover_data = param.Dict(doc="Hover callback data")
33
34 relayout_data = param.Dict(doc="Relayout callback data")
35
36 restyle_data = param.List(doc="Restyle callback data")
37
38 selected_data = param.Dict(doc="Selected callback data")
39
40 viewport = param.Dict(doc="Current viewport state")
41
42 viewport_update_policy = param.Selector(default="mouseup", doc="""
43 Policy by which the viewport parameter is updated during user interactions.
44
45 * "mouseup": updates are synchronized when mouse button is
46 released after panning
47 * "continuous": updates are synchronized continually while panning
48 * "throttle": updates are synchronized while panning, at
49 intervals determined by the viewport_update_throttle parameter
50 """, objects=["mouseup", "continuous", "throttle"])
51
52 viewport_update_throttle = param.Integer(default=200, bounds=(0, None), doc="""
53 Time interval in milliseconds at which viewport updates are
54 synchronized when viewport_update_policy is "throttle".""")
55
56 _render_count = param.Integer(default=0, doc="""
57 Number of renders, increment to trigger re-render""")
58
59 priority = 0.8
60
61 _updates = True
62
63 @classmethod
64 def applies(cls, obj):
65 return ((isinstance(obj, list) and obj and all(cls.applies(o) for o in obj)) or
66 hasattr(obj, 'to_plotly_json') or (isinstance(obj, dict)
67 and 'data' in obj and 'layout' in obj))
68
69 def __init__(self, object=None, **params):
70 super().__init__(object, **params)
71 self._figure = None
72 self._update_figure()
73
74 def _to_figure(self, obj):
75 import plotly.graph_objs as go
76 if isinstance(obj, go.Figure):
77 return obj
78 elif isinstance(obj, dict):
79 data, layout = obj['data'], obj['layout']
80 elif isinstance(obj, tuple):
81 data, layout = obj
82 else:
83 data, layout = obj, {}
84 data = data if isinstance(data, list) else [data]
85 return go.Figure(data=data, layout=layout)
86
87 @staticmethod
88 def _get_sources(json):
89 sources = []
90 traces = json.get('data', [])
91 for trace in traces:
92 data = {}
93 Plotly._get_sources_for_trace(trace, data)
94 sources.append(ColumnDataSource(data))
95 return sources
96
97 @staticmethod
98 def _get_sources_for_trace(json, data, parent_path=''):
99 for key, value in list(json.items()):
100 full_path = key if not parent_path else (parent_path + '.' + key)
101 if isinstance(value, np.ndarray):
102 # Extract numpy array
103 data[full_path] = [json.pop(key)]
104 elif isinstance(value, dict):
105 # Recurse into dictionaries:
106 Plotly._get_sources_for_trace(value, data=data, parent_path=full_path)
107 elif isinstance(value, list) and value and isinstance(value[0], dict):
108 # recurse into object arrays:
109 for i, element in enumerate(value):
110 element_path = full_path + '.' + str(i)
111 Plotly._get_sources_for_trace(
112 element, data=data, parent_path=element_path
113 )
114
115 @param.depends('object', watch=True)
116 def _update_figure(self):
117 import plotly.graph_objs as go
118
119 if (self.object is None or
120 type(self.object) is not go.Figure or
121 self.object is self._figure):
122 return
123
124 # Monkey patch the message stubs used by FigureWidget.
125 # We only patch `Figure` objects (not subclasses like FigureWidget) so
126 # we don't interfere with subclasses that override these methods.
127 fig = self.object
128 fig._send_addTraces_msg = lambda *_, **__: self.param.trigger('object')
129 fig._send_moveTraces_msg = lambda *_, **__: self.param.trigger('object')
130 fig._send_deleteTraces_msg = lambda *_, **__: self.param.trigger('object')
131 fig._send_restyle_msg = lambda *_, **__: self.param.trigger('object')
132 fig._send_relayout_msg = lambda *_, **__: self.param.trigger('object')
133 fig._send_update_msg = lambda *_, **__: self.param.trigger('object')
134 fig._send_animate_msg = lambda *_, **__: self.param.trigger('object')
135 self._figure = fig
136
137 def _update_data_sources(self, cds, trace):
138 trace_arrays = {}
139 Plotly._get_sources_for_trace(trace, trace_arrays)
140
141 update_sources = False
142 for key, new_col in trace_arrays.items():
143 new = new_col[0]
144
145 try:
146 old = cds.data.get(key)[0]
147 update_array = (
148 (type(old) != type(new)) or
149 (new.shape != old.shape) or
150 (new != old).any())
151 except Exception:
152 update_array = True
153
154 if update_array:
155 update_sources = True
156 cds.data[key] = [new]
157
158 return update_sources
159
160 @staticmethod
161 def _plotly_json_wrapper(fig):
162 """Wraps around to_plotly_json and applies necessary fixes.
163
164 For #382: Map datetime elements to strings.
165 """
166 json = fig.to_plotly_json()
167 data = json['data']
168
169 for idx in range(len(data)):
170 for key in data[idx]:
171 if isdatetime(data[idx][key]):
172 arr = data[idx][key]
173 if isinstance(arr, np.ndarray):
174 arr = arr.astype(str)
175 else:
176 arr = [str(v) for v in arr]
177 data[idx][key] = arr
178 return json
179
180 def _init_params(self):
181 viewport_params = [p for p in self.param if 'viewport' in p]
182 parameters = list(Layoutable.param)+viewport_params
183 params = {p: getattr(self, p) for p in parameters
184 if getattr(self, p) is not None}
185
186 if self.object is None:
187 json, sources = {}, []
188 else:
189 fig = self._to_figure(self.object)
190 json = self._plotly_json_wrapper(fig)
191 sources = Plotly._get_sources(json)
192
193 params['_render_count'] = self._render_count
194 params['config'] = self.config or {}
195 params['data'] = json.get('data', [])
196 params['data_sources'] = sources
197 params['layout'] = layout = json.get('layout', {})
198 if layout.get('autosize') and self.sizing_mode is self.param.sizing_mode.default:
199 params['sizing_mode'] = 'stretch_both'
200 return params
201
202 def _get_model(self, doc, root=None, parent=None, comm=None):
203 PlotlyPlot = lazy_load('panel.models.plotly', 'PlotlyPlot', isinstance(comm, JupyterComm))
204 model = PlotlyPlot(**self._init_params())
205 if root is None:
206 root = model
207 self._link_props(model, self._linkable_params, doc, root, comm)
208 self._models[root.ref['id']] = (model, parent)
209 return model
210
211 def _update(self, ref=None, model=None):
212 if self.object is None:
213 model.update(data=[], layout={})
214 model._render_count += 1
215 return
216
217 fig = self._to_figure(self.object)
218 json = self._plotly_json_wrapper(fig)
219 layout = json.get('layout')
220
221 traces = json['data']
222 new_sources = []
223 update_sources = False
224 for i, trace in enumerate(traces):
225 if i < len(model.data_sources):
226 cds = model.data_sources[i]
227 else:
228 cds = ColumnDataSource()
229 new_sources.append(cds)
230
231 update_sources = self._update_data_sources(cds, trace) or update_sources
232 try:
233 update_layout = model.layout != layout
234 except Exception:
235 update_layout = True
236
237 # Determine if model needs updates
238 if (len(model.data) != len(traces)):
239 update_data = True
240 else:
241 update_data = False
242 for new, old in zip(traces, model.data):
243 try:
244 update_data = (
245 {k: v for k, v in new.items() if k != 'uid'} !=
246 {k: v for k, v in old.items() if k != 'uid'}
247 )
248 except Exception:
249 update_data = True
250 if update_data:
251 break
252
253 updates = {}
254 if self.sizing_mode is self.param.sizing_mode.default and 'autosize' in layout:
255 autosize = layout.get('autosize')
256 if autosize and model.sizing_mode != 'stretch_both':
257 updates['sizing_mode'] = 'stretch_both'
258 elif not autosize and model.sizing_mode != 'fixed':
259 updates['sizing_mode'] = 'fixed'
260
261 if new_sources:
262 updates['data_sources'] = model.data_sources + new_sources
263
264 if update_data:
265 updates['data'] = json.get('data')
266
267 if update_layout:
268 updates['layout'] = layout
269
270 if updates:
271 model.update(**updates)
272
273 # Check if we should trigger rendering
274 if updates or update_sources:
275 model._render_count += 1
276
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/panel/pane/plotly.py b/panel/pane/plotly.py
--- a/panel/pane/plotly.py
+++ b/panel/pane/plotly.py
@@ -5,12 +5,12 @@
import numpy as np
import param
-from bokeh.models import ColumnDataSource
+from bokeh.models import ColumnDataSource, CustomJS, Tabs
from pyviz_comms import JupyterComm
from .base import PaneBase
from ..util import isdatetime, lazy_load
-from ..viewable import Layoutable
+from ..viewable import Layoutable, Viewable
@@ -206,6 +206,8 @@
root = model
self._link_props(model, self._linkable_params, doc, root, comm)
self._models[root.ref['id']] = (model, parent)
+ if _patch_tabs_plotly not in Viewable._preprocessing_hooks:
+ Viewable._preprocessing_hooks.append(_patch_tabs_plotly)
return model
def _update(self, ref=None, model=None):
@@ -273,3 +275,42 @@
# Check if we should trigger rendering
if updates or update_sources:
model._render_count += 1
+
+
+def _patch_tabs_plotly(viewable, root):
+ """
+ A preprocessing hook which ensures that any Plotly panes rendered
+ inside Tabs are only visible when the tab they are in is active.
+ This is a workaround for https://github.com/holoviz/panel/issues/804.
+ """
+ from ..models.plotly import PlotlyPlot
+
+ tabs_models = list(root.select({'type': Tabs}))
+ plotly_models = root.select({'type': PlotlyPlot})
+
+ for model in plotly_models:
+ parent_tabs = []
+ for tabs in tabs_models:
+ if tabs.select_one({'id': model.id}):
+ parent_tabs.append(tabs)
+ parent_tab = None
+ for tabs in parent_tabs:
+ if not any(tabs.select_one(pt) for pt in parent_tabs if pt is not tabs):
+ parent_tab = tabs
+ break
+ if parent_tab is None:
+ return
+ for i, tab in enumerate(parent_tab.tabs):
+ if tab.select_one({'id': model.id}):
+ break
+ updated = False
+ code = "model.visible = cb_obj.active == i;"
+ for cb in parent_tab.js_property_callbacks.get('change:active', []):
+ if cb.code == code and cb.args.get('model') is model:
+ cb.args['i'] = i
+ updated = True
+ if updated:
+ continue
+ callback = CustomJS(args={'model': model, 'i': i, 'margin': model.margin}, code=code)
+ parent_tab.js_on_change('active', callback)
+ model.visible = parent_tab.active == i
|
{"golden_diff": "diff --git a/panel/pane/plotly.py b/panel/pane/plotly.py\n--- a/panel/pane/plotly.py\n+++ b/panel/pane/plotly.py\n@@ -5,12 +5,12 @@\n import numpy as np\n import param\n \n-from bokeh.models import ColumnDataSource\n+from bokeh.models import ColumnDataSource, CustomJS, Tabs\n from pyviz_comms import JupyterComm\n \n from .base import PaneBase\n from ..util import isdatetime, lazy_load\n-from ..viewable import Layoutable\n+from ..viewable import Layoutable, Viewable\n \n \n \n@@ -206,6 +206,8 @@\n root = model\n self._link_props(model, self._linkable_params, doc, root, comm)\n self._models[root.ref['id']] = (model, parent)\n+ if _patch_tabs_plotly not in Viewable._preprocessing_hooks:\n+ Viewable._preprocessing_hooks.append(_patch_tabs_plotly)\n return model\n \n def _update(self, ref=None, model=None):\n@@ -273,3 +275,42 @@\n # Check if we should trigger rendering\n if updates or update_sources:\n model._render_count += 1\n+\n+\n+def _patch_tabs_plotly(viewable, root):\n+ \"\"\"\n+ A preprocessing hook which ensures that any Plotly panes rendered\n+ inside Tabs are only visible when the tab they are in is active.\n+ This is a workaround for https://github.com/holoviz/panel/issues/804.\n+ \"\"\"\n+ from ..models.plotly import PlotlyPlot\n+\n+ tabs_models = list(root.select({'type': Tabs}))\n+ plotly_models = root.select({'type': PlotlyPlot})\n+\n+ for model in plotly_models:\n+ parent_tabs = []\n+ for tabs in tabs_models:\n+ if tabs.select_one({'id': model.id}):\n+ parent_tabs.append(tabs)\n+ parent_tab = None\n+ for tabs in parent_tabs:\n+ if not any(tabs.select_one(pt) for pt in parent_tabs if pt is not tabs):\n+ parent_tab = tabs\n+ break\n+ if parent_tab is None:\n+ return\n+ for i, tab in enumerate(parent_tab.tabs):\n+ if tab.select_one({'id': model.id}):\n+ break\n+ updated = False\n+ code = \"model.visible = cb_obj.active == i;\"\n+ for cb in parent_tab.js_property_callbacks.get('change:active', []):\n+ if cb.code == code and cb.args.get('model') is model:\n+ cb.args['i'] = i\n+ updated = True\n+ if updated:\n+ continue\n+ callback = CustomJS(args={'model': model, 'i': i, 'margin': model.margin}, code=code)\n+ parent_tab.js_on_change('active', callback)\n+ model.visible = parent_tab.active == i\n", "issue": "unexpected Interactivity of \"layered\" plotly objects in Tabs\nMy first one of these so apologies if it's missing info / already flagged.\r\n\r\n#### Software info:\r\npython: 3.6.9\r\nPanel = 0.6.2\r\nplotly = 4.2.1\r\nbokeh = 1.3.4\r\njupyter notebook server = 6.0.0\r\nbrowser: Google chrome (and same behavior in embedded html output)\r\nOS: Windows 8.1\r\n\r\n#### Description of expected behavior and the observed behavior\r\nBelow applies to within jupyter notebook, in browser window (.show()) and in html output (.save(embed = True))\r\n\r\n__Expected behavior:__ plotly objects within panel tabs to have same interactivity as when not in tabs.\r\n\r\n__observed behavior:__ Only plotly object in \"bottom level\" (last in list of tabs, \"fig2\" tab in example) retains full interactivity (pan, zoom, select, legend trace select etc.)). All other tab \"levels\" (tabs other than the last one in tab list, \"fig1\" tab inexample) retain only legend select interactivity. Interactions with the area bound by the axis (where a crosshair is seen) in \"fig1\" results in changes to the \"fig2\" plotly object.\r\n\r\n\r\n#### Complete, minimal, self-contained example code that reproduces the issue\r\n\r\n```\r\nimport plotly.graph_objs as go\r\nimport panel as pn\r\nimport numpy as np\r\npn.extension('plotly')\r\n\r\nx = np.linspace(0,10,100)\r\ny = np.sin(x)\r\ny2 = np.cos(x)\r\n\r\ndata1 = [go.Scatter(x = x, y = y, name = 'scatter sin(x)', mode=\"markers+lines\"),\r\n go.Scatter(x = x, y = y2, name = 'scatter cos(x)', mode=\"markers+lines\")]\r\n\r\ndata2 = [go.Bar(x = x, y = y, name = 'bar sin(x)'),\r\n go.Bar(x = x, y = y2, name = 'bar cos(x)')]\r\n\r\nfig1 = go.Figure(data = data1)\r\nfig2 = go.Figure(data = data2)\r\n\r\npn.Column('## App with plotly objects in tabs:',\r\n 'Loss of interactivity on \"fig1\" tab plotly object',\r\n pn.Tabs(('fig1', pn.Pane(fig1)),\r\n ('fig2', pn.Pane(fig2)))\r\n )\r\n\r\n```\r\n#### Screenshots of issue\r\n\r\n\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nDefines a PlotlyPane which renders a plotly plot using PlotlyPlot\nbokeh model.\n\"\"\"\nimport numpy as np\nimport param\n\nfrom bokeh.models import ColumnDataSource\nfrom pyviz_comms import JupyterComm\n\nfrom .base import PaneBase\nfrom ..util import isdatetime, lazy_load\nfrom ..viewable import Layoutable\n\n\n\nclass Plotly(PaneBase):\n \"\"\"\n Plotly panes allow rendering plotly Figures and traces.\n\n For efficiency any array objects found inside a Figure are added\n to a ColumnDataSource which allows using binary transport to sync\n the figure on bokeh server and via Comms.\n \"\"\"\n\n click_data = param.Dict(doc=\"Click callback data\")\n\n clickannotation_data = param.Dict(doc=\"Clickannotation callback data\")\n\n config = param.Dict(doc=\"Config data\")\n\n hover_data = param.Dict(doc=\"Hover callback data\")\n\n relayout_data = param.Dict(doc=\"Relayout callback data\")\n\n restyle_data = param.List(doc=\"Restyle callback data\")\n\n selected_data = param.Dict(doc=\"Selected callback data\")\n\n viewport = param.Dict(doc=\"Current viewport state\")\n\n viewport_update_policy = param.Selector(default=\"mouseup\", doc=\"\"\"\n Policy by which the viewport parameter is updated during user interactions.\n\n * \"mouseup\": updates are synchronized when mouse button is\n released after panning\n * \"continuous\": updates are synchronized continually while panning\n * \"throttle\": updates are synchronized while panning, at \n intervals determined by the viewport_update_throttle parameter\n \"\"\", objects=[\"mouseup\", \"continuous\", \"throttle\"])\n\n viewport_update_throttle = param.Integer(default=200, bounds=(0, None), doc=\"\"\"\n Time interval in milliseconds at which viewport updates are\n synchronized when viewport_update_policy is \"throttle\".\"\"\")\n\n _render_count = param.Integer(default=0, doc=\"\"\"\n Number of renders, increment to trigger re-render\"\"\")\n\n priority = 0.8\n\n _updates = True\n\n @classmethod\n def applies(cls, obj):\n return ((isinstance(obj, list) and obj and all(cls.applies(o) for o in obj)) or\n hasattr(obj, 'to_plotly_json') or (isinstance(obj, dict)\n and 'data' in obj and 'layout' in obj))\n\n def __init__(self, object=None, **params):\n super().__init__(object, **params)\n self._figure = None\n self._update_figure()\n\n def _to_figure(self, obj):\n import plotly.graph_objs as go\n if isinstance(obj, go.Figure):\n return obj\n elif isinstance(obj, dict):\n data, layout = obj['data'], obj['layout']\n elif isinstance(obj, tuple):\n data, layout = obj\n else:\n data, layout = obj, {}\n data = data if isinstance(data, list) else [data]\n return go.Figure(data=data, layout=layout)\n\n @staticmethod\n def _get_sources(json):\n sources = []\n traces = json.get('data', [])\n for trace in traces:\n data = {}\n Plotly._get_sources_for_trace(trace, data)\n sources.append(ColumnDataSource(data))\n return sources\n\n @staticmethod\n def _get_sources_for_trace(json, data, parent_path=''):\n for key, value in list(json.items()):\n full_path = key if not parent_path else (parent_path + '.' + key)\n if isinstance(value, np.ndarray):\n # Extract numpy array\n data[full_path] = [json.pop(key)]\n elif isinstance(value, dict):\n # Recurse into dictionaries:\n Plotly._get_sources_for_trace(value, data=data, parent_path=full_path)\n elif isinstance(value, list) and value and isinstance(value[0], dict):\n # recurse into object arrays:\n for i, element in enumerate(value):\n element_path = full_path + '.' + str(i)\n Plotly._get_sources_for_trace(\n element, data=data, parent_path=element_path\n )\n\n @param.depends('object', watch=True)\n def _update_figure(self):\n import plotly.graph_objs as go\n\n if (self.object is None or\n type(self.object) is not go.Figure or\n self.object is self._figure):\n return\n\n # Monkey patch the message stubs used by FigureWidget.\n # We only patch `Figure` objects (not subclasses like FigureWidget) so\n # we don't interfere with subclasses that override these methods.\n fig = self.object\n fig._send_addTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_moveTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_deleteTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_restyle_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_relayout_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_update_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_animate_msg = lambda *_, **__: self.param.trigger('object')\n self._figure = fig\n\n def _update_data_sources(self, cds, trace):\n trace_arrays = {}\n Plotly._get_sources_for_trace(trace, trace_arrays)\n\n update_sources = False\n for key, new_col in trace_arrays.items():\n new = new_col[0]\n\n try:\n old = cds.data.get(key)[0]\n update_array = (\n (type(old) != type(new)) or\n (new.shape != old.shape) or\n (new != old).any())\n except Exception:\n update_array = True\n\n if update_array:\n update_sources = True\n cds.data[key] = [new]\n\n return update_sources\n\n @staticmethod\n def _plotly_json_wrapper(fig):\n \"\"\"Wraps around to_plotly_json and applies necessary fixes.\n\n For #382: Map datetime elements to strings.\n \"\"\"\n json = fig.to_plotly_json()\n data = json['data']\n\n for idx in range(len(data)):\n for key in data[idx]:\n if isdatetime(data[idx][key]):\n arr = data[idx][key]\n if isinstance(arr, np.ndarray):\n arr = arr.astype(str) \n else:\n arr = [str(v) for v in arr]\n data[idx][key] = arr\n return json\n\n def _init_params(self):\n viewport_params = [p for p in self.param if 'viewport' in p]\n parameters = list(Layoutable.param)+viewport_params\n params = {p: getattr(self, p) for p in parameters\n if getattr(self, p) is not None}\n\n if self.object is None:\n json, sources = {}, []\n else:\n fig = self._to_figure(self.object)\n json = self._plotly_json_wrapper(fig)\n sources = Plotly._get_sources(json)\n\n params['_render_count'] = self._render_count\n params['config'] = self.config or {}\n params['data'] = json.get('data', [])\n params['data_sources'] = sources\n params['layout'] = layout = json.get('layout', {})\n if layout.get('autosize') and self.sizing_mode is self.param.sizing_mode.default:\n params['sizing_mode'] = 'stretch_both'\n return params\n\n def _get_model(self, doc, root=None, parent=None, comm=None):\n PlotlyPlot = lazy_load('panel.models.plotly', 'PlotlyPlot', isinstance(comm, JupyterComm))\n model = PlotlyPlot(**self._init_params())\n if root is None:\n root = model\n self._link_props(model, self._linkable_params, doc, root, comm)\n self._models[root.ref['id']] = (model, parent)\n return model\n\n def _update(self, ref=None, model=None):\n if self.object is None:\n model.update(data=[], layout={})\n model._render_count += 1\n return\n\n fig = self._to_figure(self.object)\n json = self._plotly_json_wrapper(fig)\n layout = json.get('layout')\n\n traces = json['data']\n new_sources = []\n update_sources = False\n for i, trace in enumerate(traces):\n if i < len(model.data_sources):\n cds = model.data_sources[i]\n else:\n cds = ColumnDataSource()\n new_sources.append(cds)\n\n update_sources = self._update_data_sources(cds, trace) or update_sources\n try:\n update_layout = model.layout != layout\n except Exception:\n update_layout = True\n\n # Determine if model needs updates\n if (len(model.data) != len(traces)):\n update_data = True\n else:\n update_data = False\n for new, old in zip(traces, model.data):\n try:\n update_data = (\n {k: v for k, v in new.items() if k != 'uid'} !=\n {k: v for k, v in old.items() if k != 'uid'}\n )\n except Exception:\n update_data = True\n if update_data:\n break\n\n updates = {}\n if self.sizing_mode is self.param.sizing_mode.default and 'autosize' in layout:\n autosize = layout.get('autosize')\n if autosize and model.sizing_mode != 'stretch_both':\n updates['sizing_mode'] = 'stretch_both'\n elif not autosize and model.sizing_mode != 'fixed':\n updates['sizing_mode'] = 'fixed'\n\n if new_sources:\n updates['data_sources'] = model.data_sources + new_sources\n\n if update_data:\n updates['data'] = json.get('data')\n\n if update_layout:\n updates['layout'] = layout\n\n if updates:\n model.update(**updates)\n\n # Check if we should trigger rendering\n if updates or update_sources:\n model._render_count += 1\n", "path": "panel/pane/plotly.py"}], "after_files": [{"content": "\"\"\"\nDefines a PlotlyPane which renders a plotly plot using PlotlyPlot\nbokeh model.\n\"\"\"\nimport numpy as np\nimport param\n\nfrom bokeh.models import ColumnDataSource, CustomJS, Tabs\nfrom pyviz_comms import JupyterComm\n\nfrom .base import PaneBase\nfrom ..util import isdatetime, lazy_load\nfrom ..viewable import Layoutable, Viewable\n\n\n\nclass Plotly(PaneBase):\n \"\"\"\n Plotly panes allow rendering plotly Figures and traces.\n\n For efficiency any array objects found inside a Figure are added\n to a ColumnDataSource which allows using binary transport to sync\n the figure on bokeh server and via Comms.\n \"\"\"\n\n click_data = param.Dict(doc=\"Click callback data\")\n\n clickannotation_data = param.Dict(doc=\"Clickannotation callback data\")\n\n config = param.Dict(doc=\"Config data\")\n\n hover_data = param.Dict(doc=\"Hover callback data\")\n\n relayout_data = param.Dict(doc=\"Relayout callback data\")\n\n restyle_data = param.List(doc=\"Restyle callback data\")\n\n selected_data = param.Dict(doc=\"Selected callback data\")\n\n viewport = param.Dict(doc=\"Current viewport state\")\n\n viewport_update_policy = param.Selector(default=\"mouseup\", doc=\"\"\"\n Policy by which the viewport parameter is updated during user interactions.\n\n * \"mouseup\": updates are synchronized when mouse button is\n released after panning\n * \"continuous\": updates are synchronized continually while panning\n * \"throttle\": updates are synchronized while panning, at \n intervals determined by the viewport_update_throttle parameter\n \"\"\", objects=[\"mouseup\", \"continuous\", \"throttle\"])\n\n viewport_update_throttle = param.Integer(default=200, bounds=(0, None), doc=\"\"\"\n Time interval in milliseconds at which viewport updates are\n synchronized when viewport_update_policy is \"throttle\".\"\"\")\n\n _render_count = param.Integer(default=0, doc=\"\"\"\n Number of renders, increment to trigger re-render\"\"\")\n\n priority = 0.8\n\n _updates = True\n\n @classmethod\n def applies(cls, obj):\n return ((isinstance(obj, list) and obj and all(cls.applies(o) for o in obj)) or\n hasattr(obj, 'to_plotly_json') or (isinstance(obj, dict)\n and 'data' in obj and 'layout' in obj))\n\n def __init__(self, object=None, **params):\n super().__init__(object, **params)\n self._figure = None\n self._update_figure()\n\n def _to_figure(self, obj):\n import plotly.graph_objs as go\n if isinstance(obj, go.Figure):\n return obj\n elif isinstance(obj, dict):\n data, layout = obj['data'], obj['layout']\n elif isinstance(obj, tuple):\n data, layout = obj\n else:\n data, layout = obj, {}\n data = data if isinstance(data, list) else [data]\n return go.Figure(data=data, layout=layout)\n\n @staticmethod\n def _get_sources(json):\n sources = []\n traces = json.get('data', [])\n for trace in traces:\n data = {}\n Plotly._get_sources_for_trace(trace, data)\n sources.append(ColumnDataSource(data))\n return sources\n\n @staticmethod\n def _get_sources_for_trace(json, data, parent_path=''):\n for key, value in list(json.items()):\n full_path = key if not parent_path else (parent_path + '.' + key)\n if isinstance(value, np.ndarray):\n # Extract numpy array\n data[full_path] = [json.pop(key)]\n elif isinstance(value, dict):\n # Recurse into dictionaries:\n Plotly._get_sources_for_trace(value, data=data, parent_path=full_path)\n elif isinstance(value, list) and value and isinstance(value[0], dict):\n # recurse into object arrays:\n for i, element in enumerate(value):\n element_path = full_path + '.' + str(i)\n Plotly._get_sources_for_trace(\n element, data=data, parent_path=element_path\n )\n\n @param.depends('object', watch=True)\n def _update_figure(self):\n import plotly.graph_objs as go\n\n if (self.object is None or\n type(self.object) is not go.Figure or\n self.object is self._figure):\n return\n\n # Monkey patch the message stubs used by FigureWidget.\n # We only patch `Figure` objects (not subclasses like FigureWidget) so\n # we don't interfere with subclasses that override these methods.\n fig = self.object\n fig._send_addTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_moveTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_deleteTraces_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_restyle_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_relayout_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_update_msg = lambda *_, **__: self.param.trigger('object')\n fig._send_animate_msg = lambda *_, **__: self.param.trigger('object')\n self._figure = fig\n\n def _update_data_sources(self, cds, trace):\n trace_arrays = {}\n Plotly._get_sources_for_trace(trace, trace_arrays)\n\n update_sources = False\n for key, new_col in trace_arrays.items():\n new = new_col[0]\n\n try:\n old = cds.data.get(key)[0]\n update_array = (\n (type(old) != type(new)) or\n (new.shape != old.shape) or\n (new != old).any())\n except Exception:\n update_array = True\n\n if update_array:\n update_sources = True\n cds.data[key] = [new]\n\n return update_sources\n\n @staticmethod\n def _plotly_json_wrapper(fig):\n \"\"\"Wraps around to_plotly_json and applies necessary fixes.\n\n For #382: Map datetime elements to strings.\n \"\"\"\n json = fig.to_plotly_json()\n data = json['data']\n\n for idx in range(len(data)):\n for key in data[idx]:\n if isdatetime(data[idx][key]):\n arr = data[idx][key]\n if isinstance(arr, np.ndarray):\n arr = arr.astype(str) \n else:\n arr = [str(v) for v in arr]\n data[idx][key] = arr\n return json\n\n def _init_params(self):\n viewport_params = [p for p in self.param if 'viewport' in p]\n parameters = list(Layoutable.param)+viewport_params\n params = {p: getattr(self, p) for p in parameters\n if getattr(self, p) is not None}\n\n if self.object is None:\n json, sources = {}, []\n else:\n fig = self._to_figure(self.object)\n json = self._plotly_json_wrapper(fig)\n sources = Plotly._get_sources(json)\n\n params['_render_count'] = self._render_count\n params['config'] = self.config or {}\n params['data'] = json.get('data', [])\n params['data_sources'] = sources\n params['layout'] = layout = json.get('layout', {})\n if layout.get('autosize') and self.sizing_mode is self.param.sizing_mode.default:\n params['sizing_mode'] = 'stretch_both'\n return params\n\n def _get_model(self, doc, root=None, parent=None, comm=None):\n PlotlyPlot = lazy_load('panel.models.plotly', 'PlotlyPlot', isinstance(comm, JupyterComm))\n model = PlotlyPlot(**self._init_params())\n if root is None:\n root = model\n self._link_props(model, self._linkable_params, doc, root, comm)\n self._models[root.ref['id']] = (model, parent)\n if _patch_tabs_plotly not in Viewable._preprocessing_hooks:\n Viewable._preprocessing_hooks.append(_patch_tabs_plotly)\n return model\n\n def _update(self, ref=None, model=None):\n if self.object is None:\n model.update(data=[], layout={})\n model._render_count += 1\n return\n\n fig = self._to_figure(self.object)\n json = self._plotly_json_wrapper(fig)\n layout = json.get('layout')\n\n traces = json['data']\n new_sources = []\n update_sources = False\n for i, trace in enumerate(traces):\n if i < len(model.data_sources):\n cds = model.data_sources[i]\n else:\n cds = ColumnDataSource()\n new_sources.append(cds)\n\n update_sources = self._update_data_sources(cds, trace) or update_sources\n try:\n update_layout = model.layout != layout\n except Exception:\n update_layout = True\n\n # Determine if model needs updates\n if (len(model.data) != len(traces)):\n update_data = True\n else:\n update_data = False\n for new, old in zip(traces, model.data):\n try:\n update_data = (\n {k: v for k, v in new.items() if k != 'uid'} !=\n {k: v for k, v in old.items() if k != 'uid'}\n )\n except Exception:\n update_data = True\n if update_data:\n break\n\n updates = {}\n if self.sizing_mode is self.param.sizing_mode.default and 'autosize' in layout:\n autosize = layout.get('autosize')\n if autosize and model.sizing_mode != 'stretch_both':\n updates['sizing_mode'] = 'stretch_both'\n elif not autosize and model.sizing_mode != 'fixed':\n updates['sizing_mode'] = 'fixed'\n\n if new_sources:\n updates['data_sources'] = model.data_sources + new_sources\n\n if update_data:\n updates['data'] = json.get('data')\n\n if update_layout:\n updates['layout'] = layout\n\n if updates:\n model.update(**updates)\n\n # Check if we should trigger rendering\n if updates or update_sources:\n model._render_count += 1\n\n\ndef _patch_tabs_plotly(viewable, root):\n \"\"\"\n A preprocessing hook which ensures that any Plotly panes rendered\n inside Tabs are only visible when the tab they are in is active.\n This is a workaround for https://github.com/holoviz/panel/issues/804.\n \"\"\"\n from ..models.plotly import PlotlyPlot\n\n tabs_models = list(root.select({'type': Tabs}))\n plotly_models = root.select({'type': PlotlyPlot})\n\n for model in plotly_models:\n parent_tabs = []\n for tabs in tabs_models:\n if tabs.select_one({'id': model.id}):\n parent_tabs.append(tabs)\n parent_tab = None\n for tabs in parent_tabs:\n if not any(tabs.select_one(pt) for pt in parent_tabs if pt is not tabs):\n parent_tab = tabs\n break\n if parent_tab is None:\n return\n for i, tab in enumerate(parent_tab.tabs):\n if tab.select_one({'id': model.id}):\n break\n updated = False\n code = \"model.visible = cb_obj.active == i;\"\n for cb in parent_tab.js_property_callbacks.get('change:active', []):\n if cb.code == code and cb.args.get('model') is model:\n cb.args['i'] = i\n updated = True\n if updated:\n continue\n callback = CustomJS(args={'model': model, 'i': i, 'margin': model.margin}, code=code)\n parent_tab.js_on_change('active', callback)\n model.visible = parent_tab.active == i\n", "path": "panel/pane/plotly.py"}]}
| 3,768 | 654 |
gh_patches_debug_32920
|
rasdani/github-patches
|
git_diff
|
nipy__nipype-2490
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
UnboundLocalError: local variable 'module_name' referenced before assignment
### Summary
Discovered for myself `nipypecli` and decided to give it a try while composing cmdline invocation just following the errors it was spitting out at me and stopping when error didn't give a hint what I could have specified incorrectly:
```
$> nipypecli convert boutiques -m nipype.interfaces.ants.registration -i ANTS -o test
Traceback (most recent call last):
File "/usr/bin/nipypecli", line 11, in <module>
load_entry_point('nipype==1.0.1', 'console_scripts', 'nipypecli')()
File "/usr/lib/python2.7/dist-packages/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/usr/lib/python2.7/dist-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/lib/python2.7/dist-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/lib/python2.7/dist-packages/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/lib/python2.7/dist-packages/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/nipype/scripts/cli.py", line 254, in boutiques
verbose, ignore_template_numbers)
File "/usr/lib/python2.7/dist-packages/nipype/utils/nipype2boutiques.py", line 56, in generate_boutiques_descriptor
'command-line'] = "nipype_cmd " + module_name + " " + interface_name + " "
UnboundLocalError: local variable 'module_name' referenced before assignment
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `nipype/utils/nipype2boutiques.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 from __future__ import (print_function, division, unicode_literals,
3 absolute_import)
4
5 from builtins import str, open
6 # This tool exports a Nipype interface in the Boutiques (https://github.com/boutiques) JSON format.
7 # Boutiques tools can be imported in CBRAIN (https://github.com/aces/cbrain) among other platforms.
8 #
9 # Limitations:
10 # * List outputs are not supported.
11 # * Default values are not extracted from the documentation of the Nipype interface.
12 # * The following input types must be ignored for the output path template creation (see option -t):
13 # ** String restrictions, i.e. String inputs that accept only a restricted set of values.
14 # ** mutually exclusive inputs.
15 # * Path-templates are wrong when output files are not created in the execution directory (e.g. when a sub-directory is created).
16 # * Optional outputs, i.e. outputs that not always produced, may not be detected.
17
18 import os
19 import argparse
20 import sys
21 import tempfile
22 import simplejson as json
23
24 from ..scripts.instance import import_module
25
26
27 def generate_boutiques_descriptor(
28 module, interface_name, ignored_template_inputs, docker_image,
29 docker_index, verbose, ignore_template_numbers):
30 '''
31 Returns a JSON string containing a JSON Boutiques description of a Nipype interface.
32 Arguments:
33 * module: module where the Nipype interface is declared.
34 * interface: Nipype interface.
35 * ignored_template_inputs: a list of input names that should be ignored in the generation of output path templates.
36 * ignore_template_numbers: True if numbers must be ignored in output path creations.
37 '''
38
39 if not module:
40 raise Exception("Undefined module.")
41
42 # Retrieves Nipype interface
43 if isinstance(module, str):
44 import_module(module)
45 module_name = str(module)
46 module = sys.modules[module]
47
48 interface = getattr(module, interface_name)()
49 inputs = interface.input_spec()
50 outputs = interface.output_spec()
51
52 # Tool description
53 tool_desc = {}
54 tool_desc['name'] = interface_name
55 tool_desc[
56 'command-line'] = "nipype_cmd " + module_name + " " + interface_name + " "
57 tool_desc[
58 'description'] = interface_name + ", as implemented in Nipype (module: " + module_name + ", interface: " + interface_name + ")."
59 tool_desc['inputs'] = []
60 tool_desc['outputs'] = []
61 tool_desc['tool-version'] = interface.version
62 tool_desc['schema-version'] = '0.2-snapshot'
63 if docker_image:
64 tool_desc['docker-image'] = docker_image
65 if docker_index:
66 tool_desc['docker-index'] = docker_index
67
68 # Generates tool inputs
69 for name, spec in sorted(interface.inputs.traits(transient=None).items()):
70 input = get_boutiques_input(inputs, interface, name, spec,
71 ignored_template_inputs, verbose,
72 ignore_template_numbers)
73 tool_desc['inputs'].append(input)
74 tool_desc['command-line'] += input['command-line-key'] + " "
75 if verbose:
76 print("-> Adding input " + input['name'])
77
78 # Generates tool outputs
79 for name, spec in sorted(outputs.traits(transient=None).items()):
80 output = get_boutiques_output(name, interface, tool_desc['inputs'],
81 verbose)
82 if output['path-template'] != "":
83 tool_desc['outputs'].append(output)
84 if verbose:
85 print("-> Adding output " + output['name'])
86 elif verbose:
87 print("xx Skipping output " + output['name'] +
88 " with no path template.")
89 if tool_desc['outputs'] == []:
90 raise Exception("Tool has no output.")
91
92 # Removes all temporary values from inputs (otherwise they will
93 # appear in the JSON output)
94 for input in tool_desc['inputs']:
95 del input['tempvalue']
96
97 return json.dumps(tool_desc, indent=4, separators=(',', ': '))
98
99
100 def get_boutiques_input(inputs, interface, input_name, spec,
101 ignored_template_inputs, verbose,
102 ignore_template_numbers):
103 """
104 Returns a dictionary containing the Boutiques input corresponding to a Nipype intput.
105
106 Args:
107 * inputs: inputs of the Nipype interface.
108 * interface: Nipype interface.
109 * input_name: name of the Nipype input.
110 * spec: Nipype input spec.
111 * ignored_template_inputs: input names for which no temporary value must be generated.
112 * ignore_template_numbers: True if numbers must be ignored in output path creations.
113
114 Assumes that:
115 * Input names are unique.
116 """
117 if not spec.desc:
118 spec.desc = "No description provided."
119 spec_info = spec.full_info(inputs, input_name, None)
120
121 input = {}
122 input['id'] = input_name
123 input['name'] = input_name.replace('_', ' ').capitalize()
124 input['type'] = get_type_from_spec_info(spec_info)
125 input['list'] = is_list(spec_info)
126 input['command-line-key'] = "[" + input_name.upper(
127 ) + "]" # assumes that input names are unique
128 input['command-line-flag'] = ("--%s" % input_name + " ").strip()
129 input['tempvalue'] = None
130 input['description'] = spec_info.capitalize(
131 ) + ". " + spec.desc.capitalize()
132 if not input['description'].endswith('.'):
133 input['description'] += '.'
134 if not (hasattr(spec, "mandatory") and spec.mandatory):
135 input['optional'] = True
136 else:
137 input['optional'] = False
138 if spec.usedefault:
139 input['default-value'] = spec.default_value()[1]
140
141 # Create unique, temporary value.
142 temp_value = must_generate_value(input_name, input['type'],
143 ignored_template_inputs, spec_info, spec,
144 ignore_template_numbers)
145 if temp_value:
146 tempvalue = get_unique_value(input['type'], input_name)
147 setattr(interface.inputs, input_name, tempvalue)
148 input['tempvalue'] = tempvalue
149 if verbose:
150 print("oo Path-template creation using " + input['id'] + "=" +
151 str(tempvalue))
152
153 # Now that temp values have been generated, set Boolean types to
154 # Number (there is no Boolean type in Boutiques)
155 if input['type'] == "Boolean":
156 input['type'] = "Number"
157
158 return input
159
160
161 def get_boutiques_output(name, interface, tool_inputs, verbose=False):
162 """
163 Returns a dictionary containing the Boutiques output corresponding to a Nipype output.
164
165 Args:
166 * name: name of the Nipype output.
167 * interface: Nipype interface.
168 * tool_inputs: list of tool inputs (as produced by method get_boutiques_input).
169
170 Assumes that:
171 * Output names are unique.
172 * Input values involved in the path template are defined.
173 * Output files are written in the current directory.
174 * There is a single output value (output lists are not supported).
175 """
176 output = {}
177 output['name'] = name.replace('_', ' ').capitalize()
178 output['id'] = name
179 output['type'] = "File"
180 output['path-template'] = ""
181 output[
182 'optional'] = True # no real way to determine if an output is always produced, regardless of the input values.
183
184 # Path template creation.
185
186 output_value = interface._list_outputs()[name]
187 if output_value != "" and isinstance(
188 output_value,
189 str): # FIXME: this crashes when there are multiple output values.
190 # Go find from which input value it was built
191 for input in tool_inputs:
192 if not input['tempvalue']:
193 continue
194 input_value = input['tempvalue']
195 if input['type'] == "File":
196 # Take the base name
197 input_value = os.path.splitext(
198 os.path.basename(input_value))[0]
199 if str(input_value) in output_value:
200 output_value = os.path.basename(
201 output_value.replace(input_value,
202 input['command-line-key'])
203 ) # FIXME: this only works if output is written in the current directory
204 output['path-template'] = os.path.basename(output_value)
205 return output
206
207
208 def get_type_from_spec_info(spec_info):
209 '''
210 Returns an input type from the spec info. There must be a better
211 way to get an input type in Nipype than to parse the spec info.
212 '''
213 if ("an existing file name" in spec_info) or (
214 "input volumes" in spec_info):
215 return "File"
216 elif ("an integer" in spec_info or "a float" in spec_info):
217 return "Number"
218 elif "a boolean" in spec_info:
219 return "Boolean"
220 return "String"
221
222
223 def is_list(spec_info):
224 '''
225 Returns True if the spec info looks like it describes a list
226 parameter. There must be a better way in Nipype to check if an input
227 is a list.
228 '''
229 if "a list" in spec_info:
230 return True
231 return False
232
233
234 def get_unique_value(type, id):
235 '''
236 Returns a unique value of type 'type', for input with id 'id',
237 assuming id is unique.
238 '''
239 return {
240 "File": os.path.abspath(create_tempfile()),
241 "Boolean": True,
242 "Number": abs(hash(id)), # abs in case input param must be positive...
243 "String": id
244 }[type]
245
246
247 def create_tempfile():
248 '''
249 Creates a temp file and returns its name.
250 '''
251 fileTemp = tempfile.NamedTemporaryFile(delete=False)
252 fileTemp.write("hello")
253 fileTemp.close()
254 return fileTemp.name
255
256
257 def must_generate_value(name, type, ignored_template_inputs, spec_info, spec,
258 ignore_template_numbers):
259 '''
260 Return True if a temporary value must be generated for this input.
261 Arguments:
262 * name: input name.
263 * type: input_type.
264 * ignored_template_inputs: a list of inputs names for which no value must be generated.
265 * spec_info: spec info of the Nipype input
266 * ignore_template_numbers: True if numbers must be ignored.
267 '''
268 # Return false when type is number and numbers must be ignored.
269 if ignore_template_numbers and type == "Number":
270 return False
271 # Only generate value for the first element of mutually exclusive inputs.
272 if spec.xor and spec.xor[0] != name:
273 return False
274 # Directory types are not supported
275 if "an existing directory name" in spec_info:
276 return False
277 # Don't know how to generate a list.
278 if "a list" in spec_info or "a tuple" in spec_info:
279 return False
280 # Don't know how to generate a dictionary.
281 if "a dictionary" in spec_info:
282 return False
283 # Best guess to detect string restrictions...
284 if "' or '" in spec_info:
285 return False
286 if not ignored_template_inputs:
287 return True
288 return not (name in ignored_template_inputs)
289
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/nipype/utils/nipype2boutiques.py b/nipype/utils/nipype2boutiques.py
--- a/nipype/utils/nipype2boutiques.py
+++ b/nipype/utils/nipype2boutiques.py
@@ -2,7 +2,7 @@
from __future__ import (print_function, division, unicode_literals,
absolute_import)
-from builtins import str, open
+from builtins import str, open, bytes
# This tool exports a Nipype interface in the Boutiques (https://github.com/boutiques) JSON format.
# Boutiques tools can be imported in CBRAIN (https://github.com/aces/cbrain) among other platforms.
#
@@ -40,10 +40,12 @@
raise Exception("Undefined module.")
# Retrieves Nipype interface
- if isinstance(module, str):
+ if isinstance(module, (str, bytes)):
import_module(module)
module_name = str(module)
module = sys.modules[module]
+ else:
+ module_name = str(module.__name__)
interface = getattr(module, interface_name)()
inputs = interface.input_spec()
@@ -249,7 +251,7 @@
Creates a temp file and returns its name.
'''
fileTemp = tempfile.NamedTemporaryFile(delete=False)
- fileTemp.write("hello")
+ fileTemp.write(b"hello")
fileTemp.close()
return fileTemp.name
@@ -283,6 +285,8 @@
# Best guess to detect string restrictions...
if "' or '" in spec_info:
return False
+ if spec.default or spec.default_value():
+ return False
if not ignored_template_inputs:
return True
return not (name in ignored_template_inputs)
|
{"golden_diff": "diff --git a/nipype/utils/nipype2boutiques.py b/nipype/utils/nipype2boutiques.py\n--- a/nipype/utils/nipype2boutiques.py\n+++ b/nipype/utils/nipype2boutiques.py\n@@ -2,7 +2,7 @@\n from __future__ import (print_function, division, unicode_literals,\n absolute_import)\n \n-from builtins import str, open\n+from builtins import str, open, bytes\n # This tool exports a Nipype interface in the Boutiques (https://github.com/boutiques) JSON format.\n # Boutiques tools can be imported in CBRAIN (https://github.com/aces/cbrain) among other platforms.\n #\n@@ -40,10 +40,12 @@\n raise Exception(\"Undefined module.\")\n \n # Retrieves Nipype interface\n- if isinstance(module, str):\n+ if isinstance(module, (str, bytes)):\n import_module(module)\n module_name = str(module)\n module = sys.modules[module]\n+ else:\n+ module_name = str(module.__name__)\n \n interface = getattr(module, interface_name)()\n inputs = interface.input_spec()\n@@ -249,7 +251,7 @@\n Creates a temp file and returns its name.\n '''\n fileTemp = tempfile.NamedTemporaryFile(delete=False)\n- fileTemp.write(\"hello\")\n+ fileTemp.write(b\"hello\")\n fileTemp.close()\n return fileTemp.name\n \n@@ -283,6 +285,8 @@\n # Best guess to detect string restrictions...\n if \"' or '\" in spec_info:\n return False\n+ if spec.default or spec.default_value():\n+ return False\n if not ignored_template_inputs:\n return True\n return not (name in ignored_template_inputs)\n", "issue": "UnboundLocalError: local variable 'module_name' referenced before assignment\n### Summary\r\nDiscovered for myself `nipypecli` and decided to give it a try while composing cmdline invocation just following the errors it was spitting out at me and stopping when error didn't give a hint what I could have specified incorrectly:\r\n\r\n```\r\n$> nipypecli convert boutiques -m nipype.interfaces.ants.registration -i ANTS -o test\r\nTraceback (most recent call last):\r\n File \"/usr/bin/nipypecli\", line 11, in <module>\r\n load_entry_point('nipype==1.0.1', 'console_scripts', 'nipypecli')()\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 722, in __call__\r\n return self.main(*args, **kwargs)\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 697, in main\r\n rv = self.invoke(ctx)\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 1066, in invoke\r\n return _process_result(sub_ctx.command.invoke(sub_ctx))\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 1066, in invoke\r\n return _process_result(sub_ctx.command.invoke(sub_ctx))\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 895, in invoke\r\n return ctx.invoke(self.callback, **ctx.params)\r\n File \"/usr/lib/python2.7/dist-packages/click/core.py\", line 535, in invoke\r\n return callback(*args, **kwargs)\r\n File \"/usr/lib/python2.7/dist-packages/nipype/scripts/cli.py\", line 254, in boutiques\r\n verbose, ignore_template_numbers)\r\n File \"/usr/lib/python2.7/dist-packages/nipype/utils/nipype2boutiques.py\", line 56, in generate_boutiques_descriptor\r\n 'command-line'] = \"nipype_cmd \" + module_name + \" \" + interface_name + \" \"\r\nUnboundLocalError: local variable 'module_name' referenced before assignment\r\n```\r\n\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nfrom __future__ import (print_function, division, unicode_literals,\n absolute_import)\n\nfrom builtins import str, open\n# This tool exports a Nipype interface in the Boutiques (https://github.com/boutiques) JSON format.\n# Boutiques tools can be imported in CBRAIN (https://github.com/aces/cbrain) among other platforms.\n#\n# Limitations:\n# * List outputs are not supported.\n# * Default values are not extracted from the documentation of the Nipype interface.\n# * The following input types must be ignored for the output path template creation (see option -t):\n# ** String restrictions, i.e. String inputs that accept only a restricted set of values.\n# ** mutually exclusive inputs.\n# * Path-templates are wrong when output files are not created in the execution directory (e.g. when a sub-directory is created).\n# * Optional outputs, i.e. outputs that not always produced, may not be detected.\n\nimport os\nimport argparse\nimport sys\nimport tempfile\nimport simplejson as json\n\nfrom ..scripts.instance import import_module\n\n\ndef generate_boutiques_descriptor(\n module, interface_name, ignored_template_inputs, docker_image,\n docker_index, verbose, ignore_template_numbers):\n '''\n Returns a JSON string containing a JSON Boutiques description of a Nipype interface.\n Arguments:\n * module: module where the Nipype interface is declared.\n * interface: Nipype interface.\n * ignored_template_inputs: a list of input names that should be ignored in the generation of output path templates.\n * ignore_template_numbers: True if numbers must be ignored in output path creations.\n '''\n\n if not module:\n raise Exception(\"Undefined module.\")\n\n # Retrieves Nipype interface\n if isinstance(module, str):\n import_module(module)\n module_name = str(module)\n module = sys.modules[module]\n\n interface = getattr(module, interface_name)()\n inputs = interface.input_spec()\n outputs = interface.output_spec()\n\n # Tool description\n tool_desc = {}\n tool_desc['name'] = interface_name\n tool_desc[\n 'command-line'] = \"nipype_cmd \" + module_name + \" \" + interface_name + \" \"\n tool_desc[\n 'description'] = interface_name + \", as implemented in Nipype (module: \" + module_name + \", interface: \" + interface_name + \").\"\n tool_desc['inputs'] = []\n tool_desc['outputs'] = []\n tool_desc['tool-version'] = interface.version\n tool_desc['schema-version'] = '0.2-snapshot'\n if docker_image:\n tool_desc['docker-image'] = docker_image\n if docker_index:\n tool_desc['docker-index'] = docker_index\n\n # Generates tool inputs\n for name, spec in sorted(interface.inputs.traits(transient=None).items()):\n input = get_boutiques_input(inputs, interface, name, spec,\n ignored_template_inputs, verbose,\n ignore_template_numbers)\n tool_desc['inputs'].append(input)\n tool_desc['command-line'] += input['command-line-key'] + \" \"\n if verbose:\n print(\"-> Adding input \" + input['name'])\n\n # Generates tool outputs\n for name, spec in sorted(outputs.traits(transient=None).items()):\n output = get_boutiques_output(name, interface, tool_desc['inputs'],\n verbose)\n if output['path-template'] != \"\":\n tool_desc['outputs'].append(output)\n if verbose:\n print(\"-> Adding output \" + output['name'])\n elif verbose:\n print(\"xx Skipping output \" + output['name'] +\n \" with no path template.\")\n if tool_desc['outputs'] == []:\n raise Exception(\"Tool has no output.\")\n\n # Removes all temporary values from inputs (otherwise they will\n # appear in the JSON output)\n for input in tool_desc['inputs']:\n del input['tempvalue']\n\n return json.dumps(tool_desc, indent=4, separators=(',', ': '))\n\n\ndef get_boutiques_input(inputs, interface, input_name, spec,\n ignored_template_inputs, verbose,\n ignore_template_numbers):\n \"\"\"\n Returns a dictionary containing the Boutiques input corresponding to a Nipype intput.\n\n Args:\n * inputs: inputs of the Nipype interface.\n * interface: Nipype interface.\n * input_name: name of the Nipype input.\n * spec: Nipype input spec.\n * ignored_template_inputs: input names for which no temporary value must be generated.\n * ignore_template_numbers: True if numbers must be ignored in output path creations.\n\n Assumes that:\n * Input names are unique.\n \"\"\"\n if not spec.desc:\n spec.desc = \"No description provided.\"\n spec_info = spec.full_info(inputs, input_name, None)\n\n input = {}\n input['id'] = input_name\n input['name'] = input_name.replace('_', ' ').capitalize()\n input['type'] = get_type_from_spec_info(spec_info)\n input['list'] = is_list(spec_info)\n input['command-line-key'] = \"[\" + input_name.upper(\n ) + \"]\" # assumes that input names are unique\n input['command-line-flag'] = (\"--%s\" % input_name + \" \").strip()\n input['tempvalue'] = None\n input['description'] = spec_info.capitalize(\n ) + \". \" + spec.desc.capitalize()\n if not input['description'].endswith('.'):\n input['description'] += '.'\n if not (hasattr(spec, \"mandatory\") and spec.mandatory):\n input['optional'] = True\n else:\n input['optional'] = False\n if spec.usedefault:\n input['default-value'] = spec.default_value()[1]\n\n # Create unique, temporary value.\n temp_value = must_generate_value(input_name, input['type'],\n ignored_template_inputs, spec_info, spec,\n ignore_template_numbers)\n if temp_value:\n tempvalue = get_unique_value(input['type'], input_name)\n setattr(interface.inputs, input_name, tempvalue)\n input['tempvalue'] = tempvalue\n if verbose:\n print(\"oo Path-template creation using \" + input['id'] + \"=\" +\n str(tempvalue))\n\n # Now that temp values have been generated, set Boolean types to\n # Number (there is no Boolean type in Boutiques)\n if input['type'] == \"Boolean\":\n input['type'] = \"Number\"\n\n return input\n\n\ndef get_boutiques_output(name, interface, tool_inputs, verbose=False):\n \"\"\"\n Returns a dictionary containing the Boutiques output corresponding to a Nipype output.\n\n Args:\n * name: name of the Nipype output.\n * interface: Nipype interface.\n * tool_inputs: list of tool inputs (as produced by method get_boutiques_input).\n\n Assumes that:\n * Output names are unique.\n * Input values involved in the path template are defined.\n * Output files are written in the current directory.\n * There is a single output value (output lists are not supported).\n \"\"\"\n output = {}\n output['name'] = name.replace('_', ' ').capitalize()\n output['id'] = name\n output['type'] = \"File\"\n output['path-template'] = \"\"\n output[\n 'optional'] = True # no real way to determine if an output is always produced, regardless of the input values.\n\n # Path template creation.\n\n output_value = interface._list_outputs()[name]\n if output_value != \"\" and isinstance(\n output_value,\n str): # FIXME: this crashes when there are multiple output values.\n # Go find from which input value it was built\n for input in tool_inputs:\n if not input['tempvalue']:\n continue\n input_value = input['tempvalue']\n if input['type'] == \"File\":\n # Take the base name\n input_value = os.path.splitext(\n os.path.basename(input_value))[0]\n if str(input_value) in output_value:\n output_value = os.path.basename(\n output_value.replace(input_value,\n input['command-line-key'])\n ) # FIXME: this only works if output is written in the current directory\n output['path-template'] = os.path.basename(output_value)\n return output\n\n\ndef get_type_from_spec_info(spec_info):\n '''\n Returns an input type from the spec info. There must be a better\n way to get an input type in Nipype than to parse the spec info.\n '''\n if (\"an existing file name\" in spec_info) or (\n \"input volumes\" in spec_info):\n return \"File\"\n elif (\"an integer\" in spec_info or \"a float\" in spec_info):\n return \"Number\"\n elif \"a boolean\" in spec_info:\n return \"Boolean\"\n return \"String\"\n\n\ndef is_list(spec_info):\n '''\n Returns True if the spec info looks like it describes a list\n parameter. There must be a better way in Nipype to check if an input\n is a list.\n '''\n if \"a list\" in spec_info:\n return True\n return False\n\n\ndef get_unique_value(type, id):\n '''\n Returns a unique value of type 'type', for input with id 'id',\n assuming id is unique.\n '''\n return {\n \"File\": os.path.abspath(create_tempfile()),\n \"Boolean\": True,\n \"Number\": abs(hash(id)), # abs in case input param must be positive...\n \"String\": id\n }[type]\n\n\ndef create_tempfile():\n '''\n Creates a temp file and returns its name.\n '''\n fileTemp = tempfile.NamedTemporaryFile(delete=False)\n fileTemp.write(\"hello\")\n fileTemp.close()\n return fileTemp.name\n\n\ndef must_generate_value(name, type, ignored_template_inputs, spec_info, spec,\n ignore_template_numbers):\n '''\n Return True if a temporary value must be generated for this input.\n Arguments:\n * name: input name.\n * type: input_type.\n * ignored_template_inputs: a list of inputs names for which no value must be generated.\n * spec_info: spec info of the Nipype input\n * ignore_template_numbers: True if numbers must be ignored.\n '''\n # Return false when type is number and numbers must be ignored.\n if ignore_template_numbers and type == \"Number\":\n return False\n # Only generate value for the first element of mutually exclusive inputs.\n if spec.xor and spec.xor[0] != name:\n return False\n # Directory types are not supported\n if \"an existing directory name\" in spec_info:\n return False\n # Don't know how to generate a list.\n if \"a list\" in spec_info or \"a tuple\" in spec_info:\n return False\n # Don't know how to generate a dictionary.\n if \"a dictionary\" in spec_info:\n return False\n # Best guess to detect string restrictions...\n if \"' or '\" in spec_info:\n return False\n if not ignored_template_inputs:\n return True\n return not (name in ignored_template_inputs)\n", "path": "nipype/utils/nipype2boutiques.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\nfrom __future__ import (print_function, division, unicode_literals,\n absolute_import)\n\nfrom builtins import str, open, bytes\n# This tool exports a Nipype interface in the Boutiques (https://github.com/boutiques) JSON format.\n# Boutiques tools can be imported in CBRAIN (https://github.com/aces/cbrain) among other platforms.\n#\n# Limitations:\n# * List outputs are not supported.\n# * Default values are not extracted from the documentation of the Nipype interface.\n# * The following input types must be ignored for the output path template creation (see option -t):\n# ** String restrictions, i.e. String inputs that accept only a restricted set of values.\n# ** mutually exclusive inputs.\n# * Path-templates are wrong when output files are not created in the execution directory (e.g. when a sub-directory is created).\n# * Optional outputs, i.e. outputs that not always produced, may not be detected.\n\nimport os\nimport argparse\nimport sys\nimport tempfile\nimport simplejson as json\n\nfrom ..scripts.instance import import_module\n\n\ndef generate_boutiques_descriptor(\n module, interface_name, ignored_template_inputs, docker_image,\n docker_index, verbose, ignore_template_numbers):\n '''\n Returns a JSON string containing a JSON Boutiques description of a Nipype interface.\n Arguments:\n * module: module where the Nipype interface is declared.\n * interface: Nipype interface.\n * ignored_template_inputs: a list of input names that should be ignored in the generation of output path templates.\n * ignore_template_numbers: True if numbers must be ignored in output path creations.\n '''\n\n if not module:\n raise Exception(\"Undefined module.\")\n\n # Retrieves Nipype interface\n if isinstance(module, (str, bytes)):\n import_module(module)\n module_name = str(module)\n module = sys.modules[module]\n else:\n module_name = str(module.__name__)\n\n interface = getattr(module, interface_name)()\n inputs = interface.input_spec()\n outputs = interface.output_spec()\n\n # Tool description\n tool_desc = {}\n tool_desc['name'] = interface_name\n tool_desc[\n 'command-line'] = \"nipype_cmd \" + module_name + \" \" + interface_name + \" \"\n tool_desc[\n 'description'] = interface_name + \", as implemented in Nipype (module: \" + module_name + \", interface: \" + interface_name + \").\"\n tool_desc['inputs'] = []\n tool_desc['outputs'] = []\n tool_desc['tool-version'] = interface.version\n tool_desc['schema-version'] = '0.2-snapshot'\n if docker_image:\n tool_desc['docker-image'] = docker_image\n if docker_index:\n tool_desc['docker-index'] = docker_index\n\n # Generates tool inputs\n for name, spec in sorted(interface.inputs.traits(transient=None).items()):\n input = get_boutiques_input(inputs, interface, name, spec,\n ignored_template_inputs, verbose,\n ignore_template_numbers)\n tool_desc['inputs'].append(input)\n tool_desc['command-line'] += input['command-line-key'] + \" \"\n if verbose:\n print(\"-> Adding input \" + input['name'])\n\n # Generates tool outputs\n for name, spec in sorted(outputs.traits(transient=None).items()):\n output = get_boutiques_output(name, interface, tool_desc['inputs'],\n verbose)\n if output['path-template'] != \"\":\n tool_desc['outputs'].append(output)\n if verbose:\n print(\"-> Adding output \" + output['name'])\n elif verbose:\n print(\"xx Skipping output \" + output['name'] +\n \" with no path template.\")\n if tool_desc['outputs'] == []:\n raise Exception(\"Tool has no output.\")\n\n # Removes all temporary values from inputs (otherwise they will\n # appear in the JSON output)\n for input in tool_desc['inputs']:\n del input['tempvalue']\n\n return json.dumps(tool_desc, indent=4, separators=(',', ': '))\n\n\ndef get_boutiques_input(inputs, interface, input_name, spec,\n ignored_template_inputs, verbose,\n ignore_template_numbers):\n \"\"\"\n Returns a dictionary containing the Boutiques input corresponding to a Nipype intput.\n\n Args:\n * inputs: inputs of the Nipype interface.\n * interface: Nipype interface.\n * input_name: name of the Nipype input.\n * spec: Nipype input spec.\n * ignored_template_inputs: input names for which no temporary value must be generated.\n * ignore_template_numbers: True if numbers must be ignored in output path creations.\n\n Assumes that:\n * Input names are unique.\n \"\"\"\n if not spec.desc:\n spec.desc = \"No description provided.\"\n spec_info = spec.full_info(inputs, input_name, None)\n\n input = {}\n input['id'] = input_name\n input['name'] = input_name.replace('_', ' ').capitalize()\n input['type'] = get_type_from_spec_info(spec_info)\n input['list'] = is_list(spec_info)\n input['command-line-key'] = \"[\" + input_name.upper(\n ) + \"]\" # assumes that input names are unique\n input['command-line-flag'] = (\"--%s\" % input_name + \" \").strip()\n input['tempvalue'] = None\n input['description'] = spec_info.capitalize(\n ) + \". \" + spec.desc.capitalize()\n if not input['description'].endswith('.'):\n input['description'] += '.'\n if not (hasattr(spec, \"mandatory\") and spec.mandatory):\n input['optional'] = True\n else:\n input['optional'] = False\n if spec.usedefault:\n input['default-value'] = spec.default_value()[1]\n\n # Create unique, temporary value.\n temp_value = must_generate_value(input_name, input['type'],\n ignored_template_inputs, spec_info, spec,\n ignore_template_numbers)\n if temp_value:\n tempvalue = get_unique_value(input['type'], input_name)\n setattr(interface.inputs, input_name, tempvalue)\n input['tempvalue'] = tempvalue\n if verbose:\n print(\"oo Path-template creation using \" + input['id'] + \"=\" +\n str(tempvalue))\n\n # Now that temp values have been generated, set Boolean types to\n # Number (there is no Boolean type in Boutiques)\n if input['type'] == \"Boolean\":\n input['type'] = \"Number\"\n\n return input\n\n\ndef get_boutiques_output(name, interface, tool_inputs, verbose=False):\n \"\"\"\n Returns a dictionary containing the Boutiques output corresponding to a Nipype output.\n\n Args:\n * name: name of the Nipype output.\n * interface: Nipype interface.\n * tool_inputs: list of tool inputs (as produced by method get_boutiques_input).\n\n Assumes that:\n * Output names are unique.\n * Input values involved in the path template are defined.\n * Output files are written in the current directory.\n * There is a single output value (output lists are not supported).\n \"\"\"\n output = {}\n output['name'] = name.replace('_', ' ').capitalize()\n output['id'] = name\n output['type'] = \"File\"\n output['path-template'] = \"\"\n output[\n 'optional'] = True # no real way to determine if an output is always produced, regardless of the input values.\n\n # Path template creation.\n\n output_value = interface._list_outputs()[name]\n if output_value != \"\" and isinstance(\n output_value,\n str): # FIXME: this crashes when there are multiple output values.\n # Go find from which input value it was built\n for input in tool_inputs:\n if not input['tempvalue']:\n continue\n input_value = input['tempvalue']\n if input['type'] == \"File\":\n # Take the base name\n input_value = os.path.splitext(\n os.path.basename(input_value))[0]\n if str(input_value) in output_value:\n output_value = os.path.basename(\n output_value.replace(input_value,\n input['command-line-key'])\n ) # FIXME: this only works if output is written in the current directory\n output['path-template'] = os.path.basename(output_value)\n return output\n\n\ndef get_type_from_spec_info(spec_info):\n '''\n Returns an input type from the spec info. There must be a better\n way to get an input type in Nipype than to parse the spec info.\n '''\n if (\"an existing file name\" in spec_info) or (\n \"input volumes\" in spec_info):\n return \"File\"\n elif (\"an integer\" in spec_info or \"a float\" in spec_info):\n return \"Number\"\n elif \"a boolean\" in spec_info:\n return \"Boolean\"\n return \"String\"\n\n\ndef is_list(spec_info):\n '''\n Returns True if the spec info looks like it describes a list\n parameter. There must be a better way in Nipype to check if an input\n is a list.\n '''\n if \"a list\" in spec_info:\n return True\n return False\n\n\ndef get_unique_value(type, id):\n '''\n Returns a unique value of type 'type', for input with id 'id',\n assuming id is unique.\n '''\n return {\n \"File\": os.path.abspath(create_tempfile()),\n \"Boolean\": True,\n \"Number\": abs(hash(id)), # abs in case input param must be positive...\n \"String\": id\n }[type]\n\n\ndef create_tempfile():\n '''\n Creates a temp file and returns its name.\n '''\n fileTemp = tempfile.NamedTemporaryFile(delete=False)\n fileTemp.write(b\"hello\")\n fileTemp.close()\n return fileTemp.name\n\n\ndef must_generate_value(name, type, ignored_template_inputs, spec_info, spec,\n ignore_template_numbers):\n '''\n Return True if a temporary value must be generated for this input.\n Arguments:\n * name: input name.\n * type: input_type.\n * ignored_template_inputs: a list of inputs names for which no value must be generated.\n * spec_info: spec info of the Nipype input\n * ignore_template_numbers: True if numbers must be ignored.\n '''\n # Return false when type is number and numbers must be ignored.\n if ignore_template_numbers and type == \"Number\":\n return False\n # Only generate value for the first element of mutually exclusive inputs.\n if spec.xor and spec.xor[0] != name:\n return False\n # Directory types are not supported\n if \"an existing directory name\" in spec_info:\n return False\n # Don't know how to generate a list.\n if \"a list\" in spec_info or \"a tuple\" in spec_info:\n return False\n # Don't know how to generate a dictionary.\n if \"a dictionary\" in spec_info:\n return False\n # Best guess to detect string restrictions...\n if \"' or '\" in spec_info:\n return False\n if spec.default or spec.default_value():\n return False\n if not ignored_template_inputs:\n return True\n return not (name in ignored_template_inputs)\n", "path": "nipype/utils/nipype2boutiques.py"}]}
| 3,897 | 392 |
gh_patches_debug_5450
|
rasdani/github-patches
|
git_diff
|
mozilla__bugbug-1713
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Set up Sentry for bugbug
Adding it will help investigate issues (like the Mercurial woes).
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `http_service/bugbug_http/worker.py`
Content:
```
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3 # This Source Code Form is subject to the terms of the Mozilla Public
4 # License, v. 2.0. If a copy of the MPL was not distributed with this file,
5 # You can obtain one at http://mozilla.org/MPL/2.0/.
6
7 import os
8 import sys
9
10 from redis import Redis
11 from rq import Connection, Worker
12
13 import bugbug_http.boot
14
15
16 def main():
17 # Bootstrap the worker assets
18 bugbug_http.boot.boot_worker()
19
20 # Provide queue names to listen to as arguments to this script,
21 # similar to rq worker
22 redis_url = os.environ.get("REDIS_URL", "redis://localhost/0")
23 redis_conn = Redis.from_url(redis_url)
24 with Connection(connection=redis_conn):
25 qs = sys.argv[1:] or ["default"]
26
27 w = Worker(qs)
28 w.work()
29
30
31 if __name__ == "__main__":
32 main()
33
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/http_service/bugbug_http/worker.py b/http_service/bugbug_http/worker.py
--- a/http_service/bugbug_http/worker.py
+++ b/http_service/bugbug_http/worker.py
@@ -7,10 +7,20 @@
import os
import sys
+import sentry_sdk
from redis import Redis
from rq import Connection, Worker
+from sentry_sdk.integrations.rq import RqIntegration
import bugbug_http.boot
+from bugbug import get_bugbug_version
+
+if os.environ.get("SENTRY_DSN"):
+ sentry_sdk.init(
+ os.environ.get("SENTRY_DSN"),
+ integrations=[RqIntegration()],
+ release=get_bugbug_version(),
+ )
def main():
|
{"golden_diff": "diff --git a/http_service/bugbug_http/worker.py b/http_service/bugbug_http/worker.py\n--- a/http_service/bugbug_http/worker.py\n+++ b/http_service/bugbug_http/worker.py\n@@ -7,10 +7,20 @@\n import os\n import sys\n \n+import sentry_sdk\n from redis import Redis\n from rq import Connection, Worker\n+from sentry_sdk.integrations.rq import RqIntegration\n \n import bugbug_http.boot\n+from bugbug import get_bugbug_version\n+\n+if os.environ.get(\"SENTRY_DSN\"):\n+ sentry_sdk.init(\n+ os.environ.get(\"SENTRY_DSN\"),\n+ integrations=[RqIntegration()],\n+ release=get_bugbug_version(),\n+ )\n \n \n def main():\n", "issue": "Set up Sentry for bugbug\nAdding it will help investigate issues (like the Mercurial woes).\n", "before_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# This Source Code Form is subject to the terms of the Mozilla Public\n# License, v. 2.0. If a copy of the MPL was not distributed with this file,\n# You can obtain one at http://mozilla.org/MPL/2.0/.\n\nimport os\nimport sys\n\nfrom redis import Redis\nfrom rq import Connection, Worker\n\nimport bugbug_http.boot\n\n\ndef main():\n # Bootstrap the worker assets\n bugbug_http.boot.boot_worker()\n\n # Provide queue names to listen to as arguments to this script,\n # similar to rq worker\n redis_url = os.environ.get(\"REDIS_URL\", \"redis://localhost/0\")\n redis_conn = Redis.from_url(redis_url)\n with Connection(connection=redis_conn):\n qs = sys.argv[1:] or [\"default\"]\n\n w = Worker(qs)\n w.work()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "http_service/bugbug_http/worker.py"}], "after_files": [{"content": "#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# This Source Code Form is subject to the terms of the Mozilla Public\n# License, v. 2.0. If a copy of the MPL was not distributed with this file,\n# You can obtain one at http://mozilla.org/MPL/2.0/.\n\nimport os\nimport sys\n\nimport sentry_sdk\nfrom redis import Redis\nfrom rq import Connection, Worker\nfrom sentry_sdk.integrations.rq import RqIntegration\n\nimport bugbug_http.boot\nfrom bugbug import get_bugbug_version\n\nif os.environ.get(\"SENTRY_DSN\"):\n sentry_sdk.init(\n os.environ.get(\"SENTRY_DSN\"),\n integrations=[RqIntegration()],\n release=get_bugbug_version(),\n )\n\n\ndef main():\n # Bootstrap the worker assets\n bugbug_http.boot.boot_worker()\n\n # Provide queue names to listen to as arguments to this script,\n # similar to rq worker\n redis_url = os.environ.get(\"REDIS_URL\", \"redis://localhost/0\")\n redis_conn = Redis.from_url(redis_url)\n with Connection(connection=redis_conn):\n qs = sys.argv[1:] or [\"default\"]\n\n w = Worker(qs)\n w.work()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "http_service/bugbug_http/worker.py"}]}
| 552 | 169 |
gh_patches_debug_58376
|
rasdani/github-patches
|
git_diff
|
deis__deis-323
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
`deis run` generates ugly error if app code not yet pushed
I ran `deis run ls -la` after I had created the app, but before I had pushed the code with `git push deis master`. Here is the error I received:
```
ben$ example-python-flask > deis run ls -la
Warning: non-zero return code 255
lxc-start: No such file or directory - failed to mount '/opt/deis/runtime/slugs/hushed-sailfish-1/app' on '/usr/lib/lxc/root///app'
lxc-start: failed to setup the mount entries for '5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'
lxc-start: failed to setup the container
lxc-start: invalid sequence number 1. expected 2
lxc-start: failed to spawn '5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'
lxc-start: Device or resource busy - failed to remove cgroup '/sys/fs/cgroup/cpuset//lxc/5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `api/tasks.py`
Content:
```
1
2 from __future__ import unicode_literals
3 import importlib
4
5 from celery import task
6 from celery.canvas import group
7
8 from deis import settings
9 from provider import import_provider_module
10
11 # import user-defined config management module
12 CM = importlib.import_module(settings.CM_MODULE)
13
14
15 @task
16 def build_layer(layer):
17 provider = import_provider_module(layer.flavor.provider.type)
18 provider.build_layer(layer.flat())
19
20
21 @task
22 def destroy_layer(layer):
23 provider = import_provider_module(layer.flavor.provider.type)
24 provider.destroy_layer(layer.flat())
25 layer.delete()
26
27
28 @task
29 def build_node(node):
30 provider = import_provider_module(node.layer.flavor.provider.type)
31 provider_id, fqdn, metadata = provider.build_node(node.flat())
32 node.provider_id = provider_id
33 node.fqdn = fqdn
34 node.metadata = metadata
35 node.save()
36 CM.bootstrap_node(node.flat())
37
38
39 @task
40 def destroy_node(node):
41 provider = import_provider_module(node.layer.flavor.provider.type)
42 provider.destroy_node(node.flat())
43 CM.purge_node(node.flat())
44 node.delete()
45
46
47 @task
48 def converge_node(node):
49 output, rc = CM.converge_node(node.flat())
50 return output, rc
51
52
53 @task
54 def run_node(node, command):
55 output, rc = CM.run_node(node.flat(), command)
56 return output, rc
57
58
59 @task
60 def build_formation(formation):
61 return
62
63
64 @task
65 def destroy_formation(formation):
66 app_tasks = [destroy_app.si(a) for a in formation.app_set.all()]
67 node_tasks = [destroy_node.si(n) for n in formation.node_set.all()]
68 layer_tasks = [destroy_layer.si(l) for l in formation.layer_set.all()]
69 group(app_tasks + node_tasks).apply_async().join()
70 group(layer_tasks).apply_async().join()
71 CM.purge_formation(formation.flat())
72 formation.delete()
73
74
75 @task
76 def converge_formation(formation):
77 nodes = formation.node_set.all()
78 subtasks = []
79 for n in nodes:
80 subtask = converge_node.si(n)
81 subtasks.append(subtask)
82 group(*subtasks).apply_async().join()
83
84
85 @task
86 def build_app(app):
87 return
88
89
90 @task
91 def destroy_app(app):
92 CM.purge_app(app.flat())
93 app.delete()
94 app.formation.publish()
95
96
97 @task
98 def converge_controller():
99 CM.converge_controller()
100 return None
101
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/api/tasks.py b/api/tasks.py
--- a/api/tasks.py
+++ b/api/tasks.py
@@ -53,6 +53,8 @@
@task
def run_node(node, command):
output, rc = CM.run_node(node.flat(), command)
+ if rc != 0 and 'failed to setup the container' in output:
+ output = '\033[35mPlease run `git push deis master` first.\033[0m\n' + output
return output, rc
|
{"golden_diff": "diff --git a/api/tasks.py b/api/tasks.py\n--- a/api/tasks.py\n+++ b/api/tasks.py\n@@ -53,6 +53,8 @@\n @task\n def run_node(node, command):\n output, rc = CM.run_node(node.flat(), command)\n+ if rc != 0 and 'failed to setup the container' in output:\n+ output = '\\033[35mPlease run `git push deis master` first.\\033[0m\\n' + output\n return output, rc\n", "issue": "`deis run` generates ugly error if app code not yet pushed\nI ran `deis run ls -la` after I had created the app, but before I had pushed the code with `git push deis master`. Here is the error I received:\n\n```\nben$ example-python-flask > deis run ls -la\nWarning: non-zero return code 255\nlxc-start: No such file or directory - failed to mount '/opt/deis/runtime/slugs/hushed-sailfish-1/app' on '/usr/lib/lxc/root///app'\nlxc-start: failed to setup the mount entries for '5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'\nlxc-start: failed to setup the container\nlxc-start: invalid sequence number 1. expected 2\nlxc-start: failed to spawn '5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'\nlxc-start: Device or resource busy - failed to remove cgroup '/sys/fs/cgroup/cpuset//lxc/5f4f4d932501338fa2062d52e5893dfbd3933fa09102c67493a169a2a87ee479'\n```\n\n", "before_files": [{"content": "\nfrom __future__ import unicode_literals\nimport importlib\n\nfrom celery import task\nfrom celery.canvas import group\n\nfrom deis import settings\nfrom provider import import_provider_module\n\n# import user-defined config management module\nCM = importlib.import_module(settings.CM_MODULE)\n\n\n@task\ndef build_layer(layer):\n provider = import_provider_module(layer.flavor.provider.type)\n provider.build_layer(layer.flat())\n\n\n@task\ndef destroy_layer(layer):\n provider = import_provider_module(layer.flavor.provider.type)\n provider.destroy_layer(layer.flat())\n layer.delete()\n\n\n@task\ndef build_node(node):\n provider = import_provider_module(node.layer.flavor.provider.type)\n provider_id, fqdn, metadata = provider.build_node(node.flat())\n node.provider_id = provider_id\n node.fqdn = fqdn\n node.metadata = metadata\n node.save()\n CM.bootstrap_node(node.flat())\n\n\n@task\ndef destroy_node(node):\n provider = import_provider_module(node.layer.flavor.provider.type)\n provider.destroy_node(node.flat())\n CM.purge_node(node.flat())\n node.delete()\n\n\n@task\ndef converge_node(node):\n output, rc = CM.converge_node(node.flat())\n return output, rc\n\n\n@task\ndef run_node(node, command):\n output, rc = CM.run_node(node.flat(), command)\n return output, rc\n\n\n@task\ndef build_formation(formation):\n return\n\n\n@task\ndef destroy_formation(formation):\n app_tasks = [destroy_app.si(a) for a in formation.app_set.all()]\n node_tasks = [destroy_node.si(n) for n in formation.node_set.all()]\n layer_tasks = [destroy_layer.si(l) for l in formation.layer_set.all()]\n group(app_tasks + node_tasks).apply_async().join()\n group(layer_tasks).apply_async().join()\n CM.purge_formation(formation.flat())\n formation.delete()\n\n\n@task\ndef converge_formation(formation):\n nodes = formation.node_set.all()\n subtasks = []\n for n in nodes:\n subtask = converge_node.si(n)\n subtasks.append(subtask)\n group(*subtasks).apply_async().join()\n\n\n@task\ndef build_app(app):\n return\n\n\n@task\ndef destroy_app(app):\n CM.purge_app(app.flat())\n app.delete()\n app.formation.publish()\n\n\n@task\ndef converge_controller():\n CM.converge_controller()\n return None\n", "path": "api/tasks.py"}], "after_files": [{"content": "\nfrom __future__ import unicode_literals\nimport importlib\n\nfrom celery import task\nfrom celery.canvas import group\n\nfrom deis import settings\nfrom provider import import_provider_module\n\n# import user-defined config management module\nCM = importlib.import_module(settings.CM_MODULE)\n\n\n@task\ndef build_layer(layer):\n provider = import_provider_module(layer.flavor.provider.type)\n provider.build_layer(layer.flat())\n\n\n@task\ndef destroy_layer(layer):\n provider = import_provider_module(layer.flavor.provider.type)\n provider.destroy_layer(layer.flat())\n layer.delete()\n\n\n@task\ndef build_node(node):\n provider = import_provider_module(node.layer.flavor.provider.type)\n provider_id, fqdn, metadata = provider.build_node(node.flat())\n node.provider_id = provider_id\n node.fqdn = fqdn\n node.metadata = metadata\n node.save()\n CM.bootstrap_node(node.flat())\n\n\n@task\ndef destroy_node(node):\n provider = import_provider_module(node.layer.flavor.provider.type)\n provider.destroy_node(node.flat())\n CM.purge_node(node.flat())\n node.delete()\n\n\n@task\ndef converge_node(node):\n output, rc = CM.converge_node(node.flat())\n return output, rc\n\n\n@task\ndef run_node(node, command):\n output, rc = CM.run_node(node.flat(), command)\n if rc != 0 and 'failed to setup the container' in output:\n output = '\\033[35mPlease run `git push deis master` first.\\033[0m\\n' + output\n return output, rc\n\n\n@task\ndef build_formation(formation):\n return\n\n\n@task\ndef destroy_formation(formation):\n app_tasks = [destroy_app.si(a) for a in formation.app_set.all()]\n node_tasks = [destroy_node.si(n) for n in formation.node_set.all()]\n layer_tasks = [destroy_layer.si(l) for l in formation.layer_set.all()]\n group(app_tasks + node_tasks).apply_async().join()\n group(layer_tasks).apply_async().join()\n CM.purge_formation(formation.flat())\n formation.delete()\n\n\n@task\ndef converge_formation(formation):\n nodes = formation.node_set.all()\n subtasks = []\n for n in nodes:\n subtask = converge_node.si(n)\n subtasks.append(subtask)\n group(*subtasks).apply_async().join()\n\n\n@task\ndef build_app(app):\n return\n\n\n@task\ndef destroy_app(app):\n CM.purge_app(app.flat())\n app.delete()\n app.formation.publish()\n\n\n@task\ndef converge_controller():\n CM.converge_controller()\n return None\n", "path": "api/tasks.py"}]}
| 1,356 | 117 |
gh_patches_debug_14336
|
rasdani/github-patches
|
git_diff
|
DataBiosphere__toil-2077
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Can't install from tarball on GitHub
```
$ pip install https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz
Collecting https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz
Downloading https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz
- 7.0MB 40.4MB/s
Complete output from command python setup.py egg_info:
fatal: Not a git repository (or any of the parent directories): .git
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-l5WfLH-build/setup.py", line 152, in <module>
version = importVersion()
File "/tmp/pip-l5WfLH-build/setup.py", line 132, in importVersion
new = version_template.expand_()
File "version_template.py", line 131, in expand_
return ''.join("%s = %s\n" % (k, repr(resolve(k))) for k, v in variables.items())
File "version_template.py", line 131, in <genexpr>
return ''.join("%s = %s\n" % (k, repr(resolve(k))) for k, v in variables.items())
File "version_template.py", line 127, in resolve
v = v()
File "version_template.py", line 85, in dockerShortTag
return shortVersion()
File "version_template.py", line 46, in shortVersion
return _version(shorten=True)
File "version_template.py", line 51, in _version
currentCommit()[:7 if shorten else None],
File "version_template.py", line 111, in currentCommit
return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]
File "/usr/lib/python2.7/subprocess.py", line 223, in check_output
raise CalledProcessError(retcode, cmd, output=output)
subprocess.CalledProcessError: Command 'git log --pretty=oneline -n 1 -- $(pwd)' returned non-zero exit status 128
----------------------------------------
Command "python setup.py egg_info" failed with error code 1 in /tmp/pip-l5WfLH-build/
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `version_template.py`
Content:
```
1 # Copyright (C) 2015-2016 Regents of the University of California
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 # http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 """This script is a template for src/toil/version.py. Running it without arguments echoes all
16 globals, i.e. module attributes. Constant assignments will be echoed verbatim while callables
17 will be invoked and their result echoed as an assignment using the function name as the left-hand
18 side and the return value of the function as right-hand side. To prevent a module attribute from
19 being echoed, start or end the attribute name with an underscore. To print the value of a single
20 symbol, pass the name of that attribute to the script as a command line argument. You can also
21 import the expand_ function and invoke it directly with either no or exactly one argument."""
22
23 # Note to maintainers:
24 #
25 # - don't import at module level unless you want the imported value to be included in the output
26 # - only import from the Python standard run-time library (you can't have any dependencies)
27
28 baseVersion = '3.15.0a1'
29
30 cgcloudVersion = '1.6.0a1.dev393'
31
32
33 def version():
34 """
35 A version identifier that includes the full-legth commit SHA1 and an optional suffix to
36 indicate that the working copy is dirty.
37 """
38 return _version()
39
40
41 def shortVersion():
42 """
43 A version identifier that includes the abbreviated commit SHA1 and an optional suffix to
44 indicate that the working copy is dirty.
45 """
46 return _version(shorten=True)
47
48
49 def _version(shorten=False):
50 return '-'.join(filter(None, [distVersion(),
51 currentCommit()[:7 if shorten else None],
52 ('dirty' if dirty() else None)]))
53
54
55 def distVersion():
56 """
57 The distribution version identifying a published release on PyPI.
58 """
59 from pkg_resources import parse_version
60 build_number = buildNumber()
61 parsedBaseVersion = parse_version(baseVersion)
62 if isinstance(parsedBaseVersion, tuple):
63 raise RuntimeError("Setuptools version 8.0 or newer required. Update by running "
64 "'pip install setuptools --upgrade'")
65
66 if build_number is not None and parsedBaseVersion.is_prerelease:
67 return baseVersion + '.dev' + build_number
68 else:
69 return baseVersion
70
71
72 def dockerTag():
73 """
74 The primary tag of the Docker image for the appliance. This uniquely identifies the appliance
75 image.
76 """
77 return version()
78
79
80 def dockerShortTag():
81 """
82 A secondary, shortened form of :func:`dockerTag` with which to tag the appliance image for
83 convenience.
84 """
85 return shortVersion()
86
87
88 def dockerMinimalTag():
89 """
90 A minimal tag with which to tag the appliance image for convenience. Does not include
91 information about the git commit or working copy dirtyness.
92 """
93 return distVersion()
94
95
96 dockerRegistry = 'quay.io/ucsc_cgl'
97
98 dockerName = 'toil'
99
100
101 def buildNumber():
102 """
103 The Jenkins build number, if defined, else None.
104 """
105 import os
106 return os.getenv('BUILD_NUMBER')
107
108
109 def currentCommit():
110 from subprocess import check_output
111 return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]
112
113
114 def dirty():
115 from subprocess import call
116 return 0 != call('(git diff --exit-code '
117 '&& git diff --cached --exit-code) > /dev/null', shell=True)
118
119
120 def expand_(name=None):
121 variables = {k: v for k, v in globals().items()
122 if not k.startswith('_') and not k.endswith('_')}
123
124 def resolve(k):
125 v = variables[k]
126 if callable(v):
127 v = v()
128 return v
129
130 if name is None:
131 return ''.join("%s = %s\n" % (k, repr(resolve(k))) for k, v in variables.items())
132 else:
133 return resolve(name)
134
135
136 def _main():
137 import sys
138 sys.stdout.write(expand_(*sys.argv[1:]))
139
140
141 if __name__ == '__main__':
142 _main()
143
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/version_template.py b/version_template.py
--- a/version_template.py
+++ b/version_template.py
@@ -108,13 +108,21 @@
def currentCommit():
from subprocess import check_output
- return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]
+ try:
+ output = check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]
+ except:
+ # Return this we are not in a git environment.
+ return '000'
+ return output
def dirty():
from subprocess import call
- return 0 != call('(git diff --exit-code '
- '&& git diff --cached --exit-code) > /dev/null', shell=True)
+ try:
+ return 0 != call('(git diff --exit-code '
+ '&& git diff --cached --exit-code) > /dev/null', shell=True)
+ except:
+ return False # In case the git call fails.
def expand_(name=None):
|
{"golden_diff": "diff --git a/version_template.py b/version_template.py\n--- a/version_template.py\n+++ b/version_template.py\n@@ -108,13 +108,21 @@\n \n def currentCommit():\n from subprocess import check_output\n- return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]\n+ try:\n+ output = check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]\n+ except:\n+ # Return this we are not in a git environment.\n+ return '000'\n+ return output\n \n \n def dirty():\n from subprocess import call\n- return 0 != call('(git diff --exit-code '\n- '&& git diff --cached --exit-code) > /dev/null', shell=True)\n+ try:\n+ return 0 != call('(git diff --exit-code '\n+ '&& git diff --cached --exit-code) > /dev/null', shell=True)\n+ except:\n+ return False # In case the git call fails.\n \n \n def expand_(name=None):\n", "issue": "Can't install from tarball on GitHub\n```\r\n$ pip install https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz\r\nCollecting https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz\r\n Downloading https://github.com/BD2KGenomics/toil/archive/3.14.0.tar.gz\r\n - 7.0MB 40.4MB/s\r\n Complete output from command python setup.py egg_info:\r\n fatal: Not a git repository (or any of the parent directories): .git\r\n Traceback (most recent call last):\r\n File \"<string>\", line 1, in <module>\r\n File \"/tmp/pip-l5WfLH-build/setup.py\", line 152, in <module>\r\n version = importVersion()\r\n File \"/tmp/pip-l5WfLH-build/setup.py\", line 132, in importVersion\r\n new = version_template.expand_()\r\n File \"version_template.py\", line 131, in expand_\r\n return ''.join(\"%s = %s\\n\" % (k, repr(resolve(k))) for k, v in variables.items())\r\n File \"version_template.py\", line 131, in <genexpr>\r\n return ''.join(\"%s = %s\\n\" % (k, repr(resolve(k))) for k, v in variables.items())\r\n File \"version_template.py\", line 127, in resolve\r\n v = v()\r\n File \"version_template.py\", line 85, in dockerShortTag\r\n return shortVersion()\r\n File \"version_template.py\", line 46, in shortVersion\r\n return _version(shorten=True)\r\n File \"version_template.py\", line 51, in _version\r\n currentCommit()[:7 if shorten else None],\r\n File \"version_template.py\", line 111, in currentCommit\r\n return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]\r\n File \"/usr/lib/python2.7/subprocess.py\", line 223, in check_output\r\n raise CalledProcessError(retcode, cmd, output=output)\r\n subprocess.CalledProcessError: Command 'git log --pretty=oneline -n 1 -- $(pwd)' returned non-zero exit status 128\r\n \r\n ----------------------------------------\r\nCommand \"python setup.py egg_info\" failed with error code 1 in /tmp/pip-l5WfLH-build/\r\n```\n", "before_files": [{"content": "# Copyright (C) 2015-2016 Regents of the University of California\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"This script is a template for src/toil/version.py. Running it without arguments echoes all\nglobals, i.e. module attributes. Constant assignments will be echoed verbatim while callables\nwill be invoked and their result echoed as an assignment using the function name as the left-hand\nside and the return value of the function as right-hand side. To prevent a module attribute from\nbeing echoed, start or end the attribute name with an underscore. To print the value of a single\nsymbol, pass the name of that attribute to the script as a command line argument. You can also\nimport the expand_ function and invoke it directly with either no or exactly one argument.\"\"\"\n\n# Note to maintainers:\n#\n# - don't import at module level unless you want the imported value to be included in the output\n# - only import from the Python standard run-time library (you can't have any dependencies)\n\nbaseVersion = '3.15.0a1'\n\ncgcloudVersion = '1.6.0a1.dev393'\n\n\ndef version():\n \"\"\"\n A version identifier that includes the full-legth commit SHA1 and an optional suffix to\n indicate that the working copy is dirty.\n \"\"\"\n return _version()\n\n\ndef shortVersion():\n \"\"\"\n A version identifier that includes the abbreviated commit SHA1 and an optional suffix to\n indicate that the working copy is dirty.\n \"\"\"\n return _version(shorten=True)\n\n\ndef _version(shorten=False):\n return '-'.join(filter(None, [distVersion(),\n currentCommit()[:7 if shorten else None],\n ('dirty' if dirty() else None)]))\n\n\ndef distVersion():\n \"\"\"\n The distribution version identifying a published release on PyPI.\n \"\"\"\n from pkg_resources import parse_version\n build_number = buildNumber()\n parsedBaseVersion = parse_version(baseVersion)\n if isinstance(parsedBaseVersion, tuple):\n raise RuntimeError(\"Setuptools version 8.0 or newer required. Update by running \"\n \"'pip install setuptools --upgrade'\")\n\n if build_number is not None and parsedBaseVersion.is_prerelease:\n return baseVersion + '.dev' + build_number\n else:\n return baseVersion\n\n\ndef dockerTag():\n \"\"\"\n The primary tag of the Docker image for the appliance. This uniquely identifies the appliance\n image.\n \"\"\"\n return version()\n\n\ndef dockerShortTag():\n \"\"\"\n A secondary, shortened form of :func:`dockerTag` with which to tag the appliance image for\n convenience.\n \"\"\"\n return shortVersion()\n\n\ndef dockerMinimalTag():\n \"\"\"\n A minimal tag with which to tag the appliance image for convenience. Does not include\n information about the git commit or working copy dirtyness.\n \"\"\"\n return distVersion()\n\n\ndockerRegistry = 'quay.io/ucsc_cgl'\n\ndockerName = 'toil'\n\n\ndef buildNumber():\n \"\"\"\n The Jenkins build number, if defined, else None.\n \"\"\"\n import os\n return os.getenv('BUILD_NUMBER')\n\n\ndef currentCommit():\n from subprocess import check_output\n return check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]\n\n\ndef dirty():\n from subprocess import call\n return 0 != call('(git diff --exit-code '\n '&& git diff --cached --exit-code) > /dev/null', shell=True)\n\n\ndef expand_(name=None):\n variables = {k: v for k, v in globals().items()\n if not k.startswith('_') and not k.endswith('_')}\n\n def resolve(k):\n v = variables[k]\n if callable(v):\n v = v()\n return v\n\n if name is None:\n return ''.join(\"%s = %s\\n\" % (k, repr(resolve(k))) for k, v in variables.items())\n else:\n return resolve(name)\n\n\ndef _main():\n import sys\n sys.stdout.write(expand_(*sys.argv[1:]))\n\n\nif __name__ == '__main__':\n _main()\n", "path": "version_template.py"}], "after_files": [{"content": "# Copyright (C) 2015-2016 Regents of the University of California\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"This script is a template for src/toil/version.py. Running it without arguments echoes all\nglobals, i.e. module attributes. Constant assignments will be echoed verbatim while callables\nwill be invoked and their result echoed as an assignment using the function name as the left-hand\nside and the return value of the function as right-hand side. To prevent a module attribute from\nbeing echoed, start or end the attribute name with an underscore. To print the value of a single\nsymbol, pass the name of that attribute to the script as a command line argument. You can also\nimport the expand_ function and invoke it directly with either no or exactly one argument.\"\"\"\n\n# Note to maintainers:\n#\n# - don't import at module level unless you want the imported value to be included in the output\n# - only import from the Python standard run-time library (you can't have any dependencies)\n\nbaseVersion = '3.15.0a1'\n\ncgcloudVersion = '1.6.0a1.dev393'\n\n\ndef version():\n \"\"\"\n A version identifier that includes the full-legth commit SHA1 and an optional suffix to\n indicate that the working copy is dirty.\n \"\"\"\n return _version()\n\n\ndef shortVersion():\n \"\"\"\n A version identifier that includes the abbreviated commit SHA1 and an optional suffix to\n indicate that the working copy is dirty.\n \"\"\"\n return _version(shorten=True)\n\n\ndef _version(shorten=False):\n return '-'.join(filter(None, [distVersion(),\n currentCommit()[:7 if shorten else None],\n ('dirty' if dirty() else None)]))\n\n\ndef distVersion():\n \"\"\"\n The distribution version identifying a published release on PyPI.\n \"\"\"\n from pkg_resources import parse_version\n build_number = buildNumber()\n parsedBaseVersion = parse_version(baseVersion)\n if isinstance(parsedBaseVersion, tuple):\n raise RuntimeError(\"Setuptools version 8.0 or newer required. Update by running \"\n \"'pip install setuptools --upgrade'\")\n\n if build_number is not None and parsedBaseVersion.is_prerelease:\n return baseVersion + '.dev' + build_number\n else:\n return baseVersion\n\n\ndef dockerTag():\n \"\"\"\n The primary tag of the Docker image for the appliance. This uniquely identifies the appliance\n image.\n \"\"\"\n return version()\n\n\ndef dockerShortTag():\n \"\"\"\n A secondary, shortened form of :func:`dockerTag` with which to tag the appliance image for\n convenience.\n \"\"\"\n return shortVersion()\n\n\ndef dockerMinimalTag():\n \"\"\"\n A minimal tag with which to tag the appliance image for convenience. Does not include\n information about the git commit or working copy dirtyness.\n \"\"\"\n return distVersion()\n\n\ndockerRegistry = 'quay.io/ucsc_cgl'\n\ndockerName = 'toil'\n\n\ndef buildNumber():\n \"\"\"\n The Jenkins build number, if defined, else None.\n \"\"\"\n import os\n return os.getenv('BUILD_NUMBER')\n\n\ndef currentCommit():\n from subprocess import check_output\n try:\n output = check_output('git log --pretty=oneline -n 1 -- $(pwd)', shell=True).split()[0]\n except:\n # Return this we are not in a git environment.\n return '000'\n return output\n\n\ndef dirty():\n from subprocess import call\n try:\n return 0 != call('(git diff --exit-code '\n '&& git diff --cached --exit-code) > /dev/null', shell=True)\n except:\n return False # In case the git call fails.\n\n\ndef expand_(name=None):\n variables = {k: v for k, v in globals().items()\n if not k.startswith('_') and not k.endswith('_')}\n\n def resolve(k):\n v = variables[k]\n if callable(v):\n v = v()\n return v\n\n if name is None:\n return ''.join(\"%s = %s\\n\" % (k, repr(resolve(k))) for k, v in variables.items())\n else:\n return resolve(name)\n\n\ndef _main():\n import sys\n sys.stdout.write(expand_(*sys.argv[1:]))\n\n\nif __name__ == '__main__':\n _main()\n", "path": "version_template.py"}]}
| 2,144 | 248 |
gh_patches_debug_15911
|
rasdani/github-patches
|
git_diff
|
pallets__click-2599
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
deprecate `__version__` attribute
The `__version__` attribute is an old pattern from early in Python packaging. Setuptools eventually made it easier to use the pattern by allowing reading the value from the attribute at build time, and some other build backends have done the same.
However, there's no reason to expose this directly in code anymore. It's usually easier to use feature detection (`hasattr`, `try/except`) instead. `importlib.metadata.version("click")` can be used to get the version at runtime in a standard way, if it's really needed.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/click/__init__.py`
Content:
```
1 """
2 Click is a simple Python module inspired by the stdlib optparse to make
3 writing command line scripts fun. Unlike other modules, it's based
4 around a simple API that does not come with too much magic and is
5 composable.
6 """
7 from __future__ import annotations
8
9 from .core import Argument as Argument
10 from .core import Command as Command
11 from .core import CommandCollection as CommandCollection
12 from .core import Context as Context
13 from .core import Group as Group
14 from .core import Option as Option
15 from .core import Parameter as Parameter
16 from .decorators import argument as argument
17 from .decorators import command as command
18 from .decorators import confirmation_option as confirmation_option
19 from .decorators import group as group
20 from .decorators import help_option as help_option
21 from .decorators import make_pass_decorator as make_pass_decorator
22 from .decorators import option as option
23 from .decorators import pass_context as pass_context
24 from .decorators import pass_obj as pass_obj
25 from .decorators import password_option as password_option
26 from .decorators import version_option as version_option
27 from .exceptions import Abort as Abort
28 from .exceptions import BadArgumentUsage as BadArgumentUsage
29 from .exceptions import BadOptionUsage as BadOptionUsage
30 from .exceptions import BadParameter as BadParameter
31 from .exceptions import ClickException as ClickException
32 from .exceptions import FileError as FileError
33 from .exceptions import MissingParameter as MissingParameter
34 from .exceptions import NoSuchOption as NoSuchOption
35 from .exceptions import UsageError as UsageError
36 from .formatting import HelpFormatter as HelpFormatter
37 from .formatting import wrap_text as wrap_text
38 from .globals import get_current_context as get_current_context
39 from .termui import clear as clear
40 from .termui import confirm as confirm
41 from .termui import echo_via_pager as echo_via_pager
42 from .termui import edit as edit
43 from .termui import getchar as getchar
44 from .termui import launch as launch
45 from .termui import pause as pause
46 from .termui import progressbar as progressbar
47 from .termui import prompt as prompt
48 from .termui import secho as secho
49 from .termui import style as style
50 from .termui import unstyle as unstyle
51 from .types import BOOL as BOOL
52 from .types import Choice as Choice
53 from .types import DateTime as DateTime
54 from .types import File as File
55 from .types import FLOAT as FLOAT
56 from .types import FloatRange as FloatRange
57 from .types import INT as INT
58 from .types import IntRange as IntRange
59 from .types import ParamType as ParamType
60 from .types import Path as Path
61 from .types import STRING as STRING
62 from .types import Tuple as Tuple
63 from .types import UNPROCESSED as UNPROCESSED
64 from .types import UUID as UUID
65 from .utils import echo as echo
66 from .utils import format_filename as format_filename
67 from .utils import get_app_dir as get_app_dir
68 from .utils import get_binary_stream as get_binary_stream
69 from .utils import get_text_stream as get_text_stream
70 from .utils import open_file as open_file
71
72 __version__ = "8.2.0.dev0"
73
74
75 def __getattr__(name: str) -> object:
76 import warnings
77
78 if name == "BaseCommand":
79 from .core import _BaseCommand
80
81 warnings.warn(
82 "'BaseCommand' is deprecated and will be removed in Click 9.0. Use"
83 " 'Command' instead.",
84 DeprecationWarning,
85 stacklevel=2,
86 )
87 return _BaseCommand
88
89 if name == "MultiCommand":
90 from .core import _MultiCommand
91
92 warnings.warn(
93 "'MultiCommand' is deprecated and will be removed in Click 9.0. Use"
94 " 'Group' instead.",
95 DeprecationWarning,
96 stacklevel=2,
97 )
98 return _MultiCommand
99
100 if name == "OptionParser":
101 from .parser import _OptionParser
102
103 warnings.warn(
104 "'OptionParser' is deprecated and will be removed in Click 9.0. The"
105 " old parser is available in 'optparse'.",
106 DeprecationWarning,
107 stacklevel=2,
108 )
109 return _OptionParser
110
111 raise AttributeError(name)
112
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/click/__init__.py b/src/click/__init__.py
--- a/src/click/__init__.py
+++ b/src/click/__init__.py
@@ -69,8 +69,6 @@
from .utils import get_text_stream as get_text_stream
from .utils import open_file as open_file
-__version__ = "8.2.0.dev0"
-
def __getattr__(name: str) -> object:
import warnings
@@ -108,4 +106,17 @@
)
return _OptionParser
+ if name == "__version__":
+ import importlib.metadata
+ import warnings
+
+ warnings.warn(
+ "The '__version__' attribute is deprecated and will be removed in"
+ " Click 9.1. Use feature detection or"
+ " 'importlib.metadata.version(\"click\")' instead.",
+ DeprecationWarning,
+ stacklevel=2,
+ )
+ return importlib.metadata.version("click")
+
raise AttributeError(name)
|
{"golden_diff": "diff --git a/src/click/__init__.py b/src/click/__init__.py\n--- a/src/click/__init__.py\n+++ b/src/click/__init__.py\n@@ -69,8 +69,6 @@\n from .utils import get_text_stream as get_text_stream\n from .utils import open_file as open_file\n \n-__version__ = \"8.2.0.dev0\"\n-\n \n def __getattr__(name: str) -> object:\n import warnings\n@@ -108,4 +106,17 @@\n )\n return _OptionParser\n \n+ if name == \"__version__\":\n+ import importlib.metadata\n+ import warnings\n+\n+ warnings.warn(\n+ \"The '__version__' attribute is deprecated and will be removed in\"\n+ \" Click 9.1. Use feature detection or\"\n+ \" 'importlib.metadata.version(\\\"click\\\")' instead.\",\n+ DeprecationWarning,\n+ stacklevel=2,\n+ )\n+ return importlib.metadata.version(\"click\")\n+\n raise AttributeError(name)\n", "issue": "deprecate `__version__` attribute\nThe `__version__` attribute is an old pattern from early in Python packaging. Setuptools eventually made it easier to use the pattern by allowing reading the value from the attribute at build time, and some other build backends have done the same.\r\n\r\nHowever, there's no reason to expose this directly in code anymore. It's usually easier to use feature detection (`hasattr`, `try/except`) instead. `importlib.metadata.version(\"click\")` can be used to get the version at runtime in a standard way, if it's really needed.\n", "before_files": [{"content": "\"\"\"\nClick is a simple Python module inspired by the stdlib optparse to make\nwriting command line scripts fun. Unlike other modules, it's based\naround a simple API that does not come with too much magic and is\ncomposable.\n\"\"\"\nfrom __future__ import annotations\n\nfrom .core import Argument as Argument\nfrom .core import Command as Command\nfrom .core import CommandCollection as CommandCollection\nfrom .core import Context as Context\nfrom .core import Group as Group\nfrom .core import Option as Option\nfrom .core import Parameter as Parameter\nfrom .decorators import argument as argument\nfrom .decorators import command as command\nfrom .decorators import confirmation_option as confirmation_option\nfrom .decorators import group as group\nfrom .decorators import help_option as help_option\nfrom .decorators import make_pass_decorator as make_pass_decorator\nfrom .decorators import option as option\nfrom .decorators import pass_context as pass_context\nfrom .decorators import pass_obj as pass_obj\nfrom .decorators import password_option as password_option\nfrom .decorators import version_option as version_option\nfrom .exceptions import Abort as Abort\nfrom .exceptions import BadArgumentUsage as BadArgumentUsage\nfrom .exceptions import BadOptionUsage as BadOptionUsage\nfrom .exceptions import BadParameter as BadParameter\nfrom .exceptions import ClickException as ClickException\nfrom .exceptions import FileError as FileError\nfrom .exceptions import MissingParameter as MissingParameter\nfrom .exceptions import NoSuchOption as NoSuchOption\nfrom .exceptions import UsageError as UsageError\nfrom .formatting import HelpFormatter as HelpFormatter\nfrom .formatting import wrap_text as wrap_text\nfrom .globals import get_current_context as get_current_context\nfrom .termui import clear as clear\nfrom .termui import confirm as confirm\nfrom .termui import echo_via_pager as echo_via_pager\nfrom .termui import edit as edit\nfrom .termui import getchar as getchar\nfrom .termui import launch as launch\nfrom .termui import pause as pause\nfrom .termui import progressbar as progressbar\nfrom .termui import prompt as prompt\nfrom .termui import secho as secho\nfrom .termui import style as style\nfrom .termui import unstyle as unstyle\nfrom .types import BOOL as BOOL\nfrom .types import Choice as Choice\nfrom .types import DateTime as DateTime\nfrom .types import File as File\nfrom .types import FLOAT as FLOAT\nfrom .types import FloatRange as FloatRange\nfrom .types import INT as INT\nfrom .types import IntRange as IntRange\nfrom .types import ParamType as ParamType\nfrom .types import Path as Path\nfrom .types import STRING as STRING\nfrom .types import Tuple as Tuple\nfrom .types import UNPROCESSED as UNPROCESSED\nfrom .types import UUID as UUID\nfrom .utils import echo as echo\nfrom .utils import format_filename as format_filename\nfrom .utils import get_app_dir as get_app_dir\nfrom .utils import get_binary_stream as get_binary_stream\nfrom .utils import get_text_stream as get_text_stream\nfrom .utils import open_file as open_file\n\n__version__ = \"8.2.0.dev0\"\n\n\ndef __getattr__(name: str) -> object:\n import warnings\n\n if name == \"BaseCommand\":\n from .core import _BaseCommand\n\n warnings.warn(\n \"'BaseCommand' is deprecated and will be removed in Click 9.0. Use\"\n \" 'Command' instead.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _BaseCommand\n\n if name == \"MultiCommand\":\n from .core import _MultiCommand\n\n warnings.warn(\n \"'MultiCommand' is deprecated and will be removed in Click 9.0. Use\"\n \" 'Group' instead.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _MultiCommand\n\n if name == \"OptionParser\":\n from .parser import _OptionParser\n\n warnings.warn(\n \"'OptionParser' is deprecated and will be removed in Click 9.0. The\"\n \" old parser is available in 'optparse'.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _OptionParser\n\n raise AttributeError(name)\n", "path": "src/click/__init__.py"}], "after_files": [{"content": "\"\"\"\nClick is a simple Python module inspired by the stdlib optparse to make\nwriting command line scripts fun. Unlike other modules, it's based\naround a simple API that does not come with too much magic and is\ncomposable.\n\"\"\"\nfrom __future__ import annotations\n\nfrom .core import Argument as Argument\nfrom .core import Command as Command\nfrom .core import CommandCollection as CommandCollection\nfrom .core import Context as Context\nfrom .core import Group as Group\nfrom .core import Option as Option\nfrom .core import Parameter as Parameter\nfrom .decorators import argument as argument\nfrom .decorators import command as command\nfrom .decorators import confirmation_option as confirmation_option\nfrom .decorators import group as group\nfrom .decorators import help_option as help_option\nfrom .decorators import make_pass_decorator as make_pass_decorator\nfrom .decorators import option as option\nfrom .decorators import pass_context as pass_context\nfrom .decorators import pass_obj as pass_obj\nfrom .decorators import password_option as password_option\nfrom .decorators import version_option as version_option\nfrom .exceptions import Abort as Abort\nfrom .exceptions import BadArgumentUsage as BadArgumentUsage\nfrom .exceptions import BadOptionUsage as BadOptionUsage\nfrom .exceptions import BadParameter as BadParameter\nfrom .exceptions import ClickException as ClickException\nfrom .exceptions import FileError as FileError\nfrom .exceptions import MissingParameter as MissingParameter\nfrom .exceptions import NoSuchOption as NoSuchOption\nfrom .exceptions import UsageError as UsageError\nfrom .formatting import HelpFormatter as HelpFormatter\nfrom .formatting import wrap_text as wrap_text\nfrom .globals import get_current_context as get_current_context\nfrom .termui import clear as clear\nfrom .termui import confirm as confirm\nfrom .termui import echo_via_pager as echo_via_pager\nfrom .termui import edit as edit\nfrom .termui import getchar as getchar\nfrom .termui import launch as launch\nfrom .termui import pause as pause\nfrom .termui import progressbar as progressbar\nfrom .termui import prompt as prompt\nfrom .termui import secho as secho\nfrom .termui import style as style\nfrom .termui import unstyle as unstyle\nfrom .types import BOOL as BOOL\nfrom .types import Choice as Choice\nfrom .types import DateTime as DateTime\nfrom .types import File as File\nfrom .types import FLOAT as FLOAT\nfrom .types import FloatRange as FloatRange\nfrom .types import INT as INT\nfrom .types import IntRange as IntRange\nfrom .types import ParamType as ParamType\nfrom .types import Path as Path\nfrom .types import STRING as STRING\nfrom .types import Tuple as Tuple\nfrom .types import UNPROCESSED as UNPROCESSED\nfrom .types import UUID as UUID\nfrom .utils import echo as echo\nfrom .utils import format_filename as format_filename\nfrom .utils import get_app_dir as get_app_dir\nfrom .utils import get_binary_stream as get_binary_stream\nfrom .utils import get_text_stream as get_text_stream\nfrom .utils import open_file as open_file\n\n\ndef __getattr__(name: str) -> object:\n import warnings\n\n if name == \"BaseCommand\":\n from .core import _BaseCommand\n\n warnings.warn(\n \"'BaseCommand' is deprecated and will be removed in Click 9.0. Use\"\n \" 'Command' instead.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _BaseCommand\n\n if name == \"MultiCommand\":\n from .core import _MultiCommand\n\n warnings.warn(\n \"'MultiCommand' is deprecated and will be removed in Click 9.0. Use\"\n \" 'Group' instead.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _MultiCommand\n\n if name == \"OptionParser\":\n from .parser import _OptionParser\n\n warnings.warn(\n \"'OptionParser' is deprecated and will be removed in Click 9.0. The\"\n \" old parser is available in 'optparse'.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return _OptionParser\n\n if name == \"__version__\":\n import importlib.metadata\n import warnings\n\n warnings.warn(\n \"The '__version__' attribute is deprecated and will be removed in\"\n \" Click 9.1. Use feature detection or\"\n \" 'importlib.metadata.version(\\\"click\\\")' instead.\",\n DeprecationWarning,\n stacklevel=2,\n )\n return importlib.metadata.version(\"click\")\n\n raise AttributeError(name)\n", "path": "src/click/__init__.py"}]}
| 1,518 | 232 |
gh_patches_debug_289
|
rasdani/github-patches
|
git_diff
|
iterative__dvc-2282
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
test: s3: use moto to test multipart objects
Currently, we are unable to use it because of [this bug](https://github.com/spulec/moto/issues/2154). When it is fixed, we should switch to it from using actual s3 for unit testing. Related to https://github.com/iterative/dvc/pull/1867
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 from setuptools import setup, find_packages
2 from setuptools.command.build_py import build_py as _build_py
3 import os
4 import sys
5
6
7 # https://packaging.python.org/guides/single-sourcing-package-version/
8 pkg_dir = os.path.dirname(__file__)
9
10 # This will define __version__ implicitly
11 with open(os.path.join(pkg_dir, "dvc", "version.py")) as fobj:
12 exec(fobj.read())
13
14 version = __version__ # noqa: F821
15
16
17 # To achieve consistency between the build version and the one provided
18 # by your package during runtime, you need to **pin** the build version.
19 #
20 # This custom class will replace the version.py module with a **static**
21 # `__version__` that your package can read at runtime, assuring consistancy.
22 #
23 # References:
24 # - https://docs.python.org/3.7/distutils/extending.html
25 # - https://github.com/python/mypy
26 class build_py(_build_py):
27 def pin_version(self):
28 path = os.path.join(self.build_lib, "dvc")
29 self.mkpath(path)
30 with open(os.path.join(path, "version.py"), "w") as fobj:
31 fobj.write("# AUTOGENERATED at build time by setup.py\n")
32 fobj.write('__version__ = "{}"\n'.format(version))
33
34 def run(self):
35 self.execute(self.pin_version, ())
36 _build_py.run(self)
37
38
39 install_requires = [
40 "ply>=3.9", # See https://github.com/pyinstaller/pyinstaller/issues/1945
41 "configparser>=3.5.0",
42 "zc.lockfile>=1.2.1",
43 "future>=0.16.0",
44 "colorama>=0.3.9",
45 "configobj>=5.0.6",
46 "gitpython>=2.1.8",
47 "setuptools>=34.0.0",
48 "nanotime>=0.5.2",
49 "pyasn1>=0.4.1",
50 "schema>=0.6.7",
51 "jsonpath-ng>=1.4.3",
52 "requests>=2.22.0",
53 "grandalf==0.6",
54 "asciimatics>=1.10.0",
55 "distro>=1.3.0",
56 "appdirs>=1.4.3",
57 "treelib>=1.5.5",
58 "inflect>=2.1.0",
59 "humanize>=0.5.1",
60 "ruamel.yaml>=0.15.91",
61 "psutil==5.6.2",
62 "funcy>=1.12",
63 "pathspec>=0.5.9",
64 "shortuuid>=0.5.0",
65 "win-unicode-console>=0.5; sys_platform == 'win32'",
66 ]
67
68 if sys.version_info[0] == 2:
69 install_requires.append("networkx>=2.1,<2.3")
70 else:
71 install_requires.append("networkx>=2.1")
72
73 # Extra dependencies for remote integrations
74 gs = ["google-cloud-storage==1.13.0"]
75 s3 = ["boto3==1.9.115"]
76 azure = ["azure-storage-blob==2.0.1"]
77 oss = ["oss2==2.6.1"]
78 ssh = ["paramiko>=2.5.0"]
79 all_remotes = gs + s3 + azure + ssh + oss
80
81 # Extra dependecies to run tests
82 tests_requirements = [
83 "PyInstaller==3.4",
84 "wheel>=0.31.1",
85 "pydot>=1.2.4",
86 # Test requirements:
87 "pytest>=4.6.0",
88 "pytest-timeout>=1.3.3",
89 "pytest-cov>=2.6.1",
90 "pytest-xdist>=1.26.1",
91 "pytest-mock>=1.10.4",
92 "flaky>=3.5.3",
93 "mock>=3.0.0",
94 "xmltodict>=0.11.0",
95 "awscli>=1.16.125",
96 "google-compute-engine",
97 "pywin32; sys_platform == 'win32'",
98 "Pygments", # required by collective.checkdocs,
99 "collective.checkdocs",
100 "flake8",
101 "flake8-docstrings",
102 "pydocstyle<4.0",
103 "jaraco.windows==3.9.2",
104 "mock-ssh-server>=0.5.0",
105 ]
106
107 if (sys.version_info) >= (3, 6):
108 tests_requirements.append("black==19.3b0")
109
110 setup(
111 name="dvc",
112 version=version,
113 description="Git for data scientists - manage your code and data together",
114 long_description=open("README.rst", "r").read(),
115 author="Dmitry Petrov",
116 author_email="[email protected]",
117 download_url="https://github.com/iterative/dvc",
118 license="Apache License 2.0",
119 install_requires=install_requires,
120 extras_require={
121 "all": all_remotes,
122 "gs": gs,
123 "s3": s3,
124 "azure": azure,
125 "oss": oss,
126 "ssh": ssh,
127 # NOTE: https://github.com/inveniosoftware/troubleshooting/issues/1
128 ":python_version=='2.7'": ["futures", "pathlib2"],
129 "tests": tests_requirements,
130 },
131 keywords="data science, data version control, machine learning",
132 python_requires=">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*",
133 classifiers=[
134 "Development Status :: 4 - Beta",
135 "Programming Language :: Python :: 2",
136 "Programming Language :: Python :: 2.7",
137 "Programming Language :: Python :: 3",
138 "Programming Language :: Python :: 3.5",
139 "Programming Language :: Python :: 3.6",
140 "Programming Language :: Python :: 3.7",
141 ],
142 packages=find_packages(exclude=["tests"]),
143 include_package_data=True,
144 url="http://dataversioncontrol.com",
145 entry_points={"console_scripts": ["dvc = dvc.main:main"]},
146 cmdclass={"build_py": build_py},
147 zip_safe=False,
148 )
149
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -102,6 +102,7 @@
"pydocstyle<4.0",
"jaraco.windows==3.9.2",
"mock-ssh-server>=0.5.0",
+ "moto",
]
if (sys.version_info) >= (3, 6):
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -102,6 +102,7 @@\n \"pydocstyle<4.0\",\n \"jaraco.windows==3.9.2\",\n \"mock-ssh-server>=0.5.0\",\n+ \"moto\",\n ]\n \n if (sys.version_info) >= (3, 6):\n", "issue": "test: s3: use moto to test multipart objects\nCurrently, we are unable to use it because of [this bug](https://github.com/spulec/moto/issues/2154). When it is fixed, we should switch to it from using actual s3 for unit testing. Related to https://github.com/iterative/dvc/pull/1867\r\n\n", "before_files": [{"content": "from setuptools import setup, find_packages\nfrom setuptools.command.build_py import build_py as _build_py\nimport os\nimport sys\n\n\n# https://packaging.python.org/guides/single-sourcing-package-version/\npkg_dir = os.path.dirname(__file__)\n\n# This will define __version__ implicitly\nwith open(os.path.join(pkg_dir, \"dvc\", \"version.py\")) as fobj:\n exec(fobj.read())\n\nversion = __version__ # noqa: F821\n\n\n# To achieve consistency between the build version and the one provided\n# by your package during runtime, you need to **pin** the build version.\n#\n# This custom class will replace the version.py module with a **static**\n# `__version__` that your package can read at runtime, assuring consistancy.\n#\n# References:\n# - https://docs.python.org/3.7/distutils/extending.html\n# - https://github.com/python/mypy\nclass build_py(_build_py):\n def pin_version(self):\n path = os.path.join(self.build_lib, \"dvc\")\n self.mkpath(path)\n with open(os.path.join(path, \"version.py\"), \"w\") as fobj:\n fobj.write(\"# AUTOGENERATED at build time by setup.py\\n\")\n fobj.write('__version__ = \"{}\"\\n'.format(version))\n\n def run(self):\n self.execute(self.pin_version, ())\n _build_py.run(self)\n\n\ninstall_requires = [\n \"ply>=3.9\", # See https://github.com/pyinstaller/pyinstaller/issues/1945\n \"configparser>=3.5.0\",\n \"zc.lockfile>=1.2.1\",\n \"future>=0.16.0\",\n \"colorama>=0.3.9\",\n \"configobj>=5.0.6\",\n \"gitpython>=2.1.8\",\n \"setuptools>=34.0.0\",\n \"nanotime>=0.5.2\",\n \"pyasn1>=0.4.1\",\n \"schema>=0.6.7\",\n \"jsonpath-ng>=1.4.3\",\n \"requests>=2.22.0\",\n \"grandalf==0.6\",\n \"asciimatics>=1.10.0\",\n \"distro>=1.3.0\",\n \"appdirs>=1.4.3\",\n \"treelib>=1.5.5\",\n \"inflect>=2.1.0\",\n \"humanize>=0.5.1\",\n \"ruamel.yaml>=0.15.91\",\n \"psutil==5.6.2\",\n \"funcy>=1.12\",\n \"pathspec>=0.5.9\",\n \"shortuuid>=0.5.0\",\n \"win-unicode-console>=0.5; sys_platform == 'win32'\",\n]\n\nif sys.version_info[0] == 2:\n install_requires.append(\"networkx>=2.1,<2.3\")\nelse:\n install_requires.append(\"networkx>=2.1\")\n\n# Extra dependencies for remote integrations\ngs = [\"google-cloud-storage==1.13.0\"]\ns3 = [\"boto3==1.9.115\"]\nazure = [\"azure-storage-blob==2.0.1\"]\noss = [\"oss2==2.6.1\"]\nssh = [\"paramiko>=2.5.0\"]\nall_remotes = gs + s3 + azure + ssh + oss\n\n# Extra dependecies to run tests\ntests_requirements = [\n \"PyInstaller==3.4\",\n \"wheel>=0.31.1\",\n \"pydot>=1.2.4\",\n # Test requirements:\n \"pytest>=4.6.0\",\n \"pytest-timeout>=1.3.3\",\n \"pytest-cov>=2.6.1\",\n \"pytest-xdist>=1.26.1\",\n \"pytest-mock>=1.10.4\",\n \"flaky>=3.5.3\",\n \"mock>=3.0.0\",\n \"xmltodict>=0.11.0\",\n \"awscli>=1.16.125\",\n \"google-compute-engine\",\n \"pywin32; sys_platform == 'win32'\",\n \"Pygments\", # required by collective.checkdocs,\n \"collective.checkdocs\",\n \"flake8\",\n \"flake8-docstrings\",\n \"pydocstyle<4.0\",\n \"jaraco.windows==3.9.2\",\n \"mock-ssh-server>=0.5.0\",\n]\n\nif (sys.version_info) >= (3, 6):\n tests_requirements.append(\"black==19.3b0\")\n\nsetup(\n name=\"dvc\",\n version=version,\n description=\"Git for data scientists - manage your code and data together\",\n long_description=open(\"README.rst\", \"r\").read(),\n author=\"Dmitry Petrov\",\n author_email=\"[email protected]\",\n download_url=\"https://github.com/iterative/dvc\",\n license=\"Apache License 2.0\",\n install_requires=install_requires,\n extras_require={\n \"all\": all_remotes,\n \"gs\": gs,\n \"s3\": s3,\n \"azure\": azure,\n \"oss\": oss,\n \"ssh\": ssh,\n # NOTE: https://github.com/inveniosoftware/troubleshooting/issues/1\n \":python_version=='2.7'\": [\"futures\", \"pathlib2\"],\n \"tests\": tests_requirements,\n },\n keywords=\"data science, data version control, machine learning\",\n python_requires=\">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*\",\n classifiers=[\n \"Development Status :: 4 - Beta\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n ],\n packages=find_packages(exclude=[\"tests\"]),\n include_package_data=True,\n url=\"http://dataversioncontrol.com\",\n entry_points={\"console_scripts\": [\"dvc = dvc.main:main\"]},\n cmdclass={\"build_py\": build_py},\n zip_safe=False,\n)\n", "path": "setup.py"}], "after_files": [{"content": "from setuptools import setup, find_packages\nfrom setuptools.command.build_py import build_py as _build_py\nimport os\nimport sys\n\n\n# https://packaging.python.org/guides/single-sourcing-package-version/\npkg_dir = os.path.dirname(__file__)\n\n# This will define __version__ implicitly\nwith open(os.path.join(pkg_dir, \"dvc\", \"version.py\")) as fobj:\n exec(fobj.read())\n\nversion = __version__ # noqa: F821\n\n\n# To achieve consistency between the build version and the one provided\n# by your package during runtime, you need to **pin** the build version.\n#\n# This custom class will replace the version.py module with a **static**\n# `__version__` that your package can read at runtime, assuring consistancy.\n#\n# References:\n# - https://docs.python.org/3.7/distutils/extending.html\n# - https://github.com/python/mypy\nclass build_py(_build_py):\n def pin_version(self):\n path = os.path.join(self.build_lib, \"dvc\")\n self.mkpath(path)\n with open(os.path.join(path, \"version.py\"), \"w\") as fobj:\n fobj.write(\"# AUTOGENERATED at build time by setup.py\\n\")\n fobj.write('__version__ = \"{}\"\\n'.format(version))\n\n def run(self):\n self.execute(self.pin_version, ())\n _build_py.run(self)\n\n\ninstall_requires = [\n \"ply>=3.9\", # See https://github.com/pyinstaller/pyinstaller/issues/1945\n \"configparser>=3.5.0\",\n \"zc.lockfile>=1.2.1\",\n \"future>=0.16.0\",\n \"colorama>=0.3.9\",\n \"configobj>=5.0.6\",\n \"gitpython>=2.1.8\",\n \"setuptools>=34.0.0\",\n \"nanotime>=0.5.2\",\n \"pyasn1>=0.4.1\",\n \"schema>=0.6.7\",\n \"jsonpath-ng>=1.4.3\",\n \"requests>=2.22.0\",\n \"grandalf==0.6\",\n \"asciimatics>=1.10.0\",\n \"distro>=1.3.0\",\n \"appdirs>=1.4.3\",\n \"treelib>=1.5.5\",\n \"inflect>=2.1.0\",\n \"humanize>=0.5.1\",\n \"ruamel.yaml>=0.15.91\",\n \"psutil==5.6.2\",\n \"funcy>=1.12\",\n \"pathspec>=0.5.9\",\n \"shortuuid>=0.5.0\",\n \"win-unicode-console>=0.5; sys_platform == 'win32'\",\n]\n\nif sys.version_info[0] == 2:\n install_requires.append(\"networkx>=2.1,<2.3\")\nelse:\n install_requires.append(\"networkx>=2.1\")\n\n# Extra dependencies for remote integrations\ngs = [\"google-cloud-storage==1.13.0\"]\ns3 = [\"boto3==1.9.115\"]\nazure = [\"azure-storage-blob==2.0.1\"]\noss = [\"oss2==2.6.1\"]\nssh = [\"paramiko>=2.5.0\"]\nall_remotes = gs + s3 + azure + ssh + oss\n\n# Extra dependecies to run tests\ntests_requirements = [\n \"PyInstaller==3.4\",\n \"wheel>=0.31.1\",\n \"pydot>=1.2.4\",\n # Test requirements:\n \"pytest>=4.6.0\",\n \"pytest-timeout>=1.3.3\",\n \"pytest-cov>=2.6.1\",\n \"pytest-xdist>=1.26.1\",\n \"pytest-mock>=1.10.4\",\n \"flaky>=3.5.3\",\n \"mock>=3.0.0\",\n \"xmltodict>=0.11.0\",\n \"awscli>=1.16.125\",\n \"google-compute-engine\",\n \"pywin32; sys_platform == 'win32'\",\n \"Pygments\", # required by collective.checkdocs,\n \"collective.checkdocs\",\n \"flake8\",\n \"flake8-docstrings\",\n \"pydocstyle<4.0\",\n \"jaraco.windows==3.9.2\",\n \"mock-ssh-server>=0.5.0\",\n \"moto\",\n]\n\nif (sys.version_info) >= (3, 6):\n tests_requirements.append(\"black==19.3b0\")\n\nsetup(\n name=\"dvc\",\n version=version,\n description=\"Git for data scientists - manage your code and data together\",\n long_description=open(\"README.rst\", \"r\").read(),\n author=\"Dmitry Petrov\",\n author_email=\"[email protected]\",\n download_url=\"https://github.com/iterative/dvc\",\n license=\"Apache License 2.0\",\n install_requires=install_requires,\n extras_require={\n \"all\": all_remotes,\n \"gs\": gs,\n \"s3\": s3,\n \"azure\": azure,\n \"oss\": oss,\n \"ssh\": ssh,\n # NOTE: https://github.com/inveniosoftware/troubleshooting/issues/1\n \":python_version=='2.7'\": [\"futures\", \"pathlib2\"],\n \"tests\": tests_requirements,\n },\n keywords=\"data science, data version control, machine learning\",\n python_requires=\">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*, !=3.4.*\",\n classifiers=[\n \"Development Status :: 4 - Beta\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n ],\n packages=find_packages(exclude=[\"tests\"]),\n include_package_data=True,\n url=\"http://dataversioncontrol.com\",\n entry_points={\"console_scripts\": [\"dvc = dvc.main:main\"]},\n cmdclass={\"build_py\": build_py},\n zip_safe=False,\n)\n", "path": "setup.py"}]}
| 2,096 | 90 |
gh_patches_debug_1752
|
rasdani/github-patches
|
git_diff
|
spack__spack-4809
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
problem with xproto
the xproto could be compile properly but while installing, i come across the following problem
```
==> 'make' '-j8'
/Applications/Xcode.app/Contents/Developer/usr/bin/make all-recursive
Making all in specs
Making all in SIAddresses
make[3]: Nothing to be done for `all'.
make[3]: Nothing to be done for `all-am'.
make[2]: Nothing to be done for `all-am'.
==> 'make' '-j8' 'install'
Making install in specs
Making install in SIAddresses
make[3]: Nothing to be done for `install-exec-am'.
make[3]: Nothing to be done for `install-data-am'.
make[3]: Nothing to be done for `install-exec-am'.
/spack/var/spack/stage/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/xproto-7.0.31/install-sh -c -d '/spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto'
/spack/var/spack/stage/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/xproto-7.0.31/install-sh -c -d '/spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto'
mkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share: File exists
mkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc: File exists
mkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto: File exists
make[3]: *** [install-dist_shelfDATA] Error 1
make[3]: *** Waiting for unfinished jobs....
make[2]: *** [install-am] Error 2
make[1]: *** [install-recursive] Error 1
make: *** [install-recursive] Error 1
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `var/spack/repos/builtin/packages/xproto/package.py`
Content:
```
1 ##############################################################################
2 # Copyright (c) 2013-2016, Lawrence Livermore National Security, LLC.
3 # Produced at the Lawrence Livermore National Laboratory.
4 #
5 # This file is part of Spack.
6 # Created by Todd Gamblin, [email protected], All rights reserved.
7 # LLNL-CODE-647188
8 #
9 # For details, see https://github.com/llnl/spack
10 # Please also see the NOTICE and LICENSE files for our notice and the LGPL.
11 #
12 # This program is free software; you can redistribute it and/or modify
13 # it under the terms of the GNU Lesser General Public License (as
14 # published by the Free Software Foundation) version 2.1, February 1999.
15 #
16 # This program is distributed in the hope that it will be useful, but
17 # WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF
18 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and
19 # conditions of the GNU Lesser General Public License for more details.
20 #
21 # You should have received a copy of the GNU Lesser General Public
22 # License along with this program; if not, write to the Free Software
23 # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
24 ##############################################################################
25 from spack import *
26
27
28 class Xproto(AutotoolsPackage):
29 """X Window System Core Protocol.
30
31 This package provides the headers and specification documents defining
32 the X Window System Core Protocol, Version 11.
33
34 It also includes a number of headers that aren't purely protocol related,
35 but are depended upon by many other X Window System packages to provide
36 common definitions and porting layer."""
37
38 homepage = "http://cgit.freedesktop.org/xorg/proto/x11proto"
39 url = "https://www.x.org/archive/individual/proto/xproto-7.0.31.tar.gz"
40
41 version('7.0.31', '04b925bf9e472c80f9212615cd684f1e')
42 version('7.0.29', '16a78dd2c5ad73011105c96235f6a0af')
43
44 depends_on('[email protected]:', type='build')
45 depends_on('util-macros', type='build')
46
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/var/spack/repos/builtin/packages/xproto/package.py b/var/spack/repos/builtin/packages/xproto/package.py
--- a/var/spack/repos/builtin/packages/xproto/package.py
+++ b/var/spack/repos/builtin/packages/xproto/package.py
@@ -43,3 +43,8 @@
depends_on('[email protected]:', type='build')
depends_on('util-macros', type='build')
+
+ def install(self, spec, prefix):
+ # Installation fails in parallel
+ # See https://github.com/LLNL/spack/issues/4805
+ make('install', parallel=False)
|
{"golden_diff": "diff --git a/var/spack/repos/builtin/packages/xproto/package.py b/var/spack/repos/builtin/packages/xproto/package.py\n--- a/var/spack/repos/builtin/packages/xproto/package.py\n+++ b/var/spack/repos/builtin/packages/xproto/package.py\n@@ -43,3 +43,8 @@\n \n depends_on('[email protected]:', type='build')\n depends_on('util-macros', type='build')\n+\n+ def install(self, spec, prefix):\n+ # Installation fails in parallel\n+ # See https://github.com/LLNL/spack/issues/4805\n+ make('install', parallel=False)\n", "issue": "problem with xproto\nthe xproto could be compile properly but while installing, i come across the following problem\r\n\r\n```\r\n==> 'make' '-j8'\r\n/Applications/Xcode.app/Contents/Developer/usr/bin/make all-recursive\r\nMaking all in specs\r\nMaking all in SIAddresses\r\nmake[3]: Nothing to be done for `all'.\r\nmake[3]: Nothing to be done for `all-am'.\r\nmake[2]: Nothing to be done for `all-am'.\r\n==> 'make' '-j8' 'install'\r\nMaking install in specs\r\nMaking install in SIAddresses\r\nmake[3]: Nothing to be done for `install-exec-am'.\r\nmake[3]: Nothing to be done for `install-data-am'.\r\nmake[3]: Nothing to be done for `install-exec-am'.\r\n /spack/var/spack/stage/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/xproto-7.0.31/install-sh -c -d '/spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto'\r\n /spack/var/spack/stage/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/xproto-7.0.31/install-sh -c -d '/spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto'\r\nmkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share: File exists\r\nmkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc: File exists\r\nmkdir: /spack/opt/spack/darwin-sierra-x86_64/clang-3.9.0/xproto-7.0.31-rgclwhmbundqudnhjaavmnvuzceqlgsh/share/doc/xproto: File exists\r\nmake[3]: *** [install-dist_shelfDATA] Error 1\r\nmake[3]: *** Waiting for unfinished jobs....\r\nmake[2]: *** [install-am] Error 2\r\nmake[1]: *** [install-recursive] Error 1\r\nmake: *** [install-recursive] Error 1\r\n```\n", "before_files": [{"content": "##############################################################################\n# Copyright (c) 2013-2016, Lawrence Livermore National Security, LLC.\n# Produced at the Lawrence Livermore National Laboratory.\n#\n# This file is part of Spack.\n# Created by Todd Gamblin, [email protected], All rights reserved.\n# LLNL-CODE-647188\n#\n# For details, see https://github.com/llnl/spack\n# Please also see the NOTICE and LICENSE files for our notice and the LGPL.\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU Lesser General Public License (as\n# published by the Free Software Foundation) version 2.1, February 1999.\n#\n# This program is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and\n# conditions of the GNU Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this program; if not, write to the Free Software\n# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA\n##############################################################################\nfrom spack import *\n\n\nclass Xproto(AutotoolsPackage):\n \"\"\"X Window System Core Protocol.\n\n This package provides the headers and specification documents defining\n the X Window System Core Protocol, Version 11.\n\n It also includes a number of headers that aren't purely protocol related,\n but are depended upon by many other X Window System packages to provide\n common definitions and porting layer.\"\"\"\n\n homepage = \"http://cgit.freedesktop.org/xorg/proto/x11proto\"\n url = \"https://www.x.org/archive/individual/proto/xproto-7.0.31.tar.gz\"\n\n version('7.0.31', '04b925bf9e472c80f9212615cd684f1e')\n version('7.0.29', '16a78dd2c5ad73011105c96235f6a0af')\n\n depends_on('[email protected]:', type='build')\n depends_on('util-macros', type='build')\n", "path": "var/spack/repos/builtin/packages/xproto/package.py"}], "after_files": [{"content": "##############################################################################\n# Copyright (c) 2013-2016, Lawrence Livermore National Security, LLC.\n# Produced at the Lawrence Livermore National Laboratory.\n#\n# This file is part of Spack.\n# Created by Todd Gamblin, [email protected], All rights reserved.\n# LLNL-CODE-647188\n#\n# For details, see https://github.com/llnl/spack\n# Please also see the NOTICE and LICENSE files for our notice and the LGPL.\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU Lesser General Public License (as\n# published by the Free Software Foundation) version 2.1, February 1999.\n#\n# This program is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and\n# conditions of the GNU Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this program; if not, write to the Free Software\n# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA\n##############################################################################\nfrom spack import *\n\n\nclass Xproto(AutotoolsPackage):\n \"\"\"X Window System Core Protocol.\n\n This package provides the headers and specification documents defining\n the X Window System Core Protocol, Version 11.\n\n It also includes a number of headers that aren't purely protocol related,\n but are depended upon by many other X Window System packages to provide\n common definitions and porting layer.\"\"\"\n\n homepage = \"http://cgit.freedesktop.org/xorg/proto/x11proto\"\n url = \"https://www.x.org/archive/individual/proto/xproto-7.0.31.tar.gz\"\n\n version('7.0.31', '04b925bf9e472c80f9212615cd684f1e')\n version('7.0.29', '16a78dd2c5ad73011105c96235f6a0af')\n\n depends_on('[email protected]:', type='build')\n depends_on('util-macros', type='build')\n\n def install(self, spec, prefix):\n # Installation fails in parallel\n # See https://github.com/LLNL/spack/issues/4805\n make('install', parallel=False)\n", "path": "var/spack/repos/builtin/packages/xproto/package.py"}]}
| 1,522 | 147 |
gh_patches_debug_932
|
rasdani/github-patches
|
git_diff
|
praw-dev__praw-1304
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Sphinx stops emitting warnings if it encounters only one
**Describe the bug**
<!-- A clear and concise description of what the bug is. --> When running pre_push, if Sphinx runs into an warning, it does does print any more. When there are lots of warnings, it takes a lot of time to re-run pre_push per warning
I recommend adding the command line argument `--keep-going`. This will cause it to print all warnings.
**System Info**
- PRAW Version: Latest
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pre_push.py`
Content:
```
1 #!/usr/bin/env python3
2 """Run static analysis on the project."""
3
4 import argparse
5 import sys
6 from os import path
7 from shutil import rmtree
8 from subprocess import CalledProcessError, check_call
9 from tempfile import mkdtemp
10
11 current_directory = path.abspath(path.join(__file__, ".."))
12
13
14 def do_process(args, shell=False):
15 """Run program provided by args.
16
17 Return True on success.
18
19 Output failed message on non-zero exit and return False.
20
21 Exit if command is not found.
22 """
23 print("Running: {}".format(" ".join(args)))
24 try:
25 check_call(args, shell=shell)
26 except CalledProcessError:
27 print("\nFailed: {}".format(" ".join(args)))
28 return False
29 except Exception as exc:
30 sys.stderr.write(str(exc) + "\n")
31 sys.exit(1)
32 return True
33
34
35 def run_static():
36 """Runs the static tests.
37
38 Returns a statuscode of 0 if everything ran correctly.
39 Otherwise, it will return statuscode 1
40 """
41 success = True
42 success &= do_process(
43 [
44 sys.executable,
45 path.join(current_directory, "tools", "static_word_checks.py"),
46 "--replace",
47 ]
48 )
49 success &= do_process(["black ."], shell=True)
50 success &= do_process(["flake8", "--exclude=.eggs,build,docs"])
51 success &= do_process(["pydocstyle", "praw"])
52 # success &= do_process(["pylint", "--rcfile=.pylintrc", "praw"])
53
54 tmp_dir = mkdtemp()
55 try:
56 success &= do_process(["sphinx-build", "-W", "docs", tmp_dir])
57 finally:
58 rmtree(tmp_dir)
59
60 return success
61
62
63 def run_unit():
64 """Runs the unit-tests.
65
66 Follows the behavior of the static tests,
67 where any failed tests cause pre_push.py to fail.
68 """
69 return do_process(
70 [sys.executable, path.join(current_directory, "setup.py"), "test"]
71 )
72
73
74 def main():
75 """Runs the main function.
76
77 usage: pre_push.py [-h] [-n] [-u] [-a]
78
79 Run static and/or unit-tests
80 """
81 parser = argparse.ArgumentParser(
82 description="Run static and/or unit-tests"
83 )
84 parser.add_argument(
85 "-n",
86 "--unstatic",
87 action="store_true",
88 help="Do not run static tests (black/flake8/pydocstyle/sphinx-build)",
89 default=False,
90 )
91 parser.add_argument(
92 "-u",
93 "--unit-tests",
94 "--unit",
95 action="store_true",
96 default=False,
97 help="Run the unit tests",
98 )
99 parser.add_argument(
100 "-a",
101 "--all",
102 action="store_true",
103 default=False,
104 help="Run all of the tests (static and unit). "
105 "Overrides the unstatic argument.",
106 )
107 args = parser.parse_args()
108 success = True
109 try:
110 if not args.unstatic or args.all:
111 success &= run_static()
112 if args.all or args.unit_tests:
113 success &= run_unit()
114 except KeyboardInterrupt:
115 return int(not False)
116 return int(not success)
117
118
119 if __name__ == "__main__":
120 exit_code = main()
121 print(
122 "\npre_push.py: Success!" if not exit_code else "\npre_push.py: Fail"
123 )
124 sys.exit(exit_code)
125
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pre_push.py b/pre_push.py
--- a/pre_push.py
+++ b/pre_push.py
@@ -53,7 +53,9 @@
tmp_dir = mkdtemp()
try:
- success &= do_process(["sphinx-build", "-W", "docs", tmp_dir])
+ success &= do_process(
+ ["sphinx-build", "-W", "--keep-going", "docs", tmp_dir]
+ )
finally:
rmtree(tmp_dir)
|
{"golden_diff": "diff --git a/pre_push.py b/pre_push.py\n--- a/pre_push.py\n+++ b/pre_push.py\n@@ -53,7 +53,9 @@\n \n tmp_dir = mkdtemp()\n try:\n- success &= do_process([\"sphinx-build\", \"-W\", \"docs\", tmp_dir])\n+ success &= do_process(\n+ [\"sphinx-build\", \"-W\", \"--keep-going\", \"docs\", tmp_dir]\n+ )\n finally:\n rmtree(tmp_dir)\n", "issue": "Sphinx stops emitting warnings if it encounters only one\n**Describe the bug**\r\n<!-- A clear and concise description of what the bug is. --> When running pre_push, if Sphinx runs into an warning, it does does print any more. When there are lots of warnings, it takes a lot of time to re-run pre_push per warning\r\n\r\nI recommend adding the command line argument `--keep-going`. This will cause it to print all warnings.\r\n\r\n**System Info**\r\n - PRAW Version: Latest\r\n\n", "before_files": [{"content": "#!/usr/bin/env python3\n\"\"\"Run static analysis on the project.\"\"\"\n\nimport argparse\nimport sys\nfrom os import path\nfrom shutil import rmtree\nfrom subprocess import CalledProcessError, check_call\nfrom tempfile import mkdtemp\n\ncurrent_directory = path.abspath(path.join(__file__, \"..\"))\n\n\ndef do_process(args, shell=False):\n \"\"\"Run program provided by args.\n\n Return True on success.\n\n Output failed message on non-zero exit and return False.\n\n Exit if command is not found.\n \"\"\"\n print(\"Running: {}\".format(\" \".join(args)))\n try:\n check_call(args, shell=shell)\n except CalledProcessError:\n print(\"\\nFailed: {}\".format(\" \".join(args)))\n return False\n except Exception as exc:\n sys.stderr.write(str(exc) + \"\\n\")\n sys.exit(1)\n return True\n\n\ndef run_static():\n \"\"\"Runs the static tests.\n\n Returns a statuscode of 0 if everything ran correctly.\n Otherwise, it will return statuscode 1\n \"\"\"\n success = True\n success &= do_process(\n [\n sys.executable,\n path.join(current_directory, \"tools\", \"static_word_checks.py\"),\n \"--replace\",\n ]\n )\n success &= do_process([\"black .\"], shell=True)\n success &= do_process([\"flake8\", \"--exclude=.eggs,build,docs\"])\n success &= do_process([\"pydocstyle\", \"praw\"])\n # success &= do_process([\"pylint\", \"--rcfile=.pylintrc\", \"praw\"])\n\n tmp_dir = mkdtemp()\n try:\n success &= do_process([\"sphinx-build\", \"-W\", \"docs\", tmp_dir])\n finally:\n rmtree(tmp_dir)\n\n return success\n\n\ndef run_unit():\n \"\"\"Runs the unit-tests.\n\n Follows the behavior of the static tests,\n where any failed tests cause pre_push.py to fail.\n \"\"\"\n return do_process(\n [sys.executable, path.join(current_directory, \"setup.py\"), \"test\"]\n )\n\n\ndef main():\n \"\"\"Runs the main function.\n\n usage: pre_push.py [-h] [-n] [-u] [-a]\n\n Run static and/or unit-tests\n \"\"\"\n parser = argparse.ArgumentParser(\n description=\"Run static and/or unit-tests\"\n )\n parser.add_argument(\n \"-n\",\n \"--unstatic\",\n action=\"store_true\",\n help=\"Do not run static tests (black/flake8/pydocstyle/sphinx-build)\",\n default=False,\n )\n parser.add_argument(\n \"-u\",\n \"--unit-tests\",\n \"--unit\",\n action=\"store_true\",\n default=False,\n help=\"Run the unit tests\",\n )\n parser.add_argument(\n \"-a\",\n \"--all\",\n action=\"store_true\",\n default=False,\n help=\"Run all of the tests (static and unit). \"\n \"Overrides the unstatic argument.\",\n )\n args = parser.parse_args()\n success = True\n try:\n if not args.unstatic or args.all:\n success &= run_static()\n if args.all or args.unit_tests:\n success &= run_unit()\n except KeyboardInterrupt:\n return int(not False)\n return int(not success)\n\n\nif __name__ == \"__main__\":\n exit_code = main()\n print(\n \"\\npre_push.py: Success!\" if not exit_code else \"\\npre_push.py: Fail\"\n )\n sys.exit(exit_code)\n", "path": "pre_push.py"}], "after_files": [{"content": "#!/usr/bin/env python3\n\"\"\"Run static analysis on the project.\"\"\"\n\nimport argparse\nimport sys\nfrom os import path\nfrom shutil import rmtree\nfrom subprocess import CalledProcessError, check_call\nfrom tempfile import mkdtemp\n\ncurrent_directory = path.abspath(path.join(__file__, \"..\"))\n\n\ndef do_process(args, shell=False):\n \"\"\"Run program provided by args.\n\n Return True on success.\n\n Output failed message on non-zero exit and return False.\n\n Exit if command is not found.\n \"\"\"\n print(\"Running: {}\".format(\" \".join(args)))\n try:\n check_call(args, shell=shell)\n except CalledProcessError:\n print(\"\\nFailed: {}\".format(\" \".join(args)))\n return False\n except Exception as exc:\n sys.stderr.write(str(exc) + \"\\n\")\n sys.exit(1)\n return True\n\n\ndef run_static():\n \"\"\"Runs the static tests.\n\n Returns a statuscode of 0 if everything ran correctly.\n Otherwise, it will return statuscode 1\n \"\"\"\n success = True\n success &= do_process(\n [\n sys.executable,\n path.join(current_directory, \"tools\", \"static_word_checks.py\"),\n \"--replace\",\n ]\n )\n success &= do_process([\"black .\"], shell=True)\n success &= do_process([\"flake8\", \"--exclude=.eggs,build,docs\"])\n success &= do_process([\"pydocstyle\", \"praw\"])\n # success &= do_process([\"pylint\", \"--rcfile=.pylintrc\", \"praw\"])\n\n tmp_dir = mkdtemp()\n try:\n success &= do_process(\n [\"sphinx-build\", \"-W\", \"--keep-going\", \"docs\", tmp_dir]\n )\n finally:\n rmtree(tmp_dir)\n\n return success\n\n\ndef run_unit():\n \"\"\"Runs the unit-tests.\n\n Follows the behavior of the static tests,\n where any failed tests cause pre_push.py to fail.\n \"\"\"\n return do_process(\n [sys.executable, path.join(current_directory, \"setup.py\"), \"test\"]\n )\n\n\ndef main():\n \"\"\"Runs the main function.\n\n usage: pre_push.py [-h] [-n] [-u] [-a]\n\n Run static and/or unit-tests\n \"\"\"\n parser = argparse.ArgumentParser(\n description=\"Run static and/or unit-tests\"\n )\n parser.add_argument(\n \"-n\",\n \"--unstatic\",\n action=\"store_true\",\n help=\"Do not run static tests (black/flake8/pydocstyle/sphinx-build)\",\n default=False,\n )\n parser.add_argument(\n \"-u\",\n \"--unit-tests\",\n \"--unit\",\n action=\"store_true\",\n default=False,\n help=\"Run the unit tests\",\n )\n parser.add_argument(\n \"-a\",\n \"--all\",\n action=\"store_true\",\n default=False,\n help=\"Run all of the tests (static and unit). \"\n \"Overrides the unstatic argument.\",\n )\n args = parser.parse_args()\n success = True\n try:\n if not args.unstatic or args.all:\n success &= run_static()\n if args.all or args.unit_tests:\n success &= run_unit()\n except KeyboardInterrupt:\n return int(not False)\n return int(not success)\n\n\nif __name__ == \"__main__\":\n exit_code = main()\n print(\n \"\\npre_push.py: Success!\" if not exit_code else \"\\npre_push.py: Fail\"\n )\n sys.exit(exit_code)\n", "path": "pre_push.py"}]}
| 1,372 | 107 |
gh_patches_debug_4208
|
rasdani/github-patches
|
git_diff
|
mathesar-foundation__mathesar-1363
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Unable to set number_format to null for Money type
## Reproduce
1. Set up a Money column.
1. Submit a `PATCH` request to the columns API, e.g. `/api/db/v0/tables/15/columns/52/`
1. Send:
```json
{
"type": "MATHESAR_TYPES.MATHESAR_MONEY",
"display_options": {
"currency_symbol": "$",
"currency_symbol_location": "after-minus",
"number_format": "english"
}
}
```
Receive success. Good.
1. Now change `display_options.number_format` to `null`, sending:
```json
{
"type": "MATHESAR_TYPES.MATHESAR_MONEY",
"display_options": {
"currency_symbol": "$",
"currency_symbol_location": "after-minus",
"number_format": null
}
}
```
Expect success.
Receive:
```json
[
{
"code": 2024,
"field": "number_format",
"message": "This field may not be null.",
"detail": {}
}
]
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mathesar/api/serializers/shared_serializers.py`
Content:
```
1 from django.core.exceptions import ImproperlyConfigured
2 from rest_framework import serializers
3
4 from mathesar.api.exceptions.mixins import MathesarErrorMessageMixin
5 from mathesar.database.types import MathesarTypeIdentifier, get_mathesar_type_from_db_type
6
7
8 class ReadOnlyPolymorphicSerializerMappingMixin:
9 """
10 This serializer mixin is helpful in serializing polymorphic models,
11 by switching to correct serializer based on the mapping field value.
12 """
13
14 def __new__(cls, *args, **kwargs):
15 if cls.serializers_mapping is None:
16 raise ImproperlyConfigured(
17 '`{cls}` is missing a '
18 '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)
19 )
20 return super().__new__(cls, *args, **kwargs)
21
22 def __init__(self, *args, **kwargs):
23 super().__init__(*args, **kwargs)
24 self.serializers_cls_mapping = {}
25 serializers_mapping = self.serializers_mapping
26 self.serializers_mapping = {}
27 for identifier, serializer_cls in serializers_mapping.items():
28 if callable(serializer_cls):
29 serializer = serializer_cls(*args, **kwargs)
30 serializer.parent = self
31 else:
32 serializer = serializer_cls
33 self.serializers_mapping[identifier] = serializer
34 self.serializers_cls_mapping[identifier] = serializer_cls
35
36 def to_representation(self, instance):
37 serializer = self.serializers_mapping.get(self.get_mapping_field(), None)
38 if serializer is not None:
39 return serializer.to_representation(instance)
40 else:
41 return instance
42
43 def get_mapping_field(self):
44 mapping_field = getattr(self, "mapping_field", None)
45 if mapping_field is None:
46 raise Exception(
47 "Add a `mapping_field` to be used as a identifier"
48 "or override this method to return a identifier to identify a proper serializer"
49 )
50 return mapping_field
51
52
53 class ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):
54 def to_internal_value(self, data):
55 serializer = self.serializers_mapping.get(self.get_mapping_field())
56 if serializer is not None:
57 return serializer.to_internal_value(data=data)
58 else:
59 data = {}
60 return data
61
62
63 class MonkeyPatchPartial:
64 """
65 Work around bug #3847 in djangorestframework by monkey-patching the partial
66 attribute of the root serializer during the call to validate_empty_values.
67 https://github.com/encode/django-rest-framework/issues/3847
68 """
69
70 def __init__(self, root):
71 self._root = root
72
73 def __enter__(self):
74 self._old = getattr(self._root, 'partial')
75 setattr(self._root, 'partial', False)
76
77 def __exit__(self, *args):
78 setattr(self._root, 'partial', self._old)
79
80
81 class OverrideRootPartialMixin:
82 """
83 This mixin is used to convert a serializer into a partial serializer,
84 based on the serializer `partial` property rather than the parent's `partial` property.
85 Refer to the issue
86 https://github.com/encode/django-rest-framework/issues/3847
87 """
88
89 def run_validation(self, *args, **kwargs):
90 if not self.partial:
91 with MonkeyPatchPartial(self.root):
92 return super().run_validation(*args, **kwargs)
93 return super().run_validation(*args, **kwargs)
94
95
96 class MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):
97 def get_serializer_fields(self):
98 return self.serializers_mapping[self.get_mapping_field()].fields
99
100
101 class CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):
102 TRUE = serializers.CharField()
103 FALSE = serializers.CharField()
104
105
106 DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'
107
108
109 class BooleanDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):
110 input = serializers.ChoiceField(choices=[("dropdown", "dropdown"), ("checkbox", "checkbox")])
111 custom_labels = CustomBooleanLabelSerializer(required=False)
112
113
114 class AbstractNumberDisplayOptionSerializer(serializers.Serializer):
115 number_format = serializers.ChoiceField(required=False, choices=['english', 'german', 'french', 'hindi', 'swiss'])
116
117
118 class NumberDisplayOptionSerializer(
119 MathesarErrorMessageMixin,
120 OverrideRootPartialMixin,
121 AbstractNumberDisplayOptionSerializer
122 ):
123 show_as_percentage = serializers.BooleanField(default=False)
124
125
126 class MoneyDisplayOptionSerializer(
127 MathesarErrorMessageMixin,
128 OverrideRootPartialMixin,
129 AbstractNumberDisplayOptionSerializer
130 ):
131 currency_symbol = serializers.CharField()
132 currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])
133
134
135 class TimeFormatDisplayOptionSerializer(
136 MathesarErrorMessageMixin,
137 OverrideRootPartialMixin,
138 serializers.Serializer
139 ):
140 format = serializers.CharField(max_length=255)
141
142
143 class DateTimeFormatDisplayOptionSerializer(
144 MathesarErrorMessageMixin,
145 OverrideRootPartialMixin,
146 serializers.Serializer
147 ):
148 time_format = serializers.CharField(max_length=255)
149 date_format = serializers.CharField(max_length=255)
150
151
152 class DurationDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):
153 min = serializers.CharField(max_length=255)
154 max = serializers.CharField(max_length=255)
155 show_units = serializers.BooleanField()
156
157
158 class DisplayOptionsMappingSerializer(
159 OverrideRootPartialMixin,
160 MathesarPolymorphicErrorMixin,
161 ReadWritePolymorphicSerializerMappingMixin,
162 serializers.Serializer
163 ):
164 serializers_mapping = {
165 MathesarTypeIdentifier.BOOLEAN.value: BooleanDisplayOptionSerializer,
166 MathesarTypeIdentifier.DATETIME.value: DateTimeFormatDisplayOptionSerializer,
167 MathesarTypeIdentifier.DATE.value: TimeFormatDisplayOptionSerializer,
168 MathesarTypeIdentifier.DURATION.value: DurationDisplayOptionSerializer,
169 MathesarTypeIdentifier.MONEY.value: MoneyDisplayOptionSerializer,
170 MathesarTypeIdentifier.NUMBER.value: NumberDisplayOptionSerializer,
171 MathesarTypeIdentifier.TIME.value: TimeFormatDisplayOptionSerializer,
172 }
173
174 def get_mapping_field(self):
175 db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]
176 mathesar_type = get_mathesar_type_from_db_type(db_type)
177 return mathesar_type
178
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/mathesar/api/serializers/shared_serializers.py b/mathesar/api/serializers/shared_serializers.py
--- a/mathesar/api/serializers/shared_serializers.py
+++ b/mathesar/api/serializers/shared_serializers.py
@@ -112,7 +112,7 @@
class AbstractNumberDisplayOptionSerializer(serializers.Serializer):
- number_format = serializers.ChoiceField(required=False, choices=['english', 'german', 'french', 'hindi', 'swiss'])
+ number_format = serializers.ChoiceField(required=False, allow_null=True, choices=['english', 'german', 'french', 'hindi', 'swiss'])
class NumberDisplayOptionSerializer(
|
{"golden_diff": "diff --git a/mathesar/api/serializers/shared_serializers.py b/mathesar/api/serializers/shared_serializers.py\n--- a/mathesar/api/serializers/shared_serializers.py\n+++ b/mathesar/api/serializers/shared_serializers.py\n@@ -112,7 +112,7 @@\n \n \n class AbstractNumberDisplayOptionSerializer(serializers.Serializer):\n- number_format = serializers.ChoiceField(required=False, choices=['english', 'german', 'french', 'hindi', 'swiss'])\n+ number_format = serializers.ChoiceField(required=False, allow_null=True, choices=['english', 'german', 'french', 'hindi', 'swiss'])\n \n \n class NumberDisplayOptionSerializer(\n", "issue": "Unable to set number_format to null for Money type\n## Reproduce\n\n1. Set up a Money column.\n\n1. Submit a `PATCH` request to the columns API, e.g. `/api/db/v0/tables/15/columns/52/`\n\n1. Send:\n\n ```json\n {\n \"type\": \"MATHESAR_TYPES.MATHESAR_MONEY\",\n \"display_options\": {\n \"currency_symbol\": \"$\",\n \"currency_symbol_location\": \"after-minus\",\n \"number_format\": \"english\"\n }\n }\n ```\n\n Receive success. Good.\n\n1. Now change `display_options.number_format` to `null`, sending:\n\n ```json\n {\n \"type\": \"MATHESAR_TYPES.MATHESAR_MONEY\",\n \"display_options\": {\n \"currency_symbol\": \"$\",\n \"currency_symbol_location\": \"after-minus\",\n \"number_format\": null\n }\n }\n ```\n \n Expect success.\n\n Receive:\n\n ```json\n [\n {\n \"code\": 2024,\n \"field\": \"number_format\",\n \"message\": \"This field may not be null.\",\n \"detail\": {}\n }\n ]\n ```\n\n", "before_files": [{"content": "from django.core.exceptions import ImproperlyConfigured\nfrom rest_framework import serializers\n\nfrom mathesar.api.exceptions.mixins import MathesarErrorMessageMixin\nfrom mathesar.database.types import MathesarTypeIdentifier, get_mathesar_type_from_db_type\n\n\nclass ReadOnlyPolymorphicSerializerMappingMixin:\n \"\"\"\n This serializer mixin is helpful in serializing polymorphic models,\n by switching to correct serializer based on the mapping field value.\n \"\"\"\n\n def __new__(cls, *args, **kwargs):\n if cls.serializers_mapping is None:\n raise ImproperlyConfigured(\n '`{cls}` is missing a '\n '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)\n )\n return super().__new__(cls, *args, **kwargs)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.serializers_cls_mapping = {}\n serializers_mapping = self.serializers_mapping\n self.serializers_mapping = {}\n for identifier, serializer_cls in serializers_mapping.items():\n if callable(serializer_cls):\n serializer = serializer_cls(*args, **kwargs)\n serializer.parent = self\n else:\n serializer = serializer_cls\n self.serializers_mapping[identifier] = serializer\n self.serializers_cls_mapping[identifier] = serializer_cls\n\n def to_representation(self, instance):\n serializer = self.serializers_mapping.get(self.get_mapping_field(), None)\n if serializer is not None:\n return serializer.to_representation(instance)\n else:\n return instance\n\n def get_mapping_field(self):\n mapping_field = getattr(self, \"mapping_field\", None)\n if mapping_field is None:\n raise Exception(\n \"Add a `mapping_field` to be used as a identifier\"\n \"or override this method to return a identifier to identify a proper serializer\"\n )\n return mapping_field\n\n\nclass ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):\n def to_internal_value(self, data):\n serializer = self.serializers_mapping.get(self.get_mapping_field())\n if serializer is not None:\n return serializer.to_internal_value(data=data)\n else:\n data = {}\n return data\n\n\nclass MonkeyPatchPartial:\n \"\"\"\n Work around bug #3847 in djangorestframework by monkey-patching the partial\n attribute of the root serializer during the call to validate_empty_values.\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def __init__(self, root):\n self._root = root\n\n def __enter__(self):\n self._old = getattr(self._root, 'partial')\n setattr(self._root, 'partial', False)\n\n def __exit__(self, *args):\n setattr(self._root, 'partial', self._old)\n\n\nclass OverrideRootPartialMixin:\n \"\"\"\n This mixin is used to convert a serializer into a partial serializer,\n based on the serializer `partial` property rather than the parent's `partial` property.\n Refer to the issue\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def run_validation(self, *args, **kwargs):\n if not self.partial:\n with MonkeyPatchPartial(self.root):\n return super().run_validation(*args, **kwargs)\n return super().run_validation(*args, **kwargs)\n\n\nclass MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):\n def get_serializer_fields(self):\n return self.serializers_mapping[self.get_mapping_field()].fields\n\n\nclass CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):\n TRUE = serializers.CharField()\n FALSE = serializers.CharField()\n\n\nDISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'\n\n\nclass BooleanDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n input = serializers.ChoiceField(choices=[(\"dropdown\", \"dropdown\"), (\"checkbox\", \"checkbox\")])\n custom_labels = CustomBooleanLabelSerializer(required=False)\n\n\nclass AbstractNumberDisplayOptionSerializer(serializers.Serializer):\n number_format = serializers.ChoiceField(required=False, choices=['english', 'german', 'french', 'hindi', 'swiss'])\n\n\nclass NumberDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n AbstractNumberDisplayOptionSerializer\n):\n show_as_percentage = serializers.BooleanField(default=False)\n\n\nclass MoneyDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n AbstractNumberDisplayOptionSerializer\n):\n currency_symbol = serializers.CharField()\n currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])\n\n\nclass TimeFormatDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n serializers.Serializer\n):\n format = serializers.CharField(max_length=255)\n\n\nclass DateTimeFormatDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n serializers.Serializer\n):\n time_format = serializers.CharField(max_length=255)\n date_format = serializers.CharField(max_length=255)\n\n\nclass DurationDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n min = serializers.CharField(max_length=255)\n max = serializers.CharField(max_length=255)\n show_units = serializers.BooleanField()\n\n\nclass DisplayOptionsMappingSerializer(\n OverrideRootPartialMixin,\n MathesarPolymorphicErrorMixin,\n ReadWritePolymorphicSerializerMappingMixin,\n serializers.Serializer\n):\n serializers_mapping = {\n MathesarTypeIdentifier.BOOLEAN.value: BooleanDisplayOptionSerializer,\n MathesarTypeIdentifier.DATETIME.value: DateTimeFormatDisplayOptionSerializer,\n MathesarTypeIdentifier.DATE.value: TimeFormatDisplayOptionSerializer,\n MathesarTypeIdentifier.DURATION.value: DurationDisplayOptionSerializer,\n MathesarTypeIdentifier.MONEY.value: MoneyDisplayOptionSerializer,\n MathesarTypeIdentifier.NUMBER.value: NumberDisplayOptionSerializer,\n MathesarTypeIdentifier.TIME.value: TimeFormatDisplayOptionSerializer,\n }\n\n def get_mapping_field(self):\n db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]\n mathesar_type = get_mathesar_type_from_db_type(db_type)\n return mathesar_type\n", "path": "mathesar/api/serializers/shared_serializers.py"}], "after_files": [{"content": "from django.core.exceptions import ImproperlyConfigured\nfrom rest_framework import serializers\n\nfrom mathesar.api.exceptions.mixins import MathesarErrorMessageMixin\nfrom mathesar.database.types import MathesarTypeIdentifier, get_mathesar_type_from_db_type\n\n\nclass ReadOnlyPolymorphicSerializerMappingMixin:\n \"\"\"\n This serializer mixin is helpful in serializing polymorphic models,\n by switching to correct serializer based on the mapping field value.\n \"\"\"\n\n def __new__(cls, *args, **kwargs):\n if cls.serializers_mapping is None:\n raise ImproperlyConfigured(\n '`{cls}` is missing a '\n '`{cls}.model_serializer_mapping` attribute'.format(cls=cls.__name__)\n )\n return super().__new__(cls, *args, **kwargs)\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.serializers_cls_mapping = {}\n serializers_mapping = self.serializers_mapping\n self.serializers_mapping = {}\n for identifier, serializer_cls in serializers_mapping.items():\n if callable(serializer_cls):\n serializer = serializer_cls(*args, **kwargs)\n serializer.parent = self\n else:\n serializer = serializer_cls\n self.serializers_mapping[identifier] = serializer\n self.serializers_cls_mapping[identifier] = serializer_cls\n\n def to_representation(self, instance):\n serializer = self.serializers_mapping.get(self.get_mapping_field(), None)\n if serializer is not None:\n return serializer.to_representation(instance)\n else:\n return instance\n\n def get_mapping_field(self):\n mapping_field = getattr(self, \"mapping_field\", None)\n if mapping_field is None:\n raise Exception(\n \"Add a `mapping_field` to be used as a identifier\"\n \"or override this method to return a identifier to identify a proper serializer\"\n )\n return mapping_field\n\n\nclass ReadWritePolymorphicSerializerMappingMixin(ReadOnlyPolymorphicSerializerMappingMixin):\n def to_internal_value(self, data):\n serializer = self.serializers_mapping.get(self.get_mapping_field())\n if serializer is not None:\n return serializer.to_internal_value(data=data)\n else:\n data = {}\n return data\n\n\nclass MonkeyPatchPartial:\n \"\"\"\n Work around bug #3847 in djangorestframework by monkey-patching the partial\n attribute of the root serializer during the call to validate_empty_values.\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def __init__(self, root):\n self._root = root\n\n def __enter__(self):\n self._old = getattr(self._root, 'partial')\n setattr(self._root, 'partial', False)\n\n def __exit__(self, *args):\n setattr(self._root, 'partial', self._old)\n\n\nclass OverrideRootPartialMixin:\n \"\"\"\n This mixin is used to convert a serializer into a partial serializer,\n based on the serializer `partial` property rather than the parent's `partial` property.\n Refer to the issue\n https://github.com/encode/django-rest-framework/issues/3847\n \"\"\"\n\n def run_validation(self, *args, **kwargs):\n if not self.partial:\n with MonkeyPatchPartial(self.root):\n return super().run_validation(*args, **kwargs)\n return super().run_validation(*args, **kwargs)\n\n\nclass MathesarPolymorphicErrorMixin(MathesarErrorMessageMixin):\n def get_serializer_fields(self):\n return self.serializers_mapping[self.get_mapping_field()].fields\n\n\nclass CustomBooleanLabelSerializer(MathesarErrorMessageMixin, serializers.Serializer):\n TRUE = serializers.CharField()\n FALSE = serializers.CharField()\n\n\nDISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY = 'db_type'\n\n\nclass BooleanDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n input = serializers.ChoiceField(choices=[(\"dropdown\", \"dropdown\"), (\"checkbox\", \"checkbox\")])\n custom_labels = CustomBooleanLabelSerializer(required=False)\n\n\nclass AbstractNumberDisplayOptionSerializer(serializers.Serializer):\n number_format = serializers.ChoiceField(required=False, allow_null=True, choices=['english', 'german', 'french', 'hindi', 'swiss'])\n\n\nclass NumberDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n AbstractNumberDisplayOptionSerializer\n):\n show_as_percentage = serializers.BooleanField(default=False)\n\n\nclass MoneyDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n AbstractNumberDisplayOptionSerializer\n):\n currency_symbol = serializers.CharField()\n currency_symbol_location = serializers.ChoiceField(choices=['after-minus', 'end-with-space'])\n\n\nclass TimeFormatDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n serializers.Serializer\n):\n format = serializers.CharField(max_length=255)\n\n\nclass DateTimeFormatDisplayOptionSerializer(\n MathesarErrorMessageMixin,\n OverrideRootPartialMixin,\n serializers.Serializer\n):\n time_format = serializers.CharField(max_length=255)\n date_format = serializers.CharField(max_length=255)\n\n\nclass DurationDisplayOptionSerializer(MathesarErrorMessageMixin, OverrideRootPartialMixin, serializers.Serializer):\n min = serializers.CharField(max_length=255)\n max = serializers.CharField(max_length=255)\n show_units = serializers.BooleanField()\n\n\nclass DisplayOptionsMappingSerializer(\n OverrideRootPartialMixin,\n MathesarPolymorphicErrorMixin,\n ReadWritePolymorphicSerializerMappingMixin,\n serializers.Serializer\n):\n serializers_mapping = {\n MathesarTypeIdentifier.BOOLEAN.value: BooleanDisplayOptionSerializer,\n MathesarTypeIdentifier.DATETIME.value: DateTimeFormatDisplayOptionSerializer,\n MathesarTypeIdentifier.DATE.value: TimeFormatDisplayOptionSerializer,\n MathesarTypeIdentifier.DURATION.value: DurationDisplayOptionSerializer,\n MathesarTypeIdentifier.MONEY.value: MoneyDisplayOptionSerializer,\n MathesarTypeIdentifier.NUMBER.value: NumberDisplayOptionSerializer,\n MathesarTypeIdentifier.TIME.value: TimeFormatDisplayOptionSerializer,\n }\n\n def get_mapping_field(self):\n db_type = self.context[DISPLAY_OPTIONS_SERIALIZER_MAPPING_KEY]\n mathesar_type = get_mathesar_type_from_db_type(db_type)\n return mathesar_type\n", "path": "mathesar/api/serializers/shared_serializers.py"}]}
| 2,264 | 151 |
gh_patches_debug_314
|
rasdani/github-patches
|
git_diff
|
zulip__zulip-20788
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
"Pan and zoom" cuts off images instead of using the available space
If you have a tall image and a wide monitor (and wide browser viewport), and you try to zoom… the image stays trapped inside the same box it occupied before you even tried to zoom. If the image is super wide instead of tall, the same thing happens the other way around.
This leads to a lot of frustrating panning around, to look at the different parts of the image through this narrow keyhole, while tons of screen space next to it doesn't get used.
This is the biggest of the issues described by @vanclute in #18939. It was reported again by @alexanderglueck as #19837, and I just ran into it myself ([chat](https://chat.zulip.org/#narrow/stream/6-frontend/topic/pan.2Fzoom/near/1308717)). Here's a nice illustration from #19837:

Instead, when zooming we should use the full space available. This may be bigger than the area the image occupied when it was scaled down to fit completely in the space available, because the available box may have a different aspect ratio from the image.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `version.py`
Content:
```
1 import os
2
3 ZULIP_VERSION = "5.0-dev+git"
4
5 # Add information on number of commits and commit hash to version, if available
6 zulip_git_version_file = os.path.join(
7 os.path.dirname(os.path.abspath(__file__)), "zulip-git-version"
8 )
9 lines = [ZULIP_VERSION, ""]
10 if os.path.exists(zulip_git_version_file):
11 with open(zulip_git_version_file) as f:
12 lines = f.readlines() + ["", ""]
13 ZULIP_VERSION = lines.pop(0).strip()
14 ZULIP_MERGE_BASE = lines.pop(0).strip()
15
16 LATEST_MAJOR_VERSION = "4.0"
17 LATEST_RELEASE_VERSION = "4.9"
18 LATEST_RELEASE_ANNOUNCEMENT = "https://blog.zulip.com/2021/05/13/zulip-4-0-released/"
19
20 # Versions of the desktop app below DESKTOP_MINIMUM_VERSION will be
21 # prevented from connecting to the Zulip server. Versions above
22 # DESKTOP_MINIMUM_VERSION but below DESKTOP_WARNING_VERSION will have
23 # a banner at the top of the page asking the user to upgrade.
24 DESKTOP_MINIMUM_VERSION = "5.2.0"
25 DESKTOP_WARNING_VERSION = "5.4.3"
26
27 # Bump the API_FEATURE_LEVEL whenever an API change is made
28 # that clients might want to condition on. If we forget at
29 # the time we make the change, then bump it later as soon
30 # as we notice; clients using API_FEATURE_LEVEL will just not
31 # use the new feature/API until the bump.
32 #
33 # Changes should be accompanied by documentation explaining what the
34 # new level means in templates/zerver/api/changelog.md, as well as
35 # "**Changes**" entries in the endpoint's documentation in `zulip.yaml`.
36 API_FEATURE_LEVEL = 115
37
38 # Bump the minor PROVISION_VERSION to indicate that folks should provision
39 # only when going from an old version of the code to a newer version. Bump
40 # the major version to indicate that folks should provision in both
41 # directions.
42
43 # Typically,
44 # * adding a dependency only requires a minor version bump;
45 # * removing a dependency requires a major version bump;
46 # * upgrading a dependency requires a major version bump, unless the
47 # upgraded dependency is backwards compatible with all of our
48 # historical commits sharing the same major version, in which case a
49 # minor version bump suffices.
50
51 PROVISION_VERSION = "173.3"
52
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/version.py b/version.py
--- a/version.py
+++ b/version.py
@@ -48,4 +48,4 @@
# historical commits sharing the same major version, in which case a
# minor version bump suffices.
-PROVISION_VERSION = "173.3"
+PROVISION_VERSION = "173.4"
|
{"golden_diff": "diff --git a/version.py b/version.py\n--- a/version.py\n+++ b/version.py\n@@ -48,4 +48,4 @@\n # historical commits sharing the same major version, in which case a\n # minor version bump suffices.\n \n-PROVISION_VERSION = \"173.3\"\n+PROVISION_VERSION = \"173.4\"\n", "issue": "\"Pan and zoom\" cuts off images instead of using the available space\nIf you have a tall image and a wide monitor (and wide browser viewport), and you try to zoom\u2026 the image stays trapped inside the same box it occupied before you even tried to zoom. If the image is super wide instead of tall, the same thing happens the other way around.\r\n\r\nThis leads to a lot of frustrating panning around, to look at the different parts of the image through this narrow keyhole, while tons of screen space next to it doesn't get used.\r\n\r\nThis is the biggest of the issues described by @vanclute in #18939. It was reported again by @alexanderglueck as #19837, and I just ran into it myself ([chat](https://chat.zulip.org/#narrow/stream/6-frontend/topic/pan.2Fzoom/near/1308717)). Here's a nice illustration from #19837: \r\n\r\n\r\nInstead, when zooming we should use the full space available. This may be bigger than the area the image occupied when it was scaled down to fit completely in the space available, because the available box may have a different aspect ratio from the image.\r\n\n", "before_files": [{"content": "import os\n\nZULIP_VERSION = \"5.0-dev+git\"\n\n# Add information on number of commits and commit hash to version, if available\nzulip_git_version_file = os.path.join(\n os.path.dirname(os.path.abspath(__file__)), \"zulip-git-version\"\n)\nlines = [ZULIP_VERSION, \"\"]\nif os.path.exists(zulip_git_version_file):\n with open(zulip_git_version_file) as f:\n lines = f.readlines() + [\"\", \"\"]\nZULIP_VERSION = lines.pop(0).strip()\nZULIP_MERGE_BASE = lines.pop(0).strip()\n\nLATEST_MAJOR_VERSION = \"4.0\"\nLATEST_RELEASE_VERSION = \"4.9\"\nLATEST_RELEASE_ANNOUNCEMENT = \"https://blog.zulip.com/2021/05/13/zulip-4-0-released/\"\n\n# Versions of the desktop app below DESKTOP_MINIMUM_VERSION will be\n# prevented from connecting to the Zulip server. Versions above\n# DESKTOP_MINIMUM_VERSION but below DESKTOP_WARNING_VERSION will have\n# a banner at the top of the page asking the user to upgrade.\nDESKTOP_MINIMUM_VERSION = \"5.2.0\"\nDESKTOP_WARNING_VERSION = \"5.4.3\"\n\n# Bump the API_FEATURE_LEVEL whenever an API change is made\n# that clients might want to condition on. If we forget at\n# the time we make the change, then bump it later as soon\n# as we notice; clients using API_FEATURE_LEVEL will just not\n# use the new feature/API until the bump.\n#\n# Changes should be accompanied by documentation explaining what the\n# new level means in templates/zerver/api/changelog.md, as well as\n# \"**Changes**\" entries in the endpoint's documentation in `zulip.yaml`.\nAPI_FEATURE_LEVEL = 115\n\n# Bump the minor PROVISION_VERSION to indicate that folks should provision\n# only when going from an old version of the code to a newer version. Bump\n# the major version to indicate that folks should provision in both\n# directions.\n\n# Typically,\n# * adding a dependency only requires a minor version bump;\n# * removing a dependency requires a major version bump;\n# * upgrading a dependency requires a major version bump, unless the\n# upgraded dependency is backwards compatible with all of our\n# historical commits sharing the same major version, in which case a\n# minor version bump suffices.\n\nPROVISION_VERSION = \"173.3\"\n", "path": "version.py"}], "after_files": [{"content": "import os\n\nZULIP_VERSION = \"5.0-dev+git\"\n\n# Add information on number of commits and commit hash to version, if available\nzulip_git_version_file = os.path.join(\n os.path.dirname(os.path.abspath(__file__)), \"zulip-git-version\"\n)\nlines = [ZULIP_VERSION, \"\"]\nif os.path.exists(zulip_git_version_file):\n with open(zulip_git_version_file) as f:\n lines = f.readlines() + [\"\", \"\"]\nZULIP_VERSION = lines.pop(0).strip()\nZULIP_MERGE_BASE = lines.pop(0).strip()\n\nLATEST_MAJOR_VERSION = \"4.0\"\nLATEST_RELEASE_VERSION = \"4.9\"\nLATEST_RELEASE_ANNOUNCEMENT = \"https://blog.zulip.com/2021/05/13/zulip-4-0-released/\"\n\n# Versions of the desktop app below DESKTOP_MINIMUM_VERSION will be\n# prevented from connecting to the Zulip server. Versions above\n# DESKTOP_MINIMUM_VERSION but below DESKTOP_WARNING_VERSION will have\n# a banner at the top of the page asking the user to upgrade.\nDESKTOP_MINIMUM_VERSION = \"5.2.0\"\nDESKTOP_WARNING_VERSION = \"5.4.3\"\n\n# Bump the API_FEATURE_LEVEL whenever an API change is made\n# that clients might want to condition on. If we forget at\n# the time we make the change, then bump it later as soon\n# as we notice; clients using API_FEATURE_LEVEL will just not\n# use the new feature/API until the bump.\n#\n# Changes should be accompanied by documentation explaining what the\n# new level means in templates/zerver/api/changelog.md, as well as\n# \"**Changes**\" entries in the endpoint's documentation in `zulip.yaml`.\nAPI_FEATURE_LEVEL = 115\n\n# Bump the minor PROVISION_VERSION to indicate that folks should provision\n# only when going from an old version of the code to a newer version. Bump\n# the major version to indicate that folks should provision in both\n# directions.\n\n# Typically,\n# * adding a dependency only requires a minor version bump;\n# * removing a dependency requires a major version bump;\n# * upgrading a dependency requires a major version bump, unless the\n# upgraded dependency is backwards compatible with all of our\n# historical commits sharing the same major version, in which case a\n# minor version bump suffices.\n\nPROVISION_VERSION = \"173.4\"\n", "path": "version.py"}]}
| 1,224 | 80 |
gh_patches_debug_11329
|
rasdani/github-patches
|
git_diff
|
conan-io__conan-center-index-4675
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[package] odbc/2.3.7: sysconfdir is in Conan local cache
<!--
Please don't forget to update the issue title.
Include all applicable information to help us reproduce your problem.
-->
### Package and Environment Details (include every applicable attribute)
* Package Name/Version: **odbc/2.3.7**
### Steps to reproduce (Include if Applicable)
The odbc library has a configuration option called `--sysconfdir` that specifies the default location for configuration files. By default the value is `${prefix}/etc`. The problem is that `${prefix}` ends up being a location in the Conan local cache, for example:
```
/userhome/epederson/.conan/data/odbc/2.3.7/bincrafters/stable/package/761432f3453a17c960259b5b2605153e3bcffd7d/etc
```
This location is only valid on the machine that the package was built on.
The typical value for `sysconfdir` is `/etc`.
I will provide a PR that adds `--sysconfdir=/etc` to the configuration settings.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `recipes/odbc/all/conanfile.py`
Content:
```
1 import glob
2 import os
3 from conans import ConanFile, AutoToolsBuildEnvironment, tools
4 from conans.errors import ConanInvalidConfiguration
5
6
7 class OdbcConan(ConanFile):
8 name = "odbc"
9 description = "Package providing unixODBC"
10 topics = ("odbc", "database", "dbms", "data-access")
11 url = "https://github.com/conan-io/conan-center-index"
12 homepage = "http://www.unixodbc.org"
13 license = ("LGPL-2.1", "GPL-2.1")
14 exports_sources = "patches/**"
15 settings = "os", "compiler", "build_type", "arch"
16 options = {
17 "shared": [True, False],
18 "fPIC": [True, False],
19 "with_libiconv": [True, False]
20 }
21 default_options = {
22 "shared": False,
23 "fPIC": True,
24 "with_libiconv": True
25 }
26
27 _autotools = None
28
29 @property
30 def _source_subfolder(self):
31 return "source_subfolder"
32
33 def configure(self):
34 if self.options.shared:
35 del self.options.fPIC
36 del self.settings.compiler.libcxx
37 del self.settings.compiler.cppstd
38 if self.settings.os == "Windows":
39 raise ConanInvalidConfiguration("Windows not supported yet. Please, open an issue if you need such support")
40
41 def requirements(self):
42 if self.options.with_libiconv:
43 self.requires("libiconv/1.16")
44
45 def source(self):
46 tools.get(**self.conan_data["sources"][self.version])
47 extracted_dir = "unixODBC-%s" % self.version
48 os.rename(extracted_dir, self._source_subfolder)
49
50 def _configure_autotools(self):
51 if self._autotools:
52 return self._autotools
53 self._autotools = AutoToolsBuildEnvironment(self)
54 static_flag = "no" if self.options.shared else "yes"
55 shared_flag = "yes" if self.options.shared else "no"
56 libiconv_flag = "yes" if self.options.with_libiconv else "no"
57 args = ["--enable-static=%s" % static_flag,
58 "--enable-shared=%s" % shared_flag,
59 "--enable-ltdl-install",
60 "--enable-iconv=%s" % libiconv_flag]
61 if self.options.with_libiconv:
62 libiconv_prefix = self.deps_cpp_info["libiconv"].rootpath
63 args.append("--with-libiconv-prefix=%s" % libiconv_prefix)
64 self._autotools.configure(configure_dir=self._source_subfolder, args=args)
65 return self._autotools
66
67 def build(self):
68 for patch in self.conan_data.get("patches", {}).get(self.version, []):
69 tools.patch(**patch)
70 autotools = self._configure_autotools()
71 autotools.make()
72
73 def package(self):
74 self.copy("COPYING", src=self._source_subfolder, dst="licenses")
75 autotools = self._configure_autotools()
76 autotools.install()
77 tools.rmdir(os.path.join(self.package_folder, "share"))
78 tools.rmdir(os.path.join(self.package_folder, "etc"))
79 tools.rmdir(os.path.join(self.package_folder, "lib", "pkgconfig"))
80 for la_file in glob.glob(os.path.join(self.package_folder, "lib", "*.la")):
81 os.remove(la_file)
82
83 def package_info(self):
84 self.cpp_info.names["cmake_find_package"] = "ODBC"
85 self.cpp_info.names["cmake_find_package_multi"] = "ODBC"
86 # odbc
87 self.cpp_info.components["_odbc"].names["pkg_config"] = "odbc"
88 self.cpp_info.components["_odbc"].libs = ["odbc"]
89 self.cpp_info.components["_odbc"].requires = ["odbcltdl"]
90 if self.options.with_libiconv:
91 self.cpp_info.components["_odbc"].requires.append("libiconv::libiconv")
92 # odbcinst
93 self.cpp_info.components["odbcinst"].names["pkg_config"] = "odbcinst"
94 self.cpp_info.components["odbcinst"].libs = ["odbcinst"]
95 self.cpp_info.components["odbcinst"].requires = ["odbcltdl"]
96 # odbccr
97 self.cpp_info.components["odbccr"].names["pkg_config"] = "odbccr"
98 self.cpp_info.components["odbccr"].libs = ["odbccr"]
99
100 self.cpp_info.components["odbcltdl"].libs = ["ltdl"]
101
102 if self.settings.os == "Linux":
103 self.cpp_info.components["_odbc"].system_libs = ["pthread"]
104 self.cpp_info.components["odbcinst"].system_libs = ["pthread"]
105 self.cpp_info.components["odbcltdl"].system_libs = ["dl"]
106
107 bin_path = os.path.join(self.package_folder, "bin")
108 self.output.info("Appending PATH environment variable: {}".format(bin_path))
109 self.env_info.PATH.append(bin_path)
110
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/recipes/odbc/all/conanfile.py b/recipes/odbc/all/conanfile.py
--- a/recipes/odbc/all/conanfile.py
+++ b/recipes/odbc/all/conanfile.py
@@ -57,7 +57,8 @@
args = ["--enable-static=%s" % static_flag,
"--enable-shared=%s" % shared_flag,
"--enable-ltdl-install",
- "--enable-iconv=%s" % libiconv_flag]
+ "--enable-iconv=%s" % libiconv_flag,
+ "--sysconfdir=/etc"]
if self.options.with_libiconv:
libiconv_prefix = self.deps_cpp_info["libiconv"].rootpath
args.append("--with-libiconv-prefix=%s" % libiconv_prefix)
|
{"golden_diff": "diff --git a/recipes/odbc/all/conanfile.py b/recipes/odbc/all/conanfile.py\n--- a/recipes/odbc/all/conanfile.py\n+++ b/recipes/odbc/all/conanfile.py\n@@ -57,7 +57,8 @@\n args = [\"--enable-static=%s\" % static_flag,\n \"--enable-shared=%s\" % shared_flag,\n \"--enable-ltdl-install\",\n- \"--enable-iconv=%s\" % libiconv_flag]\n+ \"--enable-iconv=%s\" % libiconv_flag,\n+ \"--sysconfdir=/etc\"]\n if self.options.with_libiconv:\n libiconv_prefix = self.deps_cpp_info[\"libiconv\"].rootpath\n args.append(\"--with-libiconv-prefix=%s\" % libiconv_prefix)\n", "issue": "[package] odbc/2.3.7: sysconfdir is in Conan local cache\n<!-- \r\n Please don't forget to update the issue title.\r\n Include all applicable information to help us reproduce your problem.\r\n-->\r\n\r\n### Package and Environment Details (include every applicable attribute)\r\n * Package Name/Version: **odbc/2.3.7**\r\n\r\n### Steps to reproduce (Include if Applicable)\r\n\r\nThe odbc library has a configuration option called `--sysconfdir` that specifies the default location for configuration files. By default the value is `${prefix}/etc`. The problem is that `${prefix}` ends up being a location in the Conan local cache, for example:\r\n```\r\n/userhome/epederson/.conan/data/odbc/2.3.7/bincrafters/stable/package/761432f3453a17c960259b5b2605153e3bcffd7d/etc\r\n```\r\nThis location is only valid on the machine that the package was built on. \r\n\r\nThe typical value for `sysconfdir` is `/etc`. \r\n\r\nI will provide a PR that adds `--sysconfdir=/etc` to the configuration settings.\n", "before_files": [{"content": "import glob\nimport os\nfrom conans import ConanFile, AutoToolsBuildEnvironment, tools\nfrom conans.errors import ConanInvalidConfiguration\n\n\nclass OdbcConan(ConanFile):\n name = \"odbc\"\n description = \"Package providing unixODBC\"\n topics = (\"odbc\", \"database\", \"dbms\", \"data-access\")\n url = \"https://github.com/conan-io/conan-center-index\"\n homepage = \"http://www.unixodbc.org\"\n license = (\"LGPL-2.1\", \"GPL-2.1\")\n exports_sources = \"patches/**\"\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False],\n \"with_libiconv\": [True, False]\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True,\n \"with_libiconv\": True\n }\n\n _autotools = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n def configure(self):\n if self.options.shared:\n del self.options.fPIC\n del self.settings.compiler.libcxx\n del self.settings.compiler.cppstd\n if self.settings.os == \"Windows\":\n raise ConanInvalidConfiguration(\"Windows not supported yet. Please, open an issue if you need such support\")\n\n def requirements(self):\n if self.options.with_libiconv:\n self.requires(\"libiconv/1.16\")\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n extracted_dir = \"unixODBC-%s\" % self.version\n os.rename(extracted_dir, self._source_subfolder)\n\n def _configure_autotools(self):\n if self._autotools:\n return self._autotools\n self._autotools = AutoToolsBuildEnvironment(self)\n static_flag = \"no\" if self.options.shared else \"yes\"\n shared_flag = \"yes\" if self.options.shared else \"no\"\n libiconv_flag = \"yes\" if self.options.with_libiconv else \"no\"\n args = [\"--enable-static=%s\" % static_flag,\n \"--enable-shared=%s\" % shared_flag,\n \"--enable-ltdl-install\",\n \"--enable-iconv=%s\" % libiconv_flag]\n if self.options.with_libiconv:\n libiconv_prefix = self.deps_cpp_info[\"libiconv\"].rootpath\n args.append(\"--with-libiconv-prefix=%s\" % libiconv_prefix)\n self._autotools.configure(configure_dir=self._source_subfolder, args=args)\n return self._autotools\n\n def build(self):\n for patch in self.conan_data.get(\"patches\", {}).get(self.version, []):\n tools.patch(**patch)\n autotools = self._configure_autotools()\n autotools.make()\n\n def package(self):\n self.copy(\"COPYING\", src=self._source_subfolder, dst=\"licenses\")\n autotools = self._configure_autotools()\n autotools.install()\n tools.rmdir(os.path.join(self.package_folder, \"share\"))\n tools.rmdir(os.path.join(self.package_folder, \"etc\"))\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"pkgconfig\"))\n for la_file in glob.glob(os.path.join(self.package_folder, \"lib\", \"*.la\")):\n os.remove(la_file)\n\n def package_info(self):\n self.cpp_info.names[\"cmake_find_package\"] = \"ODBC\"\n self.cpp_info.names[\"cmake_find_package_multi\"] = \"ODBC\"\n # odbc\n self.cpp_info.components[\"_odbc\"].names[\"pkg_config\"] = \"odbc\"\n self.cpp_info.components[\"_odbc\"].libs = [\"odbc\"]\n self.cpp_info.components[\"_odbc\"].requires = [\"odbcltdl\"]\n if self.options.with_libiconv:\n self.cpp_info.components[\"_odbc\"].requires.append(\"libiconv::libiconv\")\n # odbcinst\n self.cpp_info.components[\"odbcinst\"].names[\"pkg_config\"] = \"odbcinst\"\n self.cpp_info.components[\"odbcinst\"].libs = [\"odbcinst\"]\n self.cpp_info.components[\"odbcinst\"].requires = [\"odbcltdl\"]\n # odbccr\n self.cpp_info.components[\"odbccr\"].names[\"pkg_config\"] = \"odbccr\"\n self.cpp_info.components[\"odbccr\"].libs = [\"odbccr\"]\n\n self.cpp_info.components[\"odbcltdl\"].libs = [\"ltdl\"]\n\n if self.settings.os == \"Linux\":\n self.cpp_info.components[\"_odbc\"].system_libs = [\"pthread\"]\n self.cpp_info.components[\"odbcinst\"].system_libs = [\"pthread\"]\n self.cpp_info.components[\"odbcltdl\"].system_libs = [\"dl\"]\n\n bin_path = os.path.join(self.package_folder, \"bin\")\n self.output.info(\"Appending PATH environment variable: {}\".format(bin_path))\n self.env_info.PATH.append(bin_path)\n", "path": "recipes/odbc/all/conanfile.py"}], "after_files": [{"content": "import glob\nimport os\nfrom conans import ConanFile, AutoToolsBuildEnvironment, tools\nfrom conans.errors import ConanInvalidConfiguration\n\n\nclass OdbcConan(ConanFile):\n name = \"odbc\"\n description = \"Package providing unixODBC\"\n topics = (\"odbc\", \"database\", \"dbms\", \"data-access\")\n url = \"https://github.com/conan-io/conan-center-index\"\n homepage = \"http://www.unixodbc.org\"\n license = (\"LGPL-2.1\", \"GPL-2.1\")\n exports_sources = \"patches/**\"\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n options = {\n \"shared\": [True, False],\n \"fPIC\": [True, False],\n \"with_libiconv\": [True, False]\n }\n default_options = {\n \"shared\": False,\n \"fPIC\": True,\n \"with_libiconv\": True\n }\n\n _autotools = None\n\n @property\n def _source_subfolder(self):\n return \"source_subfolder\"\n\n def configure(self):\n if self.options.shared:\n del self.options.fPIC\n del self.settings.compiler.libcxx\n del self.settings.compiler.cppstd\n if self.settings.os == \"Windows\":\n raise ConanInvalidConfiguration(\"Windows not supported yet. Please, open an issue if you need such support\")\n\n def requirements(self):\n if self.options.with_libiconv:\n self.requires(\"libiconv/1.16\")\n\n def source(self):\n tools.get(**self.conan_data[\"sources\"][self.version])\n extracted_dir = \"unixODBC-%s\" % self.version\n os.rename(extracted_dir, self._source_subfolder)\n\n def _configure_autotools(self):\n if self._autotools:\n return self._autotools\n self._autotools = AutoToolsBuildEnvironment(self)\n static_flag = \"no\" if self.options.shared else \"yes\"\n shared_flag = \"yes\" if self.options.shared else \"no\"\n libiconv_flag = \"yes\" if self.options.with_libiconv else \"no\"\n args = [\"--enable-static=%s\" % static_flag,\n \"--enable-shared=%s\" % shared_flag,\n \"--enable-ltdl-install\",\n \"--enable-iconv=%s\" % libiconv_flag,\n \"--sysconfdir=/etc\"]\n if self.options.with_libiconv:\n libiconv_prefix = self.deps_cpp_info[\"libiconv\"].rootpath\n args.append(\"--with-libiconv-prefix=%s\" % libiconv_prefix)\n self._autotools.configure(configure_dir=self._source_subfolder, args=args)\n return self._autotools\n\n def build(self):\n for patch in self.conan_data.get(\"patches\", {}).get(self.version, []):\n tools.patch(**patch)\n autotools = self._configure_autotools()\n autotools.make()\n\n def package(self):\n self.copy(\"COPYING\", src=self._source_subfolder, dst=\"licenses\")\n autotools = self._configure_autotools()\n autotools.install()\n tools.rmdir(os.path.join(self.package_folder, \"share\"))\n tools.rmdir(os.path.join(self.package_folder, \"etc\"))\n tools.rmdir(os.path.join(self.package_folder, \"lib\", \"pkgconfig\"))\n for la_file in glob.glob(os.path.join(self.package_folder, \"lib\", \"*.la\")):\n os.remove(la_file)\n\n def package_info(self):\n self.cpp_info.names[\"cmake_find_package\"] = \"ODBC\"\n self.cpp_info.names[\"cmake_find_package_multi\"] = \"ODBC\"\n # odbc\n self.cpp_info.components[\"_odbc\"].names[\"pkg_config\"] = \"odbc\"\n self.cpp_info.components[\"_odbc\"].libs = [\"odbc\"]\n self.cpp_info.components[\"_odbc\"].requires = [\"odbcltdl\"]\n if self.options.with_libiconv:\n self.cpp_info.components[\"_odbc\"].requires.append(\"libiconv::libiconv\")\n # odbcinst\n self.cpp_info.components[\"odbcinst\"].names[\"pkg_config\"] = \"odbcinst\"\n self.cpp_info.components[\"odbcinst\"].libs = [\"odbcinst\"]\n self.cpp_info.components[\"odbcinst\"].requires = [\"odbcltdl\"]\n # odbccr\n self.cpp_info.components[\"odbccr\"].names[\"pkg_config\"] = \"odbccr\"\n self.cpp_info.components[\"odbccr\"].libs = [\"odbccr\"]\n\n self.cpp_info.components[\"odbcltdl\"].libs = [\"ltdl\"]\n\n if self.settings.os == \"Linux\":\n self.cpp_info.components[\"_odbc\"].system_libs = [\"pthread\"]\n self.cpp_info.components[\"odbcinst\"].system_libs = [\"pthread\"]\n self.cpp_info.components[\"odbcltdl\"].system_libs = [\"dl\"]\n\n bin_path = os.path.join(self.package_folder, \"bin\")\n self.output.info(\"Appending PATH environment variable: {}\".format(bin_path))\n self.env_info.PATH.append(bin_path)\n", "path": "recipes/odbc/all/conanfile.py"}]}
| 1,859 | 182 |
gh_patches_debug_10406
|
rasdani/github-patches
|
git_diff
|
iterative__dvc-9925
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
unnecessary remote repository fetches
DVC has started cloning external repositories even on cases where it's not strictly required.
Anything that uses `outs_graph` has started cloning external repositories.
Commands like `add/plots modify/pull/push/fetch/move` are affected, and maybe potentially others.
### Reproduce
```bash
git clone https://github.com/iterative/example-get-started.git
cd example-get-started
dvc add foo
```
### Expected
No cloning occurs.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `dvc/repo/graph.py`
Content:
```
1 from typing import TYPE_CHECKING, Any, Iterator, List, Optional, Set, TypeVar
2
3 from dvc.fs import localfs
4 from dvc.utils.fs import path_isin
5
6 if TYPE_CHECKING:
7 from networkx import DiGraph
8
9 from dvc.stage import Stage
10
11 T = TypeVar("T")
12
13
14 def check_acyclic(graph: "DiGraph") -> None:
15 import networkx as nx
16
17 from dvc.exceptions import CyclicGraphError
18
19 try:
20 edges = nx.find_cycle(graph, orientation="original")
21 except nx.NetworkXNoCycle:
22 return
23
24 stages: Set["Stage"] = set()
25 for from_node, to_node, _ in edges:
26 stages.add(from_node)
27 stages.add(to_node)
28
29 raise CyclicGraphError(list(stages))
30
31
32 def get_pipeline(pipelines, node):
33 found = [i for i in pipelines if i.has_node(node)]
34 if not found:
35 return None
36
37 assert len(found) == 1
38 return found[0]
39
40
41 def get_pipelines(graph: "DiGraph"):
42 import networkx as nx
43
44 return [graph.subgraph(c).copy() for c in nx.weakly_connected_components(graph)]
45
46
47 def get_subgraph_of_nodes(
48 graph: "DiGraph", sources: Optional[List[Any]] = None, downstream: bool = False
49 ) -> "DiGraph":
50 from networkx import dfs_postorder_nodes, reverse_view
51
52 if not sources:
53 return graph
54
55 g = reverse_view(graph) if downstream else graph
56 nodes = []
57 for source in sources:
58 nodes.extend(dfs_postorder_nodes(g, source))
59 return graph.subgraph(nodes)
60
61
62 def collect_pipeline(stage: "Stage", graph: "DiGraph") -> Iterator["Stage"]:
63 import networkx as nx
64
65 pipeline = get_pipeline(get_pipelines(graph), stage)
66 if not pipeline:
67 return iter([])
68
69 return nx.dfs_postorder_nodes(pipeline, stage)
70
71
72 def collect_inside_path(path: str, graph: "DiGraph") -> List["Stage"]:
73 import networkx as nx
74
75 stages = nx.dfs_postorder_nodes(graph)
76 return [stage for stage in stages if path_isin(stage.path, path)]
77
78
79 def build_graph(stages, outs_trie=None):
80 """Generate a graph by using the given stages on the given directory
81
82 The nodes of the graph are the stage's path relative to the root.
83
84 Edges are created when the output of one stage is used as a
85 dependency in other stage.
86
87 The direction of the edges goes from the stage to its dependency:
88
89 For example, running the following:
90
91 $ dvc run -o A "echo A > A"
92 $ dvc run -d A -o B "echo B > B"
93 $ dvc run -d B -o C "echo C > C"
94
95 Will create the following graph:
96
97 ancestors <--
98 |
99 C.dvc -> B.dvc -> A.dvc
100 | |
101 | --> descendants
102 |
103 ------- pipeline ------>
104 |
105 v
106 (weakly connected components)
107
108 Args:
109 stages (list): used to build a graph from
110
111 Raises:
112 OutputDuplicationError: two outputs with the same path
113 StagePathAsOutputError: stage inside an output directory
114 OverlappingOutputPathsError: output inside output directory
115 CyclicGraphError: resulting graph has cycles
116 """
117 import networkx as nx
118
119 from dvc.exceptions import StagePathAsOutputError
120
121 from .trie import build_outs_trie
122
123 graph = nx.DiGraph()
124
125 # Use trie to efficiently find overlapping outs and deps
126 outs_trie = outs_trie or build_outs_trie(stages)
127
128 for stage in stages:
129 out = outs_trie.shortest_prefix(localfs.path.parts(stage.path)).value
130 if out:
131 raise StagePathAsOutputError(stage, str(out))
132
133 # Building graph
134 graph.add_nodes_from(stages)
135 for stage in stages:
136 if stage.is_repo_import:
137 continue
138
139 for dep in stage.deps:
140 dep_key = dep.fs.path.parts(dep.fs_path)
141 overlapping = [n.value for n in outs_trie.prefixes(dep_key)]
142 if outs_trie.has_subtrie(dep_key):
143 overlapping.extend(outs_trie.values(prefix=dep_key))
144
145 graph.add_edges_from((stage, out.stage) for out in overlapping)
146 check_acyclic(graph)
147
148 return graph
149
150
151 # NOTE: using stage graph instead of just list of stages to make sure that it
152 # has already passed all the sanity checks like cycles/overlapping outputs and
153 # so on.
154 def build_outs_graph(graph, outs_trie):
155 import networkx as nx
156
157 outs_graph = nx.DiGraph()
158
159 outs_graph.add_nodes_from(outs_trie.values())
160 for stage in graph.nodes():
161 for dep in stage.deps:
162 if dep.fs_path is None:
163 # RepoDependency don't have a path
164 continue
165 dep_key = dep.fs.path.parts(dep.fs_path)
166 overlapping = [n.value for n in outs_trie.prefixes(dep_key)]
167 if outs_trie.has_subtrie(dep_key):
168 overlapping.extend(outs_trie.values(prefix=dep_key))
169
170 for from_out in stage.outs:
171 outs_graph.add_edges_from((from_out, out) for out in overlapping)
172 return outs_graph
173
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/dvc/repo/graph.py b/dvc/repo/graph.py
--- a/dvc/repo/graph.py
+++ b/dvc/repo/graph.py
@@ -158,10 +158,9 @@
outs_graph.add_nodes_from(outs_trie.values())
for stage in graph.nodes():
+ if stage.is_repo_import:
+ continue
for dep in stage.deps:
- if dep.fs_path is None:
- # RepoDependency don't have a path
- continue
dep_key = dep.fs.path.parts(dep.fs_path)
overlapping = [n.value for n in outs_trie.prefixes(dep_key)]
if outs_trie.has_subtrie(dep_key):
|
{"golden_diff": "diff --git a/dvc/repo/graph.py b/dvc/repo/graph.py\n--- a/dvc/repo/graph.py\n+++ b/dvc/repo/graph.py\n@@ -158,10 +158,9 @@\n \n outs_graph.add_nodes_from(outs_trie.values())\n for stage in graph.nodes():\n+ if stage.is_repo_import:\n+ continue\n for dep in stage.deps:\n- if dep.fs_path is None:\n- # RepoDependency don't have a path\n- continue\n dep_key = dep.fs.path.parts(dep.fs_path)\n overlapping = [n.value for n in outs_trie.prefixes(dep_key)]\n if outs_trie.has_subtrie(dep_key):\n", "issue": "unnecessary remote repository fetches\nDVC has started cloning external repositories even on cases where it's not strictly required. \r\nAnything that uses `outs_graph` has started cloning external repositories.\r\n\r\nCommands like `add/plots modify/pull/push/fetch/move` are affected, and maybe potentially others.\r\n\r\n### Reproduce\r\n\r\n```bash\r\ngit clone https://github.com/iterative/example-get-started.git\r\ncd example-get-started\r\ndvc add foo\r\n```\r\n\r\n### Expected\r\n\r\nNo cloning occurs.\r\n\r\n\r\n\n", "before_files": [{"content": "from typing import TYPE_CHECKING, Any, Iterator, List, Optional, Set, TypeVar\n\nfrom dvc.fs import localfs\nfrom dvc.utils.fs import path_isin\n\nif TYPE_CHECKING:\n from networkx import DiGraph\n\n from dvc.stage import Stage\n\nT = TypeVar(\"T\")\n\n\ndef check_acyclic(graph: \"DiGraph\") -> None:\n import networkx as nx\n\n from dvc.exceptions import CyclicGraphError\n\n try:\n edges = nx.find_cycle(graph, orientation=\"original\")\n except nx.NetworkXNoCycle:\n return\n\n stages: Set[\"Stage\"] = set()\n for from_node, to_node, _ in edges:\n stages.add(from_node)\n stages.add(to_node)\n\n raise CyclicGraphError(list(stages))\n\n\ndef get_pipeline(pipelines, node):\n found = [i for i in pipelines if i.has_node(node)]\n if not found:\n return None\n\n assert len(found) == 1\n return found[0]\n\n\ndef get_pipelines(graph: \"DiGraph\"):\n import networkx as nx\n\n return [graph.subgraph(c).copy() for c in nx.weakly_connected_components(graph)]\n\n\ndef get_subgraph_of_nodes(\n graph: \"DiGraph\", sources: Optional[List[Any]] = None, downstream: bool = False\n) -> \"DiGraph\":\n from networkx import dfs_postorder_nodes, reverse_view\n\n if not sources:\n return graph\n\n g = reverse_view(graph) if downstream else graph\n nodes = []\n for source in sources:\n nodes.extend(dfs_postorder_nodes(g, source))\n return graph.subgraph(nodes)\n\n\ndef collect_pipeline(stage: \"Stage\", graph: \"DiGraph\") -> Iterator[\"Stage\"]:\n import networkx as nx\n\n pipeline = get_pipeline(get_pipelines(graph), stage)\n if not pipeline:\n return iter([])\n\n return nx.dfs_postorder_nodes(pipeline, stage)\n\n\ndef collect_inside_path(path: str, graph: \"DiGraph\") -> List[\"Stage\"]:\n import networkx as nx\n\n stages = nx.dfs_postorder_nodes(graph)\n return [stage for stage in stages if path_isin(stage.path, path)]\n\n\ndef build_graph(stages, outs_trie=None):\n \"\"\"Generate a graph by using the given stages on the given directory\n\n The nodes of the graph are the stage's path relative to the root.\n\n Edges are created when the output of one stage is used as a\n dependency in other stage.\n\n The direction of the edges goes from the stage to its dependency:\n\n For example, running the following:\n\n $ dvc run -o A \"echo A > A\"\n $ dvc run -d A -o B \"echo B > B\"\n $ dvc run -d B -o C \"echo C > C\"\n\n Will create the following graph:\n\n ancestors <--\n |\n C.dvc -> B.dvc -> A.dvc\n | |\n | --> descendants\n |\n ------- pipeline ------>\n |\n v\n (weakly connected components)\n\n Args:\n stages (list): used to build a graph from\n\n Raises:\n OutputDuplicationError: two outputs with the same path\n StagePathAsOutputError: stage inside an output directory\n OverlappingOutputPathsError: output inside output directory\n CyclicGraphError: resulting graph has cycles\n \"\"\"\n import networkx as nx\n\n from dvc.exceptions import StagePathAsOutputError\n\n from .trie import build_outs_trie\n\n graph = nx.DiGraph()\n\n # Use trie to efficiently find overlapping outs and deps\n outs_trie = outs_trie or build_outs_trie(stages)\n\n for stage in stages:\n out = outs_trie.shortest_prefix(localfs.path.parts(stage.path)).value\n if out:\n raise StagePathAsOutputError(stage, str(out))\n\n # Building graph\n graph.add_nodes_from(stages)\n for stage in stages:\n if stage.is_repo_import:\n continue\n\n for dep in stage.deps:\n dep_key = dep.fs.path.parts(dep.fs_path)\n overlapping = [n.value for n in outs_trie.prefixes(dep_key)]\n if outs_trie.has_subtrie(dep_key):\n overlapping.extend(outs_trie.values(prefix=dep_key))\n\n graph.add_edges_from((stage, out.stage) for out in overlapping)\n check_acyclic(graph)\n\n return graph\n\n\n# NOTE: using stage graph instead of just list of stages to make sure that it\n# has already passed all the sanity checks like cycles/overlapping outputs and\n# so on.\ndef build_outs_graph(graph, outs_trie):\n import networkx as nx\n\n outs_graph = nx.DiGraph()\n\n outs_graph.add_nodes_from(outs_trie.values())\n for stage in graph.nodes():\n for dep in stage.deps:\n if dep.fs_path is None:\n # RepoDependency don't have a path\n continue\n dep_key = dep.fs.path.parts(dep.fs_path)\n overlapping = [n.value for n in outs_trie.prefixes(dep_key)]\n if outs_trie.has_subtrie(dep_key):\n overlapping.extend(outs_trie.values(prefix=dep_key))\n\n for from_out in stage.outs:\n outs_graph.add_edges_from((from_out, out) for out in overlapping)\n return outs_graph\n", "path": "dvc/repo/graph.py"}], "after_files": [{"content": "from typing import TYPE_CHECKING, Any, Iterator, List, Optional, Set, TypeVar\n\nfrom dvc.fs import localfs\nfrom dvc.utils.fs import path_isin\n\nif TYPE_CHECKING:\n from networkx import DiGraph\n\n from dvc.stage import Stage\n\nT = TypeVar(\"T\")\n\n\ndef check_acyclic(graph: \"DiGraph\") -> None:\n import networkx as nx\n\n from dvc.exceptions import CyclicGraphError\n\n try:\n edges = nx.find_cycle(graph, orientation=\"original\")\n except nx.NetworkXNoCycle:\n return\n\n stages: Set[\"Stage\"] = set()\n for from_node, to_node, _ in edges:\n stages.add(from_node)\n stages.add(to_node)\n\n raise CyclicGraphError(list(stages))\n\n\ndef get_pipeline(pipelines, node):\n found = [i for i in pipelines if i.has_node(node)]\n if not found:\n return None\n\n assert len(found) == 1\n return found[0]\n\n\ndef get_pipelines(graph: \"DiGraph\"):\n import networkx as nx\n\n return [graph.subgraph(c).copy() for c in nx.weakly_connected_components(graph)]\n\n\ndef get_subgraph_of_nodes(\n graph: \"DiGraph\", sources: Optional[List[Any]] = None, downstream: bool = False\n) -> \"DiGraph\":\n from networkx import dfs_postorder_nodes, reverse_view\n\n if not sources:\n return graph\n\n g = reverse_view(graph) if downstream else graph\n nodes = []\n for source in sources:\n nodes.extend(dfs_postorder_nodes(g, source))\n return graph.subgraph(nodes)\n\n\ndef collect_pipeline(stage: \"Stage\", graph: \"DiGraph\") -> Iterator[\"Stage\"]:\n import networkx as nx\n\n pipeline = get_pipeline(get_pipelines(graph), stage)\n if not pipeline:\n return iter([])\n\n return nx.dfs_postorder_nodes(pipeline, stage)\n\n\ndef collect_inside_path(path: str, graph: \"DiGraph\") -> List[\"Stage\"]:\n import networkx as nx\n\n stages = nx.dfs_postorder_nodes(graph)\n return [stage for stage in stages if path_isin(stage.path, path)]\n\n\ndef build_graph(stages, outs_trie=None):\n \"\"\"Generate a graph by using the given stages on the given directory\n\n The nodes of the graph are the stage's path relative to the root.\n\n Edges are created when the output of one stage is used as a\n dependency in other stage.\n\n The direction of the edges goes from the stage to its dependency:\n\n For example, running the following:\n\n $ dvc run -o A \"echo A > A\"\n $ dvc run -d A -o B \"echo B > B\"\n $ dvc run -d B -o C \"echo C > C\"\n\n Will create the following graph:\n\n ancestors <--\n |\n C.dvc -> B.dvc -> A.dvc\n | |\n | --> descendants\n |\n ------- pipeline ------>\n |\n v\n (weakly connected components)\n\n Args:\n stages (list): used to build a graph from\n\n Raises:\n OutputDuplicationError: two outputs with the same path\n StagePathAsOutputError: stage inside an output directory\n OverlappingOutputPathsError: output inside output directory\n CyclicGraphError: resulting graph has cycles\n \"\"\"\n import networkx as nx\n\n from dvc.exceptions import StagePathAsOutputError\n\n from .trie import build_outs_trie\n\n graph = nx.DiGraph()\n\n # Use trie to efficiently find overlapping outs and deps\n outs_trie = outs_trie or build_outs_trie(stages)\n\n for stage in stages:\n out = outs_trie.shortest_prefix(localfs.path.parts(stage.path)).value\n if out:\n raise StagePathAsOutputError(stage, str(out))\n\n # Building graph\n graph.add_nodes_from(stages)\n for stage in stages:\n if stage.is_repo_import:\n continue\n\n for dep in stage.deps:\n dep_key = dep.fs.path.parts(dep.fs_path)\n overlapping = [n.value for n in outs_trie.prefixes(dep_key)]\n if outs_trie.has_subtrie(dep_key):\n overlapping.extend(outs_trie.values(prefix=dep_key))\n\n graph.add_edges_from((stage, out.stage) for out in overlapping)\n check_acyclic(graph)\n\n return graph\n\n\n# NOTE: using stage graph instead of just list of stages to make sure that it\n# has already passed all the sanity checks like cycles/overlapping outputs and\n# so on.\ndef build_outs_graph(graph, outs_trie):\n import networkx as nx\n\n outs_graph = nx.DiGraph()\n\n outs_graph.add_nodes_from(outs_trie.values())\n for stage in graph.nodes():\n if stage.is_repo_import:\n continue\n for dep in stage.deps:\n dep_key = dep.fs.path.parts(dep.fs_path)\n overlapping = [n.value for n in outs_trie.prefixes(dep_key)]\n if outs_trie.has_subtrie(dep_key):\n overlapping.extend(outs_trie.values(prefix=dep_key))\n\n for from_out in stage.outs:\n outs_graph.add_edges_from((from_out, out) for out in overlapping)\n return outs_graph\n", "path": "dvc/repo/graph.py"}]}
| 1,973 | 155 |
gh_patches_debug_7854
|
rasdani/github-patches
|
git_diff
|
coala__coala-4969
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add docstring for coala_modes.py mode_normal
This function should have a proper docstring, and indicate that `log_printer` is unused & deprecated.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `coalib/coala_modes.py`
Content:
```
1 def mode_normal(console_printer, log_printer, args, debug=False):
2 import functools
3
4 from coalib.coala_main import run_coala
5 from coalib.output.ConsoleInteraction import (
6 acquire_settings, nothing_done,
7 print_results, print_section_beginning)
8
9 partial_print_sec_beg = functools.partial(
10 print_section_beginning,
11 console_printer)
12 results, exitcode, _ = run_coala(
13 print_results=print_results,
14 acquire_settings=acquire_settings,
15 print_section_beginning=partial_print_sec_beg,
16 nothing_done=nothing_done,
17 console_printer=console_printer,
18 args=args,
19 debug=debug)
20
21 return exitcode
22
23
24 def mode_non_interactive(console_printer, args, debug=False):
25 import functools
26
27 from coalib.coala_main import run_coala
28 from coalib.output.ConsoleInteraction import (
29 print_results_no_input, print_section_beginning)
30
31 partial_print_sec_beg = functools.partial(
32 print_section_beginning,
33 console_printer)
34 results, exitcode, _ = run_coala(
35 print_results=print_results_no_input,
36 print_section_beginning=partial_print_sec_beg,
37 force_show_patch=True,
38 console_printer=console_printer,
39 args=args,
40 debug=debug)
41
42 return exitcode
43
44
45 def mode_json(args, debug=False):
46 import json
47
48 from coalib.coala_main import run_coala
49 from coalib.output.Logging import configure_json_logging
50 from coalib.output.JSONEncoder import create_json_encoder
51
52 if args.log_json:
53 log_stream = configure_json_logging()
54
55 JSONEncoder = create_json_encoder(use_relpath=args.relpath)
56
57 results, exitcode, _ = run_coala(args=args, debug=debug)
58
59 retval = {'results': results}
60
61 if args.log_json:
62 retval['logs'] = [json.loads(line) for line in
63 log_stream.getvalue().splitlines()]
64
65 if args.output:
66 filename = str(args.output[0])
67 with open(filename, 'w') as fp:
68 json.dump(retval, fp,
69 cls=JSONEncoder,
70 sort_keys=True,
71 indent=2,
72 separators=(',', ': '))
73 else:
74 print(json.dumps(retval,
75 cls=JSONEncoder,
76 sort_keys=True,
77 indent=2,
78 separators=(',', ': ')))
79
80 return 0 if args.show_bears else exitcode
81
82
83 def mode_format(args, debug=False):
84 from coalib.coala_main import run_coala
85 from coalib.output.ConsoleInteraction import print_results_formatted
86
87 _, exitcode, _ = run_coala(
88 print_results=print_results_formatted, args=args, debug=debug)
89 return exitcode
90
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/coalib/coala_modes.py b/coalib/coala_modes.py
--- a/coalib/coala_modes.py
+++ b/coalib/coala_modes.py
@@ -1,4 +1,13 @@
def mode_normal(console_printer, log_printer, args, debug=False):
+ """
+ This is the default coala mode. User interaction is allowed in this mode.
+
+ :param console_printer: Object to print messages on the console.
+ :param log_printer: Deprecated.
+ :param args: Alternative pre-parsed CLI arguments.
+ :param debug: Run in debug mode, bypassing multiprocessing,
+ and not catching any exceptions.
+ """
import functools
from coalib.coala_main import run_coala
|
{"golden_diff": "diff --git a/coalib/coala_modes.py b/coalib/coala_modes.py\n--- a/coalib/coala_modes.py\n+++ b/coalib/coala_modes.py\n@@ -1,4 +1,13 @@\n def mode_normal(console_printer, log_printer, args, debug=False):\n+ \"\"\"\n+ This is the default coala mode. User interaction is allowed in this mode.\n+\n+ :param console_printer: Object to print messages on the console.\n+ :param log_printer: Deprecated.\n+ :param args: Alternative pre-parsed CLI arguments.\n+ :param debug: Run in debug mode, bypassing multiprocessing,\n+ and not catching any exceptions.\n+ \"\"\"\n import functools\n \n from coalib.coala_main import run_coala\n", "issue": "Add docstring for coala_modes.py mode_normal\nThis function should have a proper docstring, and indicate that `log_printer` is unused & deprecated.\n", "before_files": [{"content": "def mode_normal(console_printer, log_printer, args, debug=False):\n import functools\n\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import (\n acquire_settings, nothing_done,\n print_results, print_section_beginning)\n\n partial_print_sec_beg = functools.partial(\n print_section_beginning,\n console_printer)\n results, exitcode, _ = run_coala(\n print_results=print_results,\n acquire_settings=acquire_settings,\n print_section_beginning=partial_print_sec_beg,\n nothing_done=nothing_done,\n console_printer=console_printer,\n args=args,\n debug=debug)\n\n return exitcode\n\n\ndef mode_non_interactive(console_printer, args, debug=False):\n import functools\n\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import (\n print_results_no_input, print_section_beginning)\n\n partial_print_sec_beg = functools.partial(\n print_section_beginning,\n console_printer)\n results, exitcode, _ = run_coala(\n print_results=print_results_no_input,\n print_section_beginning=partial_print_sec_beg,\n force_show_patch=True,\n console_printer=console_printer,\n args=args,\n debug=debug)\n\n return exitcode\n\n\ndef mode_json(args, debug=False):\n import json\n\n from coalib.coala_main import run_coala\n from coalib.output.Logging import configure_json_logging\n from coalib.output.JSONEncoder import create_json_encoder\n\n if args.log_json:\n log_stream = configure_json_logging()\n\n JSONEncoder = create_json_encoder(use_relpath=args.relpath)\n\n results, exitcode, _ = run_coala(args=args, debug=debug)\n\n retval = {'results': results}\n\n if args.log_json:\n retval['logs'] = [json.loads(line) for line in\n log_stream.getvalue().splitlines()]\n\n if args.output:\n filename = str(args.output[0])\n with open(filename, 'w') as fp:\n json.dump(retval, fp,\n cls=JSONEncoder,\n sort_keys=True,\n indent=2,\n separators=(',', ': '))\n else:\n print(json.dumps(retval,\n cls=JSONEncoder,\n sort_keys=True,\n indent=2,\n separators=(',', ': ')))\n\n return 0 if args.show_bears else exitcode\n\n\ndef mode_format(args, debug=False):\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import print_results_formatted\n\n _, exitcode, _ = run_coala(\n print_results=print_results_formatted, args=args, debug=debug)\n return exitcode\n", "path": "coalib/coala_modes.py"}], "after_files": [{"content": "def mode_normal(console_printer, log_printer, args, debug=False):\n \"\"\"\n This is the default coala mode. User interaction is allowed in this mode.\n\n :param console_printer: Object to print messages on the console.\n :param log_printer: Deprecated.\n :param args: Alternative pre-parsed CLI arguments.\n :param debug: Run in debug mode, bypassing multiprocessing,\n and not catching any exceptions.\n \"\"\"\n import functools\n\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import (\n acquire_settings, nothing_done,\n print_results, print_section_beginning)\n\n partial_print_sec_beg = functools.partial(\n print_section_beginning,\n console_printer)\n results, exitcode, _ = run_coala(\n print_results=print_results,\n acquire_settings=acquire_settings,\n print_section_beginning=partial_print_sec_beg,\n nothing_done=nothing_done,\n console_printer=console_printer,\n args=args,\n debug=debug)\n\n return exitcode\n\n\ndef mode_non_interactive(console_printer, args, debug=False):\n import functools\n\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import (\n print_results_no_input, print_section_beginning)\n\n partial_print_sec_beg = functools.partial(\n print_section_beginning,\n console_printer)\n results, exitcode, _ = run_coala(\n print_results=print_results_no_input,\n print_section_beginning=partial_print_sec_beg,\n force_show_patch=True,\n console_printer=console_printer,\n args=args,\n debug=debug)\n\n return exitcode\n\n\ndef mode_json(args, debug=False):\n import json\n\n from coalib.coala_main import run_coala\n from coalib.output.Logging import configure_json_logging\n from coalib.output.JSONEncoder import create_json_encoder\n\n if args.log_json:\n log_stream = configure_json_logging()\n\n JSONEncoder = create_json_encoder(use_relpath=args.relpath)\n\n results, exitcode, _ = run_coala(args=args, debug=debug)\n\n retval = {'results': results}\n\n if args.log_json:\n retval['logs'] = [json.loads(line) for line in\n log_stream.getvalue().splitlines()]\n\n if args.output:\n filename = str(args.output[0])\n with open(filename, 'w') as fp:\n json.dump(retval, fp,\n cls=JSONEncoder,\n sort_keys=True,\n indent=2,\n separators=(',', ': '))\n else:\n print(json.dumps(retval,\n cls=JSONEncoder,\n sort_keys=True,\n indent=2,\n separators=(',', ': ')))\n\n return 0 if args.show_bears else exitcode\n\n\ndef mode_format(args, debug=False):\n from coalib.coala_main import run_coala\n from coalib.output.ConsoleInteraction import print_results_formatted\n\n _, exitcode, _ = run_coala(\n print_results=print_results_formatted, args=args, debug=debug)\n return exitcode\n", "path": "coalib/coala_modes.py"}]}
| 1,043 | 169 |
gh_patches_debug_24831
|
rasdani/github-patches
|
git_diff
|
sunpy__sunpy-4388
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
32 bit tests failing
The 32 bit tests are currently failing, due to a failed attempt to install the `cryptography` library from source. See e.g. https://dev.azure.com/sunpy/sunpy/_build/results?buildId=9154&view=logs&jobId=a7b3aa55-7d57-562f-3433-7f6b2d4252da&j=2920fa54-56c6-5dfd-d3d8-7b1b694d90cb&t=bba4d609-c264-5f9c-e072-020a8a3b573f
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `sunpy/util/util.py`
Content:
```
1 """
2 This module provides general utility functions.
3 """
4 import os
5 import hashlib
6 from shutil import get_terminal_size
7 from itertools import chain, count
8 from collections import UserList
9
10 __all__ = ['unique', 'replacement_filename', 'expand_list',
11 'expand_list_generator', 'dict_keys_same', 'hash_file', "get_width"]
12
13
14 def unique(itr, key=None):
15 """
16 Return only unique elements of a sequence.
17
18 Parameters
19 ----------
20 itr : `iterable`
21 Any iterable sequence.
22 key : `function`, optional
23 A function to apply to each element in the iterable. Defaults to `None`.
24
25 Returns
26 -------
27 `set`:
28 A `set` of each unique element.
29 """
30 items = set()
31 if key is None:
32 for elem in itr:
33 if elem not in items:
34 yield elem
35 items.add(elem)
36 else:
37 for elem in itr:
38 x = key(elem)
39 if x not in items:
40 yield elem
41 items.add(x)
42
43
44 def replacement_filename(path):
45 """
46 Return a replacement path if input path is currently in use.
47
48 Enumerates until an unused filename is found, e.g., "foo.fits" becomes
49 "foo.0.fits", if that is used, "foo.1.fits" and so on.
50
51 Parameters
52 ----------
53 path : `str`
54 A string path.
55
56 Returns
57 -------
58 `str`:
59 A string path.
60 """
61 if not os.path.exists(path):
62 return path
63 else:
64 dir_, filename = os.path.split(path)
65 base, ext = os.path.splitext(filename)
66 for c in count():
67 name = base + '.' + str(c) + ext
68 newpath = os.path.join(dir_, name)
69 if not os.path.exists(newpath):
70 return newpath
71
72
73 def expand_list(inp):
74 """
75 Expand a list of lists or tuples.
76
77 Parameters
78 ----------
79 inp : `list`, `tuple`, `collections.UserList`
80 The iterable to expand.
81
82 Returns
83 -------
84 `list`
85 A flat list consisting of the entries of the input.
86
87 References
88 ----------
89 * https://stackoverflow.com/questions/2185822/expanding-elements-in-a-list/2185971#2185971
90 """
91 return [item for item in expand_list_generator(inp)]
92
93
94 def expand_list_generator(inp):
95 for item in inp:
96 if isinstance(item, (tuple, list, UserList)):
97 yield from expand_list_generator(item)
98 else:
99 yield item
100
101
102 def partial_key_match(key, dictionary):
103 """
104 Return value/values from a dictionary based on a partial key.
105
106 Each element of the partial key is matched against the keys of the dictionary and
107 if a partial match is found the value of the key is returned.
108
109 Even a partial match works here i.e even if the key matches partially a value is returned.
110
111 Parameters
112 ----------
113 key : `tuple`
114 A tuple containing the partial key.
115 dictionary: `dict`
116 The target dictionary from which we want to retrieve the value based on the partial key.
117
118 Yields
119 ------
120 value
121 The value of the matched key.
122
123 References
124 ----------
125 * https://stackoverflow.com/questions/18893624/partial-match-dictionary-keyof-tuples-in-python
126
127 Examples
128 ----------
129 >>> d = {('abc','def','ghi') : 1, ('abc', 'def', 'xyz') : 2, ('pqr', 'lmn', 'tuv') : 3}
130 >>> list(partial_key_match(('abc', 'def', None), d))
131 [1, 2]
132 """
133 for k, v in dictionary.items():
134 if all(k1 == k2 or k2 is None for k1, k2 in zip(k, key)):
135 yield v
136
137
138 def dict_keys_same(list_of_dicts):
139 """
140 Makes sure that a list of dictionaries all have the same keys.
141
142 If a key is missing, it will be added but with a value of None.
143
144 Parameters
145 ----------
146 list_of_dicts : `list` of `dict`
147 A list containing each dictonary to parse.
148
149 Returns
150 ------
151 `list`
152 The list with each dict updated.
153
154 References
155 ----------
156 * https://stackoverflow.com/questions/10482439/make-sure-all-dicts-in-a-list-have-the-same-keys
157
158 Examples
159 ----------
160 >>> l = [{'x': 42}, {'x': 23, 'y': 5}]
161 >>> dict_keys_same(l)
162 [{'x': 42, 'y': None}, {'x': 23, 'y': 5}]
163 """
164 all_keys = set(chain.from_iterable(list_of_dicts))
165 for d in list_of_dicts:
166 d.update({key: None for key in all_keys if key not in d})
167 return list_of_dicts
168
169
170 def hash_file(path):
171 """
172 Returns the SHA-256 hash of a file.
173
174 Parameters
175 ----------
176 path: `str`
177 The path of the file to be hashed.
178
179 Returns
180 -------
181 `str`
182 SHA-256 hash of the file.
183
184 References
185 ----------
186 * https://stackoverflow.com/a/22058673
187 """
188 BUF_SIZE = 65536
189 sha256 = hashlib.sha256()
190
191 with open(path, 'rb') as f:
192 while True:
193 data = f.read(BUF_SIZE)
194 if not data:
195 break
196 sha256.update(data)
197
198 return sha256.hexdigest()
199
200
201 def get_width():
202 """
203 Gets the width of the current terminal.
204 Accounts for if the 'COLUMNS' environmental variable is set.
205
206 Returns
207 -------
208 `int`
209 Width of the terminal you are in.
210 Works for IPython notebooks and normal terminals.
211 """
212 width = os.environ.get("COLUMNS", None)
213 if width:
214 width = int(width)
215 else:
216 width, _ = get_terminal_size()
217 return width
218
```
Path: `sunpy/data/data_manager/cache.py`
Content:
```
1 import os
2 from pathlib import Path
3 from datetime import datetime
4 from warnings import warn
5 from urllib.request import urlopen
6
7 import astropy.units as u
8 from astropy.time import TimeDelta
9
10 from sunpy.time import parse_time
11 from sunpy.util.exceptions import SunpyUserWarning
12 from sunpy.util.net import get_filename
13 from sunpy.util.util import hash_file, replacement_filename
14
15 __all__ = ['Cache']
16
17
18 class Cache:
19 """
20 Cache provides a way to download and cache files.
21
22 Parameters
23 ----------
24 downloader: Implementation of `~sunpy.data.data_manager.downloader.DownloaderBase`
25 Downloader object for downloading remote files.
26 storage: Implementation of `~sunpy.data.data_manager.storage.StorageProviderBase`
27 Storage to store metadata about the files.
28 cache_dir: `str` or `pathlib.Path`
29 Directory where the downloaded files will be stored.
30 expiry: `astropy.units.quantity.Quantity` or `None`, optional
31 The interval after which the cache is invalidated. If the expiry is `None`,
32 then the expiry is not checked. Defaults to 10 days.
33 """
34
35 def __init__(self, downloader, storage, cache_dir, expiry=10*u.day):
36 self._downloader = downloader
37 self._storage = storage
38 self._cache_dir = Path(cache_dir)
39 self._expiry = expiry if expiry is None else TimeDelta(expiry)
40
41 def download(self, urls, redownload=False):
42 """
43 Downloads the files from the urls.
44
45 The overall flow of this function is:
46 1. If ``redownload``: Download, update cache and return file path.
47 2. If not ``redownload``: Check cache,
48 i. If present in cache:
49 - If cache has expired, remove the entry from cache, download and add to cache
50 - If cache has not expired, return the path
51
52 Parameters
53 ----------
54 urls: `list` or `str`
55 A list of urls or a single url.
56 redownload: `bool`
57 Whether to skip cache and redownload.
58
59 Returns
60 -------
61 `pathlib.PosixPath`
62 Path to the downloaded file.
63 """
64 if isinstance(urls, str):
65 urls = [urls]
66 # Program flow
67 # 1. If redownload: Download, update cache and return file path
68 # 2. If not redownload: Check cache,
69 # i. If present in cache:
70 # - If cache expired, remove entry from cache, download and add to cache
71 # - If cache not expired, return path
72 details = None
73 for url in urls:
74 details = self._get_by_url(url)
75 if details:
76 break
77 if details:
78 if redownload or self._has_expired(details):
79 # if file is in cache and it has to be redownloaded or the cache has expired
80 # then remove the file and delete the details from the storage
81 os.remove(details['file_path'])
82 self._storage.delete_by_key('url', details['url'])
83 else:
84 return Path(details['file_path'])
85
86 file_path, file_hash, url = self._download_and_hash(urls)
87
88 self._storage.store({
89 'file_hash': file_hash,
90 'file_path': str(file_path),
91 'url': url,
92 'time': datetime.now().isoformat(),
93 })
94 return file_path
95
96 def _has_expired(self, details):
97 """
98 Whether the url corresponding to details in cache has expired or not.
99
100 Parameters
101 ----------
102 details: `dict`
103 Details detached from cache.
104
105 Returns
106 -------
107 `bool`
108 Whether the url has expired or not.
109 """
110 time = details.get("time", datetime.now().isoformat())
111
112 # TODO: Remove this once we depend on Python >=3.7
113 if hasattr(datetime, "fromisoformat"):
114 time = datetime.fromisoformat(time)
115 else:
116 time = parse_time(time).datetime
117 return self._expiry and datetime.now() - time > self._expiry
118
119 def get_by_hash(self, sha_hash):
120 """
121 Returns the details which is matched by hash if present in cache.
122
123 Parameters
124 ----------
125 sha_hash: `str`
126 SHA-256 hash of the file.
127 """
128 details = self._storage.find_by_key('file_hash', sha_hash)
129 return details
130
131 def _get_by_url(self, url):
132 """
133 Returns the details which is matched by url if present in cache.
134
135 Parameters
136 ----------
137 url: `str`
138 URL of the file.
139 """
140 details = self._storage.find_by_key('url', url)
141 return details
142
143 def _download_and_hash(self, urls):
144 """
145 Downloads the file and returns the path, hash and url it used to download.
146
147 Parameters
148 ----------
149 urls: `list`
150 List of urls.
151
152 Returns
153 -------
154 `str`, `str`, `str`
155 Path, hash and URL of the file.
156 """
157 def download(url):
158 path = self._cache_dir / get_filename(urlopen(url), url)
159 # replacement_filename returns a string and we want a Path object
160 path = Path(replacement_filename(path))
161 self._downloader.download(url, path)
162 shahash = hash_file(path)
163 return path, shahash, url
164
165 for url in urls:
166 try:
167 return download(url)
168 except Exception as e:
169 warn(e, SunpyUserWarning)
170 error = e
171 else:
172 raise RuntimeError from error
173
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/sunpy/data/data_manager/cache.py b/sunpy/data/data_manager/cache.py
--- a/sunpy/data/data_manager/cache.py
+++ b/sunpy/data/data_manager/cache.py
@@ -157,16 +157,17 @@
def download(url):
path = self._cache_dir / get_filename(urlopen(url), url)
# replacement_filename returns a string and we want a Path object
- path = Path(replacement_filename(path))
+ path = Path(replacement_filename(str(path)))
self._downloader.download(url, path)
shahash = hash_file(path)
return path, shahash, url
+ errors = []
for url in urls:
try:
return download(url)
except Exception as e:
- warn(e, SunpyUserWarning)
- error = e
+ warn(f"{e}", SunpyUserWarning)
+ errors.append(f"{e}")
else:
- raise RuntimeError from error
+ raise RuntimeError(errors)
diff --git a/sunpy/util/util.py b/sunpy/util/util.py
--- a/sunpy/util/util.py
+++ b/sunpy/util/util.py
@@ -41,7 +41,7 @@
items.add(x)
-def replacement_filename(path):
+def replacement_filename(path: str) -> str:
"""
Return a replacement path if input path is currently in use.
|
{"golden_diff": "diff --git a/sunpy/data/data_manager/cache.py b/sunpy/data/data_manager/cache.py\n--- a/sunpy/data/data_manager/cache.py\n+++ b/sunpy/data/data_manager/cache.py\n@@ -157,16 +157,17 @@\n def download(url):\n path = self._cache_dir / get_filename(urlopen(url), url)\n # replacement_filename returns a string and we want a Path object\n- path = Path(replacement_filename(path))\n+ path = Path(replacement_filename(str(path)))\n self._downloader.download(url, path)\n shahash = hash_file(path)\n return path, shahash, url\n \n+ errors = []\n for url in urls:\n try:\n return download(url)\n except Exception as e:\n- warn(e, SunpyUserWarning)\n- error = e\n+ warn(f\"{e}\", SunpyUserWarning)\n+ errors.append(f\"{e}\")\n else:\n- raise RuntimeError from error\n+ raise RuntimeError(errors)\ndiff --git a/sunpy/util/util.py b/sunpy/util/util.py\n--- a/sunpy/util/util.py\n+++ b/sunpy/util/util.py\n@@ -41,7 +41,7 @@\n items.add(x)\n \n \n-def replacement_filename(path):\n+def replacement_filename(path: str) -> str:\n \"\"\"\n Return a replacement path if input path is currently in use.\n", "issue": "32 bit tests failing\nThe 32 bit tests are currently failing, due to a failed attempt to install the `cryptography` library from source. See e.g. https://dev.azure.com/sunpy/sunpy/_build/results?buildId=9154&view=logs&jobId=a7b3aa55-7d57-562f-3433-7f6b2d4252da&j=2920fa54-56c6-5dfd-d3d8-7b1b694d90cb&t=bba4d609-c264-5f9c-e072-020a8a3b573f\n", "before_files": [{"content": "\"\"\"\nThis module provides general utility functions.\n\"\"\"\nimport os\nimport hashlib\nfrom shutil import get_terminal_size\nfrom itertools import chain, count\nfrom collections import UserList\n\n__all__ = ['unique', 'replacement_filename', 'expand_list',\n 'expand_list_generator', 'dict_keys_same', 'hash_file', \"get_width\"]\n\n\ndef unique(itr, key=None):\n \"\"\"\n Return only unique elements of a sequence.\n\n Parameters\n ----------\n itr : `iterable`\n Any iterable sequence.\n key : `function`, optional\n A function to apply to each element in the iterable. Defaults to `None`.\n\n Returns\n -------\n `set`:\n A `set` of each unique element.\n \"\"\"\n items = set()\n if key is None:\n for elem in itr:\n if elem not in items:\n yield elem\n items.add(elem)\n else:\n for elem in itr:\n x = key(elem)\n if x not in items:\n yield elem\n items.add(x)\n\n\ndef replacement_filename(path):\n \"\"\"\n Return a replacement path if input path is currently in use.\n\n Enumerates until an unused filename is found, e.g., \"foo.fits\" becomes\n \"foo.0.fits\", if that is used, \"foo.1.fits\" and so on.\n\n Parameters\n ----------\n path : `str`\n A string path.\n\n Returns\n -------\n `str`:\n A string path.\n \"\"\"\n if not os.path.exists(path):\n return path\n else:\n dir_, filename = os.path.split(path)\n base, ext = os.path.splitext(filename)\n for c in count():\n name = base + '.' + str(c) + ext\n newpath = os.path.join(dir_, name)\n if not os.path.exists(newpath):\n return newpath\n\n\ndef expand_list(inp):\n \"\"\"\n Expand a list of lists or tuples.\n\n Parameters\n ----------\n inp : `list`, `tuple`, `collections.UserList`\n The iterable to expand.\n\n Returns\n -------\n `list`\n A flat list consisting of the entries of the input.\n\n References\n ----------\n * https://stackoverflow.com/questions/2185822/expanding-elements-in-a-list/2185971#2185971\n \"\"\"\n return [item for item in expand_list_generator(inp)]\n\n\ndef expand_list_generator(inp):\n for item in inp:\n if isinstance(item, (tuple, list, UserList)):\n yield from expand_list_generator(item)\n else:\n yield item\n\n\ndef partial_key_match(key, dictionary):\n \"\"\"\n Return value/values from a dictionary based on a partial key.\n\n Each element of the partial key is matched against the keys of the dictionary and\n if a partial match is found the value of the key is returned.\n\n Even a partial match works here i.e even if the key matches partially a value is returned.\n\n Parameters\n ----------\n key : `tuple`\n A tuple containing the partial key.\n dictionary: `dict`\n The target dictionary from which we want to retrieve the value based on the partial key.\n\n Yields\n ------\n value\n The value of the matched key.\n\n References\n ----------\n * https://stackoverflow.com/questions/18893624/partial-match-dictionary-keyof-tuples-in-python\n\n Examples\n ----------\n >>> d = {('abc','def','ghi') : 1, ('abc', 'def', 'xyz') : 2, ('pqr', 'lmn', 'tuv') : 3}\n >>> list(partial_key_match(('abc', 'def', None), d))\n [1, 2]\n \"\"\"\n for k, v in dictionary.items():\n if all(k1 == k2 or k2 is None for k1, k2 in zip(k, key)):\n yield v\n\n\ndef dict_keys_same(list_of_dicts):\n \"\"\"\n Makes sure that a list of dictionaries all have the same keys.\n\n If a key is missing, it will be added but with a value of None.\n\n Parameters\n ----------\n list_of_dicts : `list` of `dict`\n A list containing each dictonary to parse.\n\n Returns\n ------\n `list`\n The list with each dict updated.\n\n References\n ----------\n * https://stackoverflow.com/questions/10482439/make-sure-all-dicts-in-a-list-have-the-same-keys\n\n Examples\n ----------\n >>> l = [{'x': 42}, {'x': 23, 'y': 5}]\n >>> dict_keys_same(l)\n [{'x': 42, 'y': None}, {'x': 23, 'y': 5}]\n \"\"\"\n all_keys = set(chain.from_iterable(list_of_dicts))\n for d in list_of_dicts:\n d.update({key: None for key in all_keys if key not in d})\n return list_of_dicts\n\n\ndef hash_file(path):\n \"\"\"\n Returns the SHA-256 hash of a file.\n\n Parameters\n ----------\n path: `str`\n The path of the file to be hashed.\n\n Returns\n -------\n `str`\n SHA-256 hash of the file.\n\n References\n ----------\n * https://stackoverflow.com/a/22058673\n \"\"\"\n BUF_SIZE = 65536\n sha256 = hashlib.sha256()\n\n with open(path, 'rb') as f:\n while True:\n data = f.read(BUF_SIZE)\n if not data:\n break\n sha256.update(data)\n\n return sha256.hexdigest()\n\n\ndef get_width():\n \"\"\"\n Gets the width of the current terminal.\n Accounts for if the 'COLUMNS' environmental variable is set.\n\n Returns\n -------\n `int`\n Width of the terminal you are in.\n Works for IPython notebooks and normal terminals.\n \"\"\"\n width = os.environ.get(\"COLUMNS\", None)\n if width:\n width = int(width)\n else:\n width, _ = get_terminal_size()\n return width\n", "path": "sunpy/util/util.py"}, {"content": "import os\nfrom pathlib import Path\nfrom datetime import datetime\nfrom warnings import warn\nfrom urllib.request import urlopen\n\nimport astropy.units as u\nfrom astropy.time import TimeDelta\n\nfrom sunpy.time import parse_time\nfrom sunpy.util.exceptions import SunpyUserWarning\nfrom sunpy.util.net import get_filename\nfrom sunpy.util.util import hash_file, replacement_filename\n\n__all__ = ['Cache']\n\n\nclass Cache:\n \"\"\"\n Cache provides a way to download and cache files.\n\n Parameters\n ----------\n downloader: Implementation of `~sunpy.data.data_manager.downloader.DownloaderBase`\n Downloader object for downloading remote files.\n storage: Implementation of `~sunpy.data.data_manager.storage.StorageProviderBase`\n Storage to store metadata about the files.\n cache_dir: `str` or `pathlib.Path`\n Directory where the downloaded files will be stored.\n expiry: `astropy.units.quantity.Quantity` or `None`, optional\n The interval after which the cache is invalidated. If the expiry is `None`,\n then the expiry is not checked. Defaults to 10 days.\n \"\"\"\n\n def __init__(self, downloader, storage, cache_dir, expiry=10*u.day):\n self._downloader = downloader\n self._storage = storage\n self._cache_dir = Path(cache_dir)\n self._expiry = expiry if expiry is None else TimeDelta(expiry)\n\n def download(self, urls, redownload=False):\n \"\"\"\n Downloads the files from the urls.\n\n The overall flow of this function is:\n 1. If ``redownload``: Download, update cache and return file path.\n 2. If not ``redownload``: Check cache,\n i. If present in cache:\n - If cache has expired, remove the entry from cache, download and add to cache\n - If cache has not expired, return the path\n\n Parameters\n ----------\n urls: `list` or `str`\n A list of urls or a single url.\n redownload: `bool`\n Whether to skip cache and redownload.\n\n Returns\n -------\n `pathlib.PosixPath`\n Path to the downloaded file.\n \"\"\"\n if isinstance(urls, str):\n urls = [urls]\n # Program flow\n # 1. If redownload: Download, update cache and return file path\n # 2. If not redownload: Check cache,\n # i. If present in cache:\n # - If cache expired, remove entry from cache, download and add to cache\n # - If cache not expired, return path\n details = None\n for url in urls:\n details = self._get_by_url(url)\n if details:\n break\n if details:\n if redownload or self._has_expired(details):\n # if file is in cache and it has to be redownloaded or the cache has expired\n # then remove the file and delete the details from the storage\n os.remove(details['file_path'])\n self._storage.delete_by_key('url', details['url'])\n else:\n return Path(details['file_path'])\n\n file_path, file_hash, url = self._download_and_hash(urls)\n\n self._storage.store({\n 'file_hash': file_hash,\n 'file_path': str(file_path),\n 'url': url,\n 'time': datetime.now().isoformat(),\n })\n return file_path\n\n def _has_expired(self, details):\n \"\"\"\n Whether the url corresponding to details in cache has expired or not.\n\n Parameters\n ----------\n details: `dict`\n Details detached from cache.\n\n Returns\n -------\n `bool`\n Whether the url has expired or not.\n \"\"\"\n time = details.get(\"time\", datetime.now().isoformat())\n\n # TODO: Remove this once we depend on Python >=3.7\n if hasattr(datetime, \"fromisoformat\"):\n time = datetime.fromisoformat(time)\n else:\n time = parse_time(time).datetime\n return self._expiry and datetime.now() - time > self._expiry\n\n def get_by_hash(self, sha_hash):\n \"\"\"\n Returns the details which is matched by hash if present in cache.\n\n Parameters\n ----------\n sha_hash: `str`\n SHA-256 hash of the file.\n \"\"\"\n details = self._storage.find_by_key('file_hash', sha_hash)\n return details\n\n def _get_by_url(self, url):\n \"\"\"\n Returns the details which is matched by url if present in cache.\n\n Parameters\n ----------\n url: `str`\n URL of the file.\n \"\"\"\n details = self._storage.find_by_key('url', url)\n return details\n\n def _download_and_hash(self, urls):\n \"\"\"\n Downloads the file and returns the path, hash and url it used to download.\n\n Parameters\n ----------\n urls: `list`\n List of urls.\n\n Returns\n -------\n `str`, `str`, `str`\n Path, hash and URL of the file.\n \"\"\"\n def download(url):\n path = self._cache_dir / get_filename(urlopen(url), url)\n # replacement_filename returns a string and we want a Path object\n path = Path(replacement_filename(path))\n self._downloader.download(url, path)\n shahash = hash_file(path)\n return path, shahash, url\n\n for url in urls:\n try:\n return download(url)\n except Exception as e:\n warn(e, SunpyUserWarning)\n error = e\n else:\n raise RuntimeError from error\n", "path": "sunpy/data/data_manager/cache.py"}], "after_files": [{"content": "\"\"\"\nThis module provides general utility functions.\n\"\"\"\nimport os\nimport hashlib\nfrom shutil import get_terminal_size\nfrom itertools import chain, count\nfrom collections import UserList\n\n__all__ = ['unique', 'replacement_filename', 'expand_list',\n 'expand_list_generator', 'dict_keys_same', 'hash_file', \"get_width\"]\n\n\ndef unique(itr, key=None):\n \"\"\"\n Return only unique elements of a sequence.\n\n Parameters\n ----------\n itr : `iterable`\n Any iterable sequence.\n key : `function`, optional\n A function to apply to each element in the iterable. Defaults to `None`.\n\n Returns\n -------\n `set`:\n A `set` of each unique element.\n \"\"\"\n items = set()\n if key is None:\n for elem in itr:\n if elem not in items:\n yield elem\n items.add(elem)\n else:\n for elem in itr:\n x = key(elem)\n if x not in items:\n yield elem\n items.add(x)\n\n\ndef replacement_filename(path: str) -> str:\n \"\"\"\n Return a replacement path if input path is currently in use.\n\n Enumerates until an unused filename is found, e.g., \"foo.fits\" becomes\n \"foo.0.fits\", if that is used, \"foo.1.fits\" and so on.\n\n Parameters\n ----------\n path : `str`\n A string path.\n\n Returns\n -------\n `str`:\n A string path.\n \"\"\"\n if not os.path.exists(path):\n return path\n else:\n dir_, filename = os.path.split(path)\n base, ext = os.path.splitext(filename)\n for c in count():\n name = base + '.' + str(c) + ext\n newpath = os.path.join(dir_, name)\n if not os.path.exists(newpath):\n return newpath\n\n\ndef expand_list(inp):\n \"\"\"\n Expand a list of lists or tuples.\n\n Parameters\n ----------\n inp : `list`, `tuple`, `collections.UserList`\n The iterable to expand.\n\n Returns\n -------\n `list`\n A flat list consisting of the entries of the input.\n\n References\n ----------\n * https://stackoverflow.com/questions/2185822/expanding-elements-in-a-list/2185971#2185971\n \"\"\"\n return [item for item in expand_list_generator(inp)]\n\n\ndef expand_list_generator(inp):\n for item in inp:\n if isinstance(item, (tuple, list, UserList)):\n yield from expand_list_generator(item)\n else:\n yield item\n\n\ndef partial_key_match(key, dictionary):\n \"\"\"\n Return value/values from a dictionary based on a partial key.\n\n Each element of the partial key is matched against the keys of the dictionary and\n if a partial match is found the value of the key is returned.\n\n Even a partial match works here i.e even if the key matches partially a value is returned.\n\n Parameters\n ----------\n key : `tuple`\n A tuple containing the partial key.\n dictionary: `dict`\n The target dictionary from which we want to retrieve the value based on the partial key.\n\n Yields\n ------\n value\n The value of the matched key.\n\n References\n ----------\n * https://stackoverflow.com/questions/18893624/partial-match-dictionary-keyof-tuples-in-python\n\n Examples\n ----------\n >>> d = {('abc','def','ghi') : 1, ('abc', 'def', 'xyz') : 2, ('pqr', 'lmn', 'tuv') : 3}\n >>> list(partial_key_match(('abc', 'def', None), d))\n [1, 2]\n \"\"\"\n for k, v in dictionary.items():\n if all(k1 == k2 or k2 is None for k1, k2 in zip(k, key)):\n yield v\n\n\ndef dict_keys_same(list_of_dicts):\n \"\"\"\n Makes sure that a list of dictionaries all have the same keys.\n\n If a key is missing, it will be added but with a value of None.\n\n Parameters\n ----------\n list_of_dicts : `list` of `dict`\n A list containing each dictonary to parse.\n\n Returns\n ------\n `list`\n The list with each dict updated.\n\n References\n ----------\n * https://stackoverflow.com/questions/10482439/make-sure-all-dicts-in-a-list-have-the-same-keys\n\n Examples\n ----------\n >>> l = [{'x': 42}, {'x': 23, 'y': 5}]\n >>> dict_keys_same(l)\n [{'x': 42, 'y': None}, {'x': 23, 'y': 5}]\n \"\"\"\n all_keys = set(chain.from_iterable(list_of_dicts))\n for d in list_of_dicts:\n d.update({key: None for key in all_keys if key not in d})\n return list_of_dicts\n\n\ndef hash_file(path):\n \"\"\"\n Returns the SHA-256 hash of a file.\n\n Parameters\n ----------\n path: `str`\n The path of the file to be hashed.\n\n Returns\n -------\n `str`\n SHA-256 hash of the file.\n\n References\n ----------\n * https://stackoverflow.com/a/22058673\n \"\"\"\n BUF_SIZE = 65536\n sha256 = hashlib.sha256()\n\n with open(path, 'rb') as f:\n while True:\n data = f.read(BUF_SIZE)\n if not data:\n break\n sha256.update(data)\n\n return sha256.hexdigest()\n\n\ndef get_width():\n \"\"\"\n Gets the width of the current terminal.\n Accounts for if the 'COLUMNS' environmental variable is set.\n\n Returns\n -------\n `int`\n Width of the terminal you are in.\n Works for IPython notebooks and normal terminals.\n \"\"\"\n width = os.environ.get(\"COLUMNS\", None)\n if width:\n width = int(width)\n else:\n width, _ = get_terminal_size()\n return width\n", "path": "sunpy/util/util.py"}, {"content": "import os\nfrom pathlib import Path\nfrom datetime import datetime\nfrom warnings import warn\nfrom urllib.request import urlopen\n\nimport astropy.units as u\nfrom astropy.time import TimeDelta\n\nfrom sunpy.time import parse_time\nfrom sunpy.util.exceptions import SunpyUserWarning\nfrom sunpy.util.net import get_filename\nfrom sunpy.util.util import hash_file, replacement_filename\n\n__all__ = ['Cache']\n\n\nclass Cache:\n \"\"\"\n Cache provides a way to download and cache files.\n\n Parameters\n ----------\n downloader: Implementation of `~sunpy.data.data_manager.downloader.DownloaderBase`\n Downloader object for downloading remote files.\n storage: Implementation of `~sunpy.data.data_manager.storage.StorageProviderBase`\n Storage to store metadata about the files.\n cache_dir: `str` or `pathlib.Path`\n Directory where the downloaded files will be stored.\n expiry: `astropy.units.quantity.Quantity` or `None`, optional\n The interval after which the cache is invalidated. If the expiry is `None`,\n then the expiry is not checked. Defaults to 10 days.\n \"\"\"\n\n def __init__(self, downloader, storage, cache_dir, expiry=10*u.day):\n self._downloader = downloader\n self._storage = storage\n self._cache_dir = Path(cache_dir)\n self._expiry = expiry if expiry is None else TimeDelta(expiry)\n\n def download(self, urls, redownload=False):\n \"\"\"\n Downloads the files from the urls.\n\n The overall flow of this function is:\n 1. If ``redownload``: Download, update cache and return file path.\n 2. If not ``redownload``: Check cache,\n i. If present in cache:\n - If cache has expired, remove the entry from cache, download and add to cache\n - If cache has not expired, return the path\n\n Parameters\n ----------\n urls: `list` or `str`\n A list of urls or a single url.\n redownload: `bool`\n Whether to skip cache and redownload.\n\n Returns\n -------\n `pathlib.PosixPath`\n Path to the downloaded file.\n \"\"\"\n if isinstance(urls, str):\n urls = [urls]\n # Program flow\n # 1. If redownload: Download, update cache and return file path\n # 2. If not redownload: Check cache,\n # i. If present in cache:\n # - If cache expired, remove entry from cache, download and add to cache\n # - If cache not expired, return path\n details = None\n for url in urls:\n details = self._get_by_url(url)\n if details:\n break\n if details:\n if redownload or self._has_expired(details):\n # if file is in cache and it has to be redownloaded or the cache has expired\n # then remove the file and delete the details from the storage\n os.remove(details['file_path'])\n self._storage.delete_by_key('url', details['url'])\n else:\n return Path(details['file_path'])\n\n file_path, file_hash, url = self._download_and_hash(urls)\n\n self._storage.store({\n 'file_hash': file_hash,\n 'file_path': str(file_path),\n 'url': url,\n 'time': datetime.now().isoformat(),\n })\n return file_path\n\n def _has_expired(self, details):\n \"\"\"\n Whether the url corresponding to details in cache has expired or not.\n\n Parameters\n ----------\n details: `dict`\n Details detached from cache.\n\n Returns\n -------\n `bool`\n Whether the url has expired or not.\n \"\"\"\n time = details.get(\"time\", datetime.now().isoformat())\n\n # TODO: Remove this once we depend on Python >=3.7\n if hasattr(datetime, \"fromisoformat\"):\n time = datetime.fromisoformat(time)\n else:\n time = parse_time(time).datetime\n return self._expiry and datetime.now() - time > self._expiry\n\n def get_by_hash(self, sha_hash):\n \"\"\"\n Returns the details which is matched by hash if present in cache.\n\n Parameters\n ----------\n sha_hash: `str`\n SHA-256 hash of the file.\n \"\"\"\n details = self._storage.find_by_key('file_hash', sha_hash)\n return details\n\n def _get_by_url(self, url):\n \"\"\"\n Returns the details which is matched by url if present in cache.\n\n Parameters\n ----------\n url: `str`\n URL of the file.\n \"\"\"\n details = self._storage.find_by_key('url', url)\n return details\n\n def _download_and_hash(self, urls):\n \"\"\"\n Downloads the file and returns the path, hash and url it used to download.\n\n Parameters\n ----------\n urls: `list`\n List of urls.\n\n Returns\n -------\n `str`, `str`, `str`\n Path, hash and URL of the file.\n \"\"\"\n def download(url):\n path = self._cache_dir / get_filename(urlopen(url), url)\n # replacement_filename returns a string and we want a Path object\n path = Path(replacement_filename(str(path)))\n self._downloader.download(url, path)\n shahash = hash_file(path)\n return path, shahash, url\n\n errors = []\n for url in urls:\n try:\n return download(url)\n except Exception as e:\n warn(f\"{e}\", SunpyUserWarning)\n errors.append(f\"{e}\")\n else:\n raise RuntimeError(errors)\n", "path": "sunpy/data/data_manager/cache.py"}]}
| 3,972 | 307 |
gh_patches_debug_22901
|
rasdani/github-patches
|
git_diff
|
getsentry__sentry-python-381
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
celery integration RecursionError
Hi there,
I upgraded sentry_sdk to 0.7.0 and started getting RecursionError if there's an issue with celery task. Sentry record doesn't contain any stack trace for that but found that error in my apm system (can attach screenshot only, text data is a real mess there). I'm running celery 4.2.1 on Ubuntu 18.
<img width="618" alt="2019-02-14 15 04 54" src="https://user-images.githubusercontent.com/7299611/52785969-89ea2180-306a-11e9-9828-56a52f3fddd6.png">
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `sentry_sdk/integrations/celery.py`
Content:
```
1 from __future__ import absolute_import
2
3 import sys
4
5 from celery.exceptions import ( # type: ignore
6 SoftTimeLimitExceeded,
7 Retry,
8 Ignore,
9 Reject,
10 )
11
12 from sentry_sdk.hub import Hub
13 from sentry_sdk.utils import capture_internal_exceptions, event_from_exception
14 from sentry_sdk.tracing import SpanContext
15 from sentry_sdk._compat import reraise
16 from sentry_sdk.integrations import Integration
17 from sentry_sdk.integrations.logging import ignore_logger
18
19
20 CELERY_CONTROL_FLOW_EXCEPTIONS = (Retry, Ignore, Reject)
21
22
23 class CeleryIntegration(Integration):
24 identifier = "celery"
25
26 def __init__(self, propagate_traces=True):
27 self.propagate_traces = propagate_traces
28
29 @staticmethod
30 def setup_once():
31 import celery.app.trace as trace # type: ignore
32
33 old_build_tracer = trace.build_tracer
34
35 def sentry_build_tracer(name, task, *args, **kwargs):
36 # Need to patch both methods because older celery sometimes
37 # short-circuits to task.run if it thinks it's safe.
38 task.__call__ = _wrap_task_call(task, task.__call__)
39 task.run = _wrap_task_call(task, task.run)
40 task.apply_async = _wrap_apply_async(task, task.apply_async)
41 return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs))
42
43 trace.build_tracer = sentry_build_tracer
44
45 _patch_worker_exit()
46
47 # This logger logs every status of every task that ran on the worker.
48 # Meaning that every task's breadcrumbs are full of stuff like "Task
49 # <foo> raised unexpected <bar>".
50 ignore_logger("celery.worker.job")
51
52
53 def _wrap_apply_async(task, f):
54 def apply_async(*args, **kwargs):
55 hub = Hub.current
56 integration = hub.get_integration(CeleryIntegration)
57 if integration is not None and integration.propagate_traces:
58 headers = None
59 for key, value in hub.iter_trace_propagation_headers():
60 if headers is None:
61 headers = dict(kwargs.get("headers") or {})
62 headers[key] = value
63 if headers is not None:
64 kwargs["headers"] = headers
65 return f(*args, **kwargs)
66
67 return apply_async
68
69
70 def _wrap_tracer(task, f):
71 # Need to wrap tracer for pushing the scope before prerun is sent, and
72 # popping it after postrun is sent.
73 #
74 # This is the reason we don't use signals for hooking in the first place.
75 # Also because in Celery 3, signal dispatch returns early if one handler
76 # crashes.
77 def _inner(*args, **kwargs):
78 hub = Hub.current
79 if hub.get_integration(CeleryIntegration) is None:
80 return f(*args, **kwargs)
81
82 with hub.push_scope() as scope:
83 scope._name = "celery"
84 scope.clear_breadcrumbs()
85 _continue_trace(args[3].get("headers") or {}, scope)
86 scope.add_event_processor(_make_event_processor(task, *args, **kwargs))
87
88 return f(*args, **kwargs)
89
90 return _inner
91
92
93 def _continue_trace(headers, scope):
94 if headers:
95 span_context = SpanContext.continue_from_headers(headers)
96 else:
97 span_context = SpanContext.start_trace()
98 scope.set_span_context(span_context)
99
100
101 def _wrap_task_call(task, f):
102 # Need to wrap task call because the exception is caught before we get to
103 # see it. Also celery's reported stacktrace is untrustworthy.
104 def _inner(*args, **kwargs):
105 try:
106 return f(*args, **kwargs)
107 except Exception:
108 exc_info = sys.exc_info()
109 with capture_internal_exceptions():
110 _capture_exception(task, exc_info)
111 reraise(*exc_info)
112
113 return _inner
114
115
116 def _make_event_processor(task, uuid, args, kwargs, request=None):
117 def event_processor(event, hint):
118 with capture_internal_exceptions():
119 event["transaction"] = task.name
120
121 with capture_internal_exceptions():
122 extra = event.setdefault("extra", {})
123 extra["celery-job"] = {
124 "task_name": task.name,
125 "args": args,
126 "kwargs": kwargs,
127 }
128
129 if "exc_info" in hint:
130 with capture_internal_exceptions():
131 if issubclass(hint["exc_info"][0], SoftTimeLimitExceeded):
132 event["fingerprint"] = [
133 "celery",
134 "SoftTimeLimitExceeded",
135 getattr(task, "name", task),
136 ]
137
138 return event
139
140 return event_processor
141
142
143 def _capture_exception(task, exc_info):
144 hub = Hub.current
145
146 if hub.get_integration(CeleryIntegration) is None:
147 return
148 if isinstance(exc_info[1], CELERY_CONTROL_FLOW_EXCEPTIONS):
149 return
150 if hasattr(task, "throws") and isinstance(exc_info[1], task.throws):
151 return
152
153 event, hint = event_from_exception(
154 exc_info,
155 client_options=hub.client.options,
156 mechanism={"type": "celery", "handled": False},
157 )
158
159 hub.capture_event(event, hint=hint)
160
161
162 def _patch_worker_exit():
163 # Need to flush queue before worker shutdown because a crashing worker will
164 # call os._exit
165 from billiard.pool import Worker # type: ignore
166
167 old_workloop = Worker.workloop
168
169 def sentry_workloop(*args, **kwargs):
170 try:
171 return old_workloop(*args, **kwargs)
172 finally:
173 with capture_internal_exceptions():
174 hub = Hub.current
175 if hub.get_integration(CeleryIntegration) is not None:
176 hub.flush()
177
178 Worker.workloop = sentry_workloop
179
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/sentry_sdk/integrations/celery.py b/sentry_sdk/integrations/celery.py
--- a/sentry_sdk/integrations/celery.py
+++ b/sentry_sdk/integrations/celery.py
@@ -33,11 +33,18 @@
old_build_tracer = trace.build_tracer
def sentry_build_tracer(name, task, *args, **kwargs):
- # Need to patch both methods because older celery sometimes
- # short-circuits to task.run if it thinks it's safe.
- task.__call__ = _wrap_task_call(task, task.__call__)
- task.run = _wrap_task_call(task, task.run)
- task.apply_async = _wrap_apply_async(task, task.apply_async)
+ if not getattr(task, "_sentry_is_patched", False):
+ # Need to patch both methods because older celery sometimes
+ # short-circuits to task.run if it thinks it's safe.
+ task.__call__ = _wrap_task_call(task, task.__call__)
+ task.run = _wrap_task_call(task, task.run)
+ task.apply_async = _wrap_apply_async(task, task.apply_async)
+
+ # `build_tracer` is apparently called for every task
+ # invocation. Can't wrap every celery task for every invocation
+ # or we will get infinitely nested wrapper functions.
+ task._sentry_is_patched = True
+
return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs))
trace.build_tracer = sentry_build_tracer
|
{"golden_diff": "diff --git a/sentry_sdk/integrations/celery.py b/sentry_sdk/integrations/celery.py\n--- a/sentry_sdk/integrations/celery.py\n+++ b/sentry_sdk/integrations/celery.py\n@@ -33,11 +33,18 @@\n old_build_tracer = trace.build_tracer\n \n def sentry_build_tracer(name, task, *args, **kwargs):\n- # Need to patch both methods because older celery sometimes\n- # short-circuits to task.run if it thinks it's safe.\n- task.__call__ = _wrap_task_call(task, task.__call__)\n- task.run = _wrap_task_call(task, task.run)\n- task.apply_async = _wrap_apply_async(task, task.apply_async)\n+ if not getattr(task, \"_sentry_is_patched\", False):\n+ # Need to patch both methods because older celery sometimes\n+ # short-circuits to task.run if it thinks it's safe.\n+ task.__call__ = _wrap_task_call(task, task.__call__)\n+ task.run = _wrap_task_call(task, task.run)\n+ task.apply_async = _wrap_apply_async(task, task.apply_async)\n+\n+ # `build_tracer` is apparently called for every task\n+ # invocation. Can't wrap every celery task for every invocation\n+ # or we will get infinitely nested wrapper functions.\n+ task._sentry_is_patched = True\n+\n return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs))\n \n trace.build_tracer = sentry_build_tracer\n", "issue": "celery integration RecursionError\nHi there,\r\nI upgraded sentry_sdk to 0.7.0 and started getting RecursionError if there's an issue with celery task. Sentry record doesn't contain any stack trace for that but found that error in my apm system (can attach screenshot only, text data is a real mess there). I'm running celery 4.2.1 on Ubuntu 18. \r\n\r\n<img width=\"618\" alt=\"2019-02-14 15 04 54\" src=\"https://user-images.githubusercontent.com/7299611/52785969-89ea2180-306a-11e9-9828-56a52f3fddd6.png\">\r\n\n", "before_files": [{"content": "from __future__ import absolute_import\n\nimport sys\n\nfrom celery.exceptions import ( # type: ignore\n SoftTimeLimitExceeded,\n Retry,\n Ignore,\n Reject,\n)\n\nfrom sentry_sdk.hub import Hub\nfrom sentry_sdk.utils import capture_internal_exceptions, event_from_exception\nfrom sentry_sdk.tracing import SpanContext\nfrom sentry_sdk._compat import reraise\nfrom sentry_sdk.integrations import Integration\nfrom sentry_sdk.integrations.logging import ignore_logger\n\n\nCELERY_CONTROL_FLOW_EXCEPTIONS = (Retry, Ignore, Reject)\n\n\nclass CeleryIntegration(Integration):\n identifier = \"celery\"\n\n def __init__(self, propagate_traces=True):\n self.propagate_traces = propagate_traces\n\n @staticmethod\n def setup_once():\n import celery.app.trace as trace # type: ignore\n\n old_build_tracer = trace.build_tracer\n\n def sentry_build_tracer(name, task, *args, **kwargs):\n # Need to patch both methods because older celery sometimes\n # short-circuits to task.run if it thinks it's safe.\n task.__call__ = _wrap_task_call(task, task.__call__)\n task.run = _wrap_task_call(task, task.run)\n task.apply_async = _wrap_apply_async(task, task.apply_async)\n return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs))\n\n trace.build_tracer = sentry_build_tracer\n\n _patch_worker_exit()\n\n # This logger logs every status of every task that ran on the worker.\n # Meaning that every task's breadcrumbs are full of stuff like \"Task\n # <foo> raised unexpected <bar>\".\n ignore_logger(\"celery.worker.job\")\n\n\ndef _wrap_apply_async(task, f):\n def apply_async(*args, **kwargs):\n hub = Hub.current\n integration = hub.get_integration(CeleryIntegration)\n if integration is not None and integration.propagate_traces:\n headers = None\n for key, value in hub.iter_trace_propagation_headers():\n if headers is None:\n headers = dict(kwargs.get(\"headers\") or {})\n headers[key] = value\n if headers is not None:\n kwargs[\"headers\"] = headers\n return f(*args, **kwargs)\n\n return apply_async\n\n\ndef _wrap_tracer(task, f):\n # Need to wrap tracer for pushing the scope before prerun is sent, and\n # popping it after postrun is sent.\n #\n # This is the reason we don't use signals for hooking in the first place.\n # Also because in Celery 3, signal dispatch returns early if one handler\n # crashes.\n def _inner(*args, **kwargs):\n hub = Hub.current\n if hub.get_integration(CeleryIntegration) is None:\n return f(*args, **kwargs)\n\n with hub.push_scope() as scope:\n scope._name = \"celery\"\n scope.clear_breadcrumbs()\n _continue_trace(args[3].get(\"headers\") or {}, scope)\n scope.add_event_processor(_make_event_processor(task, *args, **kwargs))\n\n return f(*args, **kwargs)\n\n return _inner\n\n\ndef _continue_trace(headers, scope):\n if headers:\n span_context = SpanContext.continue_from_headers(headers)\n else:\n span_context = SpanContext.start_trace()\n scope.set_span_context(span_context)\n\n\ndef _wrap_task_call(task, f):\n # Need to wrap task call because the exception is caught before we get to\n # see it. Also celery's reported stacktrace is untrustworthy.\n def _inner(*args, **kwargs):\n try:\n return f(*args, **kwargs)\n except Exception:\n exc_info = sys.exc_info()\n with capture_internal_exceptions():\n _capture_exception(task, exc_info)\n reraise(*exc_info)\n\n return _inner\n\n\ndef _make_event_processor(task, uuid, args, kwargs, request=None):\n def event_processor(event, hint):\n with capture_internal_exceptions():\n event[\"transaction\"] = task.name\n\n with capture_internal_exceptions():\n extra = event.setdefault(\"extra\", {})\n extra[\"celery-job\"] = {\n \"task_name\": task.name,\n \"args\": args,\n \"kwargs\": kwargs,\n }\n\n if \"exc_info\" in hint:\n with capture_internal_exceptions():\n if issubclass(hint[\"exc_info\"][0], SoftTimeLimitExceeded):\n event[\"fingerprint\"] = [\n \"celery\",\n \"SoftTimeLimitExceeded\",\n getattr(task, \"name\", task),\n ]\n\n return event\n\n return event_processor\n\n\ndef _capture_exception(task, exc_info):\n hub = Hub.current\n\n if hub.get_integration(CeleryIntegration) is None:\n return\n if isinstance(exc_info[1], CELERY_CONTROL_FLOW_EXCEPTIONS):\n return\n if hasattr(task, \"throws\") and isinstance(exc_info[1], task.throws):\n return\n\n event, hint = event_from_exception(\n exc_info,\n client_options=hub.client.options,\n mechanism={\"type\": \"celery\", \"handled\": False},\n )\n\n hub.capture_event(event, hint=hint)\n\n\ndef _patch_worker_exit():\n # Need to flush queue before worker shutdown because a crashing worker will\n # call os._exit\n from billiard.pool import Worker # type: ignore\n\n old_workloop = Worker.workloop\n\n def sentry_workloop(*args, **kwargs):\n try:\n return old_workloop(*args, **kwargs)\n finally:\n with capture_internal_exceptions():\n hub = Hub.current\n if hub.get_integration(CeleryIntegration) is not None:\n hub.flush()\n\n Worker.workloop = sentry_workloop\n", "path": "sentry_sdk/integrations/celery.py"}], "after_files": [{"content": "from __future__ import absolute_import\n\nimport sys\n\nfrom celery.exceptions import ( # type: ignore\n SoftTimeLimitExceeded,\n Retry,\n Ignore,\n Reject,\n)\n\nfrom sentry_sdk.hub import Hub\nfrom sentry_sdk.utils import capture_internal_exceptions, event_from_exception\nfrom sentry_sdk.tracing import SpanContext\nfrom sentry_sdk._compat import reraise\nfrom sentry_sdk.integrations import Integration\nfrom sentry_sdk.integrations.logging import ignore_logger\n\n\nCELERY_CONTROL_FLOW_EXCEPTIONS = (Retry, Ignore, Reject)\n\n\nclass CeleryIntegration(Integration):\n identifier = \"celery\"\n\n def __init__(self, propagate_traces=True):\n self.propagate_traces = propagate_traces\n\n @staticmethod\n def setup_once():\n import celery.app.trace as trace # type: ignore\n\n old_build_tracer = trace.build_tracer\n\n def sentry_build_tracer(name, task, *args, **kwargs):\n if not getattr(task, \"_sentry_is_patched\", False):\n # Need to patch both methods because older celery sometimes\n # short-circuits to task.run if it thinks it's safe.\n task.__call__ = _wrap_task_call(task, task.__call__)\n task.run = _wrap_task_call(task, task.run)\n task.apply_async = _wrap_apply_async(task, task.apply_async)\n\n # `build_tracer` is apparently called for every task\n # invocation. Can't wrap every celery task for every invocation\n # or we will get infinitely nested wrapper functions.\n task._sentry_is_patched = True\n\n return _wrap_tracer(task, old_build_tracer(name, task, *args, **kwargs))\n\n trace.build_tracer = sentry_build_tracer\n\n _patch_worker_exit()\n\n # This logger logs every status of every task that ran on the worker.\n # Meaning that every task's breadcrumbs are full of stuff like \"Task\n # <foo> raised unexpected <bar>\".\n ignore_logger(\"celery.worker.job\")\n\n\ndef _wrap_apply_async(task, f):\n def apply_async(*args, **kwargs):\n hub = Hub.current\n integration = hub.get_integration(CeleryIntegration)\n if integration is not None and integration.propagate_traces:\n headers = None\n for key, value in hub.iter_trace_propagation_headers():\n if headers is None:\n headers = dict(kwargs.get(\"headers\") or {})\n headers[key] = value\n if headers is not None:\n kwargs[\"headers\"] = headers\n return f(*args, **kwargs)\n\n return apply_async\n\n\ndef _wrap_tracer(task, f):\n # Need to wrap tracer for pushing the scope before prerun is sent, and\n # popping it after postrun is sent.\n #\n # This is the reason we don't use signals for hooking in the first place.\n # Also because in Celery 3, signal dispatch returns early if one handler\n # crashes.\n def _inner(*args, **kwargs):\n hub = Hub.current\n if hub.get_integration(CeleryIntegration) is None:\n return f(*args, **kwargs)\n\n with hub.push_scope() as scope:\n scope._name = \"celery\"\n scope.clear_breadcrumbs()\n _continue_trace(args[3].get(\"headers\") or {}, scope)\n scope.add_event_processor(_make_event_processor(task, *args, **kwargs))\n\n return f(*args, **kwargs)\n\n return _inner\n\n\ndef _continue_trace(headers, scope):\n if headers:\n span_context = SpanContext.continue_from_headers(headers)\n else:\n span_context = SpanContext.start_trace()\n scope.set_span_context(span_context)\n\n\ndef _wrap_task_call(task, f):\n # Need to wrap task call because the exception is caught before we get to\n # see it. Also celery's reported stacktrace is untrustworthy.\n def _inner(*args, **kwargs):\n try:\n return f(*args, **kwargs)\n except Exception:\n exc_info = sys.exc_info()\n with capture_internal_exceptions():\n _capture_exception(task, exc_info)\n reraise(*exc_info)\n\n return _inner\n\n\ndef _make_event_processor(task, uuid, args, kwargs, request=None):\n def event_processor(event, hint):\n with capture_internal_exceptions():\n event[\"transaction\"] = task.name\n\n with capture_internal_exceptions():\n extra = event.setdefault(\"extra\", {})\n extra[\"celery-job\"] = {\n \"task_name\": task.name,\n \"args\": args,\n \"kwargs\": kwargs,\n }\n\n if \"exc_info\" in hint:\n with capture_internal_exceptions():\n if issubclass(hint[\"exc_info\"][0], SoftTimeLimitExceeded):\n event[\"fingerprint\"] = [\n \"celery\",\n \"SoftTimeLimitExceeded\",\n getattr(task, \"name\", task),\n ]\n\n return event\n\n return event_processor\n\n\ndef _capture_exception(task, exc_info):\n hub = Hub.current\n\n if hub.get_integration(CeleryIntegration) is None:\n return\n if isinstance(exc_info[1], CELERY_CONTROL_FLOW_EXCEPTIONS):\n return\n if hasattr(task, \"throws\") and isinstance(exc_info[1], task.throws):\n return\n\n event, hint = event_from_exception(\n exc_info,\n client_options=hub.client.options,\n mechanism={\"type\": \"celery\", \"handled\": False},\n )\n\n hub.capture_event(event, hint=hint)\n\n\ndef _patch_worker_exit():\n # Need to flush queue before worker shutdown because a crashing worker will\n # call os._exit\n from billiard.pool import Worker # type: ignore\n\n old_workloop = Worker.workloop\n\n def sentry_workloop(*args, **kwargs):\n try:\n return old_workloop(*args, **kwargs)\n finally:\n with capture_internal_exceptions():\n hub = Hub.current\n if hub.get_integration(CeleryIntegration) is not None:\n hub.flush()\n\n Worker.workloop = sentry_workloop\n", "path": "sentry_sdk/integrations/celery.py"}]}
| 2,139 | 355 |
gh_patches_debug_26041
|
rasdani/github-patches
|
git_diff
|
comic__grand-challenge.org-2410
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add a time column to challenge requests table
**Is your feature request related to a problem? Please describe.**
Its not clear right away when each challenge request was submitted.
**Describe the solution you'd like**
I would like to add a column that shows the time when challenge request was submitted. It would be handy.
**Describe alternatives you've considered**
Looking at a specific challenge request page to see the time.
**Additional context**
Idea shown on the picture below. Color coding is not necessary, but could be nice.

--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `app/grandchallenge/challenges/admin.py`
Content:
```
1 from django.contrib import admin, messages
2 from django.contrib.admin import ModelAdmin
3 from django.core.exceptions import ValidationError
4
5 from grandchallenge.challenges.emails import send_challenge_status_update_email
6 from grandchallenge.challenges.models import (
7 Challenge,
8 ChallengeRequest,
9 ChallengeSeries,
10 ExternalChallenge,
11 )
12
13
14 class ChallengeAdmin(ModelAdmin):
15 readonly_fields = ("creator",)
16 autocomplete_fields = ("publications",)
17 ordering = ("-created",)
18 list_display = ("short_name", "created")
19 search_fields = ("short_name",)
20
21
22 class ExternalChallengeAdmin(ModelAdmin):
23 readonly_fields = ("creator",)
24 autocomplete_fields = ("publications",)
25
26
27 class ChallengeRequestAdmin(ModelAdmin):
28 readonly_fields = ("creator",)
29 ordering = ("-created",)
30 list_display = ("title", "short_name", "creator", "created", "status")
31 actions = ["create_challenge", "send_status_update_email"]
32
33 @admin.action(description="Create challenge for this request")
34 def create_challenge(self, request, queryset):
35 for challengerequest in queryset:
36 try:
37 challengerequest.create_challenge()
38 except ValidationError:
39 self.message_user(
40 request,
41 f"There already is a challenge with short "
42 f"name: {challengerequest.short_name}",
43 messages.WARNING,
44 )
45
46 @admin.action(description="Send status update email to requester")
47 def send_status_update_email(self, request, queryset):
48 for challengerequest in queryset:
49 if (
50 challengerequest.status
51 == challengerequest.ChallengeRequestStatusChoices.ACCEPTED
52 ):
53 try:
54 challenge = Challenge.objects.get(
55 short_name=challengerequest.short_name
56 )
57 except Challenge.DoesNotExist:
58 challenge = challengerequest.create_challenge()
59 else:
60 challenge = None
61 send_challenge_status_update_email(
62 challengerequest=challengerequest, challenge=challenge
63 )
64
65
66 admin.site.register(Challenge, ChallengeAdmin)
67 admin.site.register(ExternalChallenge, ExternalChallengeAdmin)
68 admin.site.register(ChallengeSeries)
69 admin.site.register(ChallengeRequest, ChallengeRequestAdmin)
70
```
Path: `app/grandchallenge/core/templatetags/naturaldelta.py`
Content:
```
1 import humanize
2 from django import template
3
4 register = template.Library()
5
6
7 @register.filter
8 def naturaldelta(value):
9 return humanize.naturaldelta(value, months=False)
10
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/app/grandchallenge/challenges/admin.py b/app/grandchallenge/challenges/admin.py
--- a/app/grandchallenge/challenges/admin.py
+++ b/app/grandchallenge/challenges/admin.py
@@ -27,8 +27,22 @@
class ChallengeRequestAdmin(ModelAdmin):
readonly_fields = ("creator",)
ordering = ("-created",)
- list_display = ("title", "short_name", "creator", "created", "status")
+ list_display = (
+ "title",
+ "short_name",
+ "creator",
+ "created",
+ "status",
+ "challenge_type",
+ "total_cost",
+ "budget_for_hosting_challenge",
+ )
actions = ["create_challenge", "send_status_update_email"]
+ list_filter = ["status", "challenge_type"]
+
+ @admin.display(description="Total cost")
+ def total_cost(self, obj):
+ return "%s" % (obj.budget["Total"])
@admin.action(description="Create challenge for this request")
def create_challenge(self, request, queryset):
diff --git a/app/grandchallenge/core/templatetags/naturaldelta.py b/app/grandchallenge/core/templatetags/naturaldelta.py
--- a/app/grandchallenge/core/templatetags/naturaldelta.py
+++ b/app/grandchallenge/core/templatetags/naturaldelta.py
@@ -1,5 +1,6 @@
import humanize
from django import template
+from django.utils import timezone
register = template.Library()
@@ -7,3 +8,8 @@
@register.filter
def naturaldelta(value):
return humanize.naturaldelta(value, months=False)
+
+
[email protected]
+def timedifference(value):
+ return (timezone.now() - value).days
|
{"golden_diff": "diff --git a/app/grandchallenge/challenges/admin.py b/app/grandchallenge/challenges/admin.py\n--- a/app/grandchallenge/challenges/admin.py\n+++ b/app/grandchallenge/challenges/admin.py\n@@ -27,8 +27,22 @@\n class ChallengeRequestAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n ordering = (\"-created\",)\n- list_display = (\"title\", \"short_name\", \"creator\", \"created\", \"status\")\n+ list_display = (\n+ \"title\",\n+ \"short_name\",\n+ \"creator\",\n+ \"created\",\n+ \"status\",\n+ \"challenge_type\",\n+ \"total_cost\",\n+ \"budget_for_hosting_challenge\",\n+ )\n actions = [\"create_challenge\", \"send_status_update_email\"]\n+ list_filter = [\"status\", \"challenge_type\"]\n+\n+ @admin.display(description=\"Total cost\")\n+ def total_cost(self, obj):\n+ return \"%s\" % (obj.budget[\"Total\"])\n \n @admin.action(description=\"Create challenge for this request\")\n def create_challenge(self, request, queryset):\ndiff --git a/app/grandchallenge/core/templatetags/naturaldelta.py b/app/grandchallenge/core/templatetags/naturaldelta.py\n--- a/app/grandchallenge/core/templatetags/naturaldelta.py\n+++ b/app/grandchallenge/core/templatetags/naturaldelta.py\n@@ -1,5 +1,6 @@\n import humanize\n from django import template\n+from django.utils import timezone\n \n register = template.Library()\n \n@@ -7,3 +8,8 @@\n @register.filter\n def naturaldelta(value):\n return humanize.naturaldelta(value, months=False)\n+\n+\[email protected]\n+def timedifference(value):\n+ return (timezone.now() - value).days\n", "issue": "Add a time column to challenge requests table\n**Is your feature request related to a problem? Please describe.**\r\nIts not clear right away when each challenge request was submitted.\r\n\r\n\r\n**Describe the solution you'd like**\r\nI would like to add a column that shows the time when challenge request was submitted. It would be handy. \r\n\r\n**Describe alternatives you've considered**\r\nLooking at a specific challenge request page to see the time.\r\n\r\n**Additional context**\r\nIdea shown on the picture below. Color coding is not necessary, but could be nice.\r\n\n", "before_files": [{"content": "from django.contrib import admin, messages\nfrom django.contrib.admin import ModelAdmin\nfrom django.core.exceptions import ValidationError\n\nfrom grandchallenge.challenges.emails import send_challenge_status_update_email\nfrom grandchallenge.challenges.models import (\n Challenge,\n ChallengeRequest,\n ChallengeSeries,\n ExternalChallenge,\n)\n\n\nclass ChallengeAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n autocomplete_fields = (\"publications\",)\n ordering = (\"-created\",)\n list_display = (\"short_name\", \"created\")\n search_fields = (\"short_name\",)\n\n\nclass ExternalChallengeAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n autocomplete_fields = (\"publications\",)\n\n\nclass ChallengeRequestAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n ordering = (\"-created\",)\n list_display = (\"title\", \"short_name\", \"creator\", \"created\", \"status\")\n actions = [\"create_challenge\", \"send_status_update_email\"]\n\n @admin.action(description=\"Create challenge for this request\")\n def create_challenge(self, request, queryset):\n for challengerequest in queryset:\n try:\n challengerequest.create_challenge()\n except ValidationError:\n self.message_user(\n request,\n f\"There already is a challenge with short \"\n f\"name: {challengerequest.short_name}\",\n messages.WARNING,\n )\n\n @admin.action(description=\"Send status update email to requester\")\n def send_status_update_email(self, request, queryset):\n for challengerequest in queryset:\n if (\n challengerequest.status\n == challengerequest.ChallengeRequestStatusChoices.ACCEPTED\n ):\n try:\n challenge = Challenge.objects.get(\n short_name=challengerequest.short_name\n )\n except Challenge.DoesNotExist:\n challenge = challengerequest.create_challenge()\n else:\n challenge = None\n send_challenge_status_update_email(\n challengerequest=challengerequest, challenge=challenge\n )\n\n\nadmin.site.register(Challenge, ChallengeAdmin)\nadmin.site.register(ExternalChallenge, ExternalChallengeAdmin)\nadmin.site.register(ChallengeSeries)\nadmin.site.register(ChallengeRequest, ChallengeRequestAdmin)\n", "path": "app/grandchallenge/challenges/admin.py"}, {"content": "import humanize\nfrom django import template\n\nregister = template.Library()\n\n\[email protected]\ndef naturaldelta(value):\n return humanize.naturaldelta(value, months=False)\n", "path": "app/grandchallenge/core/templatetags/naturaldelta.py"}], "after_files": [{"content": "from django.contrib import admin, messages\nfrom django.contrib.admin import ModelAdmin\nfrom django.core.exceptions import ValidationError\n\nfrom grandchallenge.challenges.emails import send_challenge_status_update_email\nfrom grandchallenge.challenges.models import (\n Challenge,\n ChallengeRequest,\n ChallengeSeries,\n ExternalChallenge,\n)\n\n\nclass ChallengeAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n autocomplete_fields = (\"publications\",)\n ordering = (\"-created\",)\n list_display = (\"short_name\", \"created\")\n search_fields = (\"short_name\",)\n\n\nclass ExternalChallengeAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n autocomplete_fields = (\"publications\",)\n\n\nclass ChallengeRequestAdmin(ModelAdmin):\n readonly_fields = (\"creator\",)\n ordering = (\"-created\",)\n list_display = (\n \"title\",\n \"short_name\",\n \"creator\",\n \"created\",\n \"status\",\n \"challenge_type\",\n \"total_cost\",\n \"budget_for_hosting_challenge\",\n )\n actions = [\"create_challenge\", \"send_status_update_email\"]\n list_filter = [\"status\", \"challenge_type\"]\n\n @admin.display(description=\"Total cost\")\n def total_cost(self, obj):\n return \"%s\" % (obj.budget[\"Total\"])\n\n @admin.action(description=\"Create challenge for this request\")\n def create_challenge(self, request, queryset):\n for challengerequest in queryset:\n try:\n challengerequest.create_challenge()\n except ValidationError:\n self.message_user(\n request,\n f\"There already is a challenge with short \"\n f\"name: {challengerequest.short_name}\",\n messages.WARNING,\n )\n\n @admin.action(description=\"Send status update email to requester\")\n def send_status_update_email(self, request, queryset):\n for challengerequest in queryset:\n if (\n challengerequest.status\n == challengerequest.ChallengeRequestStatusChoices.ACCEPTED\n ):\n try:\n challenge = Challenge.objects.get(\n short_name=challengerequest.short_name\n )\n except Challenge.DoesNotExist:\n challenge = challengerequest.create_challenge()\n else:\n challenge = None\n send_challenge_status_update_email(\n challengerequest=challengerequest, challenge=challenge\n )\n\n\nadmin.site.register(Challenge, ChallengeAdmin)\nadmin.site.register(ExternalChallenge, ExternalChallengeAdmin)\nadmin.site.register(ChallengeSeries)\nadmin.site.register(ChallengeRequest, ChallengeRequestAdmin)\n", "path": "app/grandchallenge/challenges/admin.py"}, {"content": "import humanize\nfrom django import template\nfrom django.utils import timezone\n\nregister = template.Library()\n\n\[email protected]\ndef naturaldelta(value):\n return humanize.naturaldelta(value, months=False)\n\n\[email protected]\ndef timedifference(value):\n return (timezone.now() - value).days\n", "path": "app/grandchallenge/core/templatetags/naturaldelta.py"}]}
| 1,084 | 394 |
gh_patches_debug_17471
|
rasdani/github-patches
|
git_diff
|
getredash__redash-1899
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Oracle: add support for cx_Oracle v5.3.
### Issue Summary
cx_Oracle.FIXED_UNICODE: TYPE_STRING,
Variable FIXED_UNICODE was removed with Release 5.3 of cx_Oracle: https://github.com/oracle/python-cx_Oracle/blob/ae45152339f0e9b46a93d5aea74f3bc16e775060/doc/src/releasenotes.rst#version-53-march-2017
Removed deprecated types UNICODE, FIXED_UNICODE and LONG_UNICODE (use NCHAR, FIXED_NCHAR and LONG_NCHAR instead).
but is referenced in https://github.com/getredash/redash/blob/master/redash/query_runner/oracle.py#L17
### Steps to Reproduce
Our Dockerfile:
FROM redash/redash:2.0.0.b2924
USER root
RUN apt-get update
RUN apt-get -y --no-install-recommends install alien
COPY "oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm" "/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm"
RUN alien -i "/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm"
RUN echo /usr/lib/oracle/12.2/client64/lib > /etc/ld.so.conf.d/oracle-instantclient.conf && ldconfig
RUN apt-get install -y libaio1
RUN pip install cx_Oracle --pre
EXPOSE 5000
ENTRYPOINT ["/app/bin/docker-entrypoint"]
CMD ["server"]
Output:
```
AttributeError: 'module' object has no attribute 'FIXED_UNICODE'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/gunicorn/arbiter.py", line 515, in spawn_worker
worker.init_process()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py", line 122, in init_process
self.load_wsgi()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py", line 130, in load_wsgi
self.wsgi = self.app.wsgi()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/base.py", line 67, in wsgi
self.callable = self.load()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py", line 65, in load
return self.load_wsgiapp()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py", line 52, in load_wsgiapp
return util.import_app(self.app_uri)
File "/usr/local/lib/python2.7/dist-packages/gunicorn/util.py", line 357, in import_app
__import__(module)
File "/app/redash/__init__.py", line 68, in <module>
import_query_runners(settings.QUERY_RUNNERS)
File "/app/redash/query_runner/__init__.py", line 175, in import_query_runners
__import__(runner_import)
File "/app/redash/query_runner/oracle.py", line 17, in <module>
cx_Oracle.FIXED_UNICODE: TYPE_STRING,
AttributeError: 'module' object has no attribute 'FIXED_UNICODE
```'
### Technical details:
* Redash Version: Docker Image redash/redash:2.0.0.b2924
* How did you install Redash: Dockerfile
Oracle: add support for cx_Oracle v5.3.
### Issue Summary
cx_Oracle.FIXED_UNICODE: TYPE_STRING,
Variable FIXED_UNICODE was removed with Release 5.3 of cx_Oracle: https://github.com/oracle/python-cx_Oracle/blob/ae45152339f0e9b46a93d5aea74f3bc16e775060/doc/src/releasenotes.rst#version-53-march-2017
Removed deprecated types UNICODE, FIXED_UNICODE and LONG_UNICODE (use NCHAR, FIXED_NCHAR and LONG_NCHAR instead).
but is referenced in https://github.com/getredash/redash/blob/master/redash/query_runner/oracle.py#L17
### Steps to Reproduce
Our Dockerfile:
FROM redash/redash:2.0.0.b2924
USER root
RUN apt-get update
RUN apt-get -y --no-install-recommends install alien
COPY "oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm" "/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm"
RUN alien -i "/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm"
RUN echo /usr/lib/oracle/12.2/client64/lib > /etc/ld.so.conf.d/oracle-instantclient.conf && ldconfig
RUN apt-get install -y libaio1
RUN pip install cx_Oracle --pre
EXPOSE 5000
ENTRYPOINT ["/app/bin/docker-entrypoint"]
CMD ["server"]
Output:
```
AttributeError: 'module' object has no attribute 'FIXED_UNICODE'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/gunicorn/arbiter.py", line 515, in spawn_worker
worker.init_process()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py", line 122, in init_process
self.load_wsgi()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py", line 130, in load_wsgi
self.wsgi = self.app.wsgi()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/base.py", line 67, in wsgi
self.callable = self.load()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py", line 65, in load
return self.load_wsgiapp()
File "/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py", line 52, in load_wsgiapp
return util.import_app(self.app_uri)
File "/usr/local/lib/python2.7/dist-packages/gunicorn/util.py", line 357, in import_app
__import__(module)
File "/app/redash/__init__.py", line 68, in <module>
import_query_runners(settings.QUERY_RUNNERS)
File "/app/redash/query_runner/__init__.py", line 175, in import_query_runners
__import__(runner_import)
File "/app/redash/query_runner/oracle.py", line 17, in <module>
cx_Oracle.FIXED_UNICODE: TYPE_STRING,
AttributeError: 'module' object has no attribute 'FIXED_UNICODE
```'
### Technical details:
* Redash Version: Docker Image redash/redash:2.0.0.b2924
* How did you install Redash: Dockerfile
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `redash/query_runner/oracle.py`
Content:
```
1 import json
2 import logging
3 import sys
4
5 from redash.query_runner import *
6 from redash.utils import JSONEncoder
7
8 try:
9 import cx_Oracle
10
11 TYPES_MAP = {
12 cx_Oracle.DATETIME: TYPE_DATETIME,
13 cx_Oracle.CLOB: TYPE_STRING,
14 cx_Oracle.LOB: TYPE_STRING,
15 cx_Oracle.FIXED_CHAR: TYPE_STRING,
16 cx_Oracle.FIXED_NCHAR: TYPE_STRING,
17 cx_Oracle.FIXED_UNICODE: TYPE_STRING,
18 cx_Oracle.INTERVAL: TYPE_DATETIME,
19 cx_Oracle.LONG_NCHAR: TYPE_STRING,
20 cx_Oracle.LONG_STRING: TYPE_STRING,
21 cx_Oracle.LONG_UNICODE: TYPE_STRING,
22 cx_Oracle.NATIVE_FLOAT: TYPE_FLOAT,
23 cx_Oracle.NCHAR: TYPE_STRING,
24 cx_Oracle.NUMBER: TYPE_FLOAT,
25 cx_Oracle.ROWID: TYPE_INTEGER,
26 cx_Oracle.STRING: TYPE_STRING,
27 cx_Oracle.TIMESTAMP: TYPE_DATETIME,
28 cx_Oracle.UNICODE: TYPE_STRING,
29 }
30
31
32 ENABLED = True
33 except ImportError:
34 ENABLED = False
35
36 logger = logging.getLogger(__name__)
37
38 class Oracle(BaseSQLQueryRunner):
39 noop_query = "SELECT 1 FROM dual"
40
41 @classmethod
42 def get_col_type(cls, col_type, scale):
43 if col_type == cx_Oracle.NUMBER:
44 return TYPE_FLOAT if scale > 0 else TYPE_INTEGER
45 else:
46 return TYPES_MAP.get(col_type, None)
47
48 @classmethod
49 def enabled(cls):
50 return ENABLED
51
52 @classmethod
53 def configuration_schema(cls):
54 return {
55 "type": "object",
56 "properties": {
57 "user": {
58 "type": "string"
59 },
60 "password": {
61 "type": "string"
62 },
63 "host": {
64 "type": "string"
65 },
66 "port": {
67 "type": "number"
68 },
69 "servicename": {
70 "type": "string",
71 "title": "DSN Service Name"
72 }
73 },
74 "required": ["servicename", "user", "password", "host", "port"],
75 "secret": ["password"]
76 }
77
78 @classmethod
79 def type(cls):
80 return "oracle"
81
82 def __init__(self, configuration):
83 super(Oracle, self).__init__(configuration)
84
85 dsn = cx_Oracle.makedsn(
86 self.configuration["host"],
87 self.configuration["port"],
88 service_name=self.configuration["servicename"])
89
90 self.connection_string = "{}/{}@{}".format(self.configuration["user"], self.configuration["password"], dsn)
91
92 def _get_tables(self, schema):
93 query = """
94 SELECT
95 all_tab_cols.OWNER,
96 all_tab_cols.TABLE_NAME,
97 all_tab_cols.COLUMN_NAME
98 FROM all_tab_cols
99 WHERE all_tab_cols.OWNER NOT IN('SYS','SYSTEM','ORDSYS','CTXSYS','WMSYS','MDSYS','ORDDATA','XDB','OUTLN','DMSYS','DSSYS','EXFSYS','LBACSYS','TSMSYS')
100 """
101
102 results, error = self.run_query(query, None)
103
104 if error is not None:
105 raise Exception("Failed getting schema.")
106
107 results = json.loads(results)
108
109 for row in results['rows']:
110 if row['OWNER'] != None:
111 table_name = '{}.{}'.format(row['OWNER'], row['TABLE_NAME'])
112 else:
113 table_name = row['TABLE_NAME']
114
115 if table_name not in schema:
116 schema[table_name] = {'name': table_name, 'columns': []}
117
118 schema[table_name]['columns'].append(row['COLUMN_NAME'])
119
120 return schema.values()
121
122 @classmethod
123 def _convert_number(cls, value):
124 try:
125 return int(value)
126 except:
127 return value
128
129 @classmethod
130 def output_handler(cls, cursor, name, default_type, length, precision, scale):
131 if default_type in (cx_Oracle.CLOB, cx_Oracle.LOB):
132 return cursor.var(cx_Oracle.LONG_STRING, 80000, cursor.arraysize)
133
134 if default_type in (cx_Oracle.STRING, cx_Oracle.FIXED_CHAR):
135 return cursor.var(unicode, length, cursor.arraysize)
136
137 if default_type == cx_Oracle.NUMBER:
138 if scale <= 0:
139 return cursor.var(cx_Oracle.STRING, 255, outconverter=Oracle._convert_number, arraysize=cursor.arraysize)
140
141 def run_query(self, query, user):
142 connection = cx_Oracle.connect(self.connection_string)
143 connection.outputtypehandler = Oracle.output_handler
144
145 cursor = connection.cursor()
146
147 try:
148 cursor.execute(query)
149
150 if cursor.description is not None:
151 columns = self.fetch_columns([(i[0], Oracle.get_col_type(i[1], i[5])) for i in cursor.description])
152 rows = [dict(zip((c['name'] for c in columns), row)) for row in cursor]
153
154 data = {'columns': columns, 'rows': rows}
155 error = None
156 json_data = json.dumps(data, cls=JSONEncoder)
157 else:
158 error = 'Query completed but it returned no data.'
159 json_data = None
160 except cx_Oracle.DatabaseError as err:
161 error = u"Query failed. {}.".format(err.message)
162 json_data = None
163 except KeyboardInterrupt:
164 connection.cancel()
165 error = "Query cancelled by user."
166 json_data = None
167 finally:
168 connection.close()
169
170 return json_data, error
171
172 register(Oracle)
173
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/redash/query_runner/oracle.py b/redash/query_runner/oracle.py
--- a/redash/query_runner/oracle.py
+++ b/redash/query_runner/oracle.py
@@ -14,18 +14,14 @@
cx_Oracle.LOB: TYPE_STRING,
cx_Oracle.FIXED_CHAR: TYPE_STRING,
cx_Oracle.FIXED_NCHAR: TYPE_STRING,
- cx_Oracle.FIXED_UNICODE: TYPE_STRING,
cx_Oracle.INTERVAL: TYPE_DATETIME,
- cx_Oracle.LONG_NCHAR: TYPE_STRING,
cx_Oracle.LONG_STRING: TYPE_STRING,
- cx_Oracle.LONG_UNICODE: TYPE_STRING,
cx_Oracle.NATIVE_FLOAT: TYPE_FLOAT,
cx_Oracle.NCHAR: TYPE_STRING,
cx_Oracle.NUMBER: TYPE_FLOAT,
cx_Oracle.ROWID: TYPE_INTEGER,
cx_Oracle.STRING: TYPE_STRING,
cx_Oracle.TIMESTAMP: TYPE_DATETIME,
- cx_Oracle.UNICODE: TYPE_STRING,
}
|
{"golden_diff": "diff --git a/redash/query_runner/oracle.py b/redash/query_runner/oracle.py\n--- a/redash/query_runner/oracle.py\n+++ b/redash/query_runner/oracle.py\n@@ -14,18 +14,14 @@\n cx_Oracle.LOB: TYPE_STRING,\n cx_Oracle.FIXED_CHAR: TYPE_STRING,\n cx_Oracle.FIXED_NCHAR: TYPE_STRING,\n- cx_Oracle.FIXED_UNICODE: TYPE_STRING,\n cx_Oracle.INTERVAL: TYPE_DATETIME,\n- cx_Oracle.LONG_NCHAR: TYPE_STRING,\n cx_Oracle.LONG_STRING: TYPE_STRING,\n- cx_Oracle.LONG_UNICODE: TYPE_STRING,\n cx_Oracle.NATIVE_FLOAT: TYPE_FLOAT,\n cx_Oracle.NCHAR: TYPE_STRING,\n cx_Oracle.NUMBER: TYPE_FLOAT,\n cx_Oracle.ROWID: TYPE_INTEGER,\n cx_Oracle.STRING: TYPE_STRING,\n cx_Oracle.TIMESTAMP: TYPE_DATETIME,\n- cx_Oracle.UNICODE: TYPE_STRING,\n }\n", "issue": "Oracle: add support for cx_Oracle v5.3.\n\r\n### Issue Summary\r\n\r\ncx_Oracle.FIXED_UNICODE: TYPE_STRING,\r\n\r\nVariable FIXED_UNICODE was removed with Release 5.3 of cx_Oracle: https://github.com/oracle/python-cx_Oracle/blob/ae45152339f0e9b46a93d5aea74f3bc16e775060/doc/src/releasenotes.rst#version-53-march-2017 \r\n\r\n Removed deprecated types UNICODE, FIXED_UNICODE and LONG_UNICODE (use NCHAR, FIXED_NCHAR and LONG_NCHAR instead).\r\n\r\nbut is referenced in https://github.com/getredash/redash/blob/master/redash/query_runner/oracle.py#L17\r\n\r\n\r\n\r\n### Steps to Reproduce\r\n\r\nOur Dockerfile:\r\n\r\n\tFROM redash/redash:2.0.0.b2924\r\n\r\n\tUSER root\r\n\r\n\tRUN apt-get update\r\n\r\n\tRUN apt-get -y --no-install-recommends install alien\r\n\tCOPY \"oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\" \"/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\"\r\n\tRUN alien -i \"/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\"\r\n\tRUN echo /usr/lib/oracle/12.2/client64/lib > /etc/ld.so.conf.d/oracle-instantclient.conf && ldconfig\r\n\tRUN apt-get install -y libaio1\r\n\tRUN pip install cx_Oracle --pre\r\n\t\r\n\tEXPOSE 5000\r\n\t\r\n\tENTRYPOINT [\"/app/bin/docker-entrypoint\"]\r\n\t\r\n\tCMD [\"server\"]\r\n\r\n\r\nOutput: \r\n```\r\n AttributeError: 'module' object has no attribute 'FIXED_UNICODE'\r\n Traceback (most recent call last):\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/arbiter.py\", line 515, in spawn_worker\r\n worker.init_process()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py\", line 122, in init_process\r\n self.load_wsgi()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py\", line 130, in load_wsgi\r\n self.wsgi = self.app.wsgi()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/base.py\", line 67, in wsgi\r\n self.callable = self.load()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py\", line 65, in load\r\n return self.load_wsgiapp()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py\", line 52, in load_wsgiapp\r\n return util.import_app(self.app_uri)\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/util.py\", line 357, in import_app\r\n __import__(module)\r\n File \"/app/redash/__init__.py\", line 68, in <module>\r\n import_query_runners(settings.QUERY_RUNNERS)\r\n File \"/app/redash/query_runner/__init__.py\", line 175, in import_query_runners\r\n __import__(runner_import)\r\n File \"/app/redash/query_runner/oracle.py\", line 17, in <module>\r\n cx_Oracle.FIXED_UNICODE: TYPE_STRING,\r\n AttributeError: 'module' object has no attribute 'FIXED_UNICODE\r\n```'\r\n\r\n\r\n### Technical details:\r\n\r\n* Redash Version: Docker Image redash/redash:2.0.0.b2924\r\n* How did you install Redash: Dockerfile\r\n\nOracle: add support for cx_Oracle v5.3.\n\r\n### Issue Summary\r\n\r\ncx_Oracle.FIXED_UNICODE: TYPE_STRING,\r\n\r\nVariable FIXED_UNICODE was removed with Release 5.3 of cx_Oracle: https://github.com/oracle/python-cx_Oracle/blob/ae45152339f0e9b46a93d5aea74f3bc16e775060/doc/src/releasenotes.rst#version-53-march-2017 \r\n\r\n Removed deprecated types UNICODE, FIXED_UNICODE and LONG_UNICODE (use NCHAR, FIXED_NCHAR and LONG_NCHAR instead).\r\n\r\nbut is referenced in https://github.com/getredash/redash/blob/master/redash/query_runner/oracle.py#L17\r\n\r\n\r\n\r\n### Steps to Reproduce\r\n\r\nOur Dockerfile:\r\n\r\n\tFROM redash/redash:2.0.0.b2924\r\n\r\n\tUSER root\r\n\r\n\tRUN apt-get update\r\n\r\n\tRUN apt-get -y --no-install-recommends install alien\r\n\tCOPY \"oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\" \"/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\"\r\n\tRUN alien -i \"/tmp/oracle-instantclient12.2-basiclite-12.2.0.1.0-1.x86_64.rpm\"\r\n\tRUN echo /usr/lib/oracle/12.2/client64/lib > /etc/ld.so.conf.d/oracle-instantclient.conf && ldconfig\r\n\tRUN apt-get install -y libaio1\r\n\tRUN pip install cx_Oracle --pre\r\n\t\r\n\tEXPOSE 5000\r\n\t\r\n\tENTRYPOINT [\"/app/bin/docker-entrypoint\"]\r\n\t\r\n\tCMD [\"server\"]\r\n\r\n\r\nOutput: \r\n```\r\n AttributeError: 'module' object has no attribute 'FIXED_UNICODE'\r\n Traceback (most recent call last):\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/arbiter.py\", line 515, in spawn_worker\r\n worker.init_process()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py\", line 122, in init_process\r\n self.load_wsgi()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/workers/base.py\", line 130, in load_wsgi\r\n self.wsgi = self.app.wsgi()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/base.py\", line 67, in wsgi\r\n self.callable = self.load()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py\", line 65, in load\r\n return self.load_wsgiapp()\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/app/wsgiapp.py\", line 52, in load_wsgiapp\r\n return util.import_app(self.app_uri)\r\n File \"/usr/local/lib/python2.7/dist-packages/gunicorn/util.py\", line 357, in import_app\r\n __import__(module)\r\n File \"/app/redash/__init__.py\", line 68, in <module>\r\n import_query_runners(settings.QUERY_RUNNERS)\r\n File \"/app/redash/query_runner/__init__.py\", line 175, in import_query_runners\r\n __import__(runner_import)\r\n File \"/app/redash/query_runner/oracle.py\", line 17, in <module>\r\n cx_Oracle.FIXED_UNICODE: TYPE_STRING,\r\n AttributeError: 'module' object has no attribute 'FIXED_UNICODE\r\n```'\r\n\r\n\r\n### Technical details:\r\n\r\n* Redash Version: Docker Image redash/redash:2.0.0.b2924\r\n* How did you install Redash: Dockerfile\r\n\n", "before_files": [{"content": "import json\nimport logging\nimport sys\n\nfrom redash.query_runner import *\nfrom redash.utils import JSONEncoder\n\ntry:\n import cx_Oracle\n\n TYPES_MAP = {\n cx_Oracle.DATETIME: TYPE_DATETIME,\n cx_Oracle.CLOB: TYPE_STRING,\n cx_Oracle.LOB: TYPE_STRING,\n cx_Oracle.FIXED_CHAR: TYPE_STRING,\n cx_Oracle.FIXED_NCHAR: TYPE_STRING,\n cx_Oracle.FIXED_UNICODE: TYPE_STRING,\n cx_Oracle.INTERVAL: TYPE_DATETIME,\n cx_Oracle.LONG_NCHAR: TYPE_STRING,\n cx_Oracle.LONG_STRING: TYPE_STRING,\n cx_Oracle.LONG_UNICODE: TYPE_STRING,\n cx_Oracle.NATIVE_FLOAT: TYPE_FLOAT,\n cx_Oracle.NCHAR: TYPE_STRING,\n cx_Oracle.NUMBER: TYPE_FLOAT,\n cx_Oracle.ROWID: TYPE_INTEGER,\n cx_Oracle.STRING: TYPE_STRING,\n cx_Oracle.TIMESTAMP: TYPE_DATETIME,\n cx_Oracle.UNICODE: TYPE_STRING,\n }\n\n\n ENABLED = True\nexcept ImportError:\n ENABLED = False\n\nlogger = logging.getLogger(__name__)\n\nclass Oracle(BaseSQLQueryRunner):\n noop_query = \"SELECT 1 FROM dual\"\n\n @classmethod\n def get_col_type(cls, col_type, scale):\n if col_type == cx_Oracle.NUMBER:\n return TYPE_FLOAT if scale > 0 else TYPE_INTEGER\n else:\n return TYPES_MAP.get(col_type, None)\n\n @classmethod\n def enabled(cls):\n return ENABLED\n\n @classmethod\n def configuration_schema(cls):\n return {\n \"type\": \"object\",\n \"properties\": {\n \"user\": {\n \"type\": \"string\"\n },\n \"password\": {\n \"type\": \"string\"\n },\n \"host\": {\n \"type\": \"string\"\n },\n \"port\": {\n \"type\": \"number\"\n },\n \"servicename\": {\n \"type\": \"string\",\n \"title\": \"DSN Service Name\"\n }\n },\n \"required\": [\"servicename\", \"user\", \"password\", \"host\", \"port\"],\n \"secret\": [\"password\"]\n }\n\n @classmethod\n def type(cls):\n return \"oracle\"\n\n def __init__(self, configuration):\n super(Oracle, self).__init__(configuration)\n\n dsn = cx_Oracle.makedsn(\n self.configuration[\"host\"],\n self.configuration[\"port\"],\n service_name=self.configuration[\"servicename\"])\n\n self.connection_string = \"{}/{}@{}\".format(self.configuration[\"user\"], self.configuration[\"password\"], dsn)\n\n def _get_tables(self, schema):\n query = \"\"\"\n SELECT\n all_tab_cols.OWNER,\n all_tab_cols.TABLE_NAME,\n all_tab_cols.COLUMN_NAME\n FROM all_tab_cols\n WHERE all_tab_cols.OWNER NOT IN('SYS','SYSTEM','ORDSYS','CTXSYS','WMSYS','MDSYS','ORDDATA','XDB','OUTLN','DMSYS','DSSYS','EXFSYS','LBACSYS','TSMSYS')\n \"\"\"\n\n results, error = self.run_query(query, None)\n\n if error is not None:\n raise Exception(\"Failed getting schema.\")\n\n results = json.loads(results)\n\n for row in results['rows']:\n if row['OWNER'] != None:\n table_name = '{}.{}'.format(row['OWNER'], row['TABLE_NAME'])\n else:\n table_name = row['TABLE_NAME']\n\n if table_name not in schema:\n schema[table_name] = {'name': table_name, 'columns': []}\n\n schema[table_name]['columns'].append(row['COLUMN_NAME'])\n\n return schema.values()\n\n @classmethod\n def _convert_number(cls, value):\n try:\n return int(value)\n except:\n return value\n\n @classmethod\n def output_handler(cls, cursor, name, default_type, length, precision, scale):\n if default_type in (cx_Oracle.CLOB, cx_Oracle.LOB):\n return cursor.var(cx_Oracle.LONG_STRING, 80000, cursor.arraysize)\n\n if default_type in (cx_Oracle.STRING, cx_Oracle.FIXED_CHAR):\n return cursor.var(unicode, length, cursor.arraysize)\n\n if default_type == cx_Oracle.NUMBER:\n if scale <= 0:\n return cursor.var(cx_Oracle.STRING, 255, outconverter=Oracle._convert_number, arraysize=cursor.arraysize)\n\n def run_query(self, query, user):\n connection = cx_Oracle.connect(self.connection_string)\n connection.outputtypehandler = Oracle.output_handler\n\n cursor = connection.cursor()\n\n try:\n cursor.execute(query)\n\n if cursor.description is not None:\n columns = self.fetch_columns([(i[0], Oracle.get_col_type(i[1], i[5])) for i in cursor.description])\n rows = [dict(zip((c['name'] for c in columns), row)) for row in cursor]\n\n data = {'columns': columns, 'rows': rows}\n error = None\n json_data = json.dumps(data, cls=JSONEncoder)\n else:\n error = 'Query completed but it returned no data.'\n json_data = None\n except cx_Oracle.DatabaseError as err:\n error = u\"Query failed. {}.\".format(err.message)\n json_data = None\n except KeyboardInterrupt:\n connection.cancel()\n error = \"Query cancelled by user.\"\n json_data = None\n finally:\n connection.close()\n\n return json_data, error\n\nregister(Oracle)\n", "path": "redash/query_runner/oracle.py"}], "after_files": [{"content": "import json\nimport logging\nimport sys\n\nfrom redash.query_runner import *\nfrom redash.utils import JSONEncoder\n\ntry:\n import cx_Oracle\n\n TYPES_MAP = {\n cx_Oracle.DATETIME: TYPE_DATETIME,\n cx_Oracle.CLOB: TYPE_STRING,\n cx_Oracle.LOB: TYPE_STRING,\n cx_Oracle.FIXED_CHAR: TYPE_STRING,\n cx_Oracle.FIXED_NCHAR: TYPE_STRING,\n cx_Oracle.INTERVAL: TYPE_DATETIME,\n cx_Oracle.LONG_STRING: TYPE_STRING,\n cx_Oracle.NATIVE_FLOAT: TYPE_FLOAT,\n cx_Oracle.NCHAR: TYPE_STRING,\n cx_Oracle.NUMBER: TYPE_FLOAT,\n cx_Oracle.ROWID: TYPE_INTEGER,\n cx_Oracle.STRING: TYPE_STRING,\n cx_Oracle.TIMESTAMP: TYPE_DATETIME,\n }\n\n\n ENABLED = True\nexcept ImportError:\n ENABLED = False\n\nlogger = logging.getLogger(__name__)\n\nclass Oracle(BaseSQLQueryRunner):\n noop_query = \"SELECT 1 FROM dual\"\n\n @classmethod\n def get_col_type(cls, col_type, scale):\n if col_type == cx_Oracle.NUMBER:\n return TYPE_FLOAT if scale > 0 else TYPE_INTEGER\n else:\n return TYPES_MAP.get(col_type, None)\n\n @classmethod\n def enabled(cls):\n return ENABLED\n\n @classmethod\n def configuration_schema(cls):\n return {\n \"type\": \"object\",\n \"properties\": {\n \"user\": {\n \"type\": \"string\"\n },\n \"password\": {\n \"type\": \"string\"\n },\n \"host\": {\n \"type\": \"string\"\n },\n \"port\": {\n \"type\": \"number\"\n },\n \"servicename\": {\n \"type\": \"string\",\n \"title\": \"DSN Service Name\"\n }\n },\n \"required\": [\"servicename\", \"user\", \"password\", \"host\", \"port\"],\n \"secret\": [\"password\"]\n }\n\n @classmethod\n def type(cls):\n return \"oracle\"\n\n def __init__(self, configuration):\n super(Oracle, self).__init__(configuration)\n\n dsn = cx_Oracle.makedsn(\n self.configuration[\"host\"],\n self.configuration[\"port\"],\n service_name=self.configuration[\"servicename\"])\n\n self.connection_string = \"{}/{}@{}\".format(self.configuration[\"user\"], self.configuration[\"password\"], dsn)\n\n def _get_tables(self, schema):\n query = \"\"\"\n SELECT\n all_tab_cols.OWNER,\n all_tab_cols.TABLE_NAME,\n all_tab_cols.COLUMN_NAME\n FROM all_tab_cols\n WHERE all_tab_cols.OWNER NOT IN('SYS','SYSTEM','ORDSYS','CTXSYS','WMSYS','MDSYS','ORDDATA','XDB','OUTLN','DMSYS','DSSYS','EXFSYS','LBACSYS','TSMSYS')\n \"\"\"\n\n results, error = self.run_query(query, None)\n\n if error is not None:\n raise Exception(\"Failed getting schema.\")\n\n results = json.loads(results)\n\n for row in results['rows']:\n if row['OWNER'] != None:\n table_name = '{}.{}'.format(row['OWNER'], row['TABLE_NAME'])\n else:\n table_name = row['TABLE_NAME']\n\n if table_name not in schema:\n schema[table_name] = {'name': table_name, 'columns': []}\n\n schema[table_name]['columns'].append(row['COLUMN_NAME'])\n\n return schema.values()\n\n @classmethod\n def _convert_number(cls, value):\n try:\n return int(value)\n except:\n return value\n\n @classmethod\n def output_handler(cls, cursor, name, default_type, length, precision, scale):\n if default_type in (cx_Oracle.CLOB, cx_Oracle.LOB):\n return cursor.var(cx_Oracle.LONG_STRING, 80000, cursor.arraysize)\n\n if default_type in (cx_Oracle.STRING, cx_Oracle.FIXED_CHAR):\n return cursor.var(unicode, length, cursor.arraysize)\n\n if default_type == cx_Oracle.NUMBER:\n if scale <= 0:\n return cursor.var(cx_Oracle.STRING, 255, outconverter=Oracle._convert_number, arraysize=cursor.arraysize)\n\n def run_query(self, query, user):\n connection = cx_Oracle.connect(self.connection_string)\n connection.outputtypehandler = Oracle.output_handler\n\n cursor = connection.cursor()\n\n try:\n cursor.execute(query)\n\n if cursor.description is not None:\n columns = self.fetch_columns([(i[0], Oracle.get_col_type(i[1], i[5])) for i in cursor.description])\n rows = [dict(zip((c['name'] for c in columns), row)) for row in cursor]\n\n data = {'columns': columns, 'rows': rows}\n error = None\n json_data = json.dumps(data, cls=JSONEncoder)\n else:\n error = 'Query completed but it returned no data.'\n json_data = None\n except cx_Oracle.DatabaseError as err:\n error = u\"Query failed. {}.\".format(err.message)\n json_data = None\n except KeyboardInterrupt:\n connection.cancel()\n error = \"Query cancelled by user.\"\n json_data = None\n finally:\n connection.close()\n\n return json_data, error\n\nregister(Oracle)\n", "path": "redash/query_runner/oracle.py"}]}
| 3,592 | 224 |
gh_patches_debug_23029
|
rasdani/github-patches
|
git_diff
|
pyqtgraph__pyqtgraph-1090
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Propose removal of `GraphicsScene.ExportDialog.ExporterParameters`
I propose to remove `ExportDialog.exporterParameters`. They are used to remember the export parameters of an exporter when the user switches between exporters (e.g. SVGExporter and ImageExporter).
https://github.com/pyqtgraph/pyqtgraph/blob/db483f8ec22a0d53bd0e71fac69b0045a96c6637/pyqtgraph/GraphicsScene/exportDialog.py#L127-L136
### Reasons
1. I doubt that anyone actually needs this.
2. I would like to create a PR that automatically sets the correct resolution in the ImageExporter and also keeps the correct aspect ratio when the user changes width or height in the ImageExporter (This feature was intended but is currently broken). The problem is that a simple fix would invalidate the rationale behind `ExportDialog.exporterParameters`, because changing the scene item would also update the ImageExporter parameters (and the stored parameters are not used). An elaborate fix would add a lot of clutter code (creating a dictionary of dictionaries for each item and each exporter).
The simple fix is highlighted here:
https://github.com/paulmueller/pyqtgraph/blob/b81f6d66d2b3cf4d334aa31b75b71e68cc7a3796/pyqtgraph/GraphicsScene/exportDialog.py#L101-L111
### Related Bug
As mentioned in my second point, the ImageExporter currently only displays the height and width of the scene and it does not adjust when e.g. a PlotItem is selected (which has different height, widht, aspect ratio). Furthermore, width and height are not updated to match the PlotItem aspect ratio if I change one of width or height.
### Question
Would you accept a PR that gets rid of `ExportDialog.exporterParameters` and fixes the related bug?
Propose removal of `GraphicsScene.ExportDialog.ExporterParameters`
I propose to remove `ExportDialog.exporterParameters`. They are used to remember the export parameters of an exporter when the user switches between exporters (e.g. SVGExporter and ImageExporter).
https://github.com/pyqtgraph/pyqtgraph/blob/db483f8ec22a0d53bd0e71fac69b0045a96c6637/pyqtgraph/GraphicsScene/exportDialog.py#L127-L136
### Reasons
1. I doubt that anyone actually needs this.
2. I would like to create a PR that automatically sets the correct resolution in the ImageExporter and also keeps the correct aspect ratio when the user changes width or height in the ImageExporter (This feature was intended but is currently broken). The problem is that a simple fix would invalidate the rationale behind `ExportDialog.exporterParameters`, because changing the scene item would also update the ImageExporter parameters (and the stored parameters are not used). An elaborate fix would add a lot of clutter code (creating a dictionary of dictionaries for each item and each exporter).
The simple fix is highlighted here:
https://github.com/paulmueller/pyqtgraph/blob/b81f6d66d2b3cf4d334aa31b75b71e68cc7a3796/pyqtgraph/GraphicsScene/exportDialog.py#L101-L111
### Related Bug
As mentioned in my second point, the ImageExporter currently only displays the height and width of the scene and it does not adjust when e.g. a PlotItem is selected (which has different height, widht, aspect ratio). Furthermore, width and height are not updated to match the PlotItem aspect ratio if I change one of width or height.
### Question
Would you accept a PR that gets rid of `ExportDialog.exporterParameters` and fixes the related bug?
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pyqtgraph/GraphicsScene/exportDialog.py`
Content:
```
1 from ..Qt import QtCore, QtGui, QT_LIB
2 from .. import exporters as exporters
3 from .. import functions as fn
4 from ..graphicsItems.ViewBox import ViewBox
5 from ..graphicsItems.PlotItem import PlotItem
6
7 if QT_LIB == 'PySide':
8 from . import exportDialogTemplate_pyside as exportDialogTemplate
9 elif QT_LIB == 'PySide2':
10 from . import exportDialogTemplate_pyside2 as exportDialogTemplate
11 elif QT_LIB == 'PyQt5':
12 from . import exportDialogTemplate_pyqt5 as exportDialogTemplate
13 else:
14 from . import exportDialogTemplate_pyqt as exportDialogTemplate
15
16
17 class ExportDialog(QtGui.QWidget):
18 def __init__(self, scene):
19 QtGui.QWidget.__init__(self)
20 self.setVisible(False)
21 self.setWindowTitle("Export")
22 self.shown = False
23 self.currentExporter = None
24 self.scene = scene
25
26 self.exporterParameters = {}
27
28 self.selectBox = QtGui.QGraphicsRectItem()
29 self.selectBox.setPen(fn.mkPen('y', width=3, style=QtCore.Qt.DashLine))
30 self.selectBox.hide()
31 self.scene.addItem(self.selectBox)
32
33 self.ui = exportDialogTemplate.Ui_Form()
34 self.ui.setupUi(self)
35
36 self.ui.closeBtn.clicked.connect(self.close)
37 self.ui.exportBtn.clicked.connect(self.exportClicked)
38 self.ui.copyBtn.clicked.connect(self.copyClicked)
39 self.ui.itemTree.currentItemChanged.connect(self.exportItemChanged)
40 self.ui.formatList.currentItemChanged.connect(self.exportFormatChanged)
41
42
43 def show(self, item=None):
44 if item is not None:
45 ## Select next exportable parent of the item originally clicked on
46 while not isinstance(item, ViewBox) and not isinstance(item, PlotItem) and item is not None:
47 item = item.parentItem()
48 ## if this is a ViewBox inside a PlotItem, select the parent instead.
49 if isinstance(item, ViewBox) and isinstance(item.parentItem(), PlotItem):
50 item = item.parentItem()
51 self.updateItemList(select=item)
52 self.setVisible(True)
53 self.activateWindow()
54 self.raise_()
55 self.selectBox.setVisible(True)
56
57 if not self.shown:
58 self.shown = True
59 vcenter = self.scene.getViewWidget().geometry().center()
60 self.setGeometry(vcenter.x()-self.width()/2, vcenter.y()-self.height()/2, self.width(), self.height())
61
62 def updateItemList(self, select=None):
63 self.ui.itemTree.clear()
64 si = QtGui.QTreeWidgetItem(["Entire Scene"])
65 si.gitem = self.scene
66 self.ui.itemTree.addTopLevelItem(si)
67 self.ui.itemTree.setCurrentItem(si)
68 si.setExpanded(True)
69 for child in self.scene.items():
70 if child.parentItem() is None:
71 self.updateItemTree(child, si, select=select)
72
73 def updateItemTree(self, item, treeItem, select=None):
74 si = None
75 if isinstance(item, ViewBox):
76 si = QtGui.QTreeWidgetItem(['ViewBox'])
77 elif isinstance(item, PlotItem):
78 si = QtGui.QTreeWidgetItem(['Plot'])
79
80 if si is not None:
81 si.gitem = item
82 treeItem.addChild(si)
83 treeItem = si
84 if si.gitem is select:
85 self.ui.itemTree.setCurrentItem(si)
86
87 for ch in item.childItems():
88 self.updateItemTree(ch, treeItem, select=select)
89
90
91 def exportItemChanged(self, item, prev):
92 if item is None:
93 return
94 if item.gitem is self.scene:
95 newBounds = self.scene.views()[0].viewRect()
96 else:
97 newBounds = item.gitem.sceneBoundingRect()
98 self.selectBox.setRect(newBounds)
99 self.selectBox.show()
100 self.updateFormatList()
101
102 def updateFormatList(self):
103 current = self.ui.formatList.currentItem()
104 if current is not None:
105 current = str(current.text())
106 self.ui.formatList.clear()
107 self.exporterClasses = {}
108 gotCurrent = False
109 for exp in exporters.listExporters():
110 self.ui.formatList.addItem(exp.Name)
111 self.exporterClasses[exp.Name] = exp
112 if exp.Name == current:
113 self.ui.formatList.setCurrentRow(self.ui.formatList.count()-1)
114 gotCurrent = True
115
116 if not gotCurrent:
117 self.ui.formatList.setCurrentRow(0)
118
119 def exportFormatChanged(self, item, prev):
120 if item is None:
121 self.currentExporter = None
122 self.ui.paramTree.clear()
123 return
124 expClass = self.exporterClasses[str(item.text())]
125 exp = expClass(item=self.ui.itemTree.currentItem().gitem)
126
127 if prev:
128 oldtext = str(prev.text())
129 self.exporterParameters[oldtext] = self.currentExporter.parameters()
130 newtext = str(item.text())
131 if newtext in self.exporterParameters.keys():
132 params = self.exporterParameters[newtext]
133 exp.params = params
134 else:
135 params = exp.parameters()
136 self.exporterParameters[newtext] = params
137
138 if params is None:
139 self.ui.paramTree.clear()
140 else:
141 self.ui.paramTree.setParameters(params)
142 self.currentExporter = exp
143 self.ui.copyBtn.setEnabled(exp.allowCopy)
144
145 def exportClicked(self):
146 self.selectBox.hide()
147 self.currentExporter.export()
148
149 def copyClicked(self):
150 self.selectBox.hide()
151 self.currentExporter.export(copy=True)
152
153 def close(self):
154 self.selectBox.setVisible(False)
155 self.setVisible(False)
156
157 def closeEvent(self, event):
158 self.close()
159 QtGui.QWidget.closeEvent(self, event)
160
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pyqtgraph/GraphicsScene/exportDialog.py b/pyqtgraph/GraphicsScene/exportDialog.py
--- a/pyqtgraph/GraphicsScene/exportDialog.py
+++ b/pyqtgraph/GraphicsScene/exportDialog.py
@@ -22,8 +22,6 @@
self.shown = False
self.currentExporter = None
self.scene = scene
-
- self.exporterParameters = {}
self.selectBox = QtGui.QGraphicsRectItem()
self.selectBox.setPen(fn.mkPen('y', width=3, style=QtCore.Qt.DashLine))
@@ -124,16 +122,7 @@
expClass = self.exporterClasses[str(item.text())]
exp = expClass(item=self.ui.itemTree.currentItem().gitem)
- if prev:
- oldtext = str(prev.text())
- self.exporterParameters[oldtext] = self.currentExporter.parameters()
- newtext = str(item.text())
- if newtext in self.exporterParameters.keys():
- params = self.exporterParameters[newtext]
- exp.params = params
- else:
- params = exp.parameters()
- self.exporterParameters[newtext] = params
+ params = exp.parameters()
if params is None:
self.ui.paramTree.clear()
|
{"golden_diff": "diff --git a/pyqtgraph/GraphicsScene/exportDialog.py b/pyqtgraph/GraphicsScene/exportDialog.py\n--- a/pyqtgraph/GraphicsScene/exportDialog.py\n+++ b/pyqtgraph/GraphicsScene/exportDialog.py\n@@ -22,8 +22,6 @@\n self.shown = False\n self.currentExporter = None\n self.scene = scene\n- \n- self.exporterParameters = {}\n \n self.selectBox = QtGui.QGraphicsRectItem()\n self.selectBox.setPen(fn.mkPen('y', width=3, style=QtCore.Qt.DashLine))\n@@ -124,16 +122,7 @@\n expClass = self.exporterClasses[str(item.text())]\n exp = expClass(item=self.ui.itemTree.currentItem().gitem)\n \n- if prev:\n- oldtext = str(prev.text())\n- self.exporterParameters[oldtext] = self.currentExporter.parameters()\n- newtext = str(item.text())\n- if newtext in self.exporterParameters.keys():\n- params = self.exporterParameters[newtext]\n- exp.params = params\n- else:\n- params = exp.parameters()\n- self.exporterParameters[newtext] = params\n+ params = exp.parameters()\n \n if params is None:\n self.ui.paramTree.clear()\n", "issue": "Propose removal of `GraphicsScene.ExportDialog.ExporterParameters`\nI propose to remove `ExportDialog.exporterParameters`. They are used to remember the export parameters of an exporter when the user switches between exporters (e.g. SVGExporter and ImageExporter).\r\n\r\nhttps://github.com/pyqtgraph/pyqtgraph/blob/db483f8ec22a0d53bd0e71fac69b0045a96c6637/pyqtgraph/GraphicsScene/exportDialog.py#L127-L136\r\n\r\n### Reasons\r\n\r\n1. I doubt that anyone actually needs this.\r\n\r\n2. I would like to create a PR that automatically sets the correct resolution in the ImageExporter and also keeps the correct aspect ratio when the user changes width or height in the ImageExporter (This feature was intended but is currently broken). The problem is that a simple fix would invalidate the rationale behind `ExportDialog.exporterParameters`, because changing the scene item would also update the ImageExporter parameters (and the stored parameters are not used). An elaborate fix would add a lot of clutter code (creating a dictionary of dictionaries for each item and each exporter).\r\nThe simple fix is highlighted here:\r\nhttps://github.com/paulmueller/pyqtgraph/blob/b81f6d66d2b3cf4d334aa31b75b71e68cc7a3796/pyqtgraph/GraphicsScene/exportDialog.py#L101-L111\r\n\r\n### Related Bug\r\n\r\nAs mentioned in my second point, the ImageExporter currently only displays the height and width of the scene and it does not adjust when e.g. a PlotItem is selected (which has different height, widht, aspect ratio). Furthermore, width and height are not updated to match the PlotItem aspect ratio if I change one of width or height.\r\n\r\n### Question\r\n\r\nWould you accept a PR that gets rid of `ExportDialog.exporterParameters` and fixes the related bug?\nPropose removal of `GraphicsScene.ExportDialog.ExporterParameters`\nI propose to remove `ExportDialog.exporterParameters`. They are used to remember the export parameters of an exporter when the user switches between exporters (e.g. SVGExporter and ImageExporter).\r\n\r\nhttps://github.com/pyqtgraph/pyqtgraph/blob/db483f8ec22a0d53bd0e71fac69b0045a96c6637/pyqtgraph/GraphicsScene/exportDialog.py#L127-L136\r\n\r\n### Reasons\r\n\r\n1. I doubt that anyone actually needs this.\r\n\r\n2. I would like to create a PR that automatically sets the correct resolution in the ImageExporter and also keeps the correct aspect ratio when the user changes width or height in the ImageExporter (This feature was intended but is currently broken). The problem is that a simple fix would invalidate the rationale behind `ExportDialog.exporterParameters`, because changing the scene item would also update the ImageExporter parameters (and the stored parameters are not used). An elaborate fix would add a lot of clutter code (creating a dictionary of dictionaries for each item and each exporter).\r\nThe simple fix is highlighted here:\r\nhttps://github.com/paulmueller/pyqtgraph/blob/b81f6d66d2b3cf4d334aa31b75b71e68cc7a3796/pyqtgraph/GraphicsScene/exportDialog.py#L101-L111\r\n\r\n### Related Bug\r\n\r\nAs mentioned in my second point, the ImageExporter currently only displays the height and width of the scene and it does not adjust when e.g. a PlotItem is selected (which has different height, widht, aspect ratio). Furthermore, width and height are not updated to match the PlotItem aspect ratio if I change one of width or height.\r\n\r\n### Question\r\n\r\nWould you accept a PR that gets rid of `ExportDialog.exporterParameters` and fixes the related bug?\n", "before_files": [{"content": "from ..Qt import QtCore, QtGui, QT_LIB\nfrom .. import exporters as exporters\nfrom .. import functions as fn\nfrom ..graphicsItems.ViewBox import ViewBox\nfrom ..graphicsItems.PlotItem import PlotItem\n\nif QT_LIB == 'PySide':\n from . import exportDialogTemplate_pyside as exportDialogTemplate\nelif QT_LIB == 'PySide2':\n from . import exportDialogTemplate_pyside2 as exportDialogTemplate\nelif QT_LIB == 'PyQt5':\n from . import exportDialogTemplate_pyqt5 as exportDialogTemplate\nelse:\n from . import exportDialogTemplate_pyqt as exportDialogTemplate\n\n\nclass ExportDialog(QtGui.QWidget):\n def __init__(self, scene):\n QtGui.QWidget.__init__(self)\n self.setVisible(False)\n self.setWindowTitle(\"Export\")\n self.shown = False\n self.currentExporter = None\n self.scene = scene\n \n self.exporterParameters = {}\n\n self.selectBox = QtGui.QGraphicsRectItem()\n self.selectBox.setPen(fn.mkPen('y', width=3, style=QtCore.Qt.DashLine))\n self.selectBox.hide()\n self.scene.addItem(self.selectBox)\n \n self.ui = exportDialogTemplate.Ui_Form()\n self.ui.setupUi(self)\n \n self.ui.closeBtn.clicked.connect(self.close)\n self.ui.exportBtn.clicked.connect(self.exportClicked)\n self.ui.copyBtn.clicked.connect(self.copyClicked)\n self.ui.itemTree.currentItemChanged.connect(self.exportItemChanged)\n self.ui.formatList.currentItemChanged.connect(self.exportFormatChanged)\n \n\n def show(self, item=None):\n if item is not None:\n ## Select next exportable parent of the item originally clicked on\n while not isinstance(item, ViewBox) and not isinstance(item, PlotItem) and item is not None:\n item = item.parentItem()\n ## if this is a ViewBox inside a PlotItem, select the parent instead.\n if isinstance(item, ViewBox) and isinstance(item.parentItem(), PlotItem):\n item = item.parentItem()\n self.updateItemList(select=item)\n self.setVisible(True)\n self.activateWindow()\n self.raise_()\n self.selectBox.setVisible(True)\n \n if not self.shown:\n self.shown = True\n vcenter = self.scene.getViewWidget().geometry().center()\n self.setGeometry(vcenter.x()-self.width()/2, vcenter.y()-self.height()/2, self.width(), self.height())\n \n def updateItemList(self, select=None):\n self.ui.itemTree.clear()\n si = QtGui.QTreeWidgetItem([\"Entire Scene\"])\n si.gitem = self.scene\n self.ui.itemTree.addTopLevelItem(si)\n self.ui.itemTree.setCurrentItem(si)\n si.setExpanded(True)\n for child in self.scene.items():\n if child.parentItem() is None:\n self.updateItemTree(child, si, select=select)\n \n def updateItemTree(self, item, treeItem, select=None):\n si = None\n if isinstance(item, ViewBox):\n si = QtGui.QTreeWidgetItem(['ViewBox'])\n elif isinstance(item, PlotItem):\n si = QtGui.QTreeWidgetItem(['Plot'])\n \n if si is not None:\n si.gitem = item\n treeItem.addChild(si)\n treeItem = si\n if si.gitem is select:\n self.ui.itemTree.setCurrentItem(si)\n \n for ch in item.childItems():\n self.updateItemTree(ch, treeItem, select=select)\n \n \n def exportItemChanged(self, item, prev):\n if item is None:\n return\n if item.gitem is self.scene:\n newBounds = self.scene.views()[0].viewRect()\n else:\n newBounds = item.gitem.sceneBoundingRect()\n self.selectBox.setRect(newBounds)\n self.selectBox.show()\n self.updateFormatList()\n \n def updateFormatList(self):\n current = self.ui.formatList.currentItem()\n if current is not None:\n current = str(current.text())\n self.ui.formatList.clear()\n self.exporterClasses = {}\n gotCurrent = False\n for exp in exporters.listExporters():\n self.ui.formatList.addItem(exp.Name)\n self.exporterClasses[exp.Name] = exp\n if exp.Name == current:\n self.ui.formatList.setCurrentRow(self.ui.formatList.count()-1)\n gotCurrent = True\n \n if not gotCurrent:\n self.ui.formatList.setCurrentRow(0)\n \n def exportFormatChanged(self, item, prev):\n if item is None:\n self.currentExporter = None\n self.ui.paramTree.clear()\n return\n expClass = self.exporterClasses[str(item.text())]\n exp = expClass(item=self.ui.itemTree.currentItem().gitem)\n\n if prev:\n oldtext = str(prev.text())\n self.exporterParameters[oldtext] = self.currentExporter.parameters()\n newtext = str(item.text())\n if newtext in self.exporterParameters.keys():\n params = self.exporterParameters[newtext]\n exp.params = params\n else:\n params = exp.parameters()\n self.exporterParameters[newtext] = params\n\n if params is None:\n self.ui.paramTree.clear()\n else:\n self.ui.paramTree.setParameters(params)\n self.currentExporter = exp\n self.ui.copyBtn.setEnabled(exp.allowCopy)\n \n def exportClicked(self):\n self.selectBox.hide()\n self.currentExporter.export()\n \n def copyClicked(self):\n self.selectBox.hide()\n self.currentExporter.export(copy=True)\n \n def close(self):\n self.selectBox.setVisible(False)\n self.setVisible(False)\n\n def closeEvent(self, event):\n self.close()\n QtGui.QWidget.closeEvent(self, event)\n", "path": "pyqtgraph/GraphicsScene/exportDialog.py"}], "after_files": [{"content": "from ..Qt import QtCore, QtGui, QT_LIB\nfrom .. import exporters as exporters\nfrom .. import functions as fn\nfrom ..graphicsItems.ViewBox import ViewBox\nfrom ..graphicsItems.PlotItem import PlotItem\n\nif QT_LIB == 'PySide':\n from . import exportDialogTemplate_pyside as exportDialogTemplate\nelif QT_LIB == 'PySide2':\n from . import exportDialogTemplate_pyside2 as exportDialogTemplate\nelif QT_LIB == 'PyQt5':\n from . import exportDialogTemplate_pyqt5 as exportDialogTemplate\nelse:\n from . import exportDialogTemplate_pyqt as exportDialogTemplate\n\n\nclass ExportDialog(QtGui.QWidget):\n def __init__(self, scene):\n QtGui.QWidget.__init__(self)\n self.setVisible(False)\n self.setWindowTitle(\"Export\")\n self.shown = False\n self.currentExporter = None\n self.scene = scene\n\n self.selectBox = QtGui.QGraphicsRectItem()\n self.selectBox.setPen(fn.mkPen('y', width=3, style=QtCore.Qt.DashLine))\n self.selectBox.hide()\n self.scene.addItem(self.selectBox)\n \n self.ui = exportDialogTemplate.Ui_Form()\n self.ui.setupUi(self)\n \n self.ui.closeBtn.clicked.connect(self.close)\n self.ui.exportBtn.clicked.connect(self.exportClicked)\n self.ui.copyBtn.clicked.connect(self.copyClicked)\n self.ui.itemTree.currentItemChanged.connect(self.exportItemChanged)\n self.ui.formatList.currentItemChanged.connect(self.exportFormatChanged)\n \n\n def show(self, item=None):\n if item is not None:\n ## Select next exportable parent of the item originally clicked on\n while not isinstance(item, ViewBox) and not isinstance(item, PlotItem) and item is not None:\n item = item.parentItem()\n ## if this is a ViewBox inside a PlotItem, select the parent instead.\n if isinstance(item, ViewBox) and isinstance(item.parentItem(), PlotItem):\n item = item.parentItem()\n self.updateItemList(select=item)\n self.setVisible(True)\n self.activateWindow()\n self.raise_()\n self.selectBox.setVisible(True)\n \n if not self.shown:\n self.shown = True\n vcenter = self.scene.getViewWidget().geometry().center()\n self.setGeometry(vcenter.x()-self.width()/2, vcenter.y()-self.height()/2, self.width(), self.height())\n \n def updateItemList(self, select=None):\n self.ui.itemTree.clear()\n si = QtGui.QTreeWidgetItem([\"Entire Scene\"])\n si.gitem = self.scene\n self.ui.itemTree.addTopLevelItem(si)\n self.ui.itemTree.setCurrentItem(si)\n si.setExpanded(True)\n for child in self.scene.items():\n if child.parentItem() is None:\n self.updateItemTree(child, si, select=select)\n \n def updateItemTree(self, item, treeItem, select=None):\n si = None\n if isinstance(item, ViewBox):\n si = QtGui.QTreeWidgetItem(['ViewBox'])\n elif isinstance(item, PlotItem):\n si = QtGui.QTreeWidgetItem(['Plot'])\n \n if si is not None:\n si.gitem = item\n treeItem.addChild(si)\n treeItem = si\n if si.gitem is select:\n self.ui.itemTree.setCurrentItem(si)\n \n for ch in item.childItems():\n self.updateItemTree(ch, treeItem, select=select)\n \n \n def exportItemChanged(self, item, prev):\n if item is None:\n return\n if item.gitem is self.scene:\n newBounds = self.scene.views()[0].viewRect()\n else:\n newBounds = item.gitem.sceneBoundingRect()\n self.selectBox.setRect(newBounds)\n self.selectBox.show()\n self.updateFormatList()\n \n def updateFormatList(self):\n current = self.ui.formatList.currentItem()\n if current is not None:\n current = str(current.text())\n self.ui.formatList.clear()\n self.exporterClasses = {}\n gotCurrent = False\n for exp in exporters.listExporters():\n self.ui.formatList.addItem(exp.Name)\n self.exporterClasses[exp.Name] = exp\n if exp.Name == current:\n self.ui.formatList.setCurrentRow(self.ui.formatList.count()-1)\n gotCurrent = True\n \n if not gotCurrent:\n self.ui.formatList.setCurrentRow(0)\n \n def exportFormatChanged(self, item, prev):\n if item is None:\n self.currentExporter = None\n self.ui.paramTree.clear()\n return\n expClass = self.exporterClasses[str(item.text())]\n exp = expClass(item=self.ui.itemTree.currentItem().gitem)\n\n params = exp.parameters()\n\n if params is None:\n self.ui.paramTree.clear()\n else:\n self.ui.paramTree.setParameters(params)\n self.currentExporter = exp\n self.ui.copyBtn.setEnabled(exp.allowCopy)\n \n def exportClicked(self):\n self.selectBox.hide()\n self.currentExporter.export()\n \n def copyClicked(self):\n self.selectBox.hide()\n self.currentExporter.export(copy=True)\n \n def close(self):\n self.selectBox.setVisible(False)\n self.setVisible(False)\n\n def closeEvent(self, event):\n self.close()\n QtGui.QWidget.closeEvent(self, event)\n", "path": "pyqtgraph/GraphicsScene/exportDialog.py"}]}
| 2,670 | 280 |
gh_patches_debug_32324
|
rasdani/github-patches
|
git_diff
|
modal-labs__modal-examples-695
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
remove duplicate webscraper example once #669 is merged
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `10_integrations/webscraper.py`
Content:
```
1 # ---
2 # runtimes: ["runc", "gvisor"]
3 # ---
4 import os
5
6 import modal
7
8 stub = modal.Stub("example-linkscraper")
9
10
11 playwright_image = modal.Image.debian_slim(
12 python_version="3.10"
13 ).run_commands( # Doesn't work with 3.11 yet
14 "apt-get update",
15 "apt-get install -y software-properties-common",
16 "apt-add-repository non-free",
17 "apt-add-repository contrib",
18 "pip install playwright==1.30.0",
19 "playwright install-deps chromium",
20 "playwright install chromium",
21 )
22
23
24 @stub.function(image=playwright_image)
25 async def get_links(url: str) -> set[str]:
26 from playwright.async_api import async_playwright
27
28 async with async_playwright() as p:
29 browser = await p.chromium.launch()
30 page = await browser.new_page()
31 await page.goto(url)
32 links = await page.eval_on_selector_all(
33 "a[href]", "elements => elements.map(element => element.href)"
34 )
35 await browser.close()
36
37 return set(links)
38
39
40 slack_sdk_image = modal.Image.debian_slim().pip_install("slack-sdk")
41
42
43 @stub.function(
44 image=slack_sdk_image,
45 secrets=[modal.Secret.from_name("scraper-slack-secret")],
46 )
47 def bot_token_msg(channel, message):
48 import slack_sdk
49
50 print(f"Posting {message} to #{channel}")
51 client = slack_sdk.WebClient(token=os.environ["SLACK_BOT_TOKEN"])
52 client.chat_postMessage(channel=channel, text=message)
53
54
55 @stub.function()
56 def scrape():
57 links_of_interest = ["http://modal.com"]
58
59 for links in get_links.map(links_of_interest):
60 for link in links:
61 bot_token_msg.remote("scraped-links", link)
62
63
64 @stub.function(schedule=modal.Period(days=1))
65 def daily_scrape():
66 scrape.remote()
67
68
69 @stub.local_entrypoint()
70 def run():
71 scrape.remote()
72
```
Path: `misc/webscraper.py`
Content:
```
1 # ---
2 # runtimes: ["runc", "gvisor"]
3 # ---
4 import os
5
6 import modal
7
8 stub = modal.Stub("example-linkscraper")
9
10
11 playwright_image = modal.Image.debian_slim(
12 python_version="3.10"
13 ).run_commands( # Doesn't work with 3.11 yet
14 "apt-get update",
15 "apt-get install -y software-properties-common",
16 "apt-add-repository non-free",
17 "apt-add-repository contrib",
18 "pip install playwright==1.30.0",
19 "playwright install-deps chromium",
20 "playwright install chromium",
21 )
22
23
24 @stub.function(image=playwright_image)
25 async def get_links(url: str) -> set[str]:
26 from playwright.async_api import async_playwright
27
28 async with async_playwright() as p:
29 browser = await p.chromium.launch()
30 page = await browser.new_page()
31 await page.goto(url)
32 links = await page.eval_on_selector_all(
33 "a[href]", "elements => elements.map(element => element.href)"
34 )
35 await browser.close()
36
37 return set(links)
38
39
40 slack_sdk_image = modal.Image.debian_slim().pip_install("slack-sdk")
41
42
43 @stub.function(
44 image=slack_sdk_image,
45 secrets=[modal.Secret.from_name("scraper-slack-secret")],
46 )
47 def bot_token_msg(channel, message):
48 import slack_sdk
49
50 print(f"Posting {message} to #{channel}")
51 client = slack_sdk.WebClient(token=os.environ["SLACK_BOT_TOKEN"])
52 client.chat_postMessage(channel=channel, text=message)
53
54
55 @stub.function()
56 def scrape():
57 links_of_interest = ["http://modal.com"]
58
59 for links in get_links.map(links_of_interest):
60 for link in links:
61 bot_token_msg.remote("scraped-links", link)
62
63
64 @stub.function(schedule=modal.Period(days=1))
65 def daily_scrape():
66 scrape.remote()
67
68
69 @stub.local_entrypoint()
70 def run():
71 scrape.remote()
72
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/10_integrations/webscraper.py b/10_integrations/webscraper.py
--- a/10_integrations/webscraper.py
+++ b/10_integrations/webscraper.py
@@ -15,7 +15,7 @@
"apt-get install -y software-properties-common",
"apt-add-repository non-free",
"apt-add-repository contrib",
- "pip install playwright==1.30.0",
+ "pip install playwright==1.42.0",
"playwright install-deps chromium",
"playwright install chromium",
)
diff --git a/misc/webscraper.py b/misc/webscraper.py
deleted file mode 100644
--- a/misc/webscraper.py
+++ /dev/null
@@ -1,71 +0,0 @@
-# ---
-# runtimes: ["runc", "gvisor"]
-# ---
-import os
-
-import modal
-
-stub = modal.Stub("example-linkscraper")
-
-
-playwright_image = modal.Image.debian_slim(
- python_version="3.10"
-).run_commands( # Doesn't work with 3.11 yet
- "apt-get update",
- "apt-get install -y software-properties-common",
- "apt-add-repository non-free",
- "apt-add-repository contrib",
- "pip install playwright==1.30.0",
- "playwright install-deps chromium",
- "playwright install chromium",
-)
-
-
[email protected](image=playwright_image)
-async def get_links(url: str) -> set[str]:
- from playwright.async_api import async_playwright
-
- async with async_playwright() as p:
- browser = await p.chromium.launch()
- page = await browser.new_page()
- await page.goto(url)
- links = await page.eval_on_selector_all(
- "a[href]", "elements => elements.map(element => element.href)"
- )
- await browser.close()
-
- return set(links)
-
-
-slack_sdk_image = modal.Image.debian_slim().pip_install("slack-sdk")
-
-
[email protected](
- image=slack_sdk_image,
- secrets=[modal.Secret.from_name("scraper-slack-secret")],
-)
-def bot_token_msg(channel, message):
- import slack_sdk
-
- print(f"Posting {message} to #{channel}")
- client = slack_sdk.WebClient(token=os.environ["SLACK_BOT_TOKEN"])
- client.chat_postMessage(channel=channel, text=message)
-
-
[email protected]()
-def scrape():
- links_of_interest = ["http://modal.com"]
-
- for links in get_links.map(links_of_interest):
- for link in links:
- bot_token_msg.remote("scraped-links", link)
-
-
[email protected](schedule=modal.Period(days=1))
-def daily_scrape():
- scrape.remote()
-
-
[email protected]_entrypoint()
-def run():
- scrape.remote()
|
{"golden_diff": "diff --git a/10_integrations/webscraper.py b/10_integrations/webscraper.py\n--- a/10_integrations/webscraper.py\n+++ b/10_integrations/webscraper.py\n@@ -15,7 +15,7 @@\n \"apt-get install -y software-properties-common\",\n \"apt-add-repository non-free\",\n \"apt-add-repository contrib\",\n- \"pip install playwright==1.30.0\",\n+ \"pip install playwright==1.42.0\",\n \"playwright install-deps chromium\",\n \"playwright install chromium\",\n )\ndiff --git a/misc/webscraper.py b/misc/webscraper.py\ndeleted file mode 100644\n--- a/misc/webscraper.py\n+++ /dev/null\n@@ -1,71 +0,0 @@\n-# ---\n-# runtimes: [\"runc\", \"gvisor\"]\n-# ---\n-import os\n-\n-import modal\n-\n-stub = modal.Stub(\"example-linkscraper\")\n-\n-\n-playwright_image = modal.Image.debian_slim(\n- python_version=\"3.10\"\n-).run_commands( # Doesn't work with 3.11 yet\n- \"apt-get update\",\n- \"apt-get install -y software-properties-common\",\n- \"apt-add-repository non-free\",\n- \"apt-add-repository contrib\",\n- \"pip install playwright==1.30.0\",\n- \"playwright install-deps chromium\",\n- \"playwright install chromium\",\n-)\n-\n-\[email protected](image=playwright_image)\n-async def get_links(url: str) -> set[str]:\n- from playwright.async_api import async_playwright\n-\n- async with async_playwright() as p:\n- browser = await p.chromium.launch()\n- page = await browser.new_page()\n- await page.goto(url)\n- links = await page.eval_on_selector_all(\n- \"a[href]\", \"elements => elements.map(element => element.href)\"\n- )\n- await browser.close()\n-\n- return set(links)\n-\n-\n-slack_sdk_image = modal.Image.debian_slim().pip_install(\"slack-sdk\")\n-\n-\[email protected](\n- image=slack_sdk_image,\n- secrets=[modal.Secret.from_name(\"scraper-slack-secret\")],\n-)\n-def bot_token_msg(channel, message):\n- import slack_sdk\n-\n- print(f\"Posting {message} to #{channel}\")\n- client = slack_sdk.WebClient(token=os.environ[\"SLACK_BOT_TOKEN\"])\n- client.chat_postMessage(channel=channel, text=message)\n-\n-\[email protected]()\n-def scrape():\n- links_of_interest = [\"http://modal.com\"]\n-\n- for links in get_links.map(links_of_interest):\n- for link in links:\n- bot_token_msg.remote(\"scraped-links\", link)\n-\n-\[email protected](schedule=modal.Period(days=1))\n-def daily_scrape():\n- scrape.remote()\n-\n-\[email protected]_entrypoint()\n-def run():\n- scrape.remote()\n", "issue": "remove duplicate webscraper example once #669 is merged\n\n", "before_files": [{"content": "# ---\n# runtimes: [\"runc\", \"gvisor\"]\n# ---\nimport os\n\nimport modal\n\nstub = modal.Stub(\"example-linkscraper\")\n\n\nplaywright_image = modal.Image.debian_slim(\n python_version=\"3.10\"\n).run_commands( # Doesn't work with 3.11 yet\n \"apt-get update\",\n \"apt-get install -y software-properties-common\",\n \"apt-add-repository non-free\",\n \"apt-add-repository contrib\",\n \"pip install playwright==1.30.0\",\n \"playwright install-deps chromium\",\n \"playwright install chromium\",\n)\n\n\[email protected](image=playwright_image)\nasync def get_links(url: str) -> set[str]:\n from playwright.async_api import async_playwright\n\n async with async_playwright() as p:\n browser = await p.chromium.launch()\n page = await browser.new_page()\n await page.goto(url)\n links = await page.eval_on_selector_all(\n \"a[href]\", \"elements => elements.map(element => element.href)\"\n )\n await browser.close()\n\n return set(links)\n\n\nslack_sdk_image = modal.Image.debian_slim().pip_install(\"slack-sdk\")\n\n\[email protected](\n image=slack_sdk_image,\n secrets=[modal.Secret.from_name(\"scraper-slack-secret\")],\n)\ndef bot_token_msg(channel, message):\n import slack_sdk\n\n print(f\"Posting {message} to #{channel}\")\n client = slack_sdk.WebClient(token=os.environ[\"SLACK_BOT_TOKEN\"])\n client.chat_postMessage(channel=channel, text=message)\n\n\[email protected]()\ndef scrape():\n links_of_interest = [\"http://modal.com\"]\n\n for links in get_links.map(links_of_interest):\n for link in links:\n bot_token_msg.remote(\"scraped-links\", link)\n\n\[email protected](schedule=modal.Period(days=1))\ndef daily_scrape():\n scrape.remote()\n\n\[email protected]_entrypoint()\ndef run():\n scrape.remote()\n", "path": "10_integrations/webscraper.py"}, {"content": "# ---\n# runtimes: [\"runc\", \"gvisor\"]\n# ---\nimport os\n\nimport modal\n\nstub = modal.Stub(\"example-linkscraper\")\n\n\nplaywright_image = modal.Image.debian_slim(\n python_version=\"3.10\"\n).run_commands( # Doesn't work with 3.11 yet\n \"apt-get update\",\n \"apt-get install -y software-properties-common\",\n \"apt-add-repository non-free\",\n \"apt-add-repository contrib\",\n \"pip install playwright==1.30.0\",\n \"playwright install-deps chromium\",\n \"playwright install chromium\",\n)\n\n\[email protected](image=playwright_image)\nasync def get_links(url: str) -> set[str]:\n from playwright.async_api import async_playwright\n\n async with async_playwright() as p:\n browser = await p.chromium.launch()\n page = await browser.new_page()\n await page.goto(url)\n links = await page.eval_on_selector_all(\n \"a[href]\", \"elements => elements.map(element => element.href)\"\n )\n await browser.close()\n\n return set(links)\n\n\nslack_sdk_image = modal.Image.debian_slim().pip_install(\"slack-sdk\")\n\n\[email protected](\n image=slack_sdk_image,\n secrets=[modal.Secret.from_name(\"scraper-slack-secret\")],\n)\ndef bot_token_msg(channel, message):\n import slack_sdk\n\n print(f\"Posting {message} to #{channel}\")\n client = slack_sdk.WebClient(token=os.environ[\"SLACK_BOT_TOKEN\"])\n client.chat_postMessage(channel=channel, text=message)\n\n\[email protected]()\ndef scrape():\n links_of_interest = [\"http://modal.com\"]\n\n for links in get_links.map(links_of_interest):\n for link in links:\n bot_token_msg.remote(\"scraped-links\", link)\n\n\[email protected](schedule=modal.Period(days=1))\ndef daily_scrape():\n scrape.remote()\n\n\[email protected]_entrypoint()\ndef run():\n scrape.remote()\n", "path": "misc/webscraper.py"}], "after_files": [{"content": "# ---\n# runtimes: [\"runc\", \"gvisor\"]\n# ---\nimport os\n\nimport modal\n\nstub = modal.Stub(\"example-linkscraper\")\n\n\nplaywright_image = modal.Image.debian_slim(\n python_version=\"3.10\"\n).run_commands( # Doesn't work with 3.11 yet\n \"apt-get update\",\n \"apt-get install -y software-properties-common\",\n \"apt-add-repository non-free\",\n \"apt-add-repository contrib\",\n \"pip install playwright==1.42.0\",\n \"playwright install-deps chromium\",\n \"playwright install chromium\",\n)\n\n\[email protected](image=playwright_image)\nasync def get_links(url: str) -> set[str]:\n from playwright.async_api import async_playwright\n\n async with async_playwright() as p:\n browser = await p.chromium.launch()\n page = await browser.new_page()\n await page.goto(url)\n links = await page.eval_on_selector_all(\n \"a[href]\", \"elements => elements.map(element => element.href)\"\n )\n await browser.close()\n\n return set(links)\n\n\nslack_sdk_image = modal.Image.debian_slim().pip_install(\"slack-sdk\")\n\n\[email protected](\n image=slack_sdk_image,\n secrets=[modal.Secret.from_name(\"scraper-slack-secret\")],\n)\ndef bot_token_msg(channel, message):\n import slack_sdk\n\n print(f\"Posting {message} to #{channel}\")\n client = slack_sdk.WebClient(token=os.environ[\"SLACK_BOT_TOKEN\"])\n client.chat_postMessage(channel=channel, text=message)\n\n\[email protected]()\ndef scrape():\n links_of_interest = [\"http://modal.com\"]\n\n for links in get_links.map(links_of_interest):\n for link in links:\n bot_token_msg.remote(\"scraped-links\", link)\n\n\[email protected](schedule=modal.Period(days=1))\ndef daily_scrape():\n scrape.remote()\n\n\[email protected]_entrypoint()\ndef run():\n scrape.remote()\n", "path": "10_integrations/webscraper.py"}, {"content": null, "path": "misc/webscraper.py"}]}
| 1,426 | 665 |
gh_patches_debug_14271
|
rasdani/github-patches
|
git_diff
|
getsentry__snuba-3697
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
About the sessions-subscription-results subscription issue
### Environment
- sentry | snuba version :23.1.1
https://github.com/getsentry/snuba/pull/2737 ,@lynnagara Hello, I have a question about this pr, hope to get your answer, thank you very much
- After removing subscriptions-scheduler-executor-session support in snuba, how to write data to the topic of sessions-subscription-results? Because I see that the crash rate warning code in sentry is still there and has not changed, for example
- https://github.com/getsentry/sentry/pull/28526
https://github.com/getsentry/sentry/blob/8e00dcdf463d916b9ca79ddbe13e99f161d58db1/src/sentry/snuba/query_subscription_consumer.py#L61-L61
My original question is as follows, I have enabled the organizations:incidents function in sentry and subscribed to sessions-results through the following script
```bash
sentry
run
query-subscription-consumer
--topic=sessions-subscription-results
```
Because there is no data in the sessions-subscription-results topic, the crash rate alarm cannot work
<img width="1568" alt="image" src="https://user-images.githubusercontent.com/18591662/216570393-64748a25-1cd4-4980-966c-f7665dc8482b.png">
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `snuba/cli/subscriptions_scheduler_executor.py`
Content:
```
1 import signal
2 from contextlib import contextmanager
3 from typing import Any, Iterator, Optional, Sequence
4
5 import click
6 from arroyo import configure_metrics
7 from arroyo.backends.kafka import KafkaProducer
8
9 from snuba import environment, state
10 from snuba.attribution.log import flush_attribution_producer
11 from snuba.datasets.entities.entity_key import EntityKey
12 from snuba.datasets.entities.factory import get_entity
13 from snuba.environment import setup_logging, setup_sentry
14 from snuba.subscriptions.combined_scheduler_executor import (
15 build_scheduler_executor_consumer,
16 )
17 from snuba.utils.metrics.wrapper import MetricsWrapper
18 from snuba.utils.streams.configuration_builder import build_kafka_producer_configuration
19 from snuba.utils.streams.metrics_adapter import StreamMetricsAdapter
20
21
22 @click.command()
23 @click.option(
24 "--dataset",
25 "dataset_name",
26 required=True,
27 type=click.Choice(["events", "transactions", "metrics"]),
28 help="The dataset to target.",
29 )
30 @click.option(
31 "--entity",
32 "entity_names",
33 required=True,
34 multiple=True,
35 type=click.Choice(["events", "transactions", "metrics_counters", "metrics_sets"]),
36 help="The entity to target.",
37 )
38 @click.option(
39 "--consumer-group",
40 default="snuba-subscription-executor",
41 help="Consumer group used for consuming the scheduled subscription topic/s.",
42 )
43 @click.option(
44 "--followed-consumer-group",
45 required=True,
46 help="Name of the consumer group to follow",
47 )
48 @click.option(
49 "--total-concurrent-queries",
50 default=64,
51 type=int,
52 help="Total max number of concurrent queries for all replicas. Used to calculate max_concurrent_queries.",
53 )
54 @click.option(
55 "--auto-offset-reset",
56 default="error",
57 type=click.Choice(["error", "earliest", "latest"]),
58 help="Kafka consumer auto offset reset.",
59 )
60 @click.option(
61 "--no-strict-offset-reset",
62 is_flag=True,
63 help="Forces the kafka consumer auto offset reset.",
64 )
65 @click.option("--schedule-ttl", type=int, default=60 * 5)
66 @click.option("--delay-seconds", type=int)
67 @click.option(
68 "--stale-threshold-seconds",
69 type=int,
70 help="Skip scheduling if timestamp is beyond this threshold compared to the system time",
71 )
72 @click.option("--log-level", help="Logging level to use.")
73 def subscriptions_scheduler_executor(
74 *,
75 dataset_name: str,
76 entity_names: Sequence[str],
77 consumer_group: str,
78 followed_consumer_group: str,
79 total_concurrent_queries: int,
80 auto_offset_reset: str,
81 no_strict_offset_reset: bool,
82 schedule_ttl: int,
83 delay_seconds: Optional[int],
84 stale_threshold_seconds: Optional[int],
85 log_level: Optional[str],
86 ) -> None:
87 """
88 Combined subscriptions scheduler and executor. Alternative to the separate scheduler and executor processes.
89 """
90 setup_logging(log_level)
91 setup_sentry()
92
93 metrics = MetricsWrapper(
94 environment.metrics,
95 "subscriptions.scheduler_executor",
96 tags={"dataset": dataset_name},
97 )
98
99 configure_metrics(StreamMetricsAdapter(metrics))
100
101 # Just get the result topic configuration from the first entity. Later we
102 # check they all have the same result topic anyway before building the consumer.
103 entity_key = EntityKey(entity_names[0])
104
105 storage = get_entity(entity_key).get_writable_storage()
106 assert storage is not None
107 stream_loader = storage.get_table_writer().get_stream_loader()
108 result_topic_spec = stream_loader.get_subscription_scheduled_topic_spec()
109 assert result_topic_spec is not None
110
111 producer = KafkaProducer(
112 build_kafka_producer_configuration(
113 result_topic_spec.topic,
114 override_params={"partitioner": "consistent"},
115 )
116 )
117
118 processor = build_scheduler_executor_consumer(
119 dataset_name,
120 entity_names,
121 consumer_group,
122 followed_consumer_group,
123 producer,
124 auto_offset_reset,
125 not no_strict_offset_reset,
126 schedule_ttl,
127 delay_seconds,
128 stale_threshold_seconds,
129 total_concurrent_queries,
130 metrics,
131 )
132
133 def handler(signum: int, frame: Any) -> None:
134 processor.signal_shutdown()
135
136 signal.signal(signal.SIGINT, handler)
137 signal.signal(signal.SIGTERM, handler)
138
139 with closing(producer), flush_querylog(), flush_attribution_producer():
140 processor.run()
141
142
143 @contextmanager
144 def closing(producer: KafkaProducer) -> Iterator[Optional[KafkaProducer]]:
145 try:
146 yield producer
147 finally:
148 producer.close().result()
149
150
151 @contextmanager
152 def flush_querylog() -> Iterator[None]:
153 try:
154 yield
155 finally:
156 state.flush_producer()
157
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/snuba/cli/subscriptions_scheduler_executor.py b/snuba/cli/subscriptions_scheduler_executor.py
--- a/snuba/cli/subscriptions_scheduler_executor.py
+++ b/snuba/cli/subscriptions_scheduler_executor.py
@@ -24,7 +24,7 @@
"--dataset",
"dataset_name",
required=True,
- type=click.Choice(["events", "transactions", "metrics"]),
+ type=click.Choice(["events", "transactions", "metrics", "sessions"]),
help="The dataset to target.",
)
@click.option(
@@ -32,7 +32,9 @@
"entity_names",
required=True,
multiple=True,
- type=click.Choice(["events", "transactions", "metrics_counters", "metrics_sets"]),
+ type=click.Choice(
+ ["events", "transactions", "metrics_counters", "metrics_sets", "sessions"]
+ ),
help="The entity to target.",
)
@click.option(
|
{"golden_diff": "diff --git a/snuba/cli/subscriptions_scheduler_executor.py b/snuba/cli/subscriptions_scheduler_executor.py\n--- a/snuba/cli/subscriptions_scheduler_executor.py\n+++ b/snuba/cli/subscriptions_scheduler_executor.py\n@@ -24,7 +24,7 @@\n \"--dataset\",\n \"dataset_name\",\n required=True,\n- type=click.Choice([\"events\", \"transactions\", \"metrics\"]),\n+ type=click.Choice([\"events\", \"transactions\", \"metrics\", \"sessions\"]),\n help=\"The dataset to target.\",\n )\n @click.option(\n@@ -32,7 +32,9 @@\n \"entity_names\",\n required=True,\n multiple=True,\n- type=click.Choice([\"events\", \"transactions\", \"metrics_counters\", \"metrics_sets\"]),\n+ type=click.Choice(\n+ [\"events\", \"transactions\", \"metrics_counters\", \"metrics_sets\", \"sessions\"]\n+ ),\n help=\"The entity to target.\",\n )\n @click.option(\n", "issue": "About the sessions-subscription-results subscription issue\n### Environment\r\n\r\n- sentry | snuba version \uff1a23.1.1\r\n\r\nhttps://github.com/getsentry/snuba/pull/2737 \uff0c@lynnagara Hello, I have a question about this pr, hope to get your answer, thank you very much\r\n\r\n- After removing subscriptions-scheduler-executor-session support in snuba, how to write data to the topic of sessions-subscription-results? Because I see that the crash rate warning code in sentry is still there and has not changed, for example\r\n\r\n- https://github.com/getsentry/sentry/pull/28526\r\n\r\nhttps://github.com/getsentry/sentry/blob/8e00dcdf463d916b9ca79ddbe13e99f161d58db1/src/sentry/snuba/query_subscription_consumer.py#L61-L61\r\n\r\nMy original question is as follows, I have enabled the organizations:incidents function in sentry and subscribed to sessions-results through the following script\r\n```bash\r\nsentry\r\nrun\r\nquery-subscription-consumer\r\n--topic=sessions-subscription-results\r\n```\r\nBecause there is no data in the sessions-subscription-results topic, the crash rate alarm cannot work\r\n<img width=\"1568\" alt=\"image\" src=\"https://user-images.githubusercontent.com/18591662/216570393-64748a25-1cd4-4980-966c-f7665dc8482b.png\">\r\n\r\n\r\n\r\n\r\n\r\n\n", "before_files": [{"content": "import signal\nfrom contextlib import contextmanager\nfrom typing import Any, Iterator, Optional, Sequence\n\nimport click\nfrom arroyo import configure_metrics\nfrom arroyo.backends.kafka import KafkaProducer\n\nfrom snuba import environment, state\nfrom snuba.attribution.log import flush_attribution_producer\nfrom snuba.datasets.entities.entity_key import EntityKey\nfrom snuba.datasets.entities.factory import get_entity\nfrom snuba.environment import setup_logging, setup_sentry\nfrom snuba.subscriptions.combined_scheduler_executor import (\n build_scheduler_executor_consumer,\n)\nfrom snuba.utils.metrics.wrapper import MetricsWrapper\nfrom snuba.utils.streams.configuration_builder import build_kafka_producer_configuration\nfrom snuba.utils.streams.metrics_adapter import StreamMetricsAdapter\n\n\[email protected]()\[email protected](\n \"--dataset\",\n \"dataset_name\",\n required=True,\n type=click.Choice([\"events\", \"transactions\", \"metrics\"]),\n help=\"The dataset to target.\",\n)\[email protected](\n \"--entity\",\n \"entity_names\",\n required=True,\n multiple=True,\n type=click.Choice([\"events\", \"transactions\", \"metrics_counters\", \"metrics_sets\"]),\n help=\"The entity to target.\",\n)\[email protected](\n \"--consumer-group\",\n default=\"snuba-subscription-executor\",\n help=\"Consumer group used for consuming the scheduled subscription topic/s.\",\n)\[email protected](\n \"--followed-consumer-group\",\n required=True,\n help=\"Name of the consumer group to follow\",\n)\[email protected](\n \"--total-concurrent-queries\",\n default=64,\n type=int,\n help=\"Total max number of concurrent queries for all replicas. Used to calculate max_concurrent_queries.\",\n)\[email protected](\n \"--auto-offset-reset\",\n default=\"error\",\n type=click.Choice([\"error\", \"earliest\", \"latest\"]),\n help=\"Kafka consumer auto offset reset.\",\n)\[email protected](\n \"--no-strict-offset-reset\",\n is_flag=True,\n help=\"Forces the kafka consumer auto offset reset.\",\n)\[email protected](\"--schedule-ttl\", type=int, default=60 * 5)\[email protected](\"--delay-seconds\", type=int)\[email protected](\n \"--stale-threshold-seconds\",\n type=int,\n help=\"Skip scheduling if timestamp is beyond this threshold compared to the system time\",\n)\[email protected](\"--log-level\", help=\"Logging level to use.\")\ndef subscriptions_scheduler_executor(\n *,\n dataset_name: str,\n entity_names: Sequence[str],\n consumer_group: str,\n followed_consumer_group: str,\n total_concurrent_queries: int,\n auto_offset_reset: str,\n no_strict_offset_reset: bool,\n schedule_ttl: int,\n delay_seconds: Optional[int],\n stale_threshold_seconds: Optional[int],\n log_level: Optional[str],\n) -> None:\n \"\"\"\n Combined subscriptions scheduler and executor. Alternative to the separate scheduler and executor processes.\n \"\"\"\n setup_logging(log_level)\n setup_sentry()\n\n metrics = MetricsWrapper(\n environment.metrics,\n \"subscriptions.scheduler_executor\",\n tags={\"dataset\": dataset_name},\n )\n\n configure_metrics(StreamMetricsAdapter(metrics))\n\n # Just get the result topic configuration from the first entity. Later we\n # check they all have the same result topic anyway before building the consumer.\n entity_key = EntityKey(entity_names[0])\n\n storage = get_entity(entity_key).get_writable_storage()\n assert storage is not None\n stream_loader = storage.get_table_writer().get_stream_loader()\n result_topic_spec = stream_loader.get_subscription_scheduled_topic_spec()\n assert result_topic_spec is not None\n\n producer = KafkaProducer(\n build_kafka_producer_configuration(\n result_topic_spec.topic,\n override_params={\"partitioner\": \"consistent\"},\n )\n )\n\n processor = build_scheduler_executor_consumer(\n dataset_name,\n entity_names,\n consumer_group,\n followed_consumer_group,\n producer,\n auto_offset_reset,\n not no_strict_offset_reset,\n schedule_ttl,\n delay_seconds,\n stale_threshold_seconds,\n total_concurrent_queries,\n metrics,\n )\n\n def handler(signum: int, frame: Any) -> None:\n processor.signal_shutdown()\n\n signal.signal(signal.SIGINT, handler)\n signal.signal(signal.SIGTERM, handler)\n\n with closing(producer), flush_querylog(), flush_attribution_producer():\n processor.run()\n\n\n@contextmanager\ndef closing(producer: KafkaProducer) -> Iterator[Optional[KafkaProducer]]:\n try:\n yield producer\n finally:\n producer.close().result()\n\n\n@contextmanager\ndef flush_querylog() -> Iterator[None]:\n try:\n yield\n finally:\n state.flush_producer()\n", "path": "snuba/cli/subscriptions_scheduler_executor.py"}], "after_files": [{"content": "import signal\nfrom contextlib import contextmanager\nfrom typing import Any, Iterator, Optional, Sequence\n\nimport click\nfrom arroyo import configure_metrics\nfrom arroyo.backends.kafka import KafkaProducer\n\nfrom snuba import environment, state\nfrom snuba.attribution.log import flush_attribution_producer\nfrom snuba.datasets.entities.entity_key import EntityKey\nfrom snuba.datasets.entities.factory import get_entity\nfrom snuba.environment import setup_logging, setup_sentry\nfrom snuba.subscriptions.combined_scheduler_executor import (\n build_scheduler_executor_consumer,\n)\nfrom snuba.utils.metrics.wrapper import MetricsWrapper\nfrom snuba.utils.streams.configuration_builder import build_kafka_producer_configuration\nfrom snuba.utils.streams.metrics_adapter import StreamMetricsAdapter\n\n\[email protected]()\[email protected](\n \"--dataset\",\n \"dataset_name\",\n required=True,\n type=click.Choice([\"events\", \"transactions\", \"metrics\", \"sessions\"]),\n help=\"The dataset to target.\",\n)\[email protected](\n \"--entity\",\n \"entity_names\",\n required=True,\n multiple=True,\n type=click.Choice(\n [\"events\", \"transactions\", \"metrics_counters\", \"metrics_sets\", \"sessions\"]\n ),\n help=\"The entity to target.\",\n)\[email protected](\n \"--consumer-group\",\n default=\"snuba-subscription-executor\",\n help=\"Consumer group used for consuming the scheduled subscription topic/s.\",\n)\[email protected](\n \"--followed-consumer-group\",\n required=True,\n help=\"Name of the consumer group to follow\",\n)\[email protected](\n \"--total-concurrent-queries\",\n default=64,\n type=int,\n help=\"Total max number of concurrent queries for all replicas. Used to calculate max_concurrent_queries.\",\n)\[email protected](\n \"--auto-offset-reset\",\n default=\"error\",\n type=click.Choice([\"error\", \"earliest\", \"latest\"]),\n help=\"Kafka consumer auto offset reset.\",\n)\[email protected](\n \"--no-strict-offset-reset\",\n is_flag=True,\n help=\"Forces the kafka consumer auto offset reset.\",\n)\[email protected](\"--schedule-ttl\", type=int, default=60 * 5)\[email protected](\"--delay-seconds\", type=int)\[email protected](\n \"--stale-threshold-seconds\",\n type=int,\n help=\"Skip scheduling if timestamp is beyond this threshold compared to the system time\",\n)\[email protected](\"--log-level\", help=\"Logging level to use.\")\ndef subscriptions_scheduler_executor(\n *,\n dataset_name: str,\n entity_names: Sequence[str],\n consumer_group: str,\n followed_consumer_group: str,\n total_concurrent_queries: int,\n auto_offset_reset: str,\n no_strict_offset_reset: bool,\n schedule_ttl: int,\n delay_seconds: Optional[int],\n stale_threshold_seconds: Optional[int],\n log_level: Optional[str],\n) -> None:\n \"\"\"\n Combined subscriptions scheduler and executor. Alternative to the separate scheduler and executor processes.\n \"\"\"\n setup_logging(log_level)\n setup_sentry()\n\n metrics = MetricsWrapper(\n environment.metrics,\n \"subscriptions.scheduler_executor\",\n tags={\"dataset\": dataset_name},\n )\n\n configure_metrics(StreamMetricsAdapter(metrics))\n\n # Just get the result topic configuration from the first entity. Later we\n # check they all have the same result topic anyway before building the consumer.\n entity_key = EntityKey(entity_names[0])\n\n storage = get_entity(entity_key).get_writable_storage()\n assert storage is not None\n stream_loader = storage.get_table_writer().get_stream_loader()\n result_topic_spec = stream_loader.get_subscription_scheduled_topic_spec()\n assert result_topic_spec is not None\n\n producer = KafkaProducer(\n build_kafka_producer_configuration(\n result_topic_spec.topic,\n override_params={\"partitioner\": \"consistent\"},\n )\n )\n\n processor = build_scheduler_executor_consumer(\n dataset_name,\n entity_names,\n consumer_group,\n followed_consumer_group,\n producer,\n auto_offset_reset,\n not no_strict_offset_reset,\n schedule_ttl,\n delay_seconds,\n stale_threshold_seconds,\n total_concurrent_queries,\n metrics,\n )\n\n def handler(signum: int, frame: Any) -> None:\n processor.signal_shutdown()\n\n signal.signal(signal.SIGINT, handler)\n signal.signal(signal.SIGTERM, handler)\n\n with closing(producer), flush_querylog(), flush_attribution_producer():\n processor.run()\n\n\n@contextmanager\ndef closing(producer: KafkaProducer) -> Iterator[Optional[KafkaProducer]]:\n try:\n yield producer\n finally:\n producer.close().result()\n\n\n@contextmanager\ndef flush_querylog() -> Iterator[None]:\n try:\n yield\n finally:\n state.flush_producer()\n", "path": "snuba/cli/subscriptions_scheduler_executor.py"}]}
| 1,973 | 206 |
gh_patches_debug_20728
|
rasdani/github-patches
|
git_diff
|
hpcaitech__ColossalAI-5096
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[tensor] fix some unittests
[tensor] fix some unittests
[tensor] fix some unittests
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `colossalai/inference/engine/policies/llama.py`
Content:
```
1 from functools import partial
2 from typing import List
3
4 import torch
5 from torch.nn import Module
6 from transformers.models.llama.modeling_llama import (
7 LlamaAttention,
8 LlamaDecoderLayer,
9 LlamaForCausalLM,
10 LlamaModel,
11 LlamaRMSNorm,
12 )
13
14 from colossalai.shardformer.policies.base_policy import ModulePolicyDescription, SubModuleReplacementDescription
15
16 # import colossalai
17 from colossalai.shardformer.policies.llama import LlamaForCausalLMPolicy
18
19 from ..modeling._utils import init_to_get_rotary
20 from ..modeling.llama import LlamaInferenceForwards
21
22 try:
23 from colossalai.kernel.triton import rmsnorm_forward
24
25 HAS_TRITON_RMSNORM = True
26 except:
27 print("you should install triton from https://github.com/openai/triton")
28 HAS_TRITON_RMSNORM = False
29
30
31 def get_triton_rmsnorm_forward():
32 if HAS_TRITON_RMSNORM:
33
34 def _triton_rmsnorm_forward(self: LlamaRMSNorm, hidden_states: torch.Tensor):
35 return rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)
36
37 return _triton_rmsnorm_forward
38 else:
39 return None
40
41
42 class LlamaModelInferPolicy(LlamaForCausalLMPolicy):
43 def __init__(self) -> None:
44 super().__init__()
45
46 def module_policy(self):
47 policy = super().module_policy()
48 decoder_attribute_replacement = {
49 "self_attn.hidden_size": self.model.config.hidden_size // self.shard_config.tensor_parallel_size,
50 "self_attn.num_heads": self.model.config.num_attention_heads // self.shard_config.tensor_parallel_size,
51 "self_attn.num_key_value_heads": self.model.config.num_key_value_heads
52 // self.shard_config.tensor_parallel_size,
53 }
54 if self.shard_config.extra_kwargs.get("quant", None) == "gptq":
55 from colossalai.inference.quant.gptq.cai_gptq import ColCaiQuantLinear, RowCaiQuantLinear
56
57 policy[LlamaDecoderLayer] = ModulePolicyDescription(
58 attribute_replacement=decoder_attribute_replacement,
59 sub_module_replacement=[
60 SubModuleReplacementDescription(
61 suffix="self_attn.q_proj",
62 target_module=ColCaiQuantLinear,
63 kwargs={"split_num": 1},
64 ),
65 SubModuleReplacementDescription(
66 suffix="self_attn.k_proj",
67 target_module=ColCaiQuantLinear,
68 kwargs={"split_num": 1},
69 ),
70 SubModuleReplacementDescription(
71 suffix="self_attn.v_proj",
72 target_module=ColCaiQuantLinear,
73 kwargs={"split_num": 1},
74 ),
75 SubModuleReplacementDescription(
76 suffix="self_attn.o_proj",
77 target_module=RowCaiQuantLinear,
78 kwargs={"split_num": 1},
79 ),
80 SubModuleReplacementDescription(
81 suffix="mlp.gate_proj",
82 target_module=ColCaiQuantLinear,
83 kwargs={"split_num": 1},
84 ),
85 SubModuleReplacementDescription(
86 suffix="mlp.up_proj",
87 target_module=ColCaiQuantLinear,
88 kwargs={"split_num": 1},
89 ),
90 SubModuleReplacementDescription(
91 suffix="mlp.down_proj",
92 target_module=RowCaiQuantLinear,
93 kwargs={"split_num": 1},
94 ),
95 ],
96 )
97
98 elif self.shard_config.extra_kwargs.get("quant", None) == "smoothquant":
99 from colossalai.inference.quant.smoothquant.models.llama import LlamaSmoothquantDecoderLayer
100 from colossalai.inference.quant.smoothquant.models.parallel_linear import (
101 ColW8A8BFP32OFP32Linear,
102 RowW8A8B8O8Linear,
103 RowW8A8BFP32O32LinearSiLU,
104 RowW8A8BFP32OFP32Linear,
105 )
106
107 policy[LlamaSmoothquantDecoderLayer] = ModulePolicyDescription(
108 attribute_replacement=decoder_attribute_replacement,
109 sub_module_replacement=[
110 SubModuleReplacementDescription(
111 suffix="self_attn.q_proj",
112 target_module=RowW8A8B8O8Linear,
113 kwargs={"split_num": 1},
114 ),
115 SubModuleReplacementDescription(
116 suffix="self_attn.k_proj",
117 target_module=RowW8A8B8O8Linear,
118 kwargs={"split_num": 1},
119 ),
120 SubModuleReplacementDescription(
121 suffix="self_attn.v_proj",
122 target_module=RowW8A8B8O8Linear,
123 kwargs={"split_num": 1},
124 ),
125 SubModuleReplacementDescription(
126 suffix="self_attn.o_proj",
127 target_module=ColW8A8BFP32OFP32Linear,
128 kwargs={"split_num": 1},
129 ),
130 SubModuleReplacementDescription(
131 suffix="mlp.gate_proj",
132 target_module=RowW8A8BFP32O32LinearSiLU,
133 kwargs={"split_num": 1},
134 ),
135 SubModuleReplacementDescription(
136 suffix="mlp.up_proj",
137 target_module=RowW8A8BFP32OFP32Linear,
138 kwargs={"split_num": 1},
139 ),
140 SubModuleReplacementDescription(
141 suffix="mlp.down_proj",
142 target_module=ColW8A8BFP32OFP32Linear,
143 kwargs={"split_num": 1},
144 ),
145 ],
146 )
147 self.shard_config._infer()
148
149 infer_forward = LlamaInferenceForwards.llama_model_forward
150 method_replacement = {"forward": partial(infer_forward)}
151 self.append_or_create_method_replacement(description=method_replacement, policy=policy, target_key=LlamaModel)
152
153 infer_forward = LlamaInferenceForwards.llama_decoder_layer_forward
154 method_replacement = {"forward": partial(infer_forward)}
155 self.append_or_create_method_replacement(
156 description=method_replacement, policy=policy, target_key=LlamaDecoderLayer
157 )
158
159 infer_forward = LlamaInferenceForwards.llama_flash_attn_kvcache_forward
160 method_replacement = {"forward": partial(infer_forward)}
161 self.append_or_create_method_replacement(
162 description=method_replacement, policy=policy, target_key=LlamaAttention
163 )
164
165 # set as default, in inference we also use pipeline style forward, just setting stage as 1
166 self.set_pipeline_forward(
167 model_cls=LlamaForCausalLM, new_forward=LlamaInferenceForwards.llama_causal_lm_forward, policy=policy
168 )
169
170 infer_forward = None
171 if HAS_TRITON_RMSNORM:
172 infer_forward = get_triton_rmsnorm_forward()
173
174 if infer_forward is not None:
175 method_replacement = {"forward": partial(infer_forward)}
176 self.append_or_create_method_replacement(
177 description=method_replacement, policy=policy, target_key=LlamaRMSNorm
178 )
179
180 return policy
181
182 def postprocess(self):
183 init_to_get_rotary(self.model.model)
184 return self.model
185
186 def get_held_layers(self) -> List[Module]:
187 """Get pipeline layers for current stage."""
188 assert self.pipeline_stage_manager is not None
189
190 if self.model.__class__.__name__ == "LlamaModel":
191 module = self.model
192 else:
193 module = self.model.model
194 stage_manager = self.pipeline_stage_manager
195
196 held_layers = []
197 layers_per_stage = self.distribute_layers(len(module.layers), stage_manager.num_stages)
198 if stage_manager.is_first_stage():
199 held_layers.append(module.embed_tokens)
200 held_layers.append(self.model.lm_head)
201 start_idx, end_idx = self.get_stage_index(layers_per_stage, stage_manager.stage)
202 held_layers.extend(module.layers[start_idx:end_idx])
203 if stage_manager.is_last_stage():
204 held_layers.append(module.norm)
205
206 return held_layers
207
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/colossalai/inference/engine/policies/llama.py b/colossalai/inference/engine/policies/llama.py
--- a/colossalai/inference/engine/policies/llama.py
+++ b/colossalai/inference/engine/policies/llama.py
@@ -20,23 +20,17 @@
from ..modeling.llama import LlamaInferenceForwards
try:
- from colossalai.kernel.triton import rmsnorm_forward
-
+ from lightllm.models.llama.triton_kernel.rmsnorm import rmsnorm_forward as lightllm_rmsnorm_forward
HAS_TRITON_RMSNORM = True
except:
print("you should install triton from https://github.com/openai/triton")
HAS_TRITON_RMSNORM = False
-
-def get_triton_rmsnorm_forward():
- if HAS_TRITON_RMSNORM:
-
+if HAS_TRITON_RMSNORM:
+ def get_triton_rmsnorm_forward():
def _triton_rmsnorm_forward(self: LlamaRMSNorm, hidden_states: torch.Tensor):
- return rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)
-
+ return lightllm_rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)
return _triton_rmsnorm_forward
- else:
- return None
class LlamaModelInferPolicy(LlamaForCausalLMPolicy):
|
{"golden_diff": "diff --git a/colossalai/inference/engine/policies/llama.py b/colossalai/inference/engine/policies/llama.py\n--- a/colossalai/inference/engine/policies/llama.py\n+++ b/colossalai/inference/engine/policies/llama.py\n@@ -20,23 +20,17 @@\n from ..modeling.llama import LlamaInferenceForwards\n \n try:\n- from colossalai.kernel.triton import rmsnorm_forward\n-\n+ from lightllm.models.llama.triton_kernel.rmsnorm import rmsnorm_forward as lightllm_rmsnorm_forward\n HAS_TRITON_RMSNORM = True\n except:\n print(\"you should install triton from https://github.com/openai/triton\")\n HAS_TRITON_RMSNORM = False\n \n-\n-def get_triton_rmsnorm_forward():\n- if HAS_TRITON_RMSNORM:\n-\n+if HAS_TRITON_RMSNORM:\n+ def get_triton_rmsnorm_forward():\n def _triton_rmsnorm_forward(self: LlamaRMSNorm, hidden_states: torch.Tensor):\n- return rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)\n-\n+ return lightllm_rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)\n return _triton_rmsnorm_forward\n- else:\n- return None\n \n \n class LlamaModelInferPolicy(LlamaForCausalLMPolicy):\n", "issue": "[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n[tensor] fix some unittests\n\n", "before_files": [{"content": "from functools import partial\nfrom typing import List\n\nimport torch\nfrom torch.nn import Module\nfrom transformers.models.llama.modeling_llama import (\n LlamaAttention,\n LlamaDecoderLayer,\n LlamaForCausalLM,\n LlamaModel,\n LlamaRMSNorm,\n)\n\nfrom colossalai.shardformer.policies.base_policy import ModulePolicyDescription, SubModuleReplacementDescription\n\n# import colossalai\nfrom colossalai.shardformer.policies.llama import LlamaForCausalLMPolicy\n\nfrom ..modeling._utils import init_to_get_rotary\nfrom ..modeling.llama import LlamaInferenceForwards\n\ntry:\n from colossalai.kernel.triton import rmsnorm_forward\n\n HAS_TRITON_RMSNORM = True\nexcept:\n print(\"you should install triton from https://github.com/openai/triton\")\n HAS_TRITON_RMSNORM = False\n\n\ndef get_triton_rmsnorm_forward():\n if HAS_TRITON_RMSNORM:\n\n def _triton_rmsnorm_forward(self: LlamaRMSNorm, hidden_states: torch.Tensor):\n return rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)\n\n return _triton_rmsnorm_forward\n else:\n return None\n\n\nclass LlamaModelInferPolicy(LlamaForCausalLMPolicy):\n def __init__(self) -> None:\n super().__init__()\n\n def module_policy(self):\n policy = super().module_policy()\n decoder_attribute_replacement = {\n \"self_attn.hidden_size\": self.model.config.hidden_size // self.shard_config.tensor_parallel_size,\n \"self_attn.num_heads\": self.model.config.num_attention_heads // self.shard_config.tensor_parallel_size,\n \"self_attn.num_key_value_heads\": self.model.config.num_key_value_heads\n // self.shard_config.tensor_parallel_size,\n }\n if self.shard_config.extra_kwargs.get(\"quant\", None) == \"gptq\":\n from colossalai.inference.quant.gptq.cai_gptq import ColCaiQuantLinear, RowCaiQuantLinear\n\n policy[LlamaDecoderLayer] = ModulePolicyDescription(\n attribute_replacement=decoder_attribute_replacement,\n sub_module_replacement=[\n SubModuleReplacementDescription(\n suffix=\"self_attn.q_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.k_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.v_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.o_proj\",\n target_module=RowCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.gate_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.up_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.down_proj\",\n target_module=RowCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n ],\n )\n\n elif self.shard_config.extra_kwargs.get(\"quant\", None) == \"smoothquant\":\n from colossalai.inference.quant.smoothquant.models.llama import LlamaSmoothquantDecoderLayer\n from colossalai.inference.quant.smoothquant.models.parallel_linear import (\n ColW8A8BFP32OFP32Linear,\n RowW8A8B8O8Linear,\n RowW8A8BFP32O32LinearSiLU,\n RowW8A8BFP32OFP32Linear,\n )\n\n policy[LlamaSmoothquantDecoderLayer] = ModulePolicyDescription(\n attribute_replacement=decoder_attribute_replacement,\n sub_module_replacement=[\n SubModuleReplacementDescription(\n suffix=\"self_attn.q_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.k_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.v_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.o_proj\",\n target_module=ColW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.gate_proj\",\n target_module=RowW8A8BFP32O32LinearSiLU,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.up_proj\",\n target_module=RowW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.down_proj\",\n target_module=ColW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n ],\n )\n self.shard_config._infer()\n\n infer_forward = LlamaInferenceForwards.llama_model_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(description=method_replacement, policy=policy, target_key=LlamaModel)\n\n infer_forward = LlamaInferenceForwards.llama_decoder_layer_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaDecoderLayer\n )\n\n infer_forward = LlamaInferenceForwards.llama_flash_attn_kvcache_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaAttention\n )\n\n # set as default, in inference we also use pipeline style forward, just setting stage as 1\n self.set_pipeline_forward(\n model_cls=LlamaForCausalLM, new_forward=LlamaInferenceForwards.llama_causal_lm_forward, policy=policy\n )\n\n infer_forward = None\n if HAS_TRITON_RMSNORM:\n infer_forward = get_triton_rmsnorm_forward()\n\n if infer_forward is not None:\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaRMSNorm\n )\n\n return policy\n\n def postprocess(self):\n init_to_get_rotary(self.model.model)\n return self.model\n\n def get_held_layers(self) -> List[Module]:\n \"\"\"Get pipeline layers for current stage.\"\"\"\n assert self.pipeline_stage_manager is not None\n\n if self.model.__class__.__name__ == \"LlamaModel\":\n module = self.model\n else:\n module = self.model.model\n stage_manager = self.pipeline_stage_manager\n\n held_layers = []\n layers_per_stage = self.distribute_layers(len(module.layers), stage_manager.num_stages)\n if stage_manager.is_first_stage():\n held_layers.append(module.embed_tokens)\n held_layers.append(self.model.lm_head)\n start_idx, end_idx = self.get_stage_index(layers_per_stage, stage_manager.stage)\n held_layers.extend(module.layers[start_idx:end_idx])\n if stage_manager.is_last_stage():\n held_layers.append(module.norm)\n\n return held_layers\n", "path": "colossalai/inference/engine/policies/llama.py"}], "after_files": [{"content": "from functools import partial\nfrom typing import List\n\nimport torch\nfrom torch.nn import Module\nfrom transformers.models.llama.modeling_llama import (\n LlamaAttention,\n LlamaDecoderLayer,\n LlamaForCausalLM,\n LlamaModel,\n LlamaRMSNorm,\n)\n\nfrom colossalai.shardformer.policies.base_policy import ModulePolicyDescription, SubModuleReplacementDescription\n\n# import colossalai\nfrom colossalai.shardformer.policies.llama import LlamaForCausalLMPolicy\n\nfrom ..modeling._utils import init_to_get_rotary\nfrom ..modeling.llama import LlamaInferenceForwards\n\ntry:\n from lightllm.models.llama.triton_kernel.rmsnorm import rmsnorm_forward as lightllm_rmsnorm_forward\n HAS_TRITON_RMSNORM = True\nexcept:\n print(\"you should install triton from https://github.com/openai/triton\")\n HAS_TRITON_RMSNORM = False\n\nif HAS_TRITON_RMSNORM:\n def get_triton_rmsnorm_forward():\n def _triton_rmsnorm_forward(self: LlamaRMSNorm, hidden_states: torch.Tensor):\n return lightllm_rmsnorm_forward(hidden_states, self.weight.data, self.variance_epsilon)\n return _triton_rmsnorm_forward\n\n\nclass LlamaModelInferPolicy(LlamaForCausalLMPolicy):\n def __init__(self) -> None:\n super().__init__()\n\n def module_policy(self):\n policy = super().module_policy()\n decoder_attribute_replacement = {\n \"self_attn.hidden_size\": self.model.config.hidden_size // self.shard_config.tensor_parallel_size,\n \"self_attn.num_heads\": self.model.config.num_attention_heads // self.shard_config.tensor_parallel_size,\n \"self_attn.num_key_value_heads\": self.model.config.num_key_value_heads\n // self.shard_config.tensor_parallel_size,\n }\n if self.shard_config.extra_kwargs.get(\"quant\", None) == \"gptq\":\n from colossalai.inference.quant.gptq.cai_gptq import ColCaiQuantLinear, RowCaiQuantLinear\n\n policy[LlamaDecoderLayer] = ModulePolicyDescription(\n attribute_replacement=decoder_attribute_replacement,\n sub_module_replacement=[\n SubModuleReplacementDescription(\n suffix=\"self_attn.q_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.k_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.v_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.o_proj\",\n target_module=RowCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.gate_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.up_proj\",\n target_module=ColCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.down_proj\",\n target_module=RowCaiQuantLinear,\n kwargs={\"split_num\": 1},\n ),\n ],\n )\n\n elif self.shard_config.extra_kwargs.get(\"quant\", None) == \"smoothquant\":\n from colossalai.inference.quant.smoothquant.models.llama import LlamaSmoothquantDecoderLayer\n from colossalai.inference.quant.smoothquant.models.parallel_linear import (\n ColW8A8BFP32OFP32Linear,\n RowW8A8B8O8Linear,\n RowW8A8BFP32O32LinearSiLU,\n RowW8A8BFP32OFP32Linear,\n )\n\n policy[LlamaSmoothquantDecoderLayer] = ModulePolicyDescription(\n attribute_replacement=decoder_attribute_replacement,\n sub_module_replacement=[\n SubModuleReplacementDescription(\n suffix=\"self_attn.q_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.k_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.v_proj\",\n target_module=RowW8A8B8O8Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"self_attn.o_proj\",\n target_module=ColW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.gate_proj\",\n target_module=RowW8A8BFP32O32LinearSiLU,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.up_proj\",\n target_module=RowW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n SubModuleReplacementDescription(\n suffix=\"mlp.down_proj\",\n target_module=ColW8A8BFP32OFP32Linear,\n kwargs={\"split_num\": 1},\n ),\n ],\n )\n self.shard_config._infer()\n\n infer_forward = LlamaInferenceForwards.llama_model_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(description=method_replacement, policy=policy, target_key=LlamaModel)\n\n infer_forward = LlamaInferenceForwards.llama_decoder_layer_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaDecoderLayer\n )\n\n infer_forward = LlamaInferenceForwards.llama_flash_attn_kvcache_forward\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaAttention\n )\n\n # set as default, in inference we also use pipeline style forward, just setting stage as 1\n self.set_pipeline_forward(\n model_cls=LlamaForCausalLM, new_forward=LlamaInferenceForwards.llama_causal_lm_forward, policy=policy\n )\n\n infer_forward = None\n if HAS_TRITON_RMSNORM:\n infer_forward = get_triton_rmsnorm_forward()\n\n if infer_forward is not None:\n method_replacement = {\"forward\": partial(infer_forward)}\n self.append_or_create_method_replacement(\n description=method_replacement, policy=policy, target_key=LlamaRMSNorm\n )\n\n return policy\n\n def postprocess(self):\n init_to_get_rotary(self.model.model)\n return self.model\n\n def get_held_layers(self) -> List[Module]:\n \"\"\"Get pipeline layers for current stage.\"\"\"\n assert self.pipeline_stage_manager is not None\n\n if self.model.__class__.__name__ == \"LlamaModel\":\n module = self.model\n else:\n module = self.model.model\n stage_manager = self.pipeline_stage_manager\n\n held_layers = []\n layers_per_stage = self.distribute_layers(len(module.layers), stage_manager.num_stages)\n if stage_manager.is_first_stage():\n held_layers.append(module.embed_tokens)\n held_layers.append(self.model.lm_head)\n start_idx, end_idx = self.get_stage_index(layers_per_stage, stage_manager.stage)\n held_layers.extend(module.layers[start_idx:end_idx])\n if stage_manager.is_last_stage():\n held_layers.append(module.norm)\n\n return held_layers\n", "path": "colossalai/inference/engine/policies/llama.py"}]}
| 2,522 | 321 |
gh_patches_debug_24137
|
rasdani/github-patches
|
git_diff
|
ansible-collections__community.vmware-1007
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Change service startup policy with vmware_host_service_manager without defining service state
<!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
When setting up the service in vSphere Client, the actions to set startup policy or service state are independent. However, when setting the service startup policy using the `vmware_host_service_manager` module, you have to specify the service state if you don't want to start the service, which is the default behavior.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
`vmware_host_service_manager` in community.vmware v1.11.0
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
Using the Ansible module should probably match the vSphere Client management behavior. It is not necessary to know the service state before changing the startup policy. The workaround is to use `vmware_host_service_info` module to first gather the state and then use it (which is kinda complicated way as the output values of the variable `running` from the `vmware_host_service_info` don't match the input values of the variable `state` in `vmware_host_service_manager`).
The `state` value of `unchanged` could be added (and set as default?). The current default is `start`.
Example playbook with the changes implemented:
<!--- Paste example playbooks or commands between quotes below -->
```yaml
- hosts: localhost
gather_facts: false
tasks:
- name: Disabling SSH service
community.vmware.vmware_host_service_manager:
esxi_hostname: <esxi_hostname>
service_name: 'TSM-SSH'
service_policy: 'off'
state: 'unchanged'
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `plugins/modules/vmware_host_service_manager.py`
Content:
```
1 #!/usr/bin/python
2 # -*- coding: utf-8 -*-
3
4 # Copyright: (c) 2018, Abhijeet Kasurde <[email protected]>
5 # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
6
7 from __future__ import absolute_import, division, print_function
8 __metaclass__ = type
9
10
11 DOCUMENTATION = r'''
12 ---
13 module: vmware_host_service_manager
14 short_description: Manage services on a given ESXi host
15 description:
16 - This module can be used to manage (start, stop, restart) services on a given ESXi host.
17 - If cluster_name is provided, specified service will be managed on all ESXi host belonging to that cluster.
18 - If specific esxi_hostname is provided, then specified service will be managed on given ESXi host only.
19 author:
20 - Abhijeet Kasurde (@Akasurde)
21 notes:
22 - Tested on vSphere 6.5
23 requirements:
24 - python >= 2.6
25 - PyVmomi
26 options:
27 cluster_name:
28 description:
29 - Name of the cluster.
30 - Service settings are applied to every ESXi host system/s in given cluster.
31 - If C(esxi_hostname) is not given, this parameter is required.
32 type: str
33 esxi_hostname:
34 description:
35 - ESXi hostname.
36 - Service settings are applied to this ESXi host system.
37 - If C(cluster_name) is not given, this parameter is required.
38 type: str
39 state:
40 description:
41 - Desired state of service.
42 - "State value 'start' and 'present' has same effect."
43 - "State value 'stop' and 'absent' has same effect."
44 choices: [ absent, present, restart, start, stop ]
45 type: str
46 default: 'start'
47 service_policy:
48 description:
49 - Set of valid service policy strings.
50 - If set C(on), then service should be started when the host starts up.
51 - If set C(automatic), then service should run if and only if it has open firewall ports.
52 - If set C(off), then Service should not be started when the host starts up.
53 choices: [ 'automatic', 'off', 'on' ]
54 type: str
55 service_name:
56 description:
57 - Name of Service to be managed. This is a brief identifier for the service, for example, ntpd, vxsyslogd etc.
58 - This value should be a valid ESXi service name.
59 required: True
60 type: str
61 extends_documentation_fragment:
62 - community.vmware.vmware.documentation
63
64 '''
65
66 EXAMPLES = r'''
67 - name: Start ntpd service setting for all ESXi Host in given Cluster
68 community.vmware.vmware_host_service_manager:
69 hostname: '{{ vcenter_hostname }}'
70 username: '{{ vcenter_username }}'
71 password: '{{ vcenter_password }}'
72 cluster_name: '{{ cluster_name }}'
73 service_name: ntpd
74 state: present
75 delegate_to: localhost
76
77 - name: Start ntpd setting for an ESXi Host
78 community.vmware.vmware_host_service_manager:
79 hostname: '{{ vcenter_hostname }}'
80 username: '{{ vcenter_username }}'
81 password: '{{ vcenter_password }}'
82 esxi_hostname: '{{ esxi_hostname }}'
83 service_name: ntpd
84 state: present
85 delegate_to: localhost
86
87 - name: Start ntpd setting for an ESXi Host with Service policy
88 community.vmware.vmware_host_service_manager:
89 hostname: '{{ vcenter_hostname }}'
90 username: '{{ vcenter_username }}'
91 password: '{{ vcenter_password }}'
92 esxi_hostname: '{{ esxi_hostname }}'
93 service_name: ntpd
94 service_policy: on
95 state: present
96 delegate_to: localhost
97
98 - name: Stop ntpd setting for an ESXi Host
99 community.vmware.vmware_host_service_manager:
100 hostname: '{{ vcenter_hostname }}'
101 username: '{{ vcenter_username }}'
102 password: '{{ vcenter_password }}'
103 esxi_hostname: '{{ esxi_hostname }}'
104 service_name: ntpd
105 state: absent
106 delegate_to: localhost
107 '''
108
109 RETURN = r'''#
110 '''
111
112 try:
113 from pyVmomi import vim, vmodl
114 except ImportError:
115 pass
116
117 from ansible.module_utils.basic import AnsibleModule
118 from ansible_collections.community.vmware.plugins.module_utils.vmware import vmware_argument_spec, PyVmomi
119 from ansible.module_utils._text import to_native
120
121
122 class VmwareServiceManager(PyVmomi):
123 def __init__(self, module):
124 super(VmwareServiceManager, self).__init__(module)
125 cluster_name = self.params.get('cluster_name', None)
126 esxi_host_name = self.params.get('esxi_hostname', None)
127 self.options = self.params.get('options', dict())
128 self.hosts = self.get_all_host_objs(cluster_name=cluster_name, esxi_host_name=esxi_host_name)
129 self.desired_state = self.params.get('state')
130 self.desired_policy = self.params.get('service_policy', None)
131 self.service_name = self.params.get('service_name')
132 self.results = {}
133
134 def service_ctrl(self):
135 changed = False
136 host_service_state = []
137 for host in self.hosts:
138 actual_service_state, actual_service_policy = self.check_service_state(host=host, service_name=self.service_name)
139 host_service_system = host.configManager.serviceSystem
140 if host_service_system:
141 changed_state = False
142 self.results[host.name] = dict(service_name=self.service_name,
143 actual_service_state='running' if actual_service_state else 'stopped',
144 actual_service_policy=actual_service_policy,
145 desired_service_policy=self.desired_policy,
146 desired_service_state=self.desired_state,
147 error='',
148 )
149 try:
150 if self.desired_state in ['start', 'present']:
151 if not actual_service_state:
152 if not self.module.check_mode:
153 host_service_system.StartService(id=self.service_name)
154 changed_state = True
155 elif self.desired_state in ['stop', 'absent']:
156 if actual_service_state:
157 if not self.module.check_mode:
158 host_service_system.StopService(id=self.service_name)
159 changed_state = True
160 elif self.desired_state == 'restart':
161 if not self.module.check_mode:
162 host_service_system.RestartService(id=self.service_name)
163 changed_state = True
164
165 if self.desired_policy:
166 if actual_service_policy != self.desired_policy:
167 if not self.module.check_mode:
168 host_service_system.UpdateServicePolicy(id=self.service_name,
169 policy=self.desired_policy)
170 changed_state = True
171
172 host_service_state.append(changed_state)
173 self.results[host.name].update(changed=changed_state)
174 except (vim.fault.InvalidState, vim.fault.NotFound,
175 vim.fault.HostConfigFault, vmodl.fault.InvalidArgument) as e:
176 self.results[host.name].update(changed=False,
177 error=to_native(e.msg))
178
179 if any(host_service_state):
180 changed = True
181 self.module.exit_json(changed=changed, host_service_status=self.results)
182
183 def check_service_state(self, host, service_name):
184 host_service_system = host.configManager.serviceSystem
185 if host_service_system:
186 services = host_service_system.serviceInfo.service
187 for service in services:
188 if service.key == service_name:
189 return service.running, service.policy
190
191 msg = "Failed to find '%s' service on host system '%s'" % (service_name, host.name)
192 cluster_name = self.params.get('cluster_name', None)
193 if cluster_name:
194 msg += " located on cluster '%s'" % cluster_name
195 msg += ", please check if you have specified a valid ESXi service name."
196 self.module.fail_json(msg=msg)
197
198
199 def main():
200 argument_spec = vmware_argument_spec()
201 argument_spec.update(
202 cluster_name=dict(type='str', required=False),
203 esxi_hostname=dict(type='str', required=False),
204 state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop']),
205 service_name=dict(type='str', required=True),
206 service_policy=dict(type='str', choices=['automatic', 'off', 'on']),
207 )
208
209 module = AnsibleModule(
210 argument_spec=argument_spec,
211 required_one_of=[
212 ['cluster_name', 'esxi_hostname'],
213 ],
214 supports_check_mode=True
215 )
216
217 vmware_host_service = VmwareServiceManager(module)
218 vmware_host_service.service_ctrl()
219
220
221 if __name__ == "__main__":
222 main()
223
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/plugins/modules/vmware_host_service_manager.py b/plugins/modules/vmware_host_service_manager.py
--- a/plugins/modules/vmware_host_service_manager.py
+++ b/plugins/modules/vmware_host_service_manager.py
@@ -41,7 +41,8 @@
- Desired state of service.
- "State value 'start' and 'present' has same effect."
- "State value 'stop' and 'absent' has same effect."
- choices: [ absent, present, restart, start, stop ]
+ - State value C(unchanged) is added in version 1.14.0 to allow defining startup policy without defining or changing service state.
+ choices: [ absent, present, restart, start, stop, unchanged ]
type: str
default: 'start'
service_policy:
@@ -201,7 +202,7 @@
argument_spec.update(
cluster_name=dict(type='str', required=False),
esxi_hostname=dict(type='str', required=False),
- state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop']),
+ state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop', 'unchanged']),
service_name=dict(type='str', required=True),
service_policy=dict(type='str', choices=['automatic', 'off', 'on']),
)
|
{"golden_diff": "diff --git a/plugins/modules/vmware_host_service_manager.py b/plugins/modules/vmware_host_service_manager.py\n--- a/plugins/modules/vmware_host_service_manager.py\n+++ b/plugins/modules/vmware_host_service_manager.py\n@@ -41,7 +41,8 @@\n - Desired state of service.\n - \"State value 'start' and 'present' has same effect.\"\n - \"State value 'stop' and 'absent' has same effect.\"\n- choices: [ absent, present, restart, start, stop ]\n+ - State value C(unchanged) is added in version 1.14.0 to allow defining startup policy without defining or changing service state.\n+ choices: [ absent, present, restart, start, stop, unchanged ]\n type: str\n default: 'start'\n service_policy:\n@@ -201,7 +202,7 @@\n argument_spec.update(\n cluster_name=dict(type='str', required=False),\n esxi_hostname=dict(type='str', required=False),\n- state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop']),\n+ state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop', 'unchanged']),\n service_name=dict(type='str', required=True),\n service_policy=dict(type='str', choices=['automatic', 'off', 'on']),\n )\n", "issue": "Change service startup policy with vmware_host_service_manager without defining service state\n<!--- Verify first that your feature was not already discussed on GitHub -->\r\n<!--- Complete *all* sections as described, this form is processed automatically -->\r\n\r\n##### SUMMARY\r\n<!--- Describe the new feature/improvement briefly below -->\r\nWhen setting up the service in vSphere Client, the actions to set startup policy or service state are independent. However, when setting the service startup policy using the `vmware_host_service_manager` module, you have to specify the service state if you don't want to start the service, which is the default behavior.\r\n\r\n##### ISSUE TYPE\r\n- Feature Idea\r\n\r\n##### COMPONENT NAME\r\n<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->\r\n`vmware_host_service_manager` in community.vmware v1.11.0\r\n\r\n##### ADDITIONAL INFORMATION\r\n<!--- Describe how the feature would be used, why it is needed and what it would solve -->\r\nUsing the Ansible module should probably match the vSphere Client management behavior. It is not necessary to know the service state before changing the startup policy. The workaround is to use `vmware_host_service_info` module to first gather the state and then use it (which is kinda complicated way as the output values of the variable `running` from the `vmware_host_service_info` don't match the input values of the variable `state` in `vmware_host_service_manager`).\r\n\r\nThe `state` value of `unchanged` could be added (and set as default?). The current default is `start`.\r\n\r\nExample playbook with the changes implemented:\r\n<!--- Paste example playbooks or commands between quotes below -->\r\n```yaml\r\n- hosts: localhost\r\n gather_facts: false\r\n tasks:\r\n - name: Disabling SSH service\r\n community.vmware.vmware_host_service_manager:\r\n esxi_hostname: <esxi_hostname>\r\n service_name: 'TSM-SSH'\r\n service_policy: 'off'\r\n state: 'unchanged'\r\n```\r\n\r\n<!--- HINT: You can also paste gist.github.com links for larger files -->\r\n\n", "before_files": [{"content": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n# Copyright: (c) 2018, Abhijeet Kasurde <[email protected]>\n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\n\nDOCUMENTATION = r'''\n---\nmodule: vmware_host_service_manager\nshort_description: Manage services on a given ESXi host\ndescription:\n- This module can be used to manage (start, stop, restart) services on a given ESXi host.\n- If cluster_name is provided, specified service will be managed on all ESXi host belonging to that cluster.\n- If specific esxi_hostname is provided, then specified service will be managed on given ESXi host only.\nauthor:\n- Abhijeet Kasurde (@Akasurde)\nnotes:\n- Tested on vSphere 6.5\nrequirements:\n- python >= 2.6\n- PyVmomi\noptions:\n cluster_name:\n description:\n - Name of the cluster.\n - Service settings are applied to every ESXi host system/s in given cluster.\n - If C(esxi_hostname) is not given, this parameter is required.\n type: str\n esxi_hostname:\n description:\n - ESXi hostname.\n - Service settings are applied to this ESXi host system.\n - If C(cluster_name) is not given, this parameter is required.\n type: str\n state:\n description:\n - Desired state of service.\n - \"State value 'start' and 'present' has same effect.\"\n - \"State value 'stop' and 'absent' has same effect.\"\n choices: [ absent, present, restart, start, stop ]\n type: str\n default: 'start'\n service_policy:\n description:\n - Set of valid service policy strings.\n - If set C(on), then service should be started when the host starts up.\n - If set C(automatic), then service should run if and only if it has open firewall ports.\n - If set C(off), then Service should not be started when the host starts up.\n choices: [ 'automatic', 'off', 'on' ]\n type: str\n service_name:\n description:\n - Name of Service to be managed. This is a brief identifier for the service, for example, ntpd, vxsyslogd etc.\n - This value should be a valid ESXi service name.\n required: True\n type: str\nextends_documentation_fragment:\n- community.vmware.vmware.documentation\n\n'''\n\nEXAMPLES = r'''\n- name: Start ntpd service setting for all ESXi Host in given Cluster\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n cluster_name: '{{ cluster_name }}'\n service_name: ntpd\n state: present\n delegate_to: localhost\n\n- name: Start ntpd setting for an ESXi Host\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n state: present\n delegate_to: localhost\n\n- name: Start ntpd setting for an ESXi Host with Service policy\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n service_policy: on\n state: present\n delegate_to: localhost\n\n- name: Stop ntpd setting for an ESXi Host\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n state: absent\n delegate_to: localhost\n'''\n\nRETURN = r'''#\n'''\n\ntry:\n from pyVmomi import vim, vmodl\nexcept ImportError:\n pass\n\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ansible_collections.community.vmware.plugins.module_utils.vmware import vmware_argument_spec, PyVmomi\nfrom ansible.module_utils._text import to_native\n\n\nclass VmwareServiceManager(PyVmomi):\n def __init__(self, module):\n super(VmwareServiceManager, self).__init__(module)\n cluster_name = self.params.get('cluster_name', None)\n esxi_host_name = self.params.get('esxi_hostname', None)\n self.options = self.params.get('options', dict())\n self.hosts = self.get_all_host_objs(cluster_name=cluster_name, esxi_host_name=esxi_host_name)\n self.desired_state = self.params.get('state')\n self.desired_policy = self.params.get('service_policy', None)\n self.service_name = self.params.get('service_name')\n self.results = {}\n\n def service_ctrl(self):\n changed = False\n host_service_state = []\n for host in self.hosts:\n actual_service_state, actual_service_policy = self.check_service_state(host=host, service_name=self.service_name)\n host_service_system = host.configManager.serviceSystem\n if host_service_system:\n changed_state = False\n self.results[host.name] = dict(service_name=self.service_name,\n actual_service_state='running' if actual_service_state else 'stopped',\n actual_service_policy=actual_service_policy,\n desired_service_policy=self.desired_policy,\n desired_service_state=self.desired_state,\n error='',\n )\n try:\n if self.desired_state in ['start', 'present']:\n if not actual_service_state:\n if not self.module.check_mode:\n host_service_system.StartService(id=self.service_name)\n changed_state = True\n elif self.desired_state in ['stop', 'absent']:\n if actual_service_state:\n if not self.module.check_mode:\n host_service_system.StopService(id=self.service_name)\n changed_state = True\n elif self.desired_state == 'restart':\n if not self.module.check_mode:\n host_service_system.RestartService(id=self.service_name)\n changed_state = True\n\n if self.desired_policy:\n if actual_service_policy != self.desired_policy:\n if not self.module.check_mode:\n host_service_system.UpdateServicePolicy(id=self.service_name,\n policy=self.desired_policy)\n changed_state = True\n\n host_service_state.append(changed_state)\n self.results[host.name].update(changed=changed_state)\n except (vim.fault.InvalidState, vim.fault.NotFound,\n vim.fault.HostConfigFault, vmodl.fault.InvalidArgument) as e:\n self.results[host.name].update(changed=False,\n error=to_native(e.msg))\n\n if any(host_service_state):\n changed = True\n self.module.exit_json(changed=changed, host_service_status=self.results)\n\n def check_service_state(self, host, service_name):\n host_service_system = host.configManager.serviceSystem\n if host_service_system:\n services = host_service_system.serviceInfo.service\n for service in services:\n if service.key == service_name:\n return service.running, service.policy\n\n msg = \"Failed to find '%s' service on host system '%s'\" % (service_name, host.name)\n cluster_name = self.params.get('cluster_name', None)\n if cluster_name:\n msg += \" located on cluster '%s'\" % cluster_name\n msg += \", please check if you have specified a valid ESXi service name.\"\n self.module.fail_json(msg=msg)\n\n\ndef main():\n argument_spec = vmware_argument_spec()\n argument_spec.update(\n cluster_name=dict(type='str', required=False),\n esxi_hostname=dict(type='str', required=False),\n state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop']),\n service_name=dict(type='str', required=True),\n service_policy=dict(type='str', choices=['automatic', 'off', 'on']),\n )\n\n module = AnsibleModule(\n argument_spec=argument_spec,\n required_one_of=[\n ['cluster_name', 'esxi_hostname'],\n ],\n supports_check_mode=True\n )\n\n vmware_host_service = VmwareServiceManager(module)\n vmware_host_service.service_ctrl()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "plugins/modules/vmware_host_service_manager.py"}], "after_files": [{"content": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\n# Copyright: (c) 2018, Abhijeet Kasurde <[email protected]>\n# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)\n\nfrom __future__ import absolute_import, division, print_function\n__metaclass__ = type\n\n\nDOCUMENTATION = r'''\n---\nmodule: vmware_host_service_manager\nshort_description: Manage services on a given ESXi host\ndescription:\n- This module can be used to manage (start, stop, restart) services on a given ESXi host.\n- If cluster_name is provided, specified service will be managed on all ESXi host belonging to that cluster.\n- If specific esxi_hostname is provided, then specified service will be managed on given ESXi host only.\nauthor:\n- Abhijeet Kasurde (@Akasurde)\nnotes:\n- Tested on vSphere 6.5\nrequirements:\n- python >= 2.6\n- PyVmomi\noptions:\n cluster_name:\n description:\n - Name of the cluster.\n - Service settings are applied to every ESXi host system/s in given cluster.\n - If C(esxi_hostname) is not given, this parameter is required.\n type: str\n esxi_hostname:\n description:\n - ESXi hostname.\n - Service settings are applied to this ESXi host system.\n - If C(cluster_name) is not given, this parameter is required.\n type: str\n state:\n description:\n - Desired state of service.\n - \"State value 'start' and 'present' has same effect.\"\n - \"State value 'stop' and 'absent' has same effect.\"\n - State value C(unchanged) is added in version 1.14.0 to allow defining startup policy without defining or changing service state.\n choices: [ absent, present, restart, start, stop, unchanged ]\n type: str\n default: 'start'\n service_policy:\n description:\n - Set of valid service policy strings.\n - If set C(on), then service should be started when the host starts up.\n - If set C(automatic), then service should run if and only if it has open firewall ports.\n - If set C(off), then Service should not be started when the host starts up.\n choices: [ 'automatic', 'off', 'on' ]\n type: str\n service_name:\n description:\n - Name of Service to be managed. This is a brief identifier for the service, for example, ntpd, vxsyslogd etc.\n - This value should be a valid ESXi service name.\n required: True\n type: str\nextends_documentation_fragment:\n- community.vmware.vmware.documentation\n\n'''\n\nEXAMPLES = r'''\n- name: Start ntpd service setting for all ESXi Host in given Cluster\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n cluster_name: '{{ cluster_name }}'\n service_name: ntpd\n state: present\n delegate_to: localhost\n\n- name: Start ntpd setting for an ESXi Host\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n state: present\n delegate_to: localhost\n\n- name: Start ntpd setting for an ESXi Host with Service policy\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n service_policy: on\n state: present\n delegate_to: localhost\n\n- name: Stop ntpd setting for an ESXi Host\n community.vmware.vmware_host_service_manager:\n hostname: '{{ vcenter_hostname }}'\n username: '{{ vcenter_username }}'\n password: '{{ vcenter_password }}'\n esxi_hostname: '{{ esxi_hostname }}'\n service_name: ntpd\n state: absent\n delegate_to: localhost\n'''\n\nRETURN = r'''#\n'''\n\ntry:\n from pyVmomi import vim, vmodl\nexcept ImportError:\n pass\n\nfrom ansible.module_utils.basic import AnsibleModule\nfrom ansible_collections.community.vmware.plugins.module_utils.vmware import vmware_argument_spec, PyVmomi\nfrom ansible.module_utils._text import to_native\n\n\nclass VmwareServiceManager(PyVmomi):\n def __init__(self, module):\n super(VmwareServiceManager, self).__init__(module)\n cluster_name = self.params.get('cluster_name', None)\n esxi_host_name = self.params.get('esxi_hostname', None)\n self.options = self.params.get('options', dict())\n self.hosts = self.get_all_host_objs(cluster_name=cluster_name, esxi_host_name=esxi_host_name)\n self.desired_state = self.params.get('state')\n self.desired_policy = self.params.get('service_policy', None)\n self.service_name = self.params.get('service_name')\n self.results = {}\n\n def service_ctrl(self):\n changed = False\n host_service_state = []\n for host in self.hosts:\n actual_service_state, actual_service_policy = self.check_service_state(host=host, service_name=self.service_name)\n host_service_system = host.configManager.serviceSystem\n if host_service_system:\n changed_state = False\n self.results[host.name] = dict(service_name=self.service_name,\n actual_service_state='running' if actual_service_state else 'stopped',\n actual_service_policy=actual_service_policy,\n desired_service_policy=self.desired_policy,\n desired_service_state=self.desired_state,\n error='',\n )\n try:\n if self.desired_state in ['start', 'present']:\n if not actual_service_state:\n if not self.module.check_mode:\n host_service_system.StartService(id=self.service_name)\n changed_state = True\n elif self.desired_state in ['stop', 'absent']:\n if actual_service_state:\n if not self.module.check_mode:\n host_service_system.StopService(id=self.service_name)\n changed_state = True\n elif self.desired_state == 'restart':\n if not self.module.check_mode:\n host_service_system.RestartService(id=self.service_name)\n changed_state = True\n\n if self.desired_policy:\n if actual_service_policy != self.desired_policy:\n if not self.module.check_mode:\n host_service_system.UpdateServicePolicy(id=self.service_name,\n policy=self.desired_policy)\n changed_state = True\n\n host_service_state.append(changed_state)\n self.results[host.name].update(changed=changed_state)\n except (vim.fault.InvalidState, vim.fault.NotFound,\n vim.fault.HostConfigFault, vmodl.fault.InvalidArgument) as e:\n self.results[host.name].update(changed=False,\n error=to_native(e.msg))\n\n if any(host_service_state):\n changed = True\n self.module.exit_json(changed=changed, host_service_status=self.results)\n\n def check_service_state(self, host, service_name):\n host_service_system = host.configManager.serviceSystem\n if host_service_system:\n services = host_service_system.serviceInfo.service\n for service in services:\n if service.key == service_name:\n return service.running, service.policy\n\n msg = \"Failed to find '%s' service on host system '%s'\" % (service_name, host.name)\n cluster_name = self.params.get('cluster_name', None)\n if cluster_name:\n msg += \" located on cluster '%s'\" % cluster_name\n msg += \", please check if you have specified a valid ESXi service name.\"\n self.module.fail_json(msg=msg)\n\n\ndef main():\n argument_spec = vmware_argument_spec()\n argument_spec.update(\n cluster_name=dict(type='str', required=False),\n esxi_hostname=dict(type='str', required=False),\n state=dict(type='str', default='start', choices=['absent', 'present', 'restart', 'start', 'stop', 'unchanged']),\n service_name=dict(type='str', required=True),\n service_policy=dict(type='str', choices=['automatic', 'off', 'on']),\n )\n\n module = AnsibleModule(\n argument_spec=argument_spec,\n required_one_of=[\n ['cluster_name', 'esxi_hostname'],\n ],\n supports_check_mode=True\n )\n\n vmware_host_service = VmwareServiceManager(module)\n vmware_host_service.service_ctrl()\n\n\nif __name__ == \"__main__\":\n main()\n", "path": "plugins/modules/vmware_host_service_manager.py"}]}
| 3,143 | 315 |
gh_patches_debug_8435
|
rasdani/github-patches
|
git_diff
|
akvo__akvo-rsr-2418
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
akvo.rest.fields in from_native AttributeError: 'NoneType' object has no attribute 'name'
http://sentry.support.akvo-ops.org/rsr/live/group/1017/
Bug introduced in 1aa60508350d4477c1f0964db7f3c124d9d1bc01
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `akvo/rest/fields.py`
Content:
```
1 # -*- coding: utf-8 -*-
2
3 # Akvo RSR is covered by the GNU Affero General Public License.
4 # See more details in the license.txt file located at the root folder of the Akvo RSR module.
5 # For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.
6
7
8 import base64
9 import imghdr
10 import six
11 import uuid
12
13 from django.core.files.base import ContentFile
14 from django.utils.encoding import smart_text
15 from django.utils.translation import ugettext_lazy as _
16
17 from rest_framework import serializers
18 from rest_framework.fields import ImageField
19 from sorl.thumbnail import get_thumbnail
20 from sorl.thumbnail.parsers import ThumbnailParseError
21
22
23 class NonNullCharField(serializers.CharField):
24 """ Fix fo CharField so that '' is returned if the field value is None
25 see https://github.com/tomchristie/django-rest-framework/pull/1665
26 """
27 def from_native(self, value):
28 if isinstance(value, six.string_types):
29 return value
30 if value is None:
31 return u''
32 return smart_text(value)
33
34
35 class NonNullURLField(NonNullCharField, serializers.URLField):
36 pass
37
38
39 class Base64ImageField(ImageField):
40 """ A django-rest-framework field for handling image-uploads through raw post data.
41 It uses base64 for en-/decoding the contents of the file.
42 Now also supports thumbnails of different sizes. See to_native() for more info.
43 """
44 ALLOWED_IMAGE_TYPES = (
45 'gif',
46 'jpeg',
47 'jpg',
48 'png',
49 )
50 def from_native(self, base64_data):
51 # Check if this is a base64 string
52 if isinstance(base64_data, basestring):
53 # Try to decode the file. Return validation error if it fails.
54 try:
55 decoded_file = base64.b64decode(base64_data)
56 except TypeError:
57 raise serializers.ValidationError(_(u"Please upload a valid image."))
58
59 # Generate file name:
60 file_name = str(uuid.uuid4())[:12] # 12 characters are more than enough.
61 # Get the file name extension:
62 file_extension = self.get_file_extension(file_name, decoded_file)
63 self.check_file_extension(file_extension)
64 complete_file_name = file_name + "." + file_extension
65 data = ContentFile(decoded_file, name=complete_file_name)
66 else:
67 data = base64_data
68 file_extension = self.get_file_extension(data.name, data.read())
69 self.check_file_extension(file_extension)
70 data.seek(0)
71
72 return super(Base64ImageField, self).from_native(data)
73
74 def to_native(self, value):
75 """
76 :param value: A Base64ImageField object
77 :return: a path to a thumbnail with a predetermined size, the default thumb
78 OR
79 a dict with a number of thumbnails, one of which is the default, the others being generated
80 from the query string parameters, and finally the path to the original image keyed to
81 "original".
82
83 The extended functionality, allowing the generation of one or more thumbnails from the
84 original image is triggered by including "image_thumb_name" in the query string. The value
85 for image_thumb_name is a comma separated list of identifiers for the generated thumbs.
86 The names must not be "default" or "original".
87
88 For each thumb thus specified a size must be supplied as a query param on the form
89 image_thumb_<name>_<dimension>
90 where <name> is the name of the thumb specified as one of the values for image_thumb_name
91 and <dimension> is one of "width, "height" or "max_size". width and height must be an integer
92 specifying that dimension in pixels. The image will be scaled correctly in the other
93 dimension. max_size is width and height concatenated with an "x" and sets the maximum size
94 allowed for the respective dimensions, while still maintaining the correct aspect ratio of
95 the image.
96
97 Example:
98 the querystring
99 ?image_thumb_name=big,small&image_thumb_small_width=90&image_thumb_big_max_size=300x200
100 results in the following dict being returned:
101 {
102 'original': '/full/path/to/original/image.png',
103 'default': '/full/path/to/default/thumbnail/image.png',
104 'small': '/full/path/to/small/thumbnail/image.png',
105 'big': '/full/path/to/big/thumbnail/image.png',
106 }
107 This dict will be converted as appropriate to JSON or XML
108
109 NOTE: This special functionality works best when there is only one image field in a model.
110 If there are more, things will still work (I think), but for each image all thumbs returned
111 will have the same dimensions
112 """
113 def get_thumb(request, name):
114 if name not in [u'original', u'default']:
115 try:
116 width = request.GET.get('image_thumb_{}_width'.format(name))
117 if width:
118 return get_thumbnail(value, '{}'.format(width), quality=99)
119 height = request.GET.get('image_thumb_{}_height'.format(name))
120 if height:
121 return get_thumbnail(value, 'x{}'.format(height), quality=99)
122 # yes this is redundant...code is nearly identical with the width code above
123 # but for clarity of function we keep them separate
124 max_size = request.GET.get('image_thumb_{}_max_size'.format(name))
125 if max_size:
126 return get_thumbnail(value, '{}'.format(max_size), quality=99)
127 except (ThumbnailParseError, IOError):
128 return None
129
130 # no size specification matching the name found; give up
131 return None
132
133 if value:
134 default_width = '191' # width of update images on akvo.org/seeithappen
135 try:
136 default_thumb = get_thumbnail(value, default_width, quality=99)
137 request = self.context['request']
138 except (ThumbnailParseError, IOError, KeyError):
139 return None
140
141 # look for name(s) of thumb(s)
142 image_thumb_name = request.GET.get('image_thumb_name')
143 if image_thumb_name:
144 names = image_thumb_name.split(',')
145 thumbs = {u'original': value.url, u'default': default_thumb.url}
146 for name in names:
147 thumb = get_thumb(request, name)
148 if thumb is not None:
149 thumbs[name] = thumb.url
150 return thumbs
151 return default_thumb.url
152
153 def get_file_extension(self, filename, decoded_file):
154 extension = imghdr.what(filename, decoded_file)
155 extension = "jpg" if extension == "jpeg" else extension
156 return extension
157
158 def check_file_extension(self, file_extension):
159 if file_extension not in self.ALLOWED_IMAGE_TYPES:
160 formats = {'format': ', '.join(self.ALLOWED_IMAGE_TYPES)}
161 raise serializers.ValidationError(
162 _(u"Unknown image type. Only the following types are accepted: %(format)s") % formats
163 )
164
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/akvo/rest/fields.py b/akvo/rest/fields.py
--- a/akvo/rest/fields.py
+++ b/akvo/rest/fields.py
@@ -48,8 +48,10 @@
'png',
)
def from_native(self, base64_data):
+ if base64_data is None:
+ data = base64_data
# Check if this is a base64 string
- if isinstance(base64_data, basestring):
+ elif isinstance(base64_data, basestring):
# Try to decode the file. Return validation error if it fails.
try:
decoded_file = base64.b64decode(base64_data)
|
{"golden_diff": "diff --git a/akvo/rest/fields.py b/akvo/rest/fields.py\n--- a/akvo/rest/fields.py\n+++ b/akvo/rest/fields.py\n@@ -48,8 +48,10 @@\n 'png',\n )\n def from_native(self, base64_data):\n+ if base64_data is None:\n+ data = base64_data\n # Check if this is a base64 string\n- if isinstance(base64_data, basestring):\n+ elif isinstance(base64_data, basestring):\n # Try to decode the file. Return validation error if it fails.\n try:\n decoded_file = base64.b64decode(base64_data)\n", "issue": "akvo.rest.fields in from_native AttributeError: 'NoneType' object has no attribute 'name'\nhttp://sentry.support.akvo-ops.org/rsr/live/group/1017/\n\nBug introduced in 1aa60508350d4477c1f0964db7f3c124d9d1bc01\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Akvo RSR is covered by the GNU Affero General Public License.\n# See more details in the license.txt file located at the root folder of the Akvo RSR module.\n# For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.\n\n\nimport base64\nimport imghdr\nimport six\nimport uuid\n\nfrom django.core.files.base import ContentFile\nfrom django.utils.encoding import smart_text\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom rest_framework import serializers\nfrom rest_framework.fields import ImageField\nfrom sorl.thumbnail import get_thumbnail\nfrom sorl.thumbnail.parsers import ThumbnailParseError\n\n\nclass NonNullCharField(serializers.CharField):\n \"\"\" Fix fo CharField so that '' is returned if the field value is None\n see https://github.com/tomchristie/django-rest-framework/pull/1665\n \"\"\"\n def from_native(self, value):\n if isinstance(value, six.string_types):\n return value\n if value is None:\n return u''\n return smart_text(value)\n\n\nclass NonNullURLField(NonNullCharField, serializers.URLField):\n pass\n\n\nclass Base64ImageField(ImageField):\n \"\"\" A django-rest-framework field for handling image-uploads through raw post data.\n It uses base64 for en-/decoding the contents of the file.\n Now also supports thumbnails of different sizes. See to_native() for more info.\n \"\"\"\n ALLOWED_IMAGE_TYPES = (\n 'gif',\n 'jpeg',\n 'jpg',\n 'png',\n )\n def from_native(self, base64_data):\n # Check if this is a base64 string\n if isinstance(base64_data, basestring):\n # Try to decode the file. Return validation error if it fails.\n try:\n decoded_file = base64.b64decode(base64_data)\n except TypeError:\n raise serializers.ValidationError(_(u\"Please upload a valid image.\"))\n\n # Generate file name:\n file_name = str(uuid.uuid4())[:12] # 12 characters are more than enough.\n # Get the file name extension:\n file_extension = self.get_file_extension(file_name, decoded_file)\n self.check_file_extension(file_extension)\n complete_file_name = file_name + \".\" + file_extension\n data = ContentFile(decoded_file, name=complete_file_name)\n else:\n data = base64_data\n file_extension = self.get_file_extension(data.name, data.read())\n self.check_file_extension(file_extension)\n data.seek(0)\n\n return super(Base64ImageField, self).from_native(data)\n\n def to_native(self, value):\n \"\"\"\n :param value: A Base64ImageField object\n :return: a path to a thumbnail with a predetermined size, the default thumb\n OR\n a dict with a number of thumbnails, one of which is the default, the others being generated\n from the query string parameters, and finally the path to the original image keyed to\n \"original\".\n\n The extended functionality, allowing the generation of one or more thumbnails from the\n original image is triggered by including \"image_thumb_name\" in the query string. The value\n for image_thumb_name is a comma separated list of identifiers for the generated thumbs.\n The names must not be \"default\" or \"original\".\n\n For each thumb thus specified a size must be supplied as a query param on the form\n image_thumb_<name>_<dimension>\n where <name> is the name of the thumb specified as one of the values for image_thumb_name\n and <dimension> is one of \"width, \"height\" or \"max_size\". width and height must be an integer\n specifying that dimension in pixels. The image will be scaled correctly in the other\n dimension. max_size is width and height concatenated with an \"x\" and sets the maximum size\n allowed for the respective dimensions, while still maintaining the correct aspect ratio of\n the image.\n\n Example:\n the querystring\n ?image_thumb_name=big,small&image_thumb_small_width=90&image_thumb_big_max_size=300x200\n results in the following dict being returned:\n {\n 'original': '/full/path/to/original/image.png',\n 'default': '/full/path/to/default/thumbnail/image.png',\n 'small': '/full/path/to/small/thumbnail/image.png',\n 'big': '/full/path/to/big/thumbnail/image.png',\n }\n This dict will be converted as appropriate to JSON or XML\n\n NOTE: This special functionality works best when there is only one image field in a model.\n If there are more, things will still work (I think), but for each image all thumbs returned\n will have the same dimensions\n \"\"\"\n def get_thumb(request, name):\n if name not in [u'original', u'default']:\n try:\n width = request.GET.get('image_thumb_{}_width'.format(name))\n if width:\n return get_thumbnail(value, '{}'.format(width), quality=99)\n height = request.GET.get('image_thumb_{}_height'.format(name))\n if height:\n return get_thumbnail(value, 'x{}'.format(height), quality=99)\n # yes this is redundant...code is nearly identical with the width code above\n # but for clarity of function we keep them separate\n max_size = request.GET.get('image_thumb_{}_max_size'.format(name))\n if max_size:\n return get_thumbnail(value, '{}'.format(max_size), quality=99)\n except (ThumbnailParseError, IOError):\n return None\n\n # no size specification matching the name found; give up\n return None\n\n if value:\n default_width = '191' # width of update images on akvo.org/seeithappen\n try:\n default_thumb = get_thumbnail(value, default_width, quality=99)\n request = self.context['request']\n except (ThumbnailParseError, IOError, KeyError):\n return None\n\n # look for name(s) of thumb(s)\n image_thumb_name = request.GET.get('image_thumb_name')\n if image_thumb_name:\n names = image_thumb_name.split(',')\n thumbs = {u'original': value.url, u'default': default_thumb.url}\n for name in names:\n thumb = get_thumb(request, name)\n if thumb is not None:\n thumbs[name] = thumb.url\n return thumbs\n return default_thumb.url\n\n def get_file_extension(self, filename, decoded_file):\n extension = imghdr.what(filename, decoded_file)\n extension = \"jpg\" if extension == \"jpeg\" else extension\n return extension\n\n def check_file_extension(self, file_extension):\n if file_extension not in self.ALLOWED_IMAGE_TYPES:\n formats = {'format': ', '.join(self.ALLOWED_IMAGE_TYPES)}\n raise serializers.ValidationError(\n _(u\"Unknown image type. Only the following types are accepted: %(format)s\") % formats\n )\n", "path": "akvo/rest/fields.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\n# Akvo RSR is covered by the GNU Affero General Public License.\n# See more details in the license.txt file located at the root folder of the Akvo RSR module.\n# For additional details on the GNU license please see < http://www.gnu.org/licenses/agpl.html >.\n\n\nimport base64\nimport imghdr\nimport six\nimport uuid\n\nfrom django.core.files.base import ContentFile\nfrom django.utils.encoding import smart_text\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom rest_framework import serializers\nfrom rest_framework.fields import ImageField\nfrom sorl.thumbnail import get_thumbnail\nfrom sorl.thumbnail.parsers import ThumbnailParseError\n\n\nclass NonNullCharField(serializers.CharField):\n \"\"\" Fix fo CharField so that '' is returned if the field value is None\n see https://github.com/tomchristie/django-rest-framework/pull/1665\n \"\"\"\n def from_native(self, value):\n if isinstance(value, six.string_types):\n return value\n if value is None:\n return u''\n return smart_text(value)\n\n\nclass NonNullURLField(NonNullCharField, serializers.URLField):\n pass\n\n\nclass Base64ImageField(ImageField):\n \"\"\" A django-rest-framework field for handling image-uploads through raw post data.\n It uses base64 for en-/decoding the contents of the file.\n Now also supports thumbnails of different sizes. See to_native() for more info.\n \"\"\"\n ALLOWED_IMAGE_TYPES = (\n 'gif',\n 'jpeg',\n 'jpg',\n 'png',\n )\n def from_native(self, base64_data):\n if base64_data is None:\n data = base64_data\n # Check if this is a base64 string\n elif isinstance(base64_data, basestring):\n # Try to decode the file. Return validation error if it fails.\n try:\n decoded_file = base64.b64decode(base64_data)\n except TypeError:\n raise serializers.ValidationError(_(u\"Please upload a valid image.\"))\n\n # Generate file name:\n file_name = str(uuid.uuid4())[:12] # 12 characters are more than enough.\n # Get the file name extension:\n file_extension = self.get_file_extension(file_name, decoded_file)\n self.check_file_extension(file_extension)\n complete_file_name = file_name + \".\" + file_extension\n data = ContentFile(decoded_file, name=complete_file_name)\n else:\n data = base64_data\n file_extension = self.get_file_extension(data.name, data.read())\n self.check_file_extension(file_extension)\n data.seek(0)\n\n return super(Base64ImageField, self).from_native(data)\n\n def to_native(self, value):\n \"\"\"\n :param value: A Base64ImageField object\n :return: a path to a thumbnail with a predetermined size, the default thumb\n OR\n a dict with a number of thumbnails, one of which is the default, the others being generated\n from the query string parameters, and finally the path to the original image keyed to\n \"original\".\n\n The extended functionality, allowing the generation of one or more thumbnails from the\n original image is triggered by including \"image_thumb_name\" in the query string. The value\n for image_thumb_name is a comma separated list of identifiers for the generated thumbs.\n The names must not be \"default\" or \"original\".\n\n For each thumb thus specified a size must be supplied as a query param on the form\n image_thumb_<name>_<dimension>\n where <name> is the name of the thumb specified as one of the values for image_thumb_name\n and <dimension> is one of \"width, \"height\" or \"max_size\". width and height must be an integer\n specifying that dimension in pixels. The image will be scaled correctly in the other\n dimension. max_size is width and height concatenated with an \"x\" and sets the maximum size\n allowed for the respective dimensions, while still maintaining the correct aspect ratio of\n the image.\n\n Example:\n the querystring\n ?image_thumb_name=big,small&image_thumb_small_width=90&image_thumb_big_max_size=300x200\n results in the following dict being returned:\n {\n 'original': '/full/path/to/original/image.png',\n 'default': '/full/path/to/default/thumbnail/image.png',\n 'small': '/full/path/to/small/thumbnail/image.png',\n 'big': '/full/path/to/big/thumbnail/image.png',\n }\n This dict will be converted as appropriate to JSON or XML\n\n NOTE: This special functionality works best when there is only one image field in a model.\n If there are more, things will still work (I think), but for each image all thumbs returned\n will have the same dimensions\n \"\"\"\n def get_thumb(request, name):\n if name not in [u'original', u'default']:\n try:\n width = request.GET.get('image_thumb_{}_width'.format(name))\n if width:\n return get_thumbnail(value, '{}'.format(width), quality=99)\n height = request.GET.get('image_thumb_{}_height'.format(name))\n if height:\n return get_thumbnail(value, 'x{}'.format(height), quality=99)\n # yes this is redundant...code is nearly identical with the width code above\n # but for clarity of function we keep them separate\n max_size = request.GET.get('image_thumb_{}_max_size'.format(name))\n if max_size:\n return get_thumbnail(value, '{}'.format(max_size), quality=99)\n except (ThumbnailParseError, IOError):\n return None\n\n # no size specification matching the name found; give up\n return None\n\n if value:\n default_width = '191' # width of update images on akvo.org/seeithappen\n try:\n default_thumb = get_thumbnail(value, default_width, quality=99)\n request = self.context['request']\n except (ThumbnailParseError, IOError, KeyError):\n return None\n\n # look for name(s) of thumb(s)\n image_thumb_name = request.GET.get('image_thumb_name')\n if image_thumb_name:\n names = image_thumb_name.split(',')\n thumbs = {u'original': value.url, u'default': default_thumb.url}\n for name in names:\n thumb = get_thumb(request, name)\n if thumb is not None:\n thumbs[name] = thumb.url\n return thumbs\n return default_thumb.url\n\n def get_file_extension(self, filename, decoded_file):\n extension = imghdr.what(filename, decoded_file)\n extension = \"jpg\" if extension == \"jpeg\" else extension\n return extension\n\n def check_file_extension(self, file_extension):\n if file_extension not in self.ALLOWED_IMAGE_TYPES:\n formats = {'format': ', '.join(self.ALLOWED_IMAGE_TYPES)}\n raise serializers.ValidationError(\n _(u\"Unknown image type. Only the following types are accepted: %(format)s\") % formats\n )\n", "path": "akvo/rest/fields.py"}]}
| 2,241 | 159 |
gh_patches_debug_13891
|
rasdani/github-patches
|
git_diff
|
Mailu__Mailu-1862
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Radicale password check fails (length limited?)
So far, the webdav radicale is protected with basic auth (apparently htaccess or similar).
If user password > 8 then it fails to connect.
we should remove this limitation to let use stong passwords
Radicale password check fails (length limited?)
So far, the webdav radicale is protected with basic auth (apparently htaccess or similar).
If user password > 8 then it fails to connect.
we should remove this limitation to let use stong passwords
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `core/admin/mailu/internal/views/auth.py`
Content:
```
1 from mailu import models, utils
2 from mailu.internal import internal, nginx
3 from flask import current_app as app
4
5 import flask
6 import flask_login
7 import base64
8 import ipaddress
9
10
11 @internal.route("/auth/email")
12 def nginx_authentication():
13 """ Main authentication endpoint for Nginx email server
14 """
15 limiter = utils.limiter.get_limiter(app.config["AUTH_RATELIMIT"], "auth-ip")
16 client_ip = flask.request.headers["Client-Ip"]
17 if not limiter.test(client_ip):
18 response = flask.Response()
19 response.headers['Auth-Status'] = 'Authentication rate limit from one source exceeded'
20 response.headers['Auth-Error-Code'] = '451 4.3.2'
21 if int(flask.request.headers['Auth-Login-Attempt']) < 10:
22 response.headers['Auth-Wait'] = '3'
23 return response
24 headers = nginx.handle_authentication(flask.request.headers)
25 response = flask.Response()
26 for key, value in headers.items():
27 response.headers[key] = str(value)
28 if ("Auth-Status" not in headers) or (headers["Auth-Status"] != "OK"):
29 limit_subnet = str(app.config["AUTH_RATELIMIT_SUBNET"]) != 'False'
30 subnet = ipaddress.ip_network(app.config["SUBNET"])
31 if limit_subnet or ipaddress.ip_address(client_ip) not in subnet:
32 limiter.hit(flask.request.headers["Client-Ip"])
33 return response
34
35
36 @internal.route("/auth/admin")
37 def admin_authentication():
38 """ Fails if the user is not an authenticated admin.
39 """
40 if (not flask_login.current_user.is_anonymous
41 and flask_login.current_user.global_admin
42 and flask_login.current_user.enabled):
43 return ""
44 return flask.abort(403)
45
46 @internal.route("/auth/user")
47 def user_authentication():
48 """ Fails if the user is not authenticated.
49 """
50 if (not flask_login.current_user.is_anonymous
51 and flask_login.current_user.enabled):
52 response = flask.Response()
53 response.headers["X-User"] = flask_login.current_user.get_id()
54 response.headers["X-User-Token"] = models.User.get_temp_token(flask_login.current_user.get_id())
55 return response
56 return flask.abort(403)
57
58
59 @internal.route("/auth/basic")
60 def basic_authentication():
61 """ Tries to authenticate using the Authorization header.
62 """
63 authorization = flask.request.headers.get("Authorization")
64 if authorization and authorization.startswith("Basic "):
65 encoded = authorization.replace("Basic ", "")
66 user_email, password = base64.b64decode(encoded).split(b":")
67 user = models.User.query.get(user_email.decode("utf8"))
68 if nginx.check_credentials(user, password.decode('utf-8'), flask.request.remote_addr, "web"):
69 response = flask.Response()
70 response.headers["X-User"] = user.email
71 return response
72 response = flask.Response(status=401)
73 response.headers["WWW-Authenticate"] = 'Basic realm="Login Required"'
74 return response
75
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/core/admin/mailu/internal/views/auth.py b/core/admin/mailu/internal/views/auth.py
--- a/core/admin/mailu/internal/views/auth.py
+++ b/core/admin/mailu/internal/views/auth.py
@@ -63,7 +63,7 @@
authorization = flask.request.headers.get("Authorization")
if authorization and authorization.startswith("Basic "):
encoded = authorization.replace("Basic ", "")
- user_email, password = base64.b64decode(encoded).split(b":")
+ user_email, password = base64.b64decode(encoded).split(b":", 1)
user = models.User.query.get(user_email.decode("utf8"))
if nginx.check_credentials(user, password.decode('utf-8'), flask.request.remote_addr, "web"):
response = flask.Response()
|
{"golden_diff": "diff --git a/core/admin/mailu/internal/views/auth.py b/core/admin/mailu/internal/views/auth.py\n--- a/core/admin/mailu/internal/views/auth.py\n+++ b/core/admin/mailu/internal/views/auth.py\n@@ -63,7 +63,7 @@\n authorization = flask.request.headers.get(\"Authorization\")\n if authorization and authorization.startswith(\"Basic \"):\n encoded = authorization.replace(\"Basic \", \"\")\n- user_email, password = base64.b64decode(encoded).split(b\":\")\n+ user_email, password = base64.b64decode(encoded).split(b\":\", 1)\n user = models.User.query.get(user_email.decode(\"utf8\"))\n if nginx.check_credentials(user, password.decode('utf-8'), flask.request.remote_addr, \"web\"):\n response = flask.Response()\n", "issue": "Radicale password check fails (length limited?)\nSo far, the webdav radicale is protected with basic auth (apparently htaccess or similar).\r\nIf user password > 8 then it fails to connect. \r\nwe should remove this limitation to let use stong passwords\nRadicale password check fails (length limited?)\nSo far, the webdav radicale is protected with basic auth (apparently htaccess or similar).\r\nIf user password > 8 then it fails to connect. \r\nwe should remove this limitation to let use stong passwords\n", "before_files": [{"content": "from mailu import models, utils\nfrom mailu.internal import internal, nginx\nfrom flask import current_app as app\n\nimport flask\nimport flask_login\nimport base64\nimport ipaddress\n\n\[email protected](\"/auth/email\")\ndef nginx_authentication():\n \"\"\" Main authentication endpoint for Nginx email server\n \"\"\"\n limiter = utils.limiter.get_limiter(app.config[\"AUTH_RATELIMIT\"], \"auth-ip\")\n client_ip = flask.request.headers[\"Client-Ip\"]\n if not limiter.test(client_ip):\n response = flask.Response()\n response.headers['Auth-Status'] = 'Authentication rate limit from one source exceeded'\n response.headers['Auth-Error-Code'] = '451 4.3.2'\n if int(flask.request.headers['Auth-Login-Attempt']) < 10:\n response.headers['Auth-Wait'] = '3'\n return response\n headers = nginx.handle_authentication(flask.request.headers)\n response = flask.Response()\n for key, value in headers.items():\n response.headers[key] = str(value)\n if (\"Auth-Status\" not in headers) or (headers[\"Auth-Status\"] != \"OK\"):\n limit_subnet = str(app.config[\"AUTH_RATELIMIT_SUBNET\"]) != 'False'\n subnet = ipaddress.ip_network(app.config[\"SUBNET\"])\n if limit_subnet or ipaddress.ip_address(client_ip) not in subnet:\n limiter.hit(flask.request.headers[\"Client-Ip\"])\n return response\n\n\[email protected](\"/auth/admin\")\ndef admin_authentication():\n \"\"\" Fails if the user is not an authenticated admin.\n \"\"\"\n if (not flask_login.current_user.is_anonymous\n and flask_login.current_user.global_admin\n and flask_login.current_user.enabled):\n return \"\"\n return flask.abort(403)\n\[email protected](\"/auth/user\")\ndef user_authentication():\n \"\"\" Fails if the user is not authenticated.\n \"\"\"\n if (not flask_login.current_user.is_anonymous\n and flask_login.current_user.enabled):\n response = flask.Response()\n response.headers[\"X-User\"] = flask_login.current_user.get_id()\n response.headers[\"X-User-Token\"] = models.User.get_temp_token(flask_login.current_user.get_id())\n return response\n return flask.abort(403)\n\n\[email protected](\"/auth/basic\")\ndef basic_authentication():\n \"\"\" Tries to authenticate using the Authorization header.\n \"\"\"\n authorization = flask.request.headers.get(\"Authorization\")\n if authorization and authorization.startswith(\"Basic \"):\n encoded = authorization.replace(\"Basic \", \"\")\n user_email, password = base64.b64decode(encoded).split(b\":\")\n user = models.User.query.get(user_email.decode(\"utf8\"))\n if nginx.check_credentials(user, password.decode('utf-8'), flask.request.remote_addr, \"web\"):\n response = flask.Response()\n response.headers[\"X-User\"] = user.email\n return response\n response = flask.Response(status=401)\n response.headers[\"WWW-Authenticate\"] = 'Basic realm=\"Login Required\"'\n return response\n", "path": "core/admin/mailu/internal/views/auth.py"}], "after_files": [{"content": "from mailu import models, utils\nfrom mailu.internal import internal, nginx\nfrom flask import current_app as app\n\nimport flask\nimport flask_login\nimport base64\nimport ipaddress\n\n\[email protected](\"/auth/email\")\ndef nginx_authentication():\n \"\"\" Main authentication endpoint for Nginx email server\n \"\"\"\n limiter = utils.limiter.get_limiter(app.config[\"AUTH_RATELIMIT\"], \"auth-ip\")\n client_ip = flask.request.headers[\"Client-Ip\"]\n if not limiter.test(client_ip):\n response = flask.Response()\n response.headers['Auth-Status'] = 'Authentication rate limit from one source exceeded'\n response.headers['Auth-Error-Code'] = '451 4.3.2'\n if int(flask.request.headers['Auth-Login-Attempt']) < 10:\n response.headers['Auth-Wait'] = '3'\n return response\n headers = nginx.handle_authentication(flask.request.headers)\n response = flask.Response()\n for key, value in headers.items():\n response.headers[key] = str(value)\n if (\"Auth-Status\" not in headers) or (headers[\"Auth-Status\"] != \"OK\"):\n limit_subnet = str(app.config[\"AUTH_RATELIMIT_SUBNET\"]) != 'False'\n subnet = ipaddress.ip_network(app.config[\"SUBNET\"])\n if limit_subnet or ipaddress.ip_address(client_ip) not in subnet:\n limiter.hit(flask.request.headers[\"Client-Ip\"])\n return response\n\n\[email protected](\"/auth/admin\")\ndef admin_authentication():\n \"\"\" Fails if the user is not an authenticated admin.\n \"\"\"\n if (not flask_login.current_user.is_anonymous\n and flask_login.current_user.global_admin\n and flask_login.current_user.enabled):\n return \"\"\n return flask.abort(403)\n\[email protected](\"/auth/user\")\ndef user_authentication():\n \"\"\" Fails if the user is not authenticated.\n \"\"\"\n if (not flask_login.current_user.is_anonymous\n and flask_login.current_user.enabled):\n response = flask.Response()\n response.headers[\"X-User\"] = flask_login.current_user.get_id()\n response.headers[\"X-User-Token\"] = models.User.get_temp_token(flask_login.current_user.get_id())\n return response\n return flask.abort(403)\n\n\[email protected](\"/auth/basic\")\ndef basic_authentication():\n \"\"\" Tries to authenticate using the Authorization header.\n \"\"\"\n authorization = flask.request.headers.get(\"Authorization\")\n if authorization and authorization.startswith(\"Basic \"):\n encoded = authorization.replace(\"Basic \", \"\")\n user_email, password = base64.b64decode(encoded).split(b\":\", 1)\n user = models.User.query.get(user_email.decode(\"utf8\"))\n if nginx.check_credentials(user, password.decode('utf-8'), flask.request.remote_addr, \"web\"):\n response = flask.Response()\n response.headers[\"X-User\"] = user.email\n return response\n response = flask.Response(status=401)\n response.headers[\"WWW-Authenticate\"] = 'Basic realm=\"Login Required\"'\n return response\n", "path": "core/admin/mailu/internal/views/auth.py"}]}
| 1,168 | 170 |
gh_patches_debug_5475
|
rasdani/github-patches
|
git_diff
|
mesonbuild__meson-8069
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
i18n module does not seem to respect MESON_INSTALL_QUIET environment variable
**Describe the bug**
When using `meson install --quiet`, there's still output generated by what I presume is gettext:
```
Installing /root/build/po/be.gmo to /root/dest/usr/share/locale/be/LC_MESSAGES/systemd.mo
Installing /root/build/po/[email protected] to /root/dest/usr/share/locale/be@latin/LC_MESSAGES/systemd.mo
Installing /root/build/po/bg.gmo to /root/dest/usr/share/locale/bg/LC_MESSAGES/systemd.mo
Installing /root/build/po/ca.gmo to /root/dest/usr/share/locale/ca/LC_MESSAGES/systemd.mo
Installing /root/build/po/cs.gmo to /root/dest/usr/share/locale/cs/LC_MESSAGES/systemd.mo
Installing /root/build/po/da.gmo to /root/dest/usr/share/locale/da/LC_MESSAGES/systemd.mo
Installing /root/build/po/de.gmo to /root/dest/usr/share/locale/de/LC_MESSAGES/systemd.mo
Installing /root/build/po/el.gmo to /root/dest/usr/share/locale/el/LC_MESSAGES/systemd.mo
Installing /root/build/po/es.gmo to /root/dest/usr/share/locale/es/LC_MESSAGES/systemd.mo
Installing /root/build/po/fr.gmo to /root/dest/usr/share/locale/fr/LC_MESSAGES/systemd.mo
Installing /root/build/po/gl.gmo to /root/dest/usr/share/locale/gl/LC_MESSAGES/systemd.mo
Installing /root/build/po/hr.gmo to /root/dest/usr/share/locale/hr/LC_MESSAGES/systemd.mo
Installing /root/build/po/hu.gmo to /root/dest/usr/share/locale/hu/LC_MESSAGES/systemd.mo
Installing /root/build/po/id.gmo to /root/dest/usr/share/locale/id/LC_MESSAGES/systemd.mo
Installing /root/build/po/it.gmo to /root/dest/usr/share/locale/it/LC_MESSAGES/systemd.mo
Installing /root/build/po/ja.gmo to /root/dest/usr/share/locale/ja/LC_MESSAGES/systemd.mo
Installing /root/build/po/ko.gmo to /root/dest/usr/share/locale/ko/LC_MESSAGES/systemd.mo
Installing /root/build/po/lt.gmo to /root/dest/usr/share/locale/lt/LC_MESSAGES/systemd.mo
Installing /root/build/po/pl.gmo to /root/dest/usr/share/locale/pl/LC_MESSAGES/systemd.mo
Installing /root/build/po/pt_BR.gmo to /root/dest/usr/share/locale/pt_BR/LC_MESSAGES/systemd.mo
Installing /root/build/po/ro.gmo to /root/dest/usr/share/locale/ro/LC_MESSAGES/systemd.mo
Installing /root/build/po/ru.gmo to /root/dest/usr/share/locale/ru/LC_MESSAGES/systemd.mo
Installing /root/build/po/sk.gmo to /root/dest/usr/share/locale/sk/LC_MESSAGES/systemd.mo
Installing /root/build/po/sr.gmo to /root/dest/usr/share/locale/sr/LC_MESSAGES/systemd.mo
Installing /root/build/po/sv.gmo to /root/dest/usr/share/locale/sv/LC_MESSAGES/systemd.mo
Installing /root/build/po/tr.gmo to /root/dest/usr/share/locale/tr/LC_MESSAGES/systemd.mo
Installing /root/build/po/uk.gmo to /root/dest/usr/share/locale/uk/LC_MESSAGES/systemd.mo
Installing /root/build/po/zh_CN.gmo to /root/dest/usr/share/locale/zh_CN/LC_MESSAGES/systemd.mo
Installing /root/build/po/zh_TW.gmo to /root/dest/usr/share/locale/zh_TW/LC_MESSAGES/systemd.mo
```
**To Reproduce**
```
git clone https://github.com/systemd/systemd.git
meson build
meson compile -C build
meson install -C build --quiet
```
**Expected behavior**
No output from gettext when --quiet is passed to `meson install`
**system parameters**
Should not be relevant to this bug. Let me know if I should add them. I'm running latest meson (0.56.0)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `mesonbuild/scripts/gettext.py`
Content:
```
1 # Copyright 2016 The Meson development team
2
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6
7 # http://www.apache.org/licenses/LICENSE-2.0
8
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 import os
16 import shutil
17 import argparse
18 import subprocess
19 from . import destdir_join
20 import typing as T
21
22 parser = argparse.ArgumentParser()
23 parser.add_argument('command')
24 parser.add_argument('--pkgname', default='')
25 parser.add_argument('--datadirs', default='')
26 parser.add_argument('--langs', default='')
27 parser.add_argument('--localedir', default='')
28 parser.add_argument('--subdir', default='')
29 parser.add_argument('--extra-args', default='')
30
31 def read_linguas(src_sub: str) -> T.List[str]:
32 # Syntax of this file is documented here:
33 # https://www.gnu.org/software/gettext/manual/html_node/po_002fLINGUAS.html
34 linguas = os.path.join(src_sub, 'LINGUAS')
35 try:
36 langs = []
37 with open(linguas) as f:
38 for line in f:
39 line = line.strip()
40 if line and not line.startswith('#'):
41 langs += line.split()
42 return langs
43 except (FileNotFoundError, PermissionError):
44 print('Could not find file LINGUAS in {}'.format(src_sub))
45 return []
46
47 def run_potgen(src_sub: str, pkgname: str, datadirs: str, args: T.List[str]) -> int:
48 listfile = os.path.join(src_sub, 'POTFILES.in')
49 if not os.path.exists(listfile):
50 listfile = os.path.join(src_sub, 'POTFILES')
51 if not os.path.exists(listfile):
52 print('Could not find file POTFILES in %s' % src_sub)
53 return 1
54
55 child_env = os.environ.copy()
56 if datadirs:
57 child_env['GETTEXTDATADIRS'] = datadirs
58
59 ofile = os.path.join(src_sub, pkgname + '.pot')
60 return subprocess.call(['xgettext', '--package-name=' + pkgname, '-p', src_sub, '-f', listfile,
61 '-D', os.environ['MESON_SOURCE_ROOT'], '-k_', '-o', ofile] + args,
62 env=child_env)
63
64 def gen_gmo(src_sub: str, bld_sub: str, langs: T.List[str]) -> int:
65 for l in langs:
66 subprocess.check_call(['msgfmt', os.path.join(src_sub, l + '.po'),
67 '-o', os.path.join(bld_sub, l + '.gmo')])
68 return 0
69
70 def update_po(src_sub: str, pkgname: str, langs: T.List[str]) -> int:
71 potfile = os.path.join(src_sub, pkgname + '.pot')
72 for l in langs:
73 pofile = os.path.join(src_sub, l + '.po')
74 if os.path.exists(pofile):
75 subprocess.check_call(['msgmerge', '-q', '-o', pofile, pofile, potfile])
76 else:
77 subprocess.check_call(['msginit', '--input', potfile, '--output-file', pofile, '--locale', l, '--no-translator'])
78 return 0
79
80 def do_install(src_sub: str, bld_sub: str, dest: str, pkgname: str, langs: T.List[str]) -> int:
81 for l in langs:
82 srcfile = os.path.join(bld_sub, l + '.gmo')
83 outfile = os.path.join(dest, l, 'LC_MESSAGES',
84 pkgname + '.mo')
85 tempfile = outfile + '.tmp'
86 os.makedirs(os.path.dirname(outfile), exist_ok=True)
87 shutil.copy2(srcfile, tempfile)
88 os.replace(tempfile, outfile)
89 print('Installing %s to %s' % (srcfile, outfile))
90 return 0
91
92 def run(args: T.List[str]) -> int:
93 options = parser.parse_args(args)
94 subcmd = options.command
95 langs = options.langs.split('@@') if options.langs else None
96 extra_args = options.extra_args.split('@@') if options.extra_args else []
97 subdir = os.environ.get('MESON_SUBDIR', '')
98 if options.subdir:
99 subdir = options.subdir
100 src_sub = os.path.join(os.environ['MESON_SOURCE_ROOT'], subdir)
101 bld_sub = os.path.join(os.environ['MESON_BUILD_ROOT'], subdir)
102
103 if not langs:
104 langs = read_linguas(src_sub)
105
106 if subcmd == 'pot':
107 return run_potgen(src_sub, options.pkgname, options.datadirs, extra_args)
108 elif subcmd == 'gen_gmo':
109 return gen_gmo(src_sub, bld_sub, langs)
110 elif subcmd == 'update_po':
111 if run_potgen(src_sub, options.pkgname, options.datadirs, extra_args) != 0:
112 return 1
113 return update_po(src_sub, options.pkgname, langs)
114 elif subcmd == 'install':
115 destdir = os.environ.get('DESTDIR', '')
116 dest = destdir_join(destdir, os.path.join(os.environ['MESON_INSTALL_PREFIX'],
117 options.localedir))
118 if gen_gmo(src_sub, bld_sub, langs) != 0:
119 return 1
120 do_install(src_sub, bld_sub, dest, options.pkgname, langs)
121 else:
122 print('Unknown subcommand.')
123 return 1
124 return 0
125
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/mesonbuild/scripts/gettext.py b/mesonbuild/scripts/gettext.py
--- a/mesonbuild/scripts/gettext.py
+++ b/mesonbuild/scripts/gettext.py
@@ -86,7 +86,8 @@
os.makedirs(os.path.dirname(outfile), exist_ok=True)
shutil.copy2(srcfile, tempfile)
os.replace(tempfile, outfile)
- print('Installing %s to %s' % (srcfile, outfile))
+ if not os.getenv('MESON_INSTALL_QUIET', False):
+ print('Installing %s to %s' % (srcfile, outfile))
return 0
def run(args: T.List[str]) -> int:
|
{"golden_diff": "diff --git a/mesonbuild/scripts/gettext.py b/mesonbuild/scripts/gettext.py\n--- a/mesonbuild/scripts/gettext.py\n+++ b/mesonbuild/scripts/gettext.py\n@@ -86,7 +86,8 @@\n os.makedirs(os.path.dirname(outfile), exist_ok=True)\n shutil.copy2(srcfile, tempfile)\n os.replace(tempfile, outfile)\n- print('Installing %s to %s' % (srcfile, outfile))\n+ if not os.getenv('MESON_INSTALL_QUIET', False):\n+ print('Installing %s to %s' % (srcfile, outfile))\n return 0\n \n def run(args: T.List[str]) -> int:\n", "issue": "i18n module does not seem to respect MESON_INSTALL_QUIET environment variable\n**Describe the bug**\r\n\r\nWhen using `meson install --quiet`, there's still output generated by what I presume is gettext:\r\n\r\n```\r\nInstalling /root/build/po/be.gmo to /root/dest/usr/share/locale/be/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/[email protected] to /root/dest/usr/share/locale/be@latin/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/bg.gmo to /root/dest/usr/share/locale/bg/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/ca.gmo to /root/dest/usr/share/locale/ca/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/cs.gmo to /root/dest/usr/share/locale/cs/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/da.gmo to /root/dest/usr/share/locale/da/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/de.gmo to /root/dest/usr/share/locale/de/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/el.gmo to /root/dest/usr/share/locale/el/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/es.gmo to /root/dest/usr/share/locale/es/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/fr.gmo to /root/dest/usr/share/locale/fr/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/gl.gmo to /root/dest/usr/share/locale/gl/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/hr.gmo to /root/dest/usr/share/locale/hr/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/hu.gmo to /root/dest/usr/share/locale/hu/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/id.gmo to /root/dest/usr/share/locale/id/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/it.gmo to /root/dest/usr/share/locale/it/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/ja.gmo to /root/dest/usr/share/locale/ja/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/ko.gmo to /root/dest/usr/share/locale/ko/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/lt.gmo to /root/dest/usr/share/locale/lt/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/pl.gmo to /root/dest/usr/share/locale/pl/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/pt_BR.gmo to /root/dest/usr/share/locale/pt_BR/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/ro.gmo to /root/dest/usr/share/locale/ro/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/ru.gmo to /root/dest/usr/share/locale/ru/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/sk.gmo to /root/dest/usr/share/locale/sk/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/sr.gmo to /root/dest/usr/share/locale/sr/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/sv.gmo to /root/dest/usr/share/locale/sv/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/tr.gmo to /root/dest/usr/share/locale/tr/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/uk.gmo to /root/dest/usr/share/locale/uk/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/zh_CN.gmo to /root/dest/usr/share/locale/zh_CN/LC_MESSAGES/systemd.mo\r\nInstalling /root/build/po/zh_TW.gmo to /root/dest/usr/share/locale/zh_TW/LC_MESSAGES/systemd.mo\r\n```\r\n\r\n**To Reproduce**\r\n\r\n```\r\ngit clone https://github.com/systemd/systemd.git\r\nmeson build\r\nmeson compile -C build\r\nmeson install -C build --quiet\r\n```\r\n\r\n**Expected behavior**\r\n\r\nNo output from gettext when --quiet is passed to `meson install`\r\n\r\n**system parameters**\r\n\r\nShould not be relevant to this bug. Let me know if I should add them. I'm running latest meson (0.56.0)\r\n\n", "before_files": [{"content": "# Copyright 2016 The Meson development team\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport shutil\nimport argparse\nimport subprocess\nfrom . import destdir_join\nimport typing as T\n\nparser = argparse.ArgumentParser()\nparser.add_argument('command')\nparser.add_argument('--pkgname', default='')\nparser.add_argument('--datadirs', default='')\nparser.add_argument('--langs', default='')\nparser.add_argument('--localedir', default='')\nparser.add_argument('--subdir', default='')\nparser.add_argument('--extra-args', default='')\n\ndef read_linguas(src_sub: str) -> T.List[str]:\n # Syntax of this file is documented here:\n # https://www.gnu.org/software/gettext/manual/html_node/po_002fLINGUAS.html\n linguas = os.path.join(src_sub, 'LINGUAS')\n try:\n langs = []\n with open(linguas) as f:\n for line in f:\n line = line.strip()\n if line and not line.startswith('#'):\n langs += line.split()\n return langs\n except (FileNotFoundError, PermissionError):\n print('Could not find file LINGUAS in {}'.format(src_sub))\n return []\n\ndef run_potgen(src_sub: str, pkgname: str, datadirs: str, args: T.List[str]) -> int:\n listfile = os.path.join(src_sub, 'POTFILES.in')\n if not os.path.exists(listfile):\n listfile = os.path.join(src_sub, 'POTFILES')\n if not os.path.exists(listfile):\n print('Could not find file POTFILES in %s' % src_sub)\n return 1\n\n child_env = os.environ.copy()\n if datadirs:\n child_env['GETTEXTDATADIRS'] = datadirs\n\n ofile = os.path.join(src_sub, pkgname + '.pot')\n return subprocess.call(['xgettext', '--package-name=' + pkgname, '-p', src_sub, '-f', listfile,\n '-D', os.environ['MESON_SOURCE_ROOT'], '-k_', '-o', ofile] + args,\n env=child_env)\n\ndef gen_gmo(src_sub: str, bld_sub: str, langs: T.List[str]) -> int:\n for l in langs:\n subprocess.check_call(['msgfmt', os.path.join(src_sub, l + '.po'),\n '-o', os.path.join(bld_sub, l + '.gmo')])\n return 0\n\ndef update_po(src_sub: str, pkgname: str, langs: T.List[str]) -> int:\n potfile = os.path.join(src_sub, pkgname + '.pot')\n for l in langs:\n pofile = os.path.join(src_sub, l + '.po')\n if os.path.exists(pofile):\n subprocess.check_call(['msgmerge', '-q', '-o', pofile, pofile, potfile])\n else:\n subprocess.check_call(['msginit', '--input', potfile, '--output-file', pofile, '--locale', l, '--no-translator'])\n return 0\n\ndef do_install(src_sub: str, bld_sub: str, dest: str, pkgname: str, langs: T.List[str]) -> int:\n for l in langs:\n srcfile = os.path.join(bld_sub, l + '.gmo')\n outfile = os.path.join(dest, l, 'LC_MESSAGES',\n pkgname + '.mo')\n tempfile = outfile + '.tmp'\n os.makedirs(os.path.dirname(outfile), exist_ok=True)\n shutil.copy2(srcfile, tempfile)\n os.replace(tempfile, outfile)\n print('Installing %s to %s' % (srcfile, outfile))\n return 0\n\ndef run(args: T.List[str]) -> int:\n options = parser.parse_args(args)\n subcmd = options.command\n langs = options.langs.split('@@') if options.langs else None\n extra_args = options.extra_args.split('@@') if options.extra_args else []\n subdir = os.environ.get('MESON_SUBDIR', '')\n if options.subdir:\n subdir = options.subdir\n src_sub = os.path.join(os.environ['MESON_SOURCE_ROOT'], subdir)\n bld_sub = os.path.join(os.environ['MESON_BUILD_ROOT'], subdir)\n\n if not langs:\n langs = read_linguas(src_sub)\n\n if subcmd == 'pot':\n return run_potgen(src_sub, options.pkgname, options.datadirs, extra_args)\n elif subcmd == 'gen_gmo':\n return gen_gmo(src_sub, bld_sub, langs)\n elif subcmd == 'update_po':\n if run_potgen(src_sub, options.pkgname, options.datadirs, extra_args) != 0:\n return 1\n return update_po(src_sub, options.pkgname, langs)\n elif subcmd == 'install':\n destdir = os.environ.get('DESTDIR', '')\n dest = destdir_join(destdir, os.path.join(os.environ['MESON_INSTALL_PREFIX'],\n options.localedir))\n if gen_gmo(src_sub, bld_sub, langs) != 0:\n return 1\n do_install(src_sub, bld_sub, dest, options.pkgname, langs)\n else:\n print('Unknown subcommand.')\n return 1\n return 0\n", "path": "mesonbuild/scripts/gettext.py"}], "after_files": [{"content": "# Copyright 2016 The Meson development team\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\nimport shutil\nimport argparse\nimport subprocess\nfrom . import destdir_join\nimport typing as T\n\nparser = argparse.ArgumentParser()\nparser.add_argument('command')\nparser.add_argument('--pkgname', default='')\nparser.add_argument('--datadirs', default='')\nparser.add_argument('--langs', default='')\nparser.add_argument('--localedir', default='')\nparser.add_argument('--subdir', default='')\nparser.add_argument('--extra-args', default='')\n\ndef read_linguas(src_sub: str) -> T.List[str]:\n # Syntax of this file is documented here:\n # https://www.gnu.org/software/gettext/manual/html_node/po_002fLINGUAS.html\n linguas = os.path.join(src_sub, 'LINGUAS')\n try:\n langs = []\n with open(linguas) as f:\n for line in f:\n line = line.strip()\n if line and not line.startswith('#'):\n langs += line.split()\n return langs\n except (FileNotFoundError, PermissionError):\n print('Could not find file LINGUAS in {}'.format(src_sub))\n return []\n\ndef run_potgen(src_sub: str, pkgname: str, datadirs: str, args: T.List[str]) -> int:\n listfile = os.path.join(src_sub, 'POTFILES.in')\n if not os.path.exists(listfile):\n listfile = os.path.join(src_sub, 'POTFILES')\n if not os.path.exists(listfile):\n print('Could not find file POTFILES in %s' % src_sub)\n return 1\n\n child_env = os.environ.copy()\n if datadirs:\n child_env['GETTEXTDATADIRS'] = datadirs\n\n ofile = os.path.join(src_sub, pkgname + '.pot')\n return subprocess.call(['xgettext', '--package-name=' + pkgname, '-p', src_sub, '-f', listfile,\n '-D', os.environ['MESON_SOURCE_ROOT'], '-k_', '-o', ofile] + args,\n env=child_env)\n\ndef gen_gmo(src_sub: str, bld_sub: str, langs: T.List[str]) -> int:\n for l in langs:\n subprocess.check_call(['msgfmt', os.path.join(src_sub, l + '.po'),\n '-o', os.path.join(bld_sub, l + '.gmo')])\n return 0\n\ndef update_po(src_sub: str, pkgname: str, langs: T.List[str]) -> int:\n potfile = os.path.join(src_sub, pkgname + '.pot')\n for l in langs:\n pofile = os.path.join(src_sub, l + '.po')\n if os.path.exists(pofile):\n subprocess.check_call(['msgmerge', '-q', '-o', pofile, pofile, potfile])\n else:\n subprocess.check_call(['msginit', '--input', potfile, '--output-file', pofile, '--locale', l, '--no-translator'])\n return 0\n\ndef do_install(src_sub: str, bld_sub: str, dest: str, pkgname: str, langs: T.List[str]) -> int:\n for l in langs:\n srcfile = os.path.join(bld_sub, l + '.gmo')\n outfile = os.path.join(dest, l, 'LC_MESSAGES',\n pkgname + '.mo')\n tempfile = outfile + '.tmp'\n os.makedirs(os.path.dirname(outfile), exist_ok=True)\n shutil.copy2(srcfile, tempfile)\n os.replace(tempfile, outfile)\n if not os.getenv('MESON_INSTALL_QUIET', False):\n print('Installing %s to %s' % (srcfile, outfile))\n return 0\n\ndef run(args: T.List[str]) -> int:\n options = parser.parse_args(args)\n subcmd = options.command\n langs = options.langs.split('@@') if options.langs else None\n extra_args = options.extra_args.split('@@') if options.extra_args else []\n subdir = os.environ.get('MESON_SUBDIR', '')\n if options.subdir:\n subdir = options.subdir\n src_sub = os.path.join(os.environ['MESON_SOURCE_ROOT'], subdir)\n bld_sub = os.path.join(os.environ['MESON_BUILD_ROOT'], subdir)\n\n if not langs:\n langs = read_linguas(src_sub)\n\n if subcmd == 'pot':\n return run_potgen(src_sub, options.pkgname, options.datadirs, extra_args)\n elif subcmd == 'gen_gmo':\n return gen_gmo(src_sub, bld_sub, langs)\n elif subcmd == 'update_po':\n if run_potgen(src_sub, options.pkgname, options.datadirs, extra_args) != 0:\n return 1\n return update_po(src_sub, options.pkgname, langs)\n elif subcmd == 'install':\n destdir = os.environ.get('DESTDIR', '')\n dest = destdir_join(destdir, os.path.join(os.environ['MESON_INSTALL_PREFIX'],\n options.localedir))\n if gen_gmo(src_sub, bld_sub, langs) != 0:\n return 1\n do_install(src_sub, bld_sub, dest, options.pkgname, langs)\n else:\n print('Unknown subcommand.')\n return 1\n return 0\n", "path": "mesonbuild/scripts/gettext.py"}]}
| 2,710 | 149 |
gh_patches_debug_41584
|
rasdani/github-patches
|
git_diff
|
pypa__pip-9438
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Remove support for pip<20.0 cache entries
Cf https://github.com/pypa/pip/pull/7319/files#diff-e455c1048a1016c5fa2c413f52146488R142
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/pip/_internal/cache.py`
Content:
```
1 """Cache Management
2 """
3
4 import hashlib
5 import json
6 import logging
7 import os
8
9 from pip._vendor.packaging.tags import interpreter_name, interpreter_version
10 from pip._vendor.packaging.utils import canonicalize_name
11
12 from pip._internal.exceptions import InvalidWheelFilename
13 from pip._internal.models.link import Link
14 from pip._internal.models.wheel import Wheel
15 from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds
16 from pip._internal.utils.typing import MYPY_CHECK_RUNNING
17 from pip._internal.utils.urls import path_to_url
18
19 if MYPY_CHECK_RUNNING:
20 from typing import Any, Dict, List, Optional, Set
21
22 from pip._vendor.packaging.tags import Tag
23
24 from pip._internal.models.format_control import FormatControl
25
26 logger = logging.getLogger(__name__)
27
28
29 def _hash_dict(d):
30 # type: (Dict[str, str]) -> str
31 """Return a stable sha224 of a dictionary."""
32 s = json.dumps(d, sort_keys=True, separators=(",", ":"), ensure_ascii=True)
33 return hashlib.sha224(s.encode("ascii")).hexdigest()
34
35
36 class Cache:
37 """An abstract class - provides cache directories for data from links
38
39
40 :param cache_dir: The root of the cache.
41 :param format_control: An object of FormatControl class to limit
42 binaries being read from the cache.
43 :param allowed_formats: which formats of files the cache should store.
44 ('binary' and 'source' are the only allowed values)
45 """
46
47 def __init__(self, cache_dir, format_control, allowed_formats):
48 # type: (str, FormatControl, Set[str]) -> None
49 super().__init__()
50 assert not cache_dir or os.path.isabs(cache_dir)
51 self.cache_dir = cache_dir or None
52 self.format_control = format_control
53 self.allowed_formats = allowed_formats
54
55 _valid_formats = {"source", "binary"}
56 assert self.allowed_formats.union(_valid_formats) == _valid_formats
57
58 def _get_cache_path_parts_legacy(self, link):
59 # type: (Link) -> List[str]
60 """Get parts of part that must be os.path.joined with cache_dir
61
62 Legacy cache key (pip < 20) for compatibility with older caches.
63 """
64
65 # We want to generate an url to use as our cache key, we don't want to
66 # just re-use the URL because it might have other items in the fragment
67 # and we don't care about those.
68 key_parts = [link.url_without_fragment]
69 if link.hash_name is not None and link.hash is not None:
70 key_parts.append("=".join([link.hash_name, link.hash]))
71 key_url = "#".join(key_parts)
72
73 # Encode our key url with sha224, we'll use this because it has similar
74 # security properties to sha256, but with a shorter total output (and
75 # thus less secure). However the differences don't make a lot of
76 # difference for our use case here.
77 hashed = hashlib.sha224(key_url.encode()).hexdigest()
78
79 # We want to nest the directories some to prevent having a ton of top
80 # level directories where we might run out of sub directories on some
81 # FS.
82 parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]
83
84 return parts
85
86 def _get_cache_path_parts(self, link):
87 # type: (Link) -> List[str]
88 """Get parts of part that must be os.path.joined with cache_dir
89 """
90
91 # We want to generate an url to use as our cache key, we don't want to
92 # just re-use the URL because it might have other items in the fragment
93 # and we don't care about those.
94 key_parts = {"url": link.url_without_fragment}
95 if link.hash_name is not None and link.hash is not None:
96 key_parts[link.hash_name] = link.hash
97 if link.subdirectory_fragment:
98 key_parts["subdirectory"] = link.subdirectory_fragment
99
100 # Include interpreter name, major and minor version in cache key
101 # to cope with ill-behaved sdists that build a different wheel
102 # depending on the python version their setup.py is being run on,
103 # and don't encode the difference in compatibility tags.
104 # https://github.com/pypa/pip/issues/7296
105 key_parts["interpreter_name"] = interpreter_name()
106 key_parts["interpreter_version"] = interpreter_version()
107
108 # Encode our key url with sha224, we'll use this because it has similar
109 # security properties to sha256, but with a shorter total output (and
110 # thus less secure). However the differences don't make a lot of
111 # difference for our use case here.
112 hashed = _hash_dict(key_parts)
113
114 # We want to nest the directories some to prevent having a ton of top
115 # level directories where we might run out of sub directories on some
116 # FS.
117 parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]
118
119 return parts
120
121 def _get_candidates(self, link, canonical_package_name):
122 # type: (Link, str) -> List[Any]
123 can_not_cache = (
124 not self.cache_dir or
125 not canonical_package_name or
126 not link
127 )
128 if can_not_cache:
129 return []
130
131 formats = self.format_control.get_allowed_formats(
132 canonical_package_name
133 )
134 if not self.allowed_formats.intersection(formats):
135 return []
136
137 candidates = []
138 path = self.get_path_for_link(link)
139 if os.path.isdir(path):
140 for candidate in os.listdir(path):
141 candidates.append((candidate, path))
142 # TODO remove legacy path lookup in pip>=21
143 legacy_path = self.get_path_for_link_legacy(link)
144 if os.path.isdir(legacy_path):
145 for candidate in os.listdir(legacy_path):
146 candidates.append((candidate, legacy_path))
147 return candidates
148
149 def get_path_for_link_legacy(self, link):
150 # type: (Link) -> str
151 raise NotImplementedError()
152
153 def get_path_for_link(self, link):
154 # type: (Link) -> str
155 """Return a directory to store cached items in for link.
156 """
157 raise NotImplementedError()
158
159 def get(
160 self,
161 link, # type: Link
162 package_name, # type: Optional[str]
163 supported_tags, # type: List[Tag]
164 ):
165 # type: (...) -> Link
166 """Returns a link to a cached item if it exists, otherwise returns the
167 passed link.
168 """
169 raise NotImplementedError()
170
171
172 class SimpleWheelCache(Cache):
173 """A cache of wheels for future installs.
174 """
175
176 def __init__(self, cache_dir, format_control):
177 # type: (str, FormatControl) -> None
178 super().__init__(cache_dir, format_control, {"binary"})
179
180 def get_path_for_link_legacy(self, link):
181 # type: (Link) -> str
182 parts = self._get_cache_path_parts_legacy(link)
183 assert self.cache_dir
184 return os.path.join(self.cache_dir, "wheels", *parts)
185
186 def get_path_for_link(self, link):
187 # type: (Link) -> str
188 """Return a directory to store cached wheels for link
189
190 Because there are M wheels for any one sdist, we provide a directory
191 to cache them in, and then consult that directory when looking up
192 cache hits.
193
194 We only insert things into the cache if they have plausible version
195 numbers, so that we don't contaminate the cache with things that were
196 not unique. E.g. ./package might have dozens of installs done for it
197 and build a version of 0.0...and if we built and cached a wheel, we'd
198 end up using the same wheel even if the source has been edited.
199
200 :param link: The link of the sdist for which this will cache wheels.
201 """
202 parts = self._get_cache_path_parts(link)
203 assert self.cache_dir
204 # Store wheels within the root cache_dir
205 return os.path.join(self.cache_dir, "wheels", *parts)
206
207 def get(
208 self,
209 link, # type: Link
210 package_name, # type: Optional[str]
211 supported_tags, # type: List[Tag]
212 ):
213 # type: (...) -> Link
214 candidates = []
215
216 if not package_name:
217 return link
218
219 canonical_package_name = canonicalize_name(package_name)
220 for wheel_name, wheel_dir in self._get_candidates(
221 link, canonical_package_name
222 ):
223 try:
224 wheel = Wheel(wheel_name)
225 except InvalidWheelFilename:
226 continue
227 if canonicalize_name(wheel.name) != canonical_package_name:
228 logger.debug(
229 "Ignoring cached wheel %s for %s as it "
230 "does not match the expected distribution name %s.",
231 wheel_name, link, package_name,
232 )
233 continue
234 if not wheel.supported(supported_tags):
235 # Built for a different python/arch/etc
236 continue
237 candidates.append(
238 (
239 wheel.support_index_min(supported_tags),
240 wheel_name,
241 wheel_dir,
242 )
243 )
244
245 if not candidates:
246 return link
247
248 _, wheel_name, wheel_dir = min(candidates)
249 return Link(path_to_url(os.path.join(wheel_dir, wheel_name)))
250
251
252 class EphemWheelCache(SimpleWheelCache):
253 """A SimpleWheelCache that creates it's own temporary cache directory
254 """
255
256 def __init__(self, format_control):
257 # type: (FormatControl) -> None
258 self._temp_dir = TempDirectory(
259 kind=tempdir_kinds.EPHEM_WHEEL_CACHE,
260 globally_managed=True,
261 )
262
263 super().__init__(self._temp_dir.path, format_control)
264
265
266 class CacheEntry:
267 def __init__(
268 self,
269 link, # type: Link
270 persistent, # type: bool
271 ):
272 self.link = link
273 self.persistent = persistent
274
275
276 class WheelCache(Cache):
277 """Wraps EphemWheelCache and SimpleWheelCache into a single Cache
278
279 This Cache allows for gracefully degradation, using the ephem wheel cache
280 when a certain link is not found in the simple wheel cache first.
281 """
282
283 def __init__(self, cache_dir, format_control):
284 # type: (str, FormatControl) -> None
285 super().__init__(cache_dir, format_control, {'binary'})
286 self._wheel_cache = SimpleWheelCache(cache_dir, format_control)
287 self._ephem_cache = EphemWheelCache(format_control)
288
289 def get_path_for_link_legacy(self, link):
290 # type: (Link) -> str
291 return self._wheel_cache.get_path_for_link_legacy(link)
292
293 def get_path_for_link(self, link):
294 # type: (Link) -> str
295 return self._wheel_cache.get_path_for_link(link)
296
297 def get_ephem_path_for_link(self, link):
298 # type: (Link) -> str
299 return self._ephem_cache.get_path_for_link(link)
300
301 def get(
302 self,
303 link, # type: Link
304 package_name, # type: Optional[str]
305 supported_tags, # type: List[Tag]
306 ):
307 # type: (...) -> Link
308 cache_entry = self.get_cache_entry(link, package_name, supported_tags)
309 if cache_entry is None:
310 return link
311 return cache_entry.link
312
313 def get_cache_entry(
314 self,
315 link, # type: Link
316 package_name, # type: Optional[str]
317 supported_tags, # type: List[Tag]
318 ):
319 # type: (...) -> Optional[CacheEntry]
320 """Returns a CacheEntry with a link to a cached item if it exists or
321 None. The cache entry indicates if the item was found in the persistent
322 or ephemeral cache.
323 """
324 retval = self._wheel_cache.get(
325 link=link,
326 package_name=package_name,
327 supported_tags=supported_tags,
328 )
329 if retval is not link:
330 return CacheEntry(retval, persistent=True)
331
332 retval = self._ephem_cache.get(
333 link=link,
334 package_name=package_name,
335 supported_tags=supported_tags,
336 )
337 if retval is not link:
338 return CacheEntry(retval, persistent=False)
339
340 return None
341
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/pip/_internal/cache.py b/src/pip/_internal/cache.py
--- a/src/pip/_internal/cache.py
+++ b/src/pip/_internal/cache.py
@@ -55,34 +55,6 @@
_valid_formats = {"source", "binary"}
assert self.allowed_formats.union(_valid_formats) == _valid_formats
- def _get_cache_path_parts_legacy(self, link):
- # type: (Link) -> List[str]
- """Get parts of part that must be os.path.joined with cache_dir
-
- Legacy cache key (pip < 20) for compatibility with older caches.
- """
-
- # We want to generate an url to use as our cache key, we don't want to
- # just re-use the URL because it might have other items in the fragment
- # and we don't care about those.
- key_parts = [link.url_without_fragment]
- if link.hash_name is not None and link.hash is not None:
- key_parts.append("=".join([link.hash_name, link.hash]))
- key_url = "#".join(key_parts)
-
- # Encode our key url with sha224, we'll use this because it has similar
- # security properties to sha256, but with a shorter total output (and
- # thus less secure). However the differences don't make a lot of
- # difference for our use case here.
- hashed = hashlib.sha224(key_url.encode()).hexdigest()
-
- # We want to nest the directories some to prevent having a ton of top
- # level directories where we might run out of sub directories on some
- # FS.
- parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]
-
- return parts
-
def _get_cache_path_parts(self, link):
# type: (Link) -> List[str]
"""Get parts of part that must be os.path.joined with cache_dir
@@ -139,17 +111,8 @@
if os.path.isdir(path):
for candidate in os.listdir(path):
candidates.append((candidate, path))
- # TODO remove legacy path lookup in pip>=21
- legacy_path = self.get_path_for_link_legacy(link)
- if os.path.isdir(legacy_path):
- for candidate in os.listdir(legacy_path):
- candidates.append((candidate, legacy_path))
return candidates
- def get_path_for_link_legacy(self, link):
- # type: (Link) -> str
- raise NotImplementedError()
-
def get_path_for_link(self, link):
# type: (Link) -> str
"""Return a directory to store cached items in for link.
@@ -177,12 +140,6 @@
# type: (str, FormatControl) -> None
super().__init__(cache_dir, format_control, {"binary"})
- def get_path_for_link_legacy(self, link):
- # type: (Link) -> str
- parts = self._get_cache_path_parts_legacy(link)
- assert self.cache_dir
- return os.path.join(self.cache_dir, "wheels", *parts)
-
def get_path_for_link(self, link):
# type: (Link) -> str
"""Return a directory to store cached wheels for link
@@ -286,10 +243,6 @@
self._wheel_cache = SimpleWheelCache(cache_dir, format_control)
self._ephem_cache = EphemWheelCache(format_control)
- def get_path_for_link_legacy(self, link):
- # type: (Link) -> str
- return self._wheel_cache.get_path_for_link_legacy(link)
-
def get_path_for_link(self, link):
# type: (Link) -> str
return self._wheel_cache.get_path_for_link(link)
|
{"golden_diff": "diff --git a/src/pip/_internal/cache.py b/src/pip/_internal/cache.py\n--- a/src/pip/_internal/cache.py\n+++ b/src/pip/_internal/cache.py\n@@ -55,34 +55,6 @@\n _valid_formats = {\"source\", \"binary\"}\n assert self.allowed_formats.union(_valid_formats) == _valid_formats\n \n- def _get_cache_path_parts_legacy(self, link):\n- # type: (Link) -> List[str]\n- \"\"\"Get parts of part that must be os.path.joined with cache_dir\n-\n- Legacy cache key (pip < 20) for compatibility with older caches.\n- \"\"\"\n-\n- # We want to generate an url to use as our cache key, we don't want to\n- # just re-use the URL because it might have other items in the fragment\n- # and we don't care about those.\n- key_parts = [link.url_without_fragment]\n- if link.hash_name is not None and link.hash is not None:\n- key_parts.append(\"=\".join([link.hash_name, link.hash]))\n- key_url = \"#\".join(key_parts)\n-\n- # Encode our key url with sha224, we'll use this because it has similar\n- # security properties to sha256, but with a shorter total output (and\n- # thus less secure). However the differences don't make a lot of\n- # difference for our use case here.\n- hashed = hashlib.sha224(key_url.encode()).hexdigest()\n-\n- # We want to nest the directories some to prevent having a ton of top\n- # level directories where we might run out of sub directories on some\n- # FS.\n- parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]\n-\n- return parts\n-\n def _get_cache_path_parts(self, link):\n # type: (Link) -> List[str]\n \"\"\"Get parts of part that must be os.path.joined with cache_dir\n@@ -139,17 +111,8 @@\n if os.path.isdir(path):\n for candidate in os.listdir(path):\n candidates.append((candidate, path))\n- # TODO remove legacy path lookup in pip>=21\n- legacy_path = self.get_path_for_link_legacy(link)\n- if os.path.isdir(legacy_path):\n- for candidate in os.listdir(legacy_path):\n- candidates.append((candidate, legacy_path))\n return candidates\n \n- def get_path_for_link_legacy(self, link):\n- # type: (Link) -> str\n- raise NotImplementedError()\n-\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached items in for link.\n@@ -177,12 +140,6 @@\n # type: (str, FormatControl) -> None\n super().__init__(cache_dir, format_control, {\"binary\"})\n \n- def get_path_for_link_legacy(self, link):\n- # type: (Link) -> str\n- parts = self._get_cache_path_parts_legacy(link)\n- assert self.cache_dir\n- return os.path.join(self.cache_dir, \"wheels\", *parts)\n-\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached wheels for link\n@@ -286,10 +243,6 @@\n self._wheel_cache = SimpleWheelCache(cache_dir, format_control)\n self._ephem_cache = EphemWheelCache(format_control)\n \n- def get_path_for_link_legacy(self, link):\n- # type: (Link) -> str\n- return self._wheel_cache.get_path_for_link_legacy(link)\n-\n def get_path_for_link(self, link):\n # type: (Link) -> str\n return self._wheel_cache.get_path_for_link(link)\n", "issue": "Remove support for pip<20.0 cache entries\nCf https://github.com/pypa/pip/pull/7319/files#diff-e455c1048a1016c5fa2c413f52146488R142\n", "before_files": [{"content": "\"\"\"Cache Management\n\"\"\"\n\nimport hashlib\nimport json\nimport logging\nimport os\n\nfrom pip._vendor.packaging.tags import interpreter_name, interpreter_version\nfrom pip._vendor.packaging.utils import canonicalize_name\n\nfrom pip._internal.exceptions import InvalidWheelFilename\nfrom pip._internal.models.link import Link\nfrom pip._internal.models.wheel import Wheel\nfrom pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds\nfrom pip._internal.utils.typing import MYPY_CHECK_RUNNING\nfrom pip._internal.utils.urls import path_to_url\n\nif MYPY_CHECK_RUNNING:\n from typing import Any, Dict, List, Optional, Set\n\n from pip._vendor.packaging.tags import Tag\n\n from pip._internal.models.format_control import FormatControl\n\nlogger = logging.getLogger(__name__)\n\n\ndef _hash_dict(d):\n # type: (Dict[str, str]) -> str\n \"\"\"Return a stable sha224 of a dictionary.\"\"\"\n s = json.dumps(d, sort_keys=True, separators=(\",\", \":\"), ensure_ascii=True)\n return hashlib.sha224(s.encode(\"ascii\")).hexdigest()\n\n\nclass Cache:\n \"\"\"An abstract class - provides cache directories for data from links\n\n\n :param cache_dir: The root of the cache.\n :param format_control: An object of FormatControl class to limit\n binaries being read from the cache.\n :param allowed_formats: which formats of files the cache should store.\n ('binary' and 'source' are the only allowed values)\n \"\"\"\n\n def __init__(self, cache_dir, format_control, allowed_formats):\n # type: (str, FormatControl, Set[str]) -> None\n super().__init__()\n assert not cache_dir or os.path.isabs(cache_dir)\n self.cache_dir = cache_dir or None\n self.format_control = format_control\n self.allowed_formats = allowed_formats\n\n _valid_formats = {\"source\", \"binary\"}\n assert self.allowed_formats.union(_valid_formats) == _valid_formats\n\n def _get_cache_path_parts_legacy(self, link):\n # type: (Link) -> List[str]\n \"\"\"Get parts of part that must be os.path.joined with cache_dir\n\n Legacy cache key (pip < 20) for compatibility with older caches.\n \"\"\"\n\n # We want to generate an url to use as our cache key, we don't want to\n # just re-use the URL because it might have other items in the fragment\n # and we don't care about those.\n key_parts = [link.url_without_fragment]\n if link.hash_name is not None and link.hash is not None:\n key_parts.append(\"=\".join([link.hash_name, link.hash]))\n key_url = \"#\".join(key_parts)\n\n # Encode our key url with sha224, we'll use this because it has similar\n # security properties to sha256, but with a shorter total output (and\n # thus less secure). However the differences don't make a lot of\n # difference for our use case here.\n hashed = hashlib.sha224(key_url.encode()).hexdigest()\n\n # We want to nest the directories some to prevent having a ton of top\n # level directories where we might run out of sub directories on some\n # FS.\n parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]\n\n return parts\n\n def _get_cache_path_parts(self, link):\n # type: (Link) -> List[str]\n \"\"\"Get parts of part that must be os.path.joined with cache_dir\n \"\"\"\n\n # We want to generate an url to use as our cache key, we don't want to\n # just re-use the URL because it might have other items in the fragment\n # and we don't care about those.\n key_parts = {\"url\": link.url_without_fragment}\n if link.hash_name is not None and link.hash is not None:\n key_parts[link.hash_name] = link.hash\n if link.subdirectory_fragment:\n key_parts[\"subdirectory\"] = link.subdirectory_fragment\n\n # Include interpreter name, major and minor version in cache key\n # to cope with ill-behaved sdists that build a different wheel\n # depending on the python version their setup.py is being run on,\n # and don't encode the difference in compatibility tags.\n # https://github.com/pypa/pip/issues/7296\n key_parts[\"interpreter_name\"] = interpreter_name()\n key_parts[\"interpreter_version\"] = interpreter_version()\n\n # Encode our key url with sha224, we'll use this because it has similar\n # security properties to sha256, but with a shorter total output (and\n # thus less secure). However the differences don't make a lot of\n # difference for our use case here.\n hashed = _hash_dict(key_parts)\n\n # We want to nest the directories some to prevent having a ton of top\n # level directories where we might run out of sub directories on some\n # FS.\n parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]\n\n return parts\n\n def _get_candidates(self, link, canonical_package_name):\n # type: (Link, str) -> List[Any]\n can_not_cache = (\n not self.cache_dir or\n not canonical_package_name or\n not link\n )\n if can_not_cache:\n return []\n\n formats = self.format_control.get_allowed_formats(\n canonical_package_name\n )\n if not self.allowed_formats.intersection(formats):\n return []\n\n candidates = []\n path = self.get_path_for_link(link)\n if os.path.isdir(path):\n for candidate in os.listdir(path):\n candidates.append((candidate, path))\n # TODO remove legacy path lookup in pip>=21\n legacy_path = self.get_path_for_link_legacy(link)\n if os.path.isdir(legacy_path):\n for candidate in os.listdir(legacy_path):\n candidates.append((candidate, legacy_path))\n return candidates\n\n def get_path_for_link_legacy(self, link):\n # type: (Link) -> str\n raise NotImplementedError()\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached items in for link.\n \"\"\"\n raise NotImplementedError()\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n \"\"\"Returns a link to a cached item if it exists, otherwise returns the\n passed link.\n \"\"\"\n raise NotImplementedError()\n\n\nclass SimpleWheelCache(Cache):\n \"\"\"A cache of wheels for future installs.\n \"\"\"\n\n def __init__(self, cache_dir, format_control):\n # type: (str, FormatControl) -> None\n super().__init__(cache_dir, format_control, {\"binary\"})\n\n def get_path_for_link_legacy(self, link):\n # type: (Link) -> str\n parts = self._get_cache_path_parts_legacy(link)\n assert self.cache_dir\n return os.path.join(self.cache_dir, \"wheels\", *parts)\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached wheels for link\n\n Because there are M wheels for any one sdist, we provide a directory\n to cache them in, and then consult that directory when looking up\n cache hits.\n\n We only insert things into the cache if they have plausible version\n numbers, so that we don't contaminate the cache with things that were\n not unique. E.g. ./package might have dozens of installs done for it\n and build a version of 0.0...and if we built and cached a wheel, we'd\n end up using the same wheel even if the source has been edited.\n\n :param link: The link of the sdist for which this will cache wheels.\n \"\"\"\n parts = self._get_cache_path_parts(link)\n assert self.cache_dir\n # Store wheels within the root cache_dir\n return os.path.join(self.cache_dir, \"wheels\", *parts)\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n candidates = []\n\n if not package_name:\n return link\n\n canonical_package_name = canonicalize_name(package_name)\n for wheel_name, wheel_dir in self._get_candidates(\n link, canonical_package_name\n ):\n try:\n wheel = Wheel(wheel_name)\n except InvalidWheelFilename:\n continue\n if canonicalize_name(wheel.name) != canonical_package_name:\n logger.debug(\n \"Ignoring cached wheel %s for %s as it \"\n \"does not match the expected distribution name %s.\",\n wheel_name, link, package_name,\n )\n continue\n if not wheel.supported(supported_tags):\n # Built for a different python/arch/etc\n continue\n candidates.append(\n (\n wheel.support_index_min(supported_tags),\n wheel_name,\n wheel_dir,\n )\n )\n\n if not candidates:\n return link\n\n _, wheel_name, wheel_dir = min(candidates)\n return Link(path_to_url(os.path.join(wheel_dir, wheel_name)))\n\n\nclass EphemWheelCache(SimpleWheelCache):\n \"\"\"A SimpleWheelCache that creates it's own temporary cache directory\n \"\"\"\n\n def __init__(self, format_control):\n # type: (FormatControl) -> None\n self._temp_dir = TempDirectory(\n kind=tempdir_kinds.EPHEM_WHEEL_CACHE,\n globally_managed=True,\n )\n\n super().__init__(self._temp_dir.path, format_control)\n\n\nclass CacheEntry:\n def __init__(\n self,\n link, # type: Link\n persistent, # type: bool\n ):\n self.link = link\n self.persistent = persistent\n\n\nclass WheelCache(Cache):\n \"\"\"Wraps EphemWheelCache and SimpleWheelCache into a single Cache\n\n This Cache allows for gracefully degradation, using the ephem wheel cache\n when a certain link is not found in the simple wheel cache first.\n \"\"\"\n\n def __init__(self, cache_dir, format_control):\n # type: (str, FormatControl) -> None\n super().__init__(cache_dir, format_control, {'binary'})\n self._wheel_cache = SimpleWheelCache(cache_dir, format_control)\n self._ephem_cache = EphemWheelCache(format_control)\n\n def get_path_for_link_legacy(self, link):\n # type: (Link) -> str\n return self._wheel_cache.get_path_for_link_legacy(link)\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n return self._wheel_cache.get_path_for_link(link)\n\n def get_ephem_path_for_link(self, link):\n # type: (Link) -> str\n return self._ephem_cache.get_path_for_link(link)\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n cache_entry = self.get_cache_entry(link, package_name, supported_tags)\n if cache_entry is None:\n return link\n return cache_entry.link\n\n def get_cache_entry(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Optional[CacheEntry]\n \"\"\"Returns a CacheEntry with a link to a cached item if it exists or\n None. The cache entry indicates if the item was found in the persistent\n or ephemeral cache.\n \"\"\"\n retval = self._wheel_cache.get(\n link=link,\n package_name=package_name,\n supported_tags=supported_tags,\n )\n if retval is not link:\n return CacheEntry(retval, persistent=True)\n\n retval = self._ephem_cache.get(\n link=link,\n package_name=package_name,\n supported_tags=supported_tags,\n )\n if retval is not link:\n return CacheEntry(retval, persistent=False)\n\n return None\n", "path": "src/pip/_internal/cache.py"}], "after_files": [{"content": "\"\"\"Cache Management\n\"\"\"\n\nimport hashlib\nimport json\nimport logging\nimport os\n\nfrom pip._vendor.packaging.tags import interpreter_name, interpreter_version\nfrom pip._vendor.packaging.utils import canonicalize_name\n\nfrom pip._internal.exceptions import InvalidWheelFilename\nfrom pip._internal.models.link import Link\nfrom pip._internal.models.wheel import Wheel\nfrom pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds\nfrom pip._internal.utils.typing import MYPY_CHECK_RUNNING\nfrom pip._internal.utils.urls import path_to_url\n\nif MYPY_CHECK_RUNNING:\n from typing import Any, Dict, List, Optional, Set\n\n from pip._vendor.packaging.tags import Tag\n\n from pip._internal.models.format_control import FormatControl\n\nlogger = logging.getLogger(__name__)\n\n\ndef _hash_dict(d):\n # type: (Dict[str, str]) -> str\n \"\"\"Return a stable sha224 of a dictionary.\"\"\"\n s = json.dumps(d, sort_keys=True, separators=(\",\", \":\"), ensure_ascii=True)\n return hashlib.sha224(s.encode(\"ascii\")).hexdigest()\n\n\nclass Cache:\n \"\"\"An abstract class - provides cache directories for data from links\n\n\n :param cache_dir: The root of the cache.\n :param format_control: An object of FormatControl class to limit\n binaries being read from the cache.\n :param allowed_formats: which formats of files the cache should store.\n ('binary' and 'source' are the only allowed values)\n \"\"\"\n\n def __init__(self, cache_dir, format_control, allowed_formats):\n # type: (str, FormatControl, Set[str]) -> None\n super().__init__()\n assert not cache_dir or os.path.isabs(cache_dir)\n self.cache_dir = cache_dir or None\n self.format_control = format_control\n self.allowed_formats = allowed_formats\n\n _valid_formats = {\"source\", \"binary\"}\n assert self.allowed_formats.union(_valid_formats) == _valid_formats\n\n def _get_cache_path_parts(self, link):\n # type: (Link) -> List[str]\n \"\"\"Get parts of part that must be os.path.joined with cache_dir\n \"\"\"\n\n # We want to generate an url to use as our cache key, we don't want to\n # just re-use the URL because it might have other items in the fragment\n # and we don't care about those.\n key_parts = {\"url\": link.url_without_fragment}\n if link.hash_name is not None and link.hash is not None:\n key_parts[link.hash_name] = link.hash\n if link.subdirectory_fragment:\n key_parts[\"subdirectory\"] = link.subdirectory_fragment\n\n # Include interpreter name, major and minor version in cache key\n # to cope with ill-behaved sdists that build a different wheel\n # depending on the python version their setup.py is being run on,\n # and don't encode the difference in compatibility tags.\n # https://github.com/pypa/pip/issues/7296\n key_parts[\"interpreter_name\"] = interpreter_name()\n key_parts[\"interpreter_version\"] = interpreter_version()\n\n # Encode our key url with sha224, we'll use this because it has similar\n # security properties to sha256, but with a shorter total output (and\n # thus less secure). However the differences don't make a lot of\n # difference for our use case here.\n hashed = _hash_dict(key_parts)\n\n # We want to nest the directories some to prevent having a ton of top\n # level directories where we might run out of sub directories on some\n # FS.\n parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]\n\n return parts\n\n def _get_candidates(self, link, canonical_package_name):\n # type: (Link, str) -> List[Any]\n can_not_cache = (\n not self.cache_dir or\n not canonical_package_name or\n not link\n )\n if can_not_cache:\n return []\n\n formats = self.format_control.get_allowed_formats(\n canonical_package_name\n )\n if not self.allowed_formats.intersection(formats):\n return []\n\n candidates = []\n path = self.get_path_for_link(link)\n if os.path.isdir(path):\n for candidate in os.listdir(path):\n candidates.append((candidate, path))\n return candidates\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached items in for link.\n \"\"\"\n raise NotImplementedError()\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n \"\"\"Returns a link to a cached item if it exists, otherwise returns the\n passed link.\n \"\"\"\n raise NotImplementedError()\n\n\nclass SimpleWheelCache(Cache):\n \"\"\"A cache of wheels for future installs.\n \"\"\"\n\n def __init__(self, cache_dir, format_control):\n # type: (str, FormatControl) -> None\n super().__init__(cache_dir, format_control, {\"binary\"})\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n \"\"\"Return a directory to store cached wheels for link\n\n Because there are M wheels for any one sdist, we provide a directory\n to cache them in, and then consult that directory when looking up\n cache hits.\n\n We only insert things into the cache if they have plausible version\n numbers, so that we don't contaminate the cache with things that were\n not unique. E.g. ./package might have dozens of installs done for it\n and build a version of 0.0...and if we built and cached a wheel, we'd\n end up using the same wheel even if the source has been edited.\n\n :param link: The link of the sdist for which this will cache wheels.\n \"\"\"\n parts = self._get_cache_path_parts(link)\n assert self.cache_dir\n # Store wheels within the root cache_dir\n return os.path.join(self.cache_dir, \"wheels\", *parts)\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n candidates = []\n\n if not package_name:\n return link\n\n canonical_package_name = canonicalize_name(package_name)\n for wheel_name, wheel_dir in self._get_candidates(\n link, canonical_package_name\n ):\n try:\n wheel = Wheel(wheel_name)\n except InvalidWheelFilename:\n continue\n if canonicalize_name(wheel.name) != canonical_package_name:\n logger.debug(\n \"Ignoring cached wheel %s for %s as it \"\n \"does not match the expected distribution name %s.\",\n wheel_name, link, package_name,\n )\n continue\n if not wheel.supported(supported_tags):\n # Built for a different python/arch/etc\n continue\n candidates.append(\n (\n wheel.support_index_min(supported_tags),\n wheel_name,\n wheel_dir,\n )\n )\n\n if not candidates:\n return link\n\n _, wheel_name, wheel_dir = min(candidates)\n return Link(path_to_url(os.path.join(wheel_dir, wheel_name)))\n\n\nclass EphemWheelCache(SimpleWheelCache):\n \"\"\"A SimpleWheelCache that creates it's own temporary cache directory\n \"\"\"\n\n def __init__(self, format_control):\n # type: (FormatControl) -> None\n self._temp_dir = TempDirectory(\n kind=tempdir_kinds.EPHEM_WHEEL_CACHE,\n globally_managed=True,\n )\n\n super().__init__(self._temp_dir.path, format_control)\n\n\nclass CacheEntry:\n def __init__(\n self,\n link, # type: Link\n persistent, # type: bool\n ):\n self.link = link\n self.persistent = persistent\n\n\nclass WheelCache(Cache):\n \"\"\"Wraps EphemWheelCache and SimpleWheelCache into a single Cache\n\n This Cache allows for gracefully degradation, using the ephem wheel cache\n when a certain link is not found in the simple wheel cache first.\n \"\"\"\n\n def __init__(self, cache_dir, format_control):\n # type: (str, FormatControl) -> None\n super().__init__(cache_dir, format_control, {'binary'})\n self._wheel_cache = SimpleWheelCache(cache_dir, format_control)\n self._ephem_cache = EphemWheelCache(format_control)\n\n def get_path_for_link(self, link):\n # type: (Link) -> str\n return self._wheel_cache.get_path_for_link(link)\n\n def get_ephem_path_for_link(self, link):\n # type: (Link) -> str\n return self._ephem_cache.get_path_for_link(link)\n\n def get(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Link\n cache_entry = self.get_cache_entry(link, package_name, supported_tags)\n if cache_entry is None:\n return link\n return cache_entry.link\n\n def get_cache_entry(\n self,\n link, # type: Link\n package_name, # type: Optional[str]\n supported_tags, # type: List[Tag]\n ):\n # type: (...) -> Optional[CacheEntry]\n \"\"\"Returns a CacheEntry with a link to a cached item if it exists or\n None. The cache entry indicates if the item was found in the persistent\n or ephemeral cache.\n \"\"\"\n retval = self._wheel_cache.get(\n link=link,\n package_name=package_name,\n supported_tags=supported_tags,\n )\n if retval is not link:\n return CacheEntry(retval, persistent=True)\n\n retval = self._ephem_cache.get(\n link=link,\n package_name=package_name,\n supported_tags=supported_tags,\n )\n if retval is not link:\n return CacheEntry(retval, persistent=False)\n\n return None\n", "path": "src/pip/_internal/cache.py"}]}
| 3,962 | 860 |
gh_patches_debug_2730
|
rasdani/github-patches
|
git_diff
|
microsoft__Qcodes-940
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
error when saving to drive other than current path
This is due to windows handling of drives. A minimal example:
``` python
import qcodes,os
datadir = r'd:\Temp'
qcodes.DataSet.default_io = qcodes.DiskIO(datadir)
p=qcodes.Parameter('p', set_cmd=None)
q=qcodes.Parameter('q', set_cmd=None)
ds=qcodes.Loop(p[0:10:1]).each(q).run() # fine
qcodes.DataSet.default_io = qcodes.DiskIO(r'c:\Temp')
ds=qcodes.Loop(p[0:10:1]).each(p).run() # error
```
This generates the error `ValueError: path is on mount 'd:', start on mount 'c:'`
Also see https://bugs.python.org/issue7195
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `qcodes/data/io.py`
Content:
```
1 """
2 IO managers for QCodes.
3
4 IO managers wrap whatever physical storage layer the user wants to use
5 in an interface mimicking the built-in <open> context manager, with
6 some restrictions to minimize the overhead in creating new IO managers.
7
8 The main thing these managers need to implement is the open context manager:
9
10 - Only the context manager needs to be implemented, not separate
11 open function and close methods.
12
13 - open takes the standard parameters:
14
15 - filename: (string)
16 - mode: (string) only 'r' (read), 'w' (write), and 'a' (append) are
17 expected to be implemented. As with normal file objects, the only
18 difference between write and append is that write empties the file
19 before adding new data, and append leaves the existing contents in
20 place but starts writing at the end.
21 - encoding: If a special output encoding is desired. i.e. 'utf8
22
23 - the file-like object returned should implement a minimal set of operations.
24
25 In read mode:
26 - read([size]): read to the end or at most size bytes into a string
27 - readline([size]): read until a newline or up to size bytes, into a string
28 - iter(): usually return self, but can be any iterator over lines
29 - next(): assuming iter() returns self, this yields the next line.
30
31 In write or append mode:
32 - write(s): add string s to the end of the file.
33 - writelines(seq): add a sequence of strings
34
35 IO managers should also implement:
36
37 - a join method, ala os.path.join(\*args).
38 - a list method, that returns all objects matching location
39 - a remove method, ala os.remove(path) except that it will remove directories
40 as well as files, since we're allowing "locations" to be directories
41 or files.
42 """
43
44 from contextlib import contextmanager
45 import os
46 import re
47 import shutil
48 from fnmatch import fnmatch
49
50 ALLOWED_OPEN_MODES = ('r', 'w', 'a')
51
52
53 class DiskIO:
54
55 """
56 Simple IO object to wrap disk operations with a custom base location.
57
58 Also accepts both forward and backward slashes at any point, and
59 normalizes both to the OS we are currently on.
60
61 Args:
62 base_location (str): a path to the root data folder.
63 Converted to an absolute path immediately, so even if you supply a
64 relative path, later changes to the OS working directory will not
65 affect data paths.
66 """
67
68 def __init__(self, base_location):
69 if base_location is None:
70 self.base_location = None
71 else:
72 base_location = self._normalize_slashes(base_location)
73 self.base_location = os.path.abspath(base_location)
74
75 @contextmanager
76 def open(self, filename, mode, encoding=None):
77 """
78 Mimic the interface of the built in open context manager.
79
80 Args:
81 filename (str): path relative to base_location.
82
83 mode (str): 'r' (read), 'w' (write), or 'a' (append).
84 Other open modes are not supported because we don't want
85 to force all IO managers to support others.
86
87 Returns:
88 context manager yielding the open file
89 """
90 if mode not in ALLOWED_OPEN_MODES:
91 raise ValueError('mode {} not allowed in IO managers'.format(mode))
92
93 filepath = self.to_path(filename)
94
95 # make directories if needed
96 dirpath = os.path.dirname(filepath)
97 if not os.path.exists(dirpath):
98 os.makedirs(dirpath)
99
100 # normally we'd construct this context manager with try/finally, but
101 # here we already have a context manager for open so we just wrap it
102 with open(filepath, mode, encoding=encoding) as f:
103 yield f
104
105 def _normalize_slashes(self, location):
106 # note that this is NOT os.path.join - the difference is os.path.join
107 # discards empty strings, so if you use it on a re.split absolute
108 # path you will get a relative path!
109 return os.sep.join(re.split('[\\\\/]', location))
110
111 def to_path(self, location):
112 """
113 Convert a location string into a path on the local file system.
114
115 For DiskIO this just fixes slashes and prepends the base location,
116 doing nothing active with the file. But for other io managers that
117 refer to remote storage, this method may actually fetch the file and
118 put it at a temporary local path.
119
120 Args:
121 location (str): A location string for a complete dataset or
122 a file within it.
123
124 Returns:
125 path (str): The path on disk to which this location maps.
126 """
127 location = self._normalize_slashes(location)
128 if self.base_location:
129 return os.path.join(self.base_location, location)
130 else:
131 return location
132
133 def to_location(self, path):
134 """
135 Convert a local filesystem path into a location string.
136
137 Args:
138 path (str): a path on the local file system.
139
140 Returns:
141 location (str): the location string corresponding to this path.
142 """
143 if self.base_location:
144 return os.path.relpath(path, self.base_location)
145 else:
146 return path
147
148 def __repr__(self):
149 """Show the base location in the repr."""
150 return '<DiskIO, base_location={}>'.format(repr(self.base_location))
151
152 def join(self, *args):
153 """Context-dependent os.path.join for this io manager."""
154 return os.path.join(*list(map(self._normalize_slashes, args)))
155
156 def isfile(self, location):
157 """Check whether this location matches a file."""
158 path = self.to_path(location)
159 return os.path.isfile(path)
160
161 def list(self, location, maxdepth=1, include_dirs=False):
162 """
163 Return all files that match location.
164
165 This is either files whose names match up to an arbitrary extension,
166 or any files within an exactly matching directory name.
167
168 Args:
169 location (str): the location to match.
170 May contain the usual path wildcards * and ?
171
172 maxdepth (int, optional): maximum levels of directory nesting to
173 recurse into looking for files. Default 1.
174
175 include_dirs (bool, optional): whether to allow directories in
176 the results or just files. Default False.
177
178 Returns:
179 A list of matching files and/or directories, as locations
180 relative to our base_location.
181 """
182 location = self._normalize_slashes(location)
183 search_dir, pattern = os.path.split(location)
184 path = self.to_path(search_dir)
185
186 if not os.path.isdir(path):
187 return []
188
189 matches = [fn for fn in os.listdir(path) if fnmatch(fn, pattern + '*')]
190 out = []
191
192 for match in matches:
193 matchpath = self.join(path, match)
194 if os.path.isdir(matchpath) and fnmatch(match, pattern):
195 if maxdepth > 0:
196 # exact directory match - walk down to maxdepth
197 for root, dirs, files in os.walk(matchpath, topdown=True):
198 depth = root[len(path):].count(os.path.sep)
199 if depth == maxdepth:
200 dirs[:] = [] # don't recurse any further
201
202 for fn in files + (dirs if include_dirs else []):
203 out.append(self.to_location(self.join(root, fn)))
204
205 elif include_dirs:
206 out.append(self.join(search_dir, match))
207
208 elif (os.path.isfile(matchpath) and
209 (fnmatch(match, pattern) or
210 fnmatch(os.path.splitext(match)[0], pattern))):
211 # exact filename match, or match up to an extension
212 # note that we need fnmatch(match, pattern) in addition to the
213 # splitext test to cover the case of the base filename itself
214 # containing a dot.
215 out.append(self.join(search_dir, match))
216
217 return out
218
219 def remove(self, filename):
220 """Delete a file or folder and prune the directory tree."""
221 path = self.to_path(filename)
222 if os.path.isdir(path):
223 shutil.rmtree(path)
224 else:
225 os.remove(path)
226
227 filepath = os.path.split(path)[0]
228 try:
229 os.removedirs(filepath)
230 except OSError:
231 # directory was not empty - good that we're not removing it!
232 pass
233
234 def remove_all(self, location):
235 """
236 Delete all files/directories in the dataset at this location.
237
238 Afterward prunes the directory tree.
239 """
240 for fn in self.list(location):
241 self.remove(fn)
242
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/qcodes/data/io.py b/qcodes/data/io.py
--- a/qcodes/data/io.py
+++ b/qcodes/data/io.py
@@ -141,7 +141,7 @@
location (str): the location string corresponding to this path.
"""
if self.base_location:
- return os.path.relpath(path, self.base_location)
+ return os.path.join(self.base_location, path)
else:
return path
|
{"golden_diff": "diff --git a/qcodes/data/io.py b/qcodes/data/io.py\n--- a/qcodes/data/io.py\n+++ b/qcodes/data/io.py\n@@ -141,7 +141,7 @@\n location (str): the location string corresponding to this path.\n \"\"\"\n if self.base_location:\n- return os.path.relpath(path, self.base_location)\n+ return os.path.join(self.base_location, path)\n else:\n return path\n", "issue": "error when saving to drive other than current path\nThis is due to windows handling of drives. A minimal example:\r\n``` python\r\nimport qcodes,os\r\n\r\ndatadir = r'd:\\Temp'\r\nqcodes.DataSet.default_io = qcodes.DiskIO(datadir)\r\n\r\np=qcodes.Parameter('p', set_cmd=None)\r\nq=qcodes.Parameter('q', set_cmd=None)\r\nds=qcodes.Loop(p[0:10:1]).each(q).run() # fine\r\n \r\nqcodes.DataSet.default_io = qcodes.DiskIO(r'c:\\Temp')\r\n\r\nds=qcodes.Loop(p[0:10:1]).each(p).run() # error\r\n```\r\nThis generates the error `ValueError: path is on mount 'd:', start on mount 'c:'`\r\n\r\nAlso see https://bugs.python.org/issue7195\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nIO managers for QCodes.\n\nIO managers wrap whatever physical storage layer the user wants to use\nin an interface mimicking the built-in <open> context manager, with\nsome restrictions to minimize the overhead in creating new IO managers.\n\nThe main thing these managers need to implement is the open context manager:\n\n- Only the context manager needs to be implemented, not separate\n open function and close methods.\n\n- open takes the standard parameters:\n\n - filename: (string)\n - mode: (string) only 'r' (read), 'w' (write), and 'a' (append) are\n expected to be implemented. As with normal file objects, the only\n difference between write and append is that write empties the file\n before adding new data, and append leaves the existing contents in\n place but starts writing at the end.\n - encoding: If a special output encoding is desired. i.e. 'utf8\n\n- the file-like object returned should implement a minimal set of operations.\n\n In read mode:\n - read([size]): read to the end or at most size bytes into a string\n - readline([size]): read until a newline or up to size bytes, into a string\n - iter(): usually return self, but can be any iterator over lines\n - next(): assuming iter() returns self, this yields the next line.\n\n In write or append mode:\n - write(s): add string s to the end of the file.\n - writelines(seq): add a sequence of strings\n\nIO managers should also implement:\n\n- a join method, ala os.path.join(\\*args).\n- a list method, that returns all objects matching location\n- a remove method, ala os.remove(path) except that it will remove directories\n as well as files, since we're allowing \"locations\" to be directories\n or files.\n\"\"\"\n\nfrom contextlib import contextmanager\nimport os\nimport re\nimport shutil\nfrom fnmatch import fnmatch\n\nALLOWED_OPEN_MODES = ('r', 'w', 'a')\n\n\nclass DiskIO:\n\n \"\"\"\n Simple IO object to wrap disk operations with a custom base location.\n\n Also accepts both forward and backward slashes at any point, and\n normalizes both to the OS we are currently on.\n\n Args:\n base_location (str): a path to the root data folder.\n Converted to an absolute path immediately, so even if you supply a\n relative path, later changes to the OS working directory will not\n affect data paths.\n \"\"\"\n\n def __init__(self, base_location):\n if base_location is None:\n self.base_location = None\n else:\n base_location = self._normalize_slashes(base_location)\n self.base_location = os.path.abspath(base_location)\n\n @contextmanager\n def open(self, filename, mode, encoding=None):\n \"\"\"\n Mimic the interface of the built in open context manager.\n\n Args:\n filename (str): path relative to base_location.\n\n mode (str): 'r' (read), 'w' (write), or 'a' (append).\n Other open modes are not supported because we don't want\n to force all IO managers to support others.\n\n Returns:\n context manager yielding the open file\n \"\"\"\n if mode not in ALLOWED_OPEN_MODES:\n raise ValueError('mode {} not allowed in IO managers'.format(mode))\n\n filepath = self.to_path(filename)\n\n # make directories if needed\n dirpath = os.path.dirname(filepath)\n if not os.path.exists(dirpath):\n os.makedirs(dirpath)\n\n # normally we'd construct this context manager with try/finally, but\n # here we already have a context manager for open so we just wrap it\n with open(filepath, mode, encoding=encoding) as f:\n yield f\n\n def _normalize_slashes(self, location):\n # note that this is NOT os.path.join - the difference is os.path.join\n # discards empty strings, so if you use it on a re.split absolute\n # path you will get a relative path!\n return os.sep.join(re.split('[\\\\\\\\/]', location))\n\n def to_path(self, location):\n \"\"\"\n Convert a location string into a path on the local file system.\n\n For DiskIO this just fixes slashes and prepends the base location,\n doing nothing active with the file. But for other io managers that\n refer to remote storage, this method may actually fetch the file and\n put it at a temporary local path.\n\n Args:\n location (str): A location string for a complete dataset or\n a file within it.\n\n Returns:\n path (str): The path on disk to which this location maps.\n \"\"\"\n location = self._normalize_slashes(location)\n if self.base_location:\n return os.path.join(self.base_location, location)\n else:\n return location\n\n def to_location(self, path):\n \"\"\"\n Convert a local filesystem path into a location string.\n\n Args:\n path (str): a path on the local file system.\n\n Returns:\n location (str): the location string corresponding to this path.\n \"\"\"\n if self.base_location:\n return os.path.relpath(path, self.base_location)\n else:\n return path\n\n def __repr__(self):\n \"\"\"Show the base location in the repr.\"\"\"\n return '<DiskIO, base_location={}>'.format(repr(self.base_location))\n\n def join(self, *args):\n \"\"\"Context-dependent os.path.join for this io manager.\"\"\"\n return os.path.join(*list(map(self._normalize_slashes, args)))\n\n def isfile(self, location):\n \"\"\"Check whether this location matches a file.\"\"\"\n path = self.to_path(location)\n return os.path.isfile(path)\n\n def list(self, location, maxdepth=1, include_dirs=False):\n \"\"\"\n Return all files that match location.\n\n This is either files whose names match up to an arbitrary extension,\n or any files within an exactly matching directory name.\n\n Args:\n location (str): the location to match.\n May contain the usual path wildcards * and ?\n\n maxdepth (int, optional): maximum levels of directory nesting to\n recurse into looking for files. Default 1.\n\n include_dirs (bool, optional): whether to allow directories in\n the results or just files. Default False.\n\n Returns:\n A list of matching files and/or directories, as locations\n relative to our base_location.\n \"\"\"\n location = self._normalize_slashes(location)\n search_dir, pattern = os.path.split(location)\n path = self.to_path(search_dir)\n\n if not os.path.isdir(path):\n return []\n\n matches = [fn for fn in os.listdir(path) if fnmatch(fn, pattern + '*')]\n out = []\n\n for match in matches:\n matchpath = self.join(path, match)\n if os.path.isdir(matchpath) and fnmatch(match, pattern):\n if maxdepth > 0:\n # exact directory match - walk down to maxdepth\n for root, dirs, files in os.walk(matchpath, topdown=True):\n depth = root[len(path):].count(os.path.sep)\n if depth == maxdepth:\n dirs[:] = [] # don't recurse any further\n\n for fn in files + (dirs if include_dirs else []):\n out.append(self.to_location(self.join(root, fn)))\n\n elif include_dirs:\n out.append(self.join(search_dir, match))\n\n elif (os.path.isfile(matchpath) and\n (fnmatch(match, pattern) or\n fnmatch(os.path.splitext(match)[0], pattern))):\n # exact filename match, or match up to an extension\n # note that we need fnmatch(match, pattern) in addition to the\n # splitext test to cover the case of the base filename itself\n # containing a dot.\n out.append(self.join(search_dir, match))\n\n return out\n\n def remove(self, filename):\n \"\"\"Delete a file or folder and prune the directory tree.\"\"\"\n path = self.to_path(filename)\n if os.path.isdir(path):\n shutil.rmtree(path)\n else:\n os.remove(path)\n\n filepath = os.path.split(path)[0]\n try:\n os.removedirs(filepath)\n except OSError:\n # directory was not empty - good that we're not removing it!\n pass\n\n def remove_all(self, location):\n \"\"\"\n Delete all files/directories in the dataset at this location.\n\n Afterward prunes the directory tree.\n \"\"\"\n for fn in self.list(location):\n self.remove(fn)\n", "path": "qcodes/data/io.py"}], "after_files": [{"content": "\"\"\"\nIO managers for QCodes.\n\nIO managers wrap whatever physical storage layer the user wants to use\nin an interface mimicking the built-in <open> context manager, with\nsome restrictions to minimize the overhead in creating new IO managers.\n\nThe main thing these managers need to implement is the open context manager:\n\n- Only the context manager needs to be implemented, not separate\n open function and close methods.\n\n- open takes the standard parameters:\n\n - filename: (string)\n - mode: (string) only 'r' (read), 'w' (write), and 'a' (append) are\n expected to be implemented. As with normal file objects, the only\n difference between write and append is that write empties the file\n before adding new data, and append leaves the existing contents in\n place but starts writing at the end.\n - encoding: If a special output encoding is desired. i.e. 'utf8\n\n- the file-like object returned should implement a minimal set of operations.\n\n In read mode:\n - read([size]): read to the end or at most size bytes into a string\n - readline([size]): read until a newline or up to size bytes, into a string\n - iter(): usually return self, but can be any iterator over lines\n - next(): assuming iter() returns self, this yields the next line.\n\n In write or append mode:\n - write(s): add string s to the end of the file.\n - writelines(seq): add a sequence of strings\n\nIO managers should also implement:\n\n- a join method, ala os.path.join(\\*args).\n- a list method, that returns all objects matching location\n- a remove method, ala os.remove(path) except that it will remove directories\n as well as files, since we're allowing \"locations\" to be directories\n or files.\n\"\"\"\n\nfrom contextlib import contextmanager\nimport os\nimport re\nimport shutil\nfrom fnmatch import fnmatch\n\nALLOWED_OPEN_MODES = ('r', 'w', 'a')\n\n\nclass DiskIO:\n\n \"\"\"\n Simple IO object to wrap disk operations with a custom base location.\n\n Also accepts both forward and backward slashes at any point, and\n normalizes both to the OS we are currently on.\n\n Args:\n base_location (str): a path to the root data folder.\n Converted to an absolute path immediately, so even if you supply a\n relative path, later changes to the OS working directory will not\n affect data paths.\n \"\"\"\n\n def __init__(self, base_location):\n if base_location is None:\n self.base_location = None\n else:\n base_location = self._normalize_slashes(base_location)\n self.base_location = os.path.abspath(base_location)\n\n @contextmanager\n def open(self, filename, mode, encoding=None):\n \"\"\"\n Mimic the interface of the built in open context manager.\n\n Args:\n filename (str): path relative to base_location.\n\n mode (str): 'r' (read), 'w' (write), or 'a' (append).\n Other open modes are not supported because we don't want\n to force all IO managers to support others.\n\n Returns:\n context manager yielding the open file\n \"\"\"\n if mode not in ALLOWED_OPEN_MODES:\n raise ValueError('mode {} not allowed in IO managers'.format(mode))\n\n filepath = self.to_path(filename)\n\n # make directories if needed\n dirpath = os.path.dirname(filepath)\n if not os.path.exists(dirpath):\n os.makedirs(dirpath)\n\n # normally we'd construct this context manager with try/finally, but\n # here we already have a context manager for open so we just wrap it\n with open(filepath, mode, encoding=encoding) as f:\n yield f\n\n def _normalize_slashes(self, location):\n # note that this is NOT os.path.join - the difference is os.path.join\n # discards empty strings, so if you use it on a re.split absolute\n # path you will get a relative path!\n return os.sep.join(re.split('[\\\\\\\\/]', location))\n\n def to_path(self, location):\n \"\"\"\n Convert a location string into a path on the local file system.\n\n For DiskIO this just fixes slashes and prepends the base location,\n doing nothing active with the file. But for other io managers that\n refer to remote storage, this method may actually fetch the file and\n put it at a temporary local path.\n\n Args:\n location (str): A location string for a complete dataset or\n a file within it.\n\n Returns:\n path (str): The path on disk to which this location maps.\n \"\"\"\n location = self._normalize_slashes(location)\n if self.base_location:\n return os.path.join(self.base_location, location)\n else:\n return location\n\n def to_location(self, path):\n \"\"\"\n Convert a local filesystem path into a location string.\n\n Args:\n path (str): a path on the local file system.\n\n Returns:\n location (str): the location string corresponding to this path.\n \"\"\"\n if self.base_location:\n return os.path.join(self.base_location, path)\n else:\n return path\n\n def __repr__(self):\n \"\"\"Show the base location in the repr.\"\"\"\n return '<DiskIO, base_location={}>'.format(repr(self.base_location))\n\n def join(self, *args):\n \"\"\"Context-dependent os.path.join for this io manager.\"\"\"\n return os.path.join(*list(map(self._normalize_slashes, args)))\n\n def isfile(self, location):\n \"\"\"Check whether this location matches a file.\"\"\"\n path = self.to_path(location)\n return os.path.isfile(path)\n\n def list(self, location, maxdepth=1, include_dirs=False):\n \"\"\"\n Return all files that match location.\n\n This is either files whose names match up to an arbitrary extension,\n or any files within an exactly matching directory name.\n\n Args:\n location (str): the location to match.\n May contain the usual path wildcards * and ?\n\n maxdepth (int, optional): maximum levels of directory nesting to\n recurse into looking for files. Default 1.\n\n include_dirs (bool, optional): whether to allow directories in\n the results or just files. Default False.\n\n Returns:\n A list of matching files and/or directories, as locations\n relative to our base_location.\n \"\"\"\n location = self._normalize_slashes(location)\n search_dir, pattern = os.path.split(location)\n path = self.to_path(search_dir)\n\n if not os.path.isdir(path):\n return []\n\n matches = [fn for fn in os.listdir(path) if fnmatch(fn, pattern + '*')]\n out = []\n\n for match in matches:\n matchpath = self.join(path, match)\n if os.path.isdir(matchpath) and fnmatch(match, pattern):\n if maxdepth > 0:\n # exact directory match - walk down to maxdepth\n for root, dirs, files in os.walk(matchpath, topdown=True):\n depth = root[len(path):].count(os.path.sep)\n if depth == maxdepth:\n dirs[:] = [] # don't recurse any further\n\n for fn in files + (dirs if include_dirs else []):\n out.append(self.to_location(self.join(root, fn)))\n\n elif include_dirs:\n out.append(self.join(search_dir, match))\n\n elif (os.path.isfile(matchpath) and\n (fnmatch(match, pattern) or\n fnmatch(os.path.splitext(match)[0], pattern))):\n # exact filename match, or match up to an extension\n # note that we need fnmatch(match, pattern) in addition to the\n # splitext test to cover the case of the base filename itself\n # containing a dot.\n out.append(self.join(search_dir, match))\n\n return out\n\n def remove(self, filename):\n \"\"\"Delete a file or folder and prune the directory tree.\"\"\"\n path = self.to_path(filename)\n if os.path.isdir(path):\n shutil.rmtree(path)\n else:\n os.remove(path)\n\n filepath = os.path.split(path)[0]\n try:\n os.removedirs(filepath)\n except OSError:\n # directory was not empty - good that we're not removing it!\n pass\n\n def remove_all(self, location):\n \"\"\"\n Delete all files/directories in the dataset at this location.\n\n Afterward prunes the directory tree.\n \"\"\"\n for fn in self.list(location):\n self.remove(fn)\n", "path": "qcodes/data/io.py"}]}
| 2,914 | 99 |
gh_patches_debug_12530
|
rasdani/github-patches
|
git_diff
|
opsdroid__opsdroid-1660
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Add a config option to disable registering a route for /
In the web server by default a route is registered for the base / path. This should be configurable in case a user wants to register their own.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `opsdroid/web.py`
Content:
```
1 """Submodule to handle web requests in opsdroid."""
2
3 import json
4 import logging
5 import ssl
6
7 from aiohttp import web
8
9 from opsdroid import __version__
10
11
12 _LOGGER = logging.getLogger(__name__)
13
14
15 class Web:
16 """Create class for opsdroid Web server."""
17
18 def __init__(self, opsdroid):
19 """Create web object."""
20 self.opsdroid = opsdroid
21 try:
22 self.config = self.opsdroid.config["web"]
23 except KeyError:
24 self.config = {}
25 self.base_url = self.config.get("base-url")
26 if not self.base_url:
27 self.base_url = "{proto}://{host}{port}".format(
28 proto="http" if self.get_ssl_context is None else "https",
29 host=self.get_host,
30 port=":{}".format(self.get_port)
31 if self.get_port not in (80, 443)
32 else "",
33 )
34 self.web_app = web.Application()
35 self.runner = web.AppRunner(self.web_app)
36 self.site = None
37 self.web_app.router.add_get("/", self.web_index_handler)
38 self.web_app.router.add_get("", self.web_index_handler)
39 self.web_app.router.add_get("/stats", self.web_stats_handler)
40 self.web_app.router.add_get("/stats/", self.web_stats_handler)
41
42 @property
43 def get_port(self):
44 """Return port from config or the default.
45
46 Args:
47 self: instance method
48
49 Returns:
50 int: returns value of port being used, config or default
51
52 """
53 try:
54 port = self.config["port"]
55 except KeyError:
56 if self.get_ssl_context is not None:
57 port = 8443
58 else:
59 port = 8080
60 return port
61
62 @property
63 def get_host(self):
64 """Return host from config or the default.
65
66 Args:
67 self: instance method
68
69 Returns:
70 string: returns address of host being used, config or default
71
72 """
73 try:
74 host = self.config["host"]
75 except KeyError:
76 host = "0.0.0.0"
77 return host
78
79 @property
80 def get_ssl_context(self):
81 """Return the ssl context or None.
82
83 Args:
84 self: instance method
85
86 Returns:
87 string (or NoneType): returns ssl context of None.
88
89 """
90 try:
91 ssl_config = self.config["ssl"]
92 sslcontext = ssl.SSLContext(ssl.PROTOCOL_TLSv1)
93 sslcontext.load_cert_chain(ssl_config["cert"], ssl_config["key"])
94 return sslcontext
95 except FileNotFoundError:
96 _LOGGER.error(_("Cannot find ssl cert or key."))
97 return None
98 except KeyError:
99 return None
100
101 async def start(self):
102 """Start web servers."""
103 _LOGGER.info(_(f"Started web server on {self.base_url}"))
104 await self.runner.setup()
105 self.site = web.TCPSite(
106 self.runner,
107 host=self.get_host,
108 port=self.get_port,
109 ssl_context=self.get_ssl_context,
110 )
111 await self.site.start()
112
113 async def stop(self):
114 """Stop the web server."""
115 await self.runner.cleanup()
116
117 @staticmethod
118 def build_response(status, result):
119 """Build a json response object to power the bot reponses.
120
121 Args:
122 result: serialize obj as a JSON formated stream
123
124 Returns:
125 json: returns json object with list of responses for the bot
126
127 """
128 return web.Response(text=json.dumps(result), status=status)
129
130 def register_skill(self, opsdroid, skill, webhook):
131 """Register a new skill in the web app router."""
132
133 async def wrapper(req, opsdroid=opsdroid, config=skill.config):
134 """Wrap up the aiohttp handler."""
135 webhook_token = self.config.get("webhook-token", None)
136 authorization_header = []
137 if req is not None:
138 authorization_header = req.headers.get("Authorization", "").split()
139
140 if webhook_token is not None:
141 if not (
142 len(authorization_header) == 2
143 and authorization_header[0] == "Bearer"
144 and authorization_header[1] == webhook_token
145 ):
146 _LOGGER.error(
147 _("Unauthorized to run skill %s via webhook"), webhook
148 )
149 return Web.build_response(403, {"called_skill": webhook})
150
151 _LOGGER.info(_("Running skill %s via webhook."), webhook)
152 opsdroid.stats["webhooks_called"] = opsdroid.stats["webhooks_called"] + 1
153 resp = await opsdroid.run_skill(skill, config, req)
154 if isinstance(resp, web.Response):
155 return resp
156 return Web.build_response(200, {"called_skill": webhook})
157
158 self.web_app.router.add_post(
159 "/skill/{}/{}".format(skill.config["name"], webhook), wrapper
160 )
161 self.web_app.router.add_post(
162 "/skill/{}/{}/".format(skill.config["name"], webhook), wrapper
163 )
164
165 def setup_webhooks(self, skills):
166 """Add the webhooks for the webhook skills to the router."""
167 for skill in skills:
168 for matcher in skill.matchers:
169 if "webhook" in matcher:
170 self.register_skill(self.opsdroid, skill, matcher["webhook"])
171
172 async def web_index_handler(self, request):
173 """Handle root web request to opsdroid API.
174
175 Args:
176 request: web request to the root (index)
177
178 Returns:
179 dict: returns successful status code and greeting for the root page
180
181 """
182 return self.build_response(200, {"message": "Welcome to the opsdroid API"})
183
184 async def web_stats_handler(self, request):
185 """Handle stats request.
186
187 Args:
188 request: web request to render opsdroid stats
189
190 Returns:
191 dict: returns successful status code and dictionary with
192 stats requested
193
194 """
195 stats = self.opsdroid.stats
196 try:
197 stats["average_response_time"] = (
198 stats["total_response_time"] / stats["total_responses"]
199 )
200 except ZeroDivisionError:
201 stats["average_response_time"] = 0
202
203 return self.build_response(
204 200,
205 {
206 "version": __version__,
207 "messages": {
208 "total_parsed": stats["messages_parsed"],
209 "webhooks_called": stats["webhooks_called"],
210 "total_response_time": stats["total_response_time"],
211 "total_responses": stats["total_responses"],
212 "average_response_time": stats["average_response_time"],
213 },
214 "modules": {
215 "skills": len(self.opsdroid.skills),
216 "connectors": len(self.opsdroid.connectors),
217 "databases": len(self.opsdroid.memory.databases),
218 },
219 },
220 )
221
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/opsdroid/web.py b/opsdroid/web.py
--- a/opsdroid/web.py
+++ b/opsdroid/web.py
@@ -34,8 +34,10 @@
self.web_app = web.Application()
self.runner = web.AppRunner(self.web_app)
self.site = None
- self.web_app.router.add_get("/", self.web_index_handler)
- self.web_app.router.add_get("", self.web_index_handler)
+ if not self.config.get("disable_web_index_handler_in_root", False):
+ self.web_app.router.add_get("/", self.web_index_handler)
+ self.web_app.router.add_get("", self.web_index_handler)
+
self.web_app.router.add_get("/stats", self.web_stats_handler)
self.web_app.router.add_get("/stats/", self.web_stats_handler)
|
{"golden_diff": "diff --git a/opsdroid/web.py b/opsdroid/web.py\n--- a/opsdroid/web.py\n+++ b/opsdroid/web.py\n@@ -34,8 +34,10 @@\n self.web_app = web.Application()\n self.runner = web.AppRunner(self.web_app)\n self.site = None\n- self.web_app.router.add_get(\"/\", self.web_index_handler)\n- self.web_app.router.add_get(\"\", self.web_index_handler)\n+ if not self.config.get(\"disable_web_index_handler_in_root\", False):\n+ self.web_app.router.add_get(\"/\", self.web_index_handler)\n+ self.web_app.router.add_get(\"\", self.web_index_handler)\n+\n self.web_app.router.add_get(\"/stats\", self.web_stats_handler)\n self.web_app.router.add_get(\"/stats/\", self.web_stats_handler)\n", "issue": "Add a config option to disable registering a route for /\nIn the web server by default a route is registered for the base / path. This should be configurable in case a user wants to register their own.\n", "before_files": [{"content": "\"\"\"Submodule to handle web requests in opsdroid.\"\"\"\n\nimport json\nimport logging\nimport ssl\n\nfrom aiohttp import web\n\nfrom opsdroid import __version__\n\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass Web:\n \"\"\"Create class for opsdroid Web server.\"\"\"\n\n def __init__(self, opsdroid):\n \"\"\"Create web object.\"\"\"\n self.opsdroid = opsdroid\n try:\n self.config = self.opsdroid.config[\"web\"]\n except KeyError:\n self.config = {}\n self.base_url = self.config.get(\"base-url\")\n if not self.base_url:\n self.base_url = \"{proto}://{host}{port}\".format(\n proto=\"http\" if self.get_ssl_context is None else \"https\",\n host=self.get_host,\n port=\":{}\".format(self.get_port)\n if self.get_port not in (80, 443)\n else \"\",\n )\n self.web_app = web.Application()\n self.runner = web.AppRunner(self.web_app)\n self.site = None\n self.web_app.router.add_get(\"/\", self.web_index_handler)\n self.web_app.router.add_get(\"\", self.web_index_handler)\n self.web_app.router.add_get(\"/stats\", self.web_stats_handler)\n self.web_app.router.add_get(\"/stats/\", self.web_stats_handler)\n\n @property\n def get_port(self):\n \"\"\"Return port from config or the default.\n\n Args:\n self: instance method\n\n Returns:\n int: returns value of port being used, config or default\n\n \"\"\"\n try:\n port = self.config[\"port\"]\n except KeyError:\n if self.get_ssl_context is not None:\n port = 8443\n else:\n port = 8080\n return port\n\n @property\n def get_host(self):\n \"\"\"Return host from config or the default.\n\n Args:\n self: instance method\n\n Returns:\n string: returns address of host being used, config or default\n\n \"\"\"\n try:\n host = self.config[\"host\"]\n except KeyError:\n host = \"0.0.0.0\"\n return host\n\n @property\n def get_ssl_context(self):\n \"\"\"Return the ssl context or None.\n\n Args:\n self: instance method\n\n Returns:\n string (or NoneType): returns ssl context of None.\n\n \"\"\"\n try:\n ssl_config = self.config[\"ssl\"]\n sslcontext = ssl.SSLContext(ssl.PROTOCOL_TLSv1)\n sslcontext.load_cert_chain(ssl_config[\"cert\"], ssl_config[\"key\"])\n return sslcontext\n except FileNotFoundError:\n _LOGGER.error(_(\"Cannot find ssl cert or key.\"))\n return None\n except KeyError:\n return None\n\n async def start(self):\n \"\"\"Start web servers.\"\"\"\n _LOGGER.info(_(f\"Started web server on {self.base_url}\"))\n await self.runner.setup()\n self.site = web.TCPSite(\n self.runner,\n host=self.get_host,\n port=self.get_port,\n ssl_context=self.get_ssl_context,\n )\n await self.site.start()\n\n async def stop(self):\n \"\"\"Stop the web server.\"\"\"\n await self.runner.cleanup()\n\n @staticmethod\n def build_response(status, result):\n \"\"\"Build a json response object to power the bot reponses.\n\n Args:\n result: serialize obj as a JSON formated stream\n\n Returns:\n json: returns json object with list of responses for the bot\n\n \"\"\"\n return web.Response(text=json.dumps(result), status=status)\n\n def register_skill(self, opsdroid, skill, webhook):\n \"\"\"Register a new skill in the web app router.\"\"\"\n\n async def wrapper(req, opsdroid=opsdroid, config=skill.config):\n \"\"\"Wrap up the aiohttp handler.\"\"\"\n webhook_token = self.config.get(\"webhook-token\", None)\n authorization_header = []\n if req is not None:\n authorization_header = req.headers.get(\"Authorization\", \"\").split()\n\n if webhook_token is not None:\n if not (\n len(authorization_header) == 2\n and authorization_header[0] == \"Bearer\"\n and authorization_header[1] == webhook_token\n ):\n _LOGGER.error(\n _(\"Unauthorized to run skill %s via webhook\"), webhook\n )\n return Web.build_response(403, {\"called_skill\": webhook})\n\n _LOGGER.info(_(\"Running skill %s via webhook.\"), webhook)\n opsdroid.stats[\"webhooks_called\"] = opsdroid.stats[\"webhooks_called\"] + 1\n resp = await opsdroid.run_skill(skill, config, req)\n if isinstance(resp, web.Response):\n return resp\n return Web.build_response(200, {\"called_skill\": webhook})\n\n self.web_app.router.add_post(\n \"/skill/{}/{}\".format(skill.config[\"name\"], webhook), wrapper\n )\n self.web_app.router.add_post(\n \"/skill/{}/{}/\".format(skill.config[\"name\"], webhook), wrapper\n )\n\n def setup_webhooks(self, skills):\n \"\"\"Add the webhooks for the webhook skills to the router.\"\"\"\n for skill in skills:\n for matcher in skill.matchers:\n if \"webhook\" in matcher:\n self.register_skill(self.opsdroid, skill, matcher[\"webhook\"])\n\n async def web_index_handler(self, request):\n \"\"\"Handle root web request to opsdroid API.\n\n Args:\n request: web request to the root (index)\n\n Returns:\n dict: returns successful status code and greeting for the root page\n\n \"\"\"\n return self.build_response(200, {\"message\": \"Welcome to the opsdroid API\"})\n\n async def web_stats_handler(self, request):\n \"\"\"Handle stats request.\n\n Args:\n request: web request to render opsdroid stats\n\n Returns:\n dict: returns successful status code and dictionary with\n stats requested\n\n \"\"\"\n stats = self.opsdroid.stats\n try:\n stats[\"average_response_time\"] = (\n stats[\"total_response_time\"] / stats[\"total_responses\"]\n )\n except ZeroDivisionError:\n stats[\"average_response_time\"] = 0\n\n return self.build_response(\n 200,\n {\n \"version\": __version__,\n \"messages\": {\n \"total_parsed\": stats[\"messages_parsed\"],\n \"webhooks_called\": stats[\"webhooks_called\"],\n \"total_response_time\": stats[\"total_response_time\"],\n \"total_responses\": stats[\"total_responses\"],\n \"average_response_time\": stats[\"average_response_time\"],\n },\n \"modules\": {\n \"skills\": len(self.opsdroid.skills),\n \"connectors\": len(self.opsdroid.connectors),\n \"databases\": len(self.opsdroid.memory.databases),\n },\n },\n )\n", "path": "opsdroid/web.py"}], "after_files": [{"content": "\"\"\"Submodule to handle web requests in opsdroid.\"\"\"\n\nimport json\nimport logging\nimport ssl\n\nfrom aiohttp import web\n\nfrom opsdroid import __version__\n\n\n_LOGGER = logging.getLogger(__name__)\n\n\nclass Web:\n \"\"\"Create class for opsdroid Web server.\"\"\"\n\n def __init__(self, opsdroid):\n \"\"\"Create web object.\"\"\"\n self.opsdroid = opsdroid\n try:\n self.config = self.opsdroid.config[\"web\"]\n except KeyError:\n self.config = {}\n self.base_url = self.config.get(\"base-url\")\n if not self.base_url:\n self.base_url = \"{proto}://{host}{port}\".format(\n proto=\"http\" if self.get_ssl_context is None else \"https\",\n host=self.get_host,\n port=\":{}\".format(self.get_port)\n if self.get_port not in (80, 443)\n else \"\",\n )\n self.web_app = web.Application()\n self.runner = web.AppRunner(self.web_app)\n self.site = None\n if not self.config.get(\"disable_web_index_handler_in_root\", False):\n self.web_app.router.add_get(\"/\", self.web_index_handler)\n self.web_app.router.add_get(\"\", self.web_index_handler)\n\n self.web_app.router.add_get(\"/stats\", self.web_stats_handler)\n self.web_app.router.add_get(\"/stats/\", self.web_stats_handler)\n\n @property\n def get_port(self):\n \"\"\"Return port from config or the default.\n\n Args:\n self: instance method\n\n Returns:\n int: returns value of port being used, config or default\n\n \"\"\"\n try:\n port = self.config[\"port\"]\n except KeyError:\n if self.get_ssl_context is not None:\n port = 8443\n else:\n port = 8080\n return port\n\n @property\n def get_host(self):\n \"\"\"Return host from config or the default.\n\n Args:\n self: instance method\n\n Returns:\n string: returns address of host being used, config or default\n\n \"\"\"\n try:\n host = self.config[\"host\"]\n except KeyError:\n host = \"0.0.0.0\"\n return host\n\n @property\n def get_ssl_context(self):\n \"\"\"Return the ssl context or None.\n\n Args:\n self: instance method\n\n Returns:\n string (or NoneType): returns ssl context of None.\n\n \"\"\"\n try:\n ssl_config = self.config[\"ssl\"]\n sslcontext = ssl.SSLContext(ssl.PROTOCOL_TLSv1)\n sslcontext.load_cert_chain(ssl_config[\"cert\"], ssl_config[\"key\"])\n return sslcontext\n except FileNotFoundError:\n _LOGGER.error(_(\"Cannot find ssl cert or key.\"))\n return None\n except KeyError:\n return None\n\n async def start(self):\n \"\"\"Start web servers.\"\"\"\n _LOGGER.info(_(f\"Started web server on {self.base_url}\"))\n await self.runner.setup()\n self.site = web.TCPSite(\n self.runner,\n host=self.get_host,\n port=self.get_port,\n ssl_context=self.get_ssl_context,\n )\n await self.site.start()\n\n async def stop(self):\n \"\"\"Stop the web server.\"\"\"\n await self.runner.cleanup()\n\n @staticmethod\n def build_response(status, result):\n \"\"\"Build a json response object to power the bot reponses.\n\n Args:\n result: serialize obj as a JSON formated stream\n\n Returns:\n json: returns json object with list of responses for the bot\n\n \"\"\"\n return web.Response(text=json.dumps(result), status=status)\n\n def register_skill(self, opsdroid, skill, webhook):\n \"\"\"Register a new skill in the web app router.\"\"\"\n\n async def wrapper(req, opsdroid=opsdroid, config=skill.config):\n \"\"\"Wrap up the aiohttp handler.\"\"\"\n webhook_token = self.config.get(\"webhook-token\", None)\n authorization_header = []\n if req is not None:\n authorization_header = req.headers.get(\"Authorization\", \"\").split()\n\n if webhook_token is not None:\n if not (\n len(authorization_header) == 2\n and authorization_header[0] == \"Bearer\"\n and authorization_header[1] == webhook_token\n ):\n _LOGGER.error(\n _(\"Unauthorized to run skill %s via webhook\"), webhook\n )\n return Web.build_response(403, {\"called_skill\": webhook})\n\n _LOGGER.info(_(\"Running skill %s via webhook.\"), webhook)\n opsdroid.stats[\"webhooks_called\"] = opsdroid.stats[\"webhooks_called\"] + 1\n resp = await opsdroid.run_skill(skill, config, req)\n if isinstance(resp, web.Response):\n return resp\n return Web.build_response(200, {\"called_skill\": webhook})\n\n self.web_app.router.add_post(\n \"/skill/{}/{}\".format(skill.config[\"name\"], webhook), wrapper\n )\n self.web_app.router.add_post(\n \"/skill/{}/{}/\".format(skill.config[\"name\"], webhook), wrapper\n )\n\n def setup_webhooks(self, skills):\n \"\"\"Add the webhooks for the webhook skills to the router.\"\"\"\n for skill in skills:\n for matcher in skill.matchers:\n if \"webhook\" in matcher:\n self.register_skill(self.opsdroid, skill, matcher[\"webhook\"])\n\n async def web_index_handler(self, request):\n \"\"\"Handle root web request to opsdroid API.\n\n Args:\n request: web request to the root (index)\n\n Returns:\n dict: returns successful status code and greeting for the root page\n\n \"\"\"\n return self.build_response(200, {\"message\": \"Welcome to the opsdroid API\"})\n\n async def web_stats_handler(self, request):\n \"\"\"Handle stats request.\n\n Args:\n request: web request to render opsdroid stats\n\n Returns:\n dict: returns successful status code and dictionary with\n stats requested\n\n \"\"\"\n stats = self.opsdroid.stats\n try:\n stats[\"average_response_time\"] = (\n stats[\"total_response_time\"] / stats[\"total_responses\"]\n )\n except ZeroDivisionError:\n stats[\"average_response_time\"] = 0\n\n return self.build_response(\n 200,\n {\n \"version\": __version__,\n \"messages\": {\n \"total_parsed\": stats[\"messages_parsed\"],\n \"webhooks_called\": stats[\"webhooks_called\"],\n \"total_response_time\": stats[\"total_response_time\"],\n \"total_responses\": stats[\"total_responses\"],\n \"average_response_time\": stats[\"average_response_time\"],\n },\n \"modules\": {\n \"skills\": len(self.opsdroid.skills),\n \"connectors\": len(self.opsdroid.connectors),\n \"databases\": len(self.opsdroid.memory.databases),\n },\n },\n )\n", "path": "opsdroid/web.py"}]}
| 2,328 | 178 |
gh_patches_debug_11556
|
rasdani/github-patches
|
git_diff
|
zestedesavoir__zds-site-6174
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Doublon pour les logs en couleur dans l'env de dev
Sur mon environnement de dev, les logs en couleur apparaissent en double : dans leur version couleur et puis dans leur version standard.

OS : Ubuntu Budgie 20.04
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `zds/settings/dev.py`
Content:
```
1 from colorlog import ColoredFormatter
2
3 from .abstract_base import *
4
5 DEBUG = True
6
7 # NOTE: Can be removed once Django 3 is used
8 ALLOWED_HOSTS = [".localhost", "127.0.0.1", "[::1]"]
9
10 INSTALLED_APPS += (
11 "debug_toolbar",
12 "django_extensions",
13 )
14
15 MIDDLEWARE = (
16 "debug_toolbar.middleware.DebugToolbarMiddleware",
17 "zds.middlewares.nocacheindevmiddleware.NoCacheInDevMiddleware",
18 ) + MIDDLEWARE
19
20 LOGGING = {
21 "version": 1,
22 "disable_existing_loggers": False,
23 "formatters": {
24 "verbose": {
25 "()": ColoredFormatter,
26 "format": "%(log_color)s %(levelname)s %(reset)s %(bold_black)s%(name)s%(reset)s %(message)s",
27 "log_colors": {
28 "DEBUG": "fg_white,bg_black",
29 "INFO": "fg_black,bg_bold_white",
30 "WARNING": "fg_black,bg_bold_yellow",
31 "ERROR": "fg_bold_white,bg_bold_red",
32 "CRITICAL": "fg_bold_white,bg_bold_red",
33 },
34 },
35 "django.server": {
36 "()": ColoredFormatter,
37 "format": "%(log_color)s%(message)s",
38 "log_colors": {
39 "INFO": "bold_black",
40 "WARNING": "bold_yellow",
41 "ERROR": "bold_red",
42 "CRITICAL": "bold_red",
43 },
44 },
45 },
46 "handlers": {
47 "console": {
48 "level": "DEBUG",
49 "class": "logging.StreamHandler",
50 "formatter": "verbose",
51 },
52 "django.server": {
53 "level": "DEBUG",
54 "class": "logging.StreamHandler",
55 "formatter": "django.server",
56 },
57 },
58 "loggers": {
59 "django": {
60 "level": "INFO",
61 "handlers": ["console"],
62 },
63 "django.server": {
64 "level": "INFO",
65 "handlers": ["django.server"],
66 "propagate": False,
67 },
68 "zds": {
69 "level": "INFO",
70 "handlers": ["console"],
71 },
72 },
73 }
74
75 ZDS_APP["site"]["url"] = "http://127.0.0.1:8000"
76 ZDS_APP["site"]["dns"] = "127.0.0.1:8000"
77
78 ZDS_APP["very_top_banner"] = {
79 "background_color": "#666",
80 "border_color": "#353535",
81 "color": "white",
82 "message": "Version locale",
83 "slug": "version-locale",
84 }
85
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/zds/settings/dev.py b/zds/settings/dev.py
--- a/zds/settings/dev.py
+++ b/zds/settings/dev.py
@@ -55,10 +55,12 @@
"formatter": "django.server",
},
},
+ "root": {
+ "handlers": ["console"],
+ },
"loggers": {
"django": {
"level": "INFO",
- "handlers": ["console"],
},
"django.server": {
"level": "INFO",
@@ -67,7 +69,6 @@
},
"zds": {
"level": "INFO",
- "handlers": ["console"],
},
},
}
|
{"golden_diff": "diff --git a/zds/settings/dev.py b/zds/settings/dev.py\n--- a/zds/settings/dev.py\n+++ b/zds/settings/dev.py\n@@ -55,10 +55,12 @@\n \"formatter\": \"django.server\",\n },\n },\n+ \"root\": {\n+ \"handlers\": [\"console\"],\n+ },\n \"loggers\": {\n \"django\": {\n \"level\": \"INFO\",\n- \"handlers\": [\"console\"],\n },\n \"django.server\": {\n \"level\": \"INFO\",\n@@ -67,7 +69,6 @@\n },\n \"zds\": {\n \"level\": \"INFO\",\n- \"handlers\": [\"console\"],\n },\n },\n }\n", "issue": "Doublon pour les logs en couleur dans l'env de dev\nSur mon environnement de dev, les logs en couleur apparaissent en double : dans leur version couleur et puis dans leur version standard.\r\n\r\n\r\n\r\nOS : Ubuntu Budgie 20.04\n", "before_files": [{"content": "from colorlog import ColoredFormatter\n\nfrom .abstract_base import *\n\nDEBUG = True\n\n# NOTE: Can be removed once Django 3 is used\nALLOWED_HOSTS = [\".localhost\", \"127.0.0.1\", \"[::1]\"]\n\nINSTALLED_APPS += (\n \"debug_toolbar\",\n \"django_extensions\",\n)\n\nMIDDLEWARE = (\n \"debug_toolbar.middleware.DebugToolbarMiddleware\",\n \"zds.middlewares.nocacheindevmiddleware.NoCacheInDevMiddleware\",\n) + MIDDLEWARE\n\nLOGGING = {\n \"version\": 1,\n \"disable_existing_loggers\": False,\n \"formatters\": {\n \"verbose\": {\n \"()\": ColoredFormatter,\n \"format\": \"%(log_color)s %(levelname)s %(reset)s %(bold_black)s%(name)s%(reset)s %(message)s\",\n \"log_colors\": {\n \"DEBUG\": \"fg_white,bg_black\",\n \"INFO\": \"fg_black,bg_bold_white\",\n \"WARNING\": \"fg_black,bg_bold_yellow\",\n \"ERROR\": \"fg_bold_white,bg_bold_red\",\n \"CRITICAL\": \"fg_bold_white,bg_bold_red\",\n },\n },\n \"django.server\": {\n \"()\": ColoredFormatter,\n \"format\": \"%(log_color)s%(message)s\",\n \"log_colors\": {\n \"INFO\": \"bold_black\",\n \"WARNING\": \"bold_yellow\",\n \"ERROR\": \"bold_red\",\n \"CRITICAL\": \"bold_red\",\n },\n },\n },\n \"handlers\": {\n \"console\": {\n \"level\": \"DEBUG\",\n \"class\": \"logging.StreamHandler\",\n \"formatter\": \"verbose\",\n },\n \"django.server\": {\n \"level\": \"DEBUG\",\n \"class\": \"logging.StreamHandler\",\n \"formatter\": \"django.server\",\n },\n },\n \"loggers\": {\n \"django\": {\n \"level\": \"INFO\",\n \"handlers\": [\"console\"],\n },\n \"django.server\": {\n \"level\": \"INFO\",\n \"handlers\": [\"django.server\"],\n \"propagate\": False,\n },\n \"zds\": {\n \"level\": \"INFO\",\n \"handlers\": [\"console\"],\n },\n },\n}\n\nZDS_APP[\"site\"][\"url\"] = \"http://127.0.0.1:8000\"\nZDS_APP[\"site\"][\"dns\"] = \"127.0.0.1:8000\"\n\nZDS_APP[\"very_top_banner\"] = {\n \"background_color\": \"#666\",\n \"border_color\": \"#353535\",\n \"color\": \"white\",\n \"message\": \"Version locale\",\n \"slug\": \"version-locale\",\n}\n", "path": "zds/settings/dev.py"}], "after_files": [{"content": "from colorlog import ColoredFormatter\n\nfrom .abstract_base import *\n\nDEBUG = True\n\n# NOTE: Can be removed once Django 3 is used\nALLOWED_HOSTS = [\".localhost\", \"127.0.0.1\", \"[::1]\"]\n\nINSTALLED_APPS += (\n \"debug_toolbar\",\n \"django_extensions\",\n)\n\nMIDDLEWARE = (\n \"debug_toolbar.middleware.DebugToolbarMiddleware\",\n \"zds.middlewares.nocacheindevmiddleware.NoCacheInDevMiddleware\",\n) + MIDDLEWARE\n\nLOGGING = {\n \"version\": 1,\n \"disable_existing_loggers\": False,\n \"formatters\": {\n \"verbose\": {\n \"()\": ColoredFormatter,\n \"format\": \"%(log_color)s %(levelname)s %(reset)s %(bold_black)s%(name)s%(reset)s %(message)s\",\n \"log_colors\": {\n \"DEBUG\": \"fg_white,bg_black\",\n \"INFO\": \"fg_black,bg_bold_white\",\n \"WARNING\": \"fg_black,bg_bold_yellow\",\n \"ERROR\": \"fg_bold_white,bg_bold_red\",\n \"CRITICAL\": \"fg_bold_white,bg_bold_red\",\n },\n },\n \"django.server\": {\n \"()\": ColoredFormatter,\n \"format\": \"%(log_color)s%(message)s\",\n \"log_colors\": {\n \"INFO\": \"bold_black\",\n \"WARNING\": \"bold_yellow\",\n \"ERROR\": \"bold_red\",\n \"CRITICAL\": \"bold_red\",\n },\n },\n },\n \"handlers\": {\n \"console\": {\n \"level\": \"DEBUG\",\n \"class\": \"logging.StreamHandler\",\n \"formatter\": \"verbose\",\n },\n \"django.server\": {\n \"level\": \"DEBUG\",\n \"class\": \"logging.StreamHandler\",\n \"formatter\": \"django.server\",\n },\n },\n \"root\": {\n \"handlers\": [\"console\"],\n },\n \"loggers\": {\n \"django\": {\n \"level\": \"INFO\",\n },\n \"django.server\": {\n \"level\": \"INFO\",\n \"handlers\": [\"django.server\"],\n \"propagate\": False,\n },\n \"zds\": {\n \"level\": \"INFO\",\n },\n },\n}\n\nZDS_APP[\"site\"][\"url\"] = \"http://127.0.0.1:8000\"\nZDS_APP[\"site\"][\"dns\"] = \"127.0.0.1:8000\"\n\nZDS_APP[\"very_top_banner\"] = {\n \"background_color\": \"#666\",\n \"border_color\": \"#353535\",\n \"color\": \"white\",\n \"message\": \"Version locale\",\n \"slug\": \"version-locale\",\n}\n", "path": "zds/settings/dev.py"}]}
| 1,123 | 156 |
gh_patches_debug_242
|
rasdani/github-patches
|
git_diff
|
sanic-org__sanic-1559
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
2 failed tests when tox is not used (missing fixture "benchmark")
`pytest-benchmark` is not present in `tests_require`, so there are 2 failed tests in `tests/benchmark/test_route_resolution_benchmark.py` when tox is not used.
This requirement is present in `tox.ini` so tox and Travis CI are working fine.
(I don't know what's a better fix — disable the benchmark tests or add `pytest-benchmark` to `tests_require`, so I didn't create a PR)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 """
2 Sanic
3 """
4 import codecs
5 import os
6 import re
7 import sys
8 from distutils.util import strtobool
9
10 from setuptools import setup
11 from setuptools.command.test import test as TestCommand
12
13
14 class PyTest(TestCommand):
15 """
16 Provide a Test runner to be used from setup.py to run unit tests
17 """
18
19 user_options = [("pytest-args=", "a", "Arguments to pass to pytest")]
20
21 def initialize_options(self):
22 TestCommand.initialize_options(self)
23 self.pytest_args = ""
24
25 def run_tests(self):
26 import shlex
27 import pytest
28
29 errno = pytest.main(shlex.split(self.pytest_args))
30 sys.exit(errno)
31
32
33 def open_local(paths, mode="r", encoding="utf8"):
34 path = os.path.join(os.path.abspath(os.path.dirname(__file__)), *paths)
35
36 return codecs.open(path, mode, encoding)
37
38
39 with open_local(["sanic", "__init__.py"], encoding="latin1") as fp:
40 try:
41 version = re.findall(
42 r"^__version__ = \"([^']+)\"\r?$", fp.read(), re.M
43 )[0]
44 except IndexError:
45 raise RuntimeError("Unable to determine version.")
46
47 with open_local(["README.rst"]) as rm:
48 long_description = rm.read()
49
50 setup_kwargs = {
51 "name": "sanic",
52 "version": version,
53 "url": "http://github.com/channelcat/sanic/",
54 "license": "MIT",
55 "author": "Channel Cat",
56 "author_email": "[email protected]",
57 "description": (
58 "A microframework based on uvloop, httptools, and learnings of flask"
59 ),
60 "long_description": long_description,
61 "packages": ["sanic"],
62 "platforms": "any",
63 "classifiers": [
64 "Development Status :: 4 - Beta",
65 "Environment :: Web Environment",
66 "License :: OSI Approved :: MIT License",
67 "Programming Language :: Python :: 3.5",
68 "Programming Language :: Python :: 3.6",
69 "Programming Language :: Python :: 3.7",
70 ],
71 }
72
73 env_dependency = (
74 '; sys_platform != "win32" ' 'and implementation_name == "cpython"'
75 )
76 ujson = "ujson>=1.35" + env_dependency
77 uvloop = "uvloop>=0.5.3" + env_dependency
78
79 requirements = [
80 "httptools>=0.0.10",
81 uvloop,
82 ujson,
83 "aiofiles>=0.3.0",
84 "websockets>=6.0,<7.0",
85 "multidict>=4.0,<5.0",
86 ]
87
88 tests_require = [
89 "pytest==4.1.0",
90 "multidict>=4.0,<5.0",
91 "gunicorn",
92 "pytest-cov",
93 "aiohttp>=2.3.0,<=3.2.1",
94 "beautifulsoup4",
95 uvloop,
96 ujson,
97 "pytest-sanic",
98 "pytest-sugar",
99 ]
100
101 if strtobool(os.environ.get("SANIC_NO_UJSON", "no")):
102 print("Installing without uJSON")
103 requirements.remove(ujson)
104 tests_require.remove(ujson)
105
106 # 'nt' means windows OS
107 if strtobool(os.environ.get("SANIC_NO_UVLOOP", "no")):
108 print("Installing without uvLoop")
109 requirements.remove(uvloop)
110 tests_require.remove(uvloop)
111
112 extras_require = {
113 "test": tests_require,
114 "dev": tests_require + ["aiofiles", "tox", "black", "flake8"],
115 "docs": [
116 "sphinx",
117 "sphinx_rtd_theme",
118 "recommonmark",
119 "sphinxcontrib-asyncio",
120 "docutils",
121 "pygments"
122 ],
123 }
124
125 setup_kwargs["install_requires"] = requirements
126 setup_kwargs["tests_require"] = tests_require
127 setup_kwargs["extras_require"] = extras_require
128 setup_kwargs["cmdclass"] = {"test": PyTest}
129 setup(**setup_kwargs)
130
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -96,6 +96,7 @@
ujson,
"pytest-sanic",
"pytest-sugar",
+ "pytest-benchmark",
]
if strtobool(os.environ.get("SANIC_NO_UJSON", "no")):
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -96,6 +96,7 @@\n ujson,\n \"pytest-sanic\",\n \"pytest-sugar\",\n+ \"pytest-benchmark\",\n ]\n \n if strtobool(os.environ.get(\"SANIC_NO_UJSON\", \"no\")):\n", "issue": "2 failed tests when tox is not used (missing fixture \"benchmark\")\n`pytest-benchmark` is not present in `tests_require`, so there are 2 failed tests in `tests/benchmark/test_route_resolution_benchmark.py` when tox is not used.\r\n\r\nThis requirement is present in `tox.ini` so tox and Travis CI are working fine.\r\n\r\n(I don't know what's a better fix \u2014 disable the benchmark tests or add `pytest-benchmark` to `tests_require`, so I didn't create a PR)\n", "before_files": [{"content": "\"\"\"\nSanic\n\"\"\"\nimport codecs\nimport os\nimport re\nimport sys\nfrom distutils.util import strtobool\n\nfrom setuptools import setup\nfrom setuptools.command.test import test as TestCommand\n\n\nclass PyTest(TestCommand):\n \"\"\"\n Provide a Test runner to be used from setup.py to run unit tests\n \"\"\"\n\n user_options = [(\"pytest-args=\", \"a\", \"Arguments to pass to pytest\")]\n\n def initialize_options(self):\n TestCommand.initialize_options(self)\n self.pytest_args = \"\"\n\n def run_tests(self):\n import shlex\n import pytest\n\n errno = pytest.main(shlex.split(self.pytest_args))\n sys.exit(errno)\n\n\ndef open_local(paths, mode=\"r\", encoding=\"utf8\"):\n path = os.path.join(os.path.abspath(os.path.dirname(__file__)), *paths)\n\n return codecs.open(path, mode, encoding)\n\n\nwith open_local([\"sanic\", \"__init__.py\"], encoding=\"latin1\") as fp:\n try:\n version = re.findall(\n r\"^__version__ = \\\"([^']+)\\\"\\r?$\", fp.read(), re.M\n )[0]\n except IndexError:\n raise RuntimeError(\"Unable to determine version.\")\n\nwith open_local([\"README.rst\"]) as rm:\n long_description = rm.read()\n\nsetup_kwargs = {\n \"name\": \"sanic\",\n \"version\": version,\n \"url\": \"http://github.com/channelcat/sanic/\",\n \"license\": \"MIT\",\n \"author\": \"Channel Cat\",\n \"author_email\": \"[email protected]\",\n \"description\": (\n \"A microframework based on uvloop, httptools, and learnings of flask\"\n ),\n \"long_description\": long_description,\n \"packages\": [\"sanic\"],\n \"platforms\": \"any\",\n \"classifiers\": [\n \"Development Status :: 4 - Beta\",\n \"Environment :: Web Environment\",\n \"License :: OSI Approved :: MIT License\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n ],\n}\n\nenv_dependency = (\n '; sys_platform != \"win32\" ' 'and implementation_name == \"cpython\"'\n)\nujson = \"ujson>=1.35\" + env_dependency\nuvloop = \"uvloop>=0.5.3\" + env_dependency\n\nrequirements = [\n \"httptools>=0.0.10\",\n uvloop,\n ujson,\n \"aiofiles>=0.3.0\",\n \"websockets>=6.0,<7.0\",\n \"multidict>=4.0,<5.0\",\n]\n\ntests_require = [\n \"pytest==4.1.0\",\n \"multidict>=4.0,<5.0\",\n \"gunicorn\",\n \"pytest-cov\",\n \"aiohttp>=2.3.0,<=3.2.1\",\n \"beautifulsoup4\",\n uvloop,\n ujson,\n \"pytest-sanic\",\n \"pytest-sugar\",\n]\n\nif strtobool(os.environ.get(\"SANIC_NO_UJSON\", \"no\")):\n print(\"Installing without uJSON\")\n requirements.remove(ujson)\n tests_require.remove(ujson)\n\n# 'nt' means windows OS\nif strtobool(os.environ.get(\"SANIC_NO_UVLOOP\", \"no\")):\n print(\"Installing without uvLoop\")\n requirements.remove(uvloop)\n tests_require.remove(uvloop)\n\nextras_require = {\n \"test\": tests_require,\n \"dev\": tests_require + [\"aiofiles\", \"tox\", \"black\", \"flake8\"],\n \"docs\": [\n \"sphinx\",\n \"sphinx_rtd_theme\",\n \"recommonmark\",\n \"sphinxcontrib-asyncio\",\n \"docutils\",\n \"pygments\"\n ],\n}\n\nsetup_kwargs[\"install_requires\"] = requirements\nsetup_kwargs[\"tests_require\"] = tests_require\nsetup_kwargs[\"extras_require\"] = extras_require\nsetup_kwargs[\"cmdclass\"] = {\"test\": PyTest}\nsetup(**setup_kwargs)\n", "path": "setup.py"}], "after_files": [{"content": "\"\"\"\nSanic\n\"\"\"\nimport codecs\nimport os\nimport re\nimport sys\nfrom distutils.util import strtobool\n\nfrom setuptools import setup\nfrom setuptools.command.test import test as TestCommand\n\n\nclass PyTest(TestCommand):\n \"\"\"\n Provide a Test runner to be used from setup.py to run unit tests\n \"\"\"\n\n user_options = [(\"pytest-args=\", \"a\", \"Arguments to pass to pytest\")]\n\n def initialize_options(self):\n TestCommand.initialize_options(self)\n self.pytest_args = \"\"\n\n def run_tests(self):\n import shlex\n import pytest\n\n errno = pytest.main(shlex.split(self.pytest_args))\n sys.exit(errno)\n\n\ndef open_local(paths, mode=\"r\", encoding=\"utf8\"):\n path = os.path.join(os.path.abspath(os.path.dirname(__file__)), *paths)\n\n return codecs.open(path, mode, encoding)\n\n\nwith open_local([\"sanic\", \"__init__.py\"], encoding=\"latin1\") as fp:\n try:\n version = re.findall(\n r\"^__version__ = \\\"([^']+)\\\"\\r?$\", fp.read(), re.M\n )[0]\n except IndexError:\n raise RuntimeError(\"Unable to determine version.\")\n\nwith open_local([\"README.rst\"]) as rm:\n long_description = rm.read()\n\nsetup_kwargs = {\n \"name\": \"sanic\",\n \"version\": version,\n \"url\": \"http://github.com/channelcat/sanic/\",\n \"license\": \"MIT\",\n \"author\": \"Channel Cat\",\n \"author_email\": \"[email protected]\",\n \"description\": (\n \"A microframework based on uvloop, httptools, and learnings of flask\"\n ),\n \"long_description\": long_description,\n \"packages\": [\"sanic\"],\n \"platforms\": \"any\",\n \"classifiers\": [\n \"Development Status :: 4 - Beta\",\n \"Environment :: Web Environment\",\n \"License :: OSI Approved :: MIT License\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n ],\n}\n\nenv_dependency = (\n '; sys_platform != \"win32\" ' 'and implementation_name == \"cpython\"'\n)\nujson = \"ujson>=1.35\" + env_dependency\nuvloop = \"uvloop>=0.5.3\" + env_dependency\n\nrequirements = [\n \"httptools>=0.0.10\",\n uvloop,\n ujson,\n \"aiofiles>=0.3.0\",\n \"websockets>=6.0,<7.0\",\n \"multidict>=4.0,<5.0\",\n]\n\ntests_require = [\n \"pytest==4.1.0\",\n \"multidict>=4.0,<5.0\",\n \"gunicorn\",\n \"pytest-cov\",\n \"aiohttp>=2.3.0,<=3.2.1\",\n \"beautifulsoup4\",\n uvloop,\n ujson,\n \"pytest-sanic\",\n \"pytest-sugar\",\n \"pytest-benchmark\",\n]\n\nif strtobool(os.environ.get(\"SANIC_NO_UJSON\", \"no\")):\n print(\"Installing without uJSON\")\n requirements.remove(ujson)\n tests_require.remove(ujson)\n\n# 'nt' means windows OS\nif strtobool(os.environ.get(\"SANIC_NO_UVLOOP\", \"no\")):\n print(\"Installing without uvLoop\")\n requirements.remove(uvloop)\n tests_require.remove(uvloop)\n\nextras_require = {\n \"test\": tests_require,\n \"dev\": tests_require + [\"aiofiles\", \"tox\", \"black\", \"flake8\"],\n \"docs\": [\n \"sphinx\",\n \"sphinx_rtd_theme\",\n \"recommonmark\",\n \"sphinxcontrib-asyncio\",\n \"docutils\",\n \"pygments\"\n ],\n}\n\nsetup_kwargs[\"install_requires\"] = requirements\nsetup_kwargs[\"tests_require\"] = tests_require\nsetup_kwargs[\"extras_require\"] = extras_require\nsetup_kwargs[\"cmdclass\"] = {\"test\": PyTest}\nsetup(**setup_kwargs)\n", "path": "setup.py"}]}
| 1,535 | 75 |
gh_patches_debug_38426
|
rasdani/github-patches
|
git_diff
|
hylang__hy-2453
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Remove the special mangling rule for question marks
This was one mangling detail that I disagreed with @gilch about years ago, but I belatedly think he was right. The rule that `foo?` mangles to `is_foo` (instead of `hyx_fooXquestion_markX`) should probably go. What's convinced me is wanting to use question marks on some names and then deciding not to because of mangling:
- In [Infinitesimal Quest 2 + ε](https://github.com/hylang/simalq), items have a Boolean attribute `destroy-on-pickup` that indicate whether the item is destroyed when it's picked up. I could call it `destroy-on-pickup?`, but that would mangle to the nonsensical `is_destroy_on_pickup`.
- Likewise, `attack?`, meaning "should this monster attack?", would be misleadingly mangled to `is_attack`.
Another concern is that Hy's core macro `is-not` can be written `not?`, which looks like a unary predicate for some sort of object called a "not", instead of a binary function that indicates whether `a` is not `b`.
One motivation for the question-mark rule was that it could provide a Lispy alias for a Python name: if a library has a function `is_foo`, you can call it as `foo`. But this purpose was always undermined by [Python core's inconsistency between `isfoo` and `is_foo`](https://github.com/python/cpython/issues/73274).
Finally, there's something to be said for making mangling a little simpler, since it's pretty complicated.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `hy/reader/mangling.py`
Content:
```
1 import re
2 import unicodedata
3
4 MANGLE_DELIM = "X"
5
6 normalizes_to_underscore = "_︳︴﹍﹎﹏_"
7
8
9 def mangle(s):
10 """Stringify the argument (with :class:`str`, not :func:`repr` or
11 :hy:func:`hy.repr`) and convert it to a valid Python identifier according
12 to :ref:`Hy's mangling rules <mangling>`. ::
13
14 (hy.mangle 'foo-bar?) ; => "is_foo_bar"
15 (hy.mangle "🦑") ; => "hyx_squid"
16
17 If the stringified argument is already both legal as a Python identifier
18 and normalized according to Unicode normalization form KC (NFKC), it will
19 be returned unchanged. Thus, ``hy.mangle`` is idempotent. ::
20
21 (setv x '♦-->♠)
22 (= (hy.mangle (hy.mangle x)) (hy.mangle x)) ; => True
23
24 Generally, the stringifed input is expected to be parsable as a symbol. As
25 a convenience, it can also have the syntax of a :ref:`dotted identifier
26 <dotted-identifiers>`, and ``hy.mangle`` will mangle the dot-delimited
27 parts separately. ::
28
29 (hy.mangle "a.b?.c!.d") ; => "a.is_b.hyx_cXexclamation_markX.d"
30 """
31
32 assert s
33 s = str(s)
34
35 if "." in s and s.strip("."):
36 return ".".join(mangle(x) if x else "" for x in s.split("."))
37
38 # Step 1: Remove and save leading underscores
39 s2 = s.lstrip(normalizes_to_underscore)
40 leading_underscores = "_" * (len(s) - len(s2))
41 s = s2
42
43 # Step 2: Convert hyphens without introducing a new leading underscore
44 s = s[0] + s[1:].replace("-", "_") if s else s
45
46 # Step 3: Convert trailing `?` to leading `is_`
47 if s.endswith("?"):
48 s = "is_" + s[:-1]
49
50 # Step 4: Convert invalid characters or reserved words
51 if not (leading_underscores + s).isidentifier():
52 # Replace illegal characters with their Unicode character
53 # names, or hexadecimal if they don't have one.
54 s = "hyx_" + "".join(
55 c if c != MANGLE_DELIM and ("S" + c).isidentifier()
56 # We prepend the "S" because some characters aren't
57 # allowed at the start of an identifier.
58 else "{0}{1}{0}".format(
59 MANGLE_DELIM,
60 unicodedata.name(c, "").lower().replace("-", "H").replace(" ", "_")
61 or "U{:x}".format(ord(c)),
62 )
63 for c in s
64 )
65
66 # Step 5: Add back leading underscores
67 s = leading_underscores + s
68
69 # Normalize Unicode per PEP 3131.
70 s = unicodedata.normalize("NFKC", s)
71
72 assert s.isidentifier()
73 return s
74
75
76 def unmangle(s):
77 """Stringify the argument and try to convert it to a pretty unmangled
78 form. See :ref:`Hy's mangling rules <mangling>`.
79
80 Unmangling may not round-trip, because different Hy symbol names can mangle
81 to the same Python identifier. In particular, Python itself already
82 considers distinct strings that have the same normalized form (according to
83 NFKC), such as ``hello`` and ``𝔥𝔢𝔩𝔩𝔬``, to be the same identifier.
84
85 Examples:
86 ::
87
88 => (hy.unmangle 'foo_bar)
89 "foo-bar"
90
91 => (hy.unmangle 'is_foo_bar)
92 "foo-bar?"
93
94 => (hy.unmangle 'hyx_XasteriskX)
95 "*"
96
97 => (hy.unmangle '_hyx_is_fooXsolidusXa)
98 "_foo/a?"
99
100 => (hy.unmangle 'hyx_XhyphenHminusX_XgreaterHthan_signX)
101 "-->"
102
103 => (hy.unmangle 'hyx_XlessHthan_signX__)
104 "<--"
105
106 => (hy.unmangle '__dunder_name__)
107 "__dunder-name__"
108
109 """
110
111 s = str(s)
112
113 prefix = ""
114 suffix = ""
115 m = re.fullmatch(r"(_+)(.*?)(_*)", s, re.DOTALL)
116 if m:
117 prefix, s, suffix = m.groups()
118
119 if s.startswith("hyx_"):
120 s = re.sub(
121 "{0}(U)?([_a-z0-9H]+?){0}".format(MANGLE_DELIM),
122 lambda mo: chr(int(mo.group(2), base=16))
123 if mo.group(1)
124 else unicodedata.lookup(
125 mo.group(2).replace("_", " ").replace("H", "-").upper()
126 ),
127 s[len("hyx_") :],
128 )
129 if s.startswith("is_"):
130 s = s[len("is_") :] + "?"
131 s = s.replace("_", "-")
132
133 return prefix + s + suffix
134
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/hy/reader/mangling.py b/hy/reader/mangling.py
--- a/hy/reader/mangling.py
+++ b/hy/reader/mangling.py
@@ -11,7 +11,7 @@
:hy:func:`hy.repr`) and convert it to a valid Python identifier according
to :ref:`Hy's mangling rules <mangling>`. ::
- (hy.mangle 'foo-bar?) ; => "is_foo_bar"
+ (hy.mangle 'foo-bar) ; => "foo_bar"
(hy.mangle "🦑") ; => "hyx_squid"
If the stringified argument is already both legal as a Python identifier
@@ -26,7 +26,7 @@
<dotted-identifiers>`, and ``hy.mangle`` will mangle the dot-delimited
parts separately. ::
- (hy.mangle "a.b?.c!.d") ; => "a.is_b.hyx_cXexclamation_markX.d"
+ (hy.mangle "a.c!.d") ; => "a.hyx_cXexclamation_markX.d"
"""
assert s
@@ -35,19 +35,15 @@
if "." in s and s.strip("."):
return ".".join(mangle(x) if x else "" for x in s.split("."))
- # Step 1: Remove and save leading underscores
+ # Remove and save leading underscores
s2 = s.lstrip(normalizes_to_underscore)
leading_underscores = "_" * (len(s) - len(s2))
s = s2
- # Step 2: Convert hyphens without introducing a new leading underscore
+ # Convert hyphens without introducing a new leading underscore
s = s[0] + s[1:].replace("-", "_") if s else s
- # Step 3: Convert trailing `?` to leading `is_`
- if s.endswith("?"):
- s = "is_" + s[:-1]
-
- # Step 4: Convert invalid characters or reserved words
+ # Convert invalid characters or reserved words
if not (leading_underscores + s).isidentifier():
# Replace illegal characters with their Unicode character
# names, or hexadecimal if they don't have one.
@@ -88,15 +84,9 @@
=> (hy.unmangle 'foo_bar)
"foo-bar"
- => (hy.unmangle 'is_foo_bar)
- "foo-bar?"
-
=> (hy.unmangle 'hyx_XasteriskX)
"*"
- => (hy.unmangle '_hyx_is_fooXsolidusXa)
- "_foo/a?"
-
=> (hy.unmangle 'hyx_XhyphenHminusX_XgreaterHthan_signX)
"-->"
@@ -126,8 +116,6 @@
),
s[len("hyx_") :],
)
- if s.startswith("is_"):
- s = s[len("is_") :] + "?"
s = s.replace("_", "-")
return prefix + s + suffix
|
{"golden_diff": "diff --git a/hy/reader/mangling.py b/hy/reader/mangling.py\n--- a/hy/reader/mangling.py\n+++ b/hy/reader/mangling.py\n@@ -11,7 +11,7 @@\n :hy:func:`hy.repr`) and convert it to a valid Python identifier according\n to :ref:`Hy's mangling rules <mangling>`. ::\n \n- (hy.mangle 'foo-bar?) ; => \"is_foo_bar\"\n+ (hy.mangle 'foo-bar) ; => \"foo_bar\"\n (hy.mangle \"\ud83e\udd91\") ; => \"hyx_squid\"\n \n If the stringified argument is already both legal as a Python identifier\n@@ -26,7 +26,7 @@\n <dotted-identifiers>`, and ``hy.mangle`` will mangle the dot-delimited\n parts separately. ::\n \n- (hy.mangle \"a.b?.c!.d\") ; => \"a.is_b.hyx_cXexclamation_markX.d\"\n+ (hy.mangle \"a.c!.d\") ; => \"a.hyx_cXexclamation_markX.d\"\n \"\"\"\n \n assert s\n@@ -35,19 +35,15 @@\n if \".\" in s and s.strip(\".\"):\n return \".\".join(mangle(x) if x else \"\" for x in s.split(\".\"))\n \n- # Step 1: Remove and save leading underscores\n+ # Remove and save leading underscores\n s2 = s.lstrip(normalizes_to_underscore)\n leading_underscores = \"_\" * (len(s) - len(s2))\n s = s2\n \n- # Step 2: Convert hyphens without introducing a new leading underscore\n+ # Convert hyphens without introducing a new leading underscore\n s = s[0] + s[1:].replace(\"-\", \"_\") if s else s\n \n- # Step 3: Convert trailing `?` to leading `is_`\n- if s.endswith(\"?\"):\n- s = \"is_\" + s[:-1]\n-\n- # Step 4: Convert invalid characters or reserved words\n+ # Convert invalid characters or reserved words\n if not (leading_underscores + s).isidentifier():\n # Replace illegal characters with their Unicode character\n # names, or hexadecimal if they don't have one.\n@@ -88,15 +84,9 @@\n => (hy.unmangle 'foo_bar)\n \"foo-bar\"\n \n- => (hy.unmangle 'is_foo_bar)\n- \"foo-bar?\"\n-\n => (hy.unmangle 'hyx_XasteriskX)\n \"*\"\n \n- => (hy.unmangle '_hyx_is_fooXsolidusXa)\n- \"_foo/a?\"\n-\n => (hy.unmangle 'hyx_XhyphenHminusX_XgreaterHthan_signX)\n \"-->\"\n \n@@ -126,8 +116,6 @@\n ),\n s[len(\"hyx_\") :],\n )\n- if s.startswith(\"is_\"):\n- s = s[len(\"is_\") :] + \"?\"\n s = s.replace(\"_\", \"-\")\n \n return prefix + s + suffix\n", "issue": "Remove the special mangling rule for question marks\nThis was one mangling detail that I disagreed with @gilch about years ago, but I belatedly think he was right. The rule that `foo?` mangles to `is_foo` (instead of `hyx_fooXquestion_markX`) should probably go. What's convinced me is wanting to use question marks on some names and then deciding not to because of mangling:\r\n\r\n- In [Infinitesimal Quest 2 + \u03b5](https://github.com/hylang/simalq), items have a Boolean attribute `destroy-on-pickup` that indicate whether the item is destroyed when it's picked up. I could call it `destroy-on-pickup?`, but that would mangle to the nonsensical `is_destroy_on_pickup`.\r\n- Likewise, `attack?`, meaning \"should this monster attack?\", would be misleadingly mangled to `is_attack`.\r\n\r\nAnother concern is that Hy's core macro `is-not` can be written `not?`, which looks like a unary predicate for some sort of object called a \"not\", instead of a binary function that indicates whether `a` is not `b`.\r\n\r\nOne motivation for the question-mark rule was that it could provide a Lispy alias for a Python name: if a library has a function `is_foo`, you can call it as `foo`. But this purpose was always undermined by [Python core's inconsistency between `isfoo` and `is_foo`](https://github.com/python/cpython/issues/73274).\r\n\r\nFinally, there's something to be said for making mangling a little simpler, since it's pretty complicated.\n", "before_files": [{"content": "import re\nimport unicodedata\n\nMANGLE_DELIM = \"X\"\n\nnormalizes_to_underscore = \"_\ufe33\ufe34\ufe4d\ufe4e\ufe4f\uff3f\"\n\n\ndef mangle(s):\n \"\"\"Stringify the argument (with :class:`str`, not :func:`repr` or\n :hy:func:`hy.repr`) and convert it to a valid Python identifier according\n to :ref:`Hy's mangling rules <mangling>`. ::\n\n (hy.mangle 'foo-bar?) ; => \"is_foo_bar\"\n (hy.mangle \"\ud83e\udd91\") ; => \"hyx_squid\"\n\n If the stringified argument is already both legal as a Python identifier\n and normalized according to Unicode normalization form KC (NFKC), it will\n be returned unchanged. Thus, ``hy.mangle`` is idempotent. ::\n\n (setv x '\u2666-->\u2660)\n (= (hy.mangle (hy.mangle x)) (hy.mangle x)) ; => True\n\n Generally, the stringifed input is expected to be parsable as a symbol. As\n a convenience, it can also have the syntax of a :ref:`dotted identifier\n <dotted-identifiers>`, and ``hy.mangle`` will mangle the dot-delimited\n parts separately. ::\n\n (hy.mangle \"a.b?.c!.d\") ; => \"a.is_b.hyx_cXexclamation_markX.d\"\n \"\"\"\n\n assert s\n s = str(s)\n\n if \".\" in s and s.strip(\".\"):\n return \".\".join(mangle(x) if x else \"\" for x in s.split(\".\"))\n\n # Step 1: Remove and save leading underscores\n s2 = s.lstrip(normalizes_to_underscore)\n leading_underscores = \"_\" * (len(s) - len(s2))\n s = s2\n\n # Step 2: Convert hyphens without introducing a new leading underscore\n s = s[0] + s[1:].replace(\"-\", \"_\") if s else s\n\n # Step 3: Convert trailing `?` to leading `is_`\n if s.endswith(\"?\"):\n s = \"is_\" + s[:-1]\n\n # Step 4: Convert invalid characters or reserved words\n if not (leading_underscores + s).isidentifier():\n # Replace illegal characters with their Unicode character\n # names, or hexadecimal if they don't have one.\n s = \"hyx_\" + \"\".join(\n c if c != MANGLE_DELIM and (\"S\" + c).isidentifier()\n # We prepend the \"S\" because some characters aren't\n # allowed at the start of an identifier.\n else \"{0}{1}{0}\".format(\n MANGLE_DELIM,\n unicodedata.name(c, \"\").lower().replace(\"-\", \"H\").replace(\" \", \"_\")\n or \"U{:x}\".format(ord(c)),\n )\n for c in s\n )\n\n # Step 5: Add back leading underscores\n s = leading_underscores + s\n\n # Normalize Unicode per PEP 3131.\n s = unicodedata.normalize(\"NFKC\", s)\n\n assert s.isidentifier()\n return s\n\n\ndef unmangle(s):\n \"\"\"Stringify the argument and try to convert it to a pretty unmangled\n form. See :ref:`Hy's mangling rules <mangling>`.\n\n Unmangling may not round-trip, because different Hy symbol names can mangle\n to the same Python identifier. In particular, Python itself already\n considers distinct strings that have the same normalized form (according to\n NFKC), such as ``hello`` and ``\ud835\udd25\ud835\udd22\ud835\udd29\ud835\udd29\ud835\udd2c``, to be the same identifier.\n\n Examples:\n ::\n\n => (hy.unmangle 'foo_bar)\n \"foo-bar\"\n\n => (hy.unmangle 'is_foo_bar)\n \"foo-bar?\"\n\n => (hy.unmangle 'hyx_XasteriskX)\n \"*\"\n\n => (hy.unmangle '_hyx_is_fooXsolidusXa)\n \"_foo/a?\"\n\n => (hy.unmangle 'hyx_XhyphenHminusX_XgreaterHthan_signX)\n \"-->\"\n\n => (hy.unmangle 'hyx_XlessHthan_signX__)\n \"<--\"\n\n => (hy.unmangle '__dunder_name__)\n \"__dunder-name__\"\n\n \"\"\"\n\n s = str(s)\n\n prefix = \"\"\n suffix = \"\"\n m = re.fullmatch(r\"(_+)(.*?)(_*)\", s, re.DOTALL)\n if m:\n prefix, s, suffix = m.groups()\n\n if s.startswith(\"hyx_\"):\n s = re.sub(\n \"{0}(U)?([_a-z0-9H]+?){0}\".format(MANGLE_DELIM),\n lambda mo: chr(int(mo.group(2), base=16))\n if mo.group(1)\n else unicodedata.lookup(\n mo.group(2).replace(\"_\", \" \").replace(\"H\", \"-\").upper()\n ),\n s[len(\"hyx_\") :],\n )\n if s.startswith(\"is_\"):\n s = s[len(\"is_\") :] + \"?\"\n s = s.replace(\"_\", \"-\")\n\n return prefix + s + suffix\n", "path": "hy/reader/mangling.py"}], "after_files": [{"content": "import re\nimport unicodedata\n\nMANGLE_DELIM = \"X\"\n\nnormalizes_to_underscore = \"_\ufe33\ufe34\ufe4d\ufe4e\ufe4f\uff3f\"\n\n\ndef mangle(s):\n \"\"\"Stringify the argument (with :class:`str`, not :func:`repr` or\n :hy:func:`hy.repr`) and convert it to a valid Python identifier according\n to :ref:`Hy's mangling rules <mangling>`. ::\n\n (hy.mangle 'foo-bar) ; => \"foo_bar\"\n (hy.mangle \"\ud83e\udd91\") ; => \"hyx_squid\"\n\n If the stringified argument is already both legal as a Python identifier\n and normalized according to Unicode normalization form KC (NFKC), it will\n be returned unchanged. Thus, ``hy.mangle`` is idempotent. ::\n\n (setv x '\u2666-->\u2660)\n (= (hy.mangle (hy.mangle x)) (hy.mangle x)) ; => True\n\n Generally, the stringifed input is expected to be parsable as a symbol. As\n a convenience, it can also have the syntax of a :ref:`dotted identifier\n <dotted-identifiers>`, and ``hy.mangle`` will mangle the dot-delimited\n parts separately. ::\n\n (hy.mangle \"a.c!.d\") ; => \"a.hyx_cXexclamation_markX.d\"\n \"\"\"\n\n assert s\n s = str(s)\n\n if \".\" in s and s.strip(\".\"):\n return \".\".join(mangle(x) if x else \"\" for x in s.split(\".\"))\n\n # Remove and save leading underscores\n s2 = s.lstrip(normalizes_to_underscore)\n leading_underscores = \"_\" * (len(s) - len(s2))\n s = s2\n\n # Convert hyphens without introducing a new leading underscore\n s = s[0] + s[1:].replace(\"-\", \"_\") if s else s\n\n # Convert invalid characters or reserved words\n if not (leading_underscores + s).isidentifier():\n # Replace illegal characters with their Unicode character\n # names, or hexadecimal if they don't have one.\n s = \"hyx_\" + \"\".join(\n c if c != MANGLE_DELIM and (\"S\" + c).isidentifier()\n # We prepend the \"S\" because some characters aren't\n # allowed at the start of an identifier.\n else \"{0}{1}{0}\".format(\n MANGLE_DELIM,\n unicodedata.name(c, \"\").lower().replace(\"-\", \"H\").replace(\" \", \"_\")\n or \"U{:x}\".format(ord(c)),\n )\n for c in s\n )\n\n # Step 5: Add back leading underscores\n s = leading_underscores + s\n\n # Normalize Unicode per PEP 3131.\n s = unicodedata.normalize(\"NFKC\", s)\n\n assert s.isidentifier()\n return s\n\n\ndef unmangle(s):\n \"\"\"Stringify the argument and try to convert it to a pretty unmangled\n form. See :ref:`Hy's mangling rules <mangling>`.\n\n Unmangling may not round-trip, because different Hy symbol names can mangle\n to the same Python identifier. In particular, Python itself already\n considers distinct strings that have the same normalized form (according to\n NFKC), such as ``hello`` and ``\ud835\udd25\ud835\udd22\ud835\udd29\ud835\udd29\ud835\udd2c``, to be the same identifier.\n\n Examples:\n ::\n\n => (hy.unmangle 'foo_bar)\n \"foo-bar\"\n\n => (hy.unmangle 'hyx_XasteriskX)\n \"*\"\n\n => (hy.unmangle 'hyx_XhyphenHminusX_XgreaterHthan_signX)\n \"-->\"\n\n => (hy.unmangle 'hyx_XlessHthan_signX__)\n \"<--\"\n\n => (hy.unmangle '__dunder_name__)\n \"__dunder-name__\"\n\n \"\"\"\n\n s = str(s)\n\n prefix = \"\"\n suffix = \"\"\n m = re.fullmatch(r\"(_+)(.*?)(_*)\", s, re.DOTALL)\n if m:\n prefix, s, suffix = m.groups()\n\n if s.startswith(\"hyx_\"):\n s = re.sub(\n \"{0}(U)?([_a-z0-9H]+?){0}\".format(MANGLE_DELIM),\n lambda mo: chr(int(mo.group(2), base=16))\n if mo.group(1)\n else unicodedata.lookup(\n mo.group(2).replace(\"_\", \" \").replace(\"H\", \"-\").upper()\n ),\n s[len(\"hyx_\") :],\n )\n s = s.replace(\"_\", \"-\")\n\n return prefix + s + suffix\n", "path": "hy/reader/mangling.py"}]}
| 2,084 | 713 |
gh_patches_debug_16692
|
rasdani/github-patches
|
git_diff
|
azavea__raster-vision-678
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Update setup.rst to reference raster-vision-cloudformation
The setup instructions currently reference raster-vision-aws.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `docs/conf.py`
Content:
```
1 from pallets_sphinx_themes import ProjectLink, get_version
2
3 # -*- coding: utf-8 -*-
4 #
5 # Configuration file for the Sphinx documentation builder.
6 #
7 # This file does only contain a selection of the most common options. For a
8 # full list see the documentation:
9 # http://www.sphinx-doc.org/en/stable/config
10
11 # -- Path setup --------------------------------------------------------------
12
13 # If extensions (or modules to document with autodoc) are in another directory,
14 # add these directories to sys.path here. If the directory is relative to the
15 # documentation root, use os.path.abspath to make it absolute, like shown here.
16 #
17 # import os
18 # import sys
19 # sys.path.insert(0, os.path.abspath('.'))
20
21
22 # -- Project information -----------------------------------------------------
23
24 project = 'Raster Vision'
25 copyright = '2018, Azavea'
26 author = 'Azavea'
27
28 # The short X.Y version
29 version = '0.8'
30 # The full version, including alpha/beta/rc tags
31 release = '0.8.0'
32
33
34 # -- General configuration ---------------------------------------------------
35
36 # If your documentation needs a minimal Sphinx version, state it here.
37 #
38 # needs_sphinx = '1.0'
39
40 # Add any Sphinx extension module names here, as strings. They can be
41 # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
42 # ones.
43 extensions = [
44 'sphinx.ext.autodoc',
45 'sphinx.ext.intersphinx',
46 'pallets_sphinx_themes',
47 'sphinxcontrib.programoutput'
48 ]
49
50 intersphinx_mapping = {'python': ('https://docs.python.org/3/', None)}
51
52 # Add any paths that contain templates here, relative to this directory.
53 templates_path = ['_templates']
54
55 # The suffix(es) of source filenames.
56 # You can specify multiple suffix as a list of string:
57 #
58 # source_suffix = ['.rst', '.md']
59 source_suffix = '.rst'
60
61 # The master toctree document.
62 master_doc = 'index'
63
64 # The language for content autogenerated by Sphinx. Refer to documentation
65 # for a list of supported languages.
66 #
67 # This is also used if you do content translation via gettext catalogs.
68 # Usually you set "language" from the command line for these cases.
69 language = None
70
71 # List of patterns, relative to source directory, that match files and
72 # directories to ignore when looking for source files.
73 # This pattern also affects html_static_path and html_extra_path .
74 exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store', 'README.md']
75
76 # The name of the Pygments (syntax highlighting) style to use.
77 # pygments_style = 'sphinx'
78
79 # HTML -----------------------------------------------------------------
80
81 html_theme = 'click'
82 html_theme_options = {'index_sidebar_logo': False}
83 html_context = {
84 'project_links': [
85 ProjectLink('Quickstart', 'quickstart.html'),
86 ProjectLink('Documentation TOC', 'index.html#documentation'),
87 ProjectLink('API Reference TOC', 'index.html#api-reference'),
88 ProjectLink('Project Website', 'https://rastervision.io/'),
89 ProjectLink('PyPI releases', 'https://pypi.org/project/rastervision/'),
90 ProjectLink('GitHub', 'https://github.com/azavea/raster-vision'),
91 ProjectLink('Gitter Channel', 'https://gitter.im/azavea/raster-vision'),
92 ProjectLink('Raster Vision Examples', 'https://github.com/azavea/raster-vision-examples'),
93 ProjectLink('QGIS Plugin', 'https://github.com/azavea/raster-vision-qgis'),
94 ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-aws'),
95 ProjectLink('Issue Tracker', 'https://github.com/azavea/raster-vision/issues/'),
96 ProjectLink('CHANGELOG', 'changelog.html'),
97 ProjectLink('Azavea', 'https://www.azavea.com/'),
98 ],
99 'css_files': [
100 '_static/rastervision.css',
101 'https://media.readthedocs.org/css/badge_only.css'
102 ]
103 }
104 html_sidebars = {
105 'index': ['project.html', 'versions.html', 'searchbox.html'],
106 '**': ['project.html', 'localtoc.html', 'relations.html', 'versions.html', 'searchbox.html'],
107 }
108 singlehtml_sidebars = {'index': ['project.html', 'versions.html', 'localtoc.html']}
109 html_static_path = ['_static']
110 html_favicon = '_static/raster-vision-icon.png'
111 html_logo = '_static/raster-vision-logo.png'
112 html_title = 'Raster Vision Documentation ({})'.format(version)
113 html_show_sourcelink = False
114 html_domain_indices = False
115 html_experimental_html5_writer = True
116
117 # -- Options for HTMLHelp output ---------------------------------------------
118
119 # Output file base name for HTML help builder.
120 htmlhelp_basename = 'RasterVisiondoc'
121
122
123 # -- Options for LaTeX output ------------------------------------------------
124
125 latex_elements = {
126 # The paper size ('letterpaper' or 'a4paper').
127 #
128 # 'papersize': 'letterpaper',
129
130 # The font size ('10pt', '11pt' or '12pt').
131 #
132 # 'pointsize': '10pt',
133
134 # Additional stuff for the LaTeX preamble.
135 #
136 # 'preamble': '',
137
138 # Latex figure (float) alignment
139 #
140 # 'figure_align': 'htbp',
141 }
142
143 # Grouping the document tree into LaTeX files. List of tuples
144 # (source start file, target name, title,
145 # author, documentclass [howto, manual, or own class]).
146 latex_documents = [
147 (master_doc, 'RasterVision.tex', 'Raster Vision Documentation',
148 'Azavea', 'manual'),
149 ]
150
151
152 # -- Options for manual page output ------------------------------------------
153
154 # One entry per manual page. List of tuples
155 # (source start file, name, description, authors, manual section).
156 man_pages = [
157 (master_doc, 'RasterVisoin-{}.tex', html_title,
158 [author], 'manual')
159 ]
160
161
162 # -- Options for Texinfo output ----------------------------------------------
163
164 # Grouping the document tree into Texinfo files. List of tuples
165 # (source start file, target name, title, author,
166 # dir menu entry, description, category)
167 texinfo_documents = [
168 (master_doc, 'RasterVision', 'Raster Vision Documentation',
169 author, 'RasterVision', 'One line description of project.',
170 'Miscellaneous'),
171 ]
172
173
174 # -- Extension configuration -------------------------------------------------
175
176 programoutput_prompt_template = '> {command}\n{output}'
177
178 # -- Options for todo extension ----------------------------------------------
179
180 # If true, `todo` and `todoList` produce output, else they produce nothing.
181 todo_include_todos = True
182
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/docs/conf.py b/docs/conf.py
--- a/docs/conf.py
+++ b/docs/conf.py
@@ -91,7 +91,7 @@
ProjectLink('Gitter Channel', 'https://gitter.im/azavea/raster-vision'),
ProjectLink('Raster Vision Examples', 'https://github.com/azavea/raster-vision-examples'),
ProjectLink('QGIS Plugin', 'https://github.com/azavea/raster-vision-qgis'),
- ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-aws'),
+ ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-cloudformation'),
ProjectLink('Issue Tracker', 'https://github.com/azavea/raster-vision/issues/'),
ProjectLink('CHANGELOG', 'changelog.html'),
ProjectLink('Azavea', 'https://www.azavea.com/'),
|
{"golden_diff": "diff --git a/docs/conf.py b/docs/conf.py\n--- a/docs/conf.py\n+++ b/docs/conf.py\n@@ -91,7 +91,7 @@\n ProjectLink('Gitter Channel', 'https://gitter.im/azavea/raster-vision'),\n ProjectLink('Raster Vision Examples', 'https://github.com/azavea/raster-vision-examples'),\n ProjectLink('QGIS Plugin', 'https://github.com/azavea/raster-vision-qgis'),\n- ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-aws'),\n+ ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-cloudformation'),\n ProjectLink('Issue Tracker', 'https://github.com/azavea/raster-vision/issues/'),\n ProjectLink('CHANGELOG', 'changelog.html'),\n ProjectLink('Azavea', 'https://www.azavea.com/'),\n", "issue": "Update setup.rst to reference raster-vision-cloudformation\nThe setup instructions currently reference raster-vision-aws.\n", "before_files": [{"content": "from pallets_sphinx_themes import ProjectLink, get_version\n\n# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\n# import os\n# import sys\n# sys.path.insert(0, os.path.abspath('.'))\n\n\n# -- Project information -----------------------------------------------------\n\nproject = 'Raster Vision'\ncopyright = '2018, Azavea'\nauthor = 'Azavea'\n\n# The short X.Y version\nversion = '0.8'\n# The full version, including alpha/beta/rc tags\nrelease = '0.8.0'\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n 'sphinx.ext.autodoc',\n 'sphinx.ext.intersphinx',\n 'pallets_sphinx_themes',\n 'sphinxcontrib.programoutput'\n]\n\nintersphinx_mapping = {'python': ('https://docs.python.org/3/', None)}\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = ['_templates']\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = '.rst'\n\n# The master toctree document.\nmaster_doc = 'index'\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = ['_build', 'Thumbs.db', '.DS_Store', 'README.md']\n\n# The name of the Pygments (syntax highlighting) style to use.\n# pygments_style = 'sphinx'\n\n# HTML -----------------------------------------------------------------\n\nhtml_theme = 'click'\nhtml_theme_options = {'index_sidebar_logo': False}\nhtml_context = {\n 'project_links': [\n ProjectLink('Quickstart', 'quickstart.html'),\n ProjectLink('Documentation TOC', 'index.html#documentation'),\n ProjectLink('API Reference TOC', 'index.html#api-reference'),\n ProjectLink('Project Website', 'https://rastervision.io/'),\n ProjectLink('PyPI releases', 'https://pypi.org/project/rastervision/'),\n ProjectLink('GitHub', 'https://github.com/azavea/raster-vision'),\n ProjectLink('Gitter Channel', 'https://gitter.im/azavea/raster-vision'),\n ProjectLink('Raster Vision Examples', 'https://github.com/azavea/raster-vision-examples'),\n ProjectLink('QGIS Plugin', 'https://github.com/azavea/raster-vision-qgis'),\n ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-aws'),\n ProjectLink('Issue Tracker', 'https://github.com/azavea/raster-vision/issues/'),\n ProjectLink('CHANGELOG', 'changelog.html'),\n ProjectLink('Azavea', 'https://www.azavea.com/'),\n ],\n 'css_files': [\n '_static/rastervision.css',\n 'https://media.readthedocs.org/css/badge_only.css'\n ]\n}\nhtml_sidebars = {\n 'index': ['project.html', 'versions.html', 'searchbox.html'],\n '**': ['project.html', 'localtoc.html', 'relations.html', 'versions.html', 'searchbox.html'],\n}\nsinglehtml_sidebars = {'index': ['project.html', 'versions.html', 'localtoc.html']}\nhtml_static_path = ['_static']\nhtml_favicon = '_static/raster-vision-icon.png'\nhtml_logo = '_static/raster-vision-logo.png'\nhtml_title = 'Raster Vision Documentation ({})'.format(version)\nhtml_show_sourcelink = False\nhtml_domain_indices = False\nhtml_experimental_html5_writer = True\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = 'RasterVisiondoc'\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (master_doc, 'RasterVision.tex', 'Raster Vision Documentation',\n 'Azavea', 'manual'),\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [\n (master_doc, 'RasterVisoin-{}.tex', html_title,\n [author], 'manual')\n]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (master_doc, 'RasterVision', 'Raster Vision Documentation',\n author, 'RasterVision', 'One line description of project.',\n 'Miscellaneous'),\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\nprogramoutput_prompt_template = '> {command}\\n{output}'\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n", "path": "docs/conf.py"}], "after_files": [{"content": "from pallets_sphinx_themes import ProjectLink, get_version\n\n# -*- coding: utf-8 -*-\n#\n# Configuration file for the Sphinx documentation builder.\n#\n# This file does only contain a selection of the most common options. For a\n# full list see the documentation:\n# http://www.sphinx-doc.org/en/stable/config\n\n# -- Path setup --------------------------------------------------------------\n\n# If extensions (or modules to document with autodoc) are in another directory,\n# add these directories to sys.path here. If the directory is relative to the\n# documentation root, use os.path.abspath to make it absolute, like shown here.\n#\n# import os\n# import sys\n# sys.path.insert(0, os.path.abspath('.'))\n\n\n# -- Project information -----------------------------------------------------\n\nproject = 'Raster Vision'\ncopyright = '2018, Azavea'\nauthor = 'Azavea'\n\n# The short X.Y version\nversion = '0.8'\n# The full version, including alpha/beta/rc tags\nrelease = '0.8.0'\n\n\n# -- General configuration ---------------------------------------------------\n\n# If your documentation needs a minimal Sphinx version, state it here.\n#\n# needs_sphinx = '1.0'\n\n# Add any Sphinx extension module names here, as strings. They can be\n# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom\n# ones.\nextensions = [\n 'sphinx.ext.autodoc',\n 'sphinx.ext.intersphinx',\n 'pallets_sphinx_themes',\n 'sphinxcontrib.programoutput'\n]\n\nintersphinx_mapping = {'python': ('https://docs.python.org/3/', None)}\n\n# Add any paths that contain templates here, relative to this directory.\ntemplates_path = ['_templates']\n\n# The suffix(es) of source filenames.\n# You can specify multiple suffix as a list of string:\n#\n# source_suffix = ['.rst', '.md']\nsource_suffix = '.rst'\n\n# The master toctree document.\nmaster_doc = 'index'\n\n# The language for content autogenerated by Sphinx. Refer to documentation\n# for a list of supported languages.\n#\n# This is also used if you do content translation via gettext catalogs.\n# Usually you set \"language\" from the command line for these cases.\nlanguage = None\n\n# List of patterns, relative to source directory, that match files and\n# directories to ignore when looking for source files.\n# This pattern also affects html_static_path and html_extra_path .\nexclude_patterns = ['_build', 'Thumbs.db', '.DS_Store', 'README.md']\n\n# The name of the Pygments (syntax highlighting) style to use.\n# pygments_style = 'sphinx'\n\n# HTML -----------------------------------------------------------------\n\nhtml_theme = 'click'\nhtml_theme_options = {'index_sidebar_logo': False}\nhtml_context = {\n 'project_links': [\n ProjectLink('Quickstart', 'quickstart.html'),\n ProjectLink('Documentation TOC', 'index.html#documentation'),\n ProjectLink('API Reference TOC', 'index.html#api-reference'),\n ProjectLink('Project Website', 'https://rastervision.io/'),\n ProjectLink('PyPI releases', 'https://pypi.org/project/rastervision/'),\n ProjectLink('GitHub', 'https://github.com/azavea/raster-vision'),\n ProjectLink('Gitter Channel', 'https://gitter.im/azavea/raster-vision'),\n ProjectLink('Raster Vision Examples', 'https://github.com/azavea/raster-vision-examples'),\n ProjectLink('QGIS Plugin', 'https://github.com/azavea/raster-vision-qgis'),\n ProjectLink('AWS Batch Setup', 'https://github.com/azavea/raster-vision-cloudformation'),\n ProjectLink('Issue Tracker', 'https://github.com/azavea/raster-vision/issues/'),\n ProjectLink('CHANGELOG', 'changelog.html'),\n ProjectLink('Azavea', 'https://www.azavea.com/'),\n ],\n 'css_files': [\n '_static/rastervision.css',\n 'https://media.readthedocs.org/css/badge_only.css'\n ]\n}\nhtml_sidebars = {\n 'index': ['project.html', 'versions.html', 'searchbox.html'],\n '**': ['project.html', 'localtoc.html', 'relations.html', 'versions.html', 'searchbox.html'],\n}\nsinglehtml_sidebars = {'index': ['project.html', 'versions.html', 'localtoc.html']}\nhtml_static_path = ['_static']\nhtml_favicon = '_static/raster-vision-icon.png'\nhtml_logo = '_static/raster-vision-logo.png'\nhtml_title = 'Raster Vision Documentation ({})'.format(version)\nhtml_show_sourcelink = False\nhtml_domain_indices = False\nhtml_experimental_html5_writer = True\n\n# -- Options for HTMLHelp output ---------------------------------------------\n\n# Output file base name for HTML help builder.\nhtmlhelp_basename = 'RasterVisiondoc'\n\n\n# -- Options for LaTeX output ------------------------------------------------\n\nlatex_elements = {\n # The paper size ('letterpaper' or 'a4paper').\n #\n # 'papersize': 'letterpaper',\n\n # The font size ('10pt', '11pt' or '12pt').\n #\n # 'pointsize': '10pt',\n\n # Additional stuff for the LaTeX preamble.\n #\n # 'preamble': '',\n\n # Latex figure (float) alignment\n #\n # 'figure_align': 'htbp',\n}\n\n# Grouping the document tree into LaTeX files. List of tuples\n# (source start file, target name, title,\n# author, documentclass [howto, manual, or own class]).\nlatex_documents = [\n (master_doc, 'RasterVision.tex', 'Raster Vision Documentation',\n 'Azavea', 'manual'),\n]\n\n\n# -- Options for manual page output ------------------------------------------\n\n# One entry per manual page. List of tuples\n# (source start file, name, description, authors, manual section).\nman_pages = [\n (master_doc, 'RasterVisoin-{}.tex', html_title,\n [author], 'manual')\n]\n\n\n# -- Options for Texinfo output ----------------------------------------------\n\n# Grouping the document tree into Texinfo files. List of tuples\n# (source start file, target name, title, author,\n# dir menu entry, description, category)\ntexinfo_documents = [\n (master_doc, 'RasterVision', 'Raster Vision Documentation',\n author, 'RasterVision', 'One line description of project.',\n 'Miscellaneous'),\n]\n\n\n# -- Extension configuration -------------------------------------------------\n\nprogramoutput_prompt_template = '> {command}\\n{output}'\n\n# -- Options for todo extension ----------------------------------------------\n\n# If true, `todo` and `todoList` produce output, else they produce nothing.\ntodo_include_todos = True\n", "path": "docs/conf.py"}]}
| 2,177 | 211 |
gh_patches_debug_3484
|
rasdani/github-patches
|
git_diff
|
dotkom__onlineweb4-2115
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
It's not possible to delete event relations in django admin
## What kind of an issue is this?
- [x] Bug report
## What is the expected behaviour?
When you mark an event relation for deletion it should be deleted when you save the object.
## What is the current behaviour?
The object is saved, but the relation is not removed.
## How do you reproduce this problem?
* Create an event
* Add a relation (e.g. Company or Attendance)
* Save
* Go to the admin panel for the event
* Mark the relation for deletion (click the delete checkbox by the relation)
* Save
* The relation is still there
## Other information
This is most likely due to us overwriting the `save_formset()` method in EventAdmin[0] and not taking care of deleting objects marked for deleting in the formset [1].
We're not doing anything custom in the method, so there's really no reason to override it. The code could actually be removed.
[0]: https://github.com/dotkom/onlineweb4/blob/develop/apps/events/admin.py#L160
[1]: https://docs.djangoproject.com/en/1.11/ref/contrib/admin/#django.contrib.admin.ModelAdmin.save_formset
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `apps/events/admin.py`
Content:
```
1 # -*- coding: utf-8 -*-
2
3 from django.contrib import admin, messages
4 from django.utils.translation import ugettext as _
5 from guardian.admin import GuardedModelAdmin
6 from reversion.admin import VersionAdmin
7
8 from apps.events.models import (AttendanceEvent, Attendee, CompanyEvent, Event, Extras,
9 FieldOfStudyRule, GradeRule, GroupRestriction, Reservation,
10 Reservee, RuleBundle, UserGroupRule)
11 from apps.feedback.admin import FeedbackRelationInline
12
13
14 class AttendeeInline(admin.TabularInline):
15 model = Attendee
16 extra = 1
17 classes = ('grp-collapse grp-open',) # style
18 inline_classes = ('grp-collapse grp-open',) # style
19
20
21 class CompanyInline(admin.TabularInline):
22 model = CompanyEvent
23 max_num = 20
24 extra = 0
25 classes = ('grp-collapse grp-open',) # style
26 inline_classes = ('grp-collapse grp-open',) # style
27
28
29 class RuleBundleInline(admin.TabularInline):
30 model = RuleBundle
31 extra = 1
32 max_num = 20
33 classes = ('grp-collapse grp-open',) # style
34 inline_classes = ('grp-collapse grp-open',) # style
35
36
37 class ExtrasInline(admin.TabularInline):
38 model = Extras
39 extra = 1
40 max_num = 20
41 classes = ('grp-collapse grp-open',) # style
42 inline_classes = ('grp-collapse grp-open',) # style
43
44
45 class GroupRestrictionInline(admin.TabularInline):
46 model = GroupRestriction
47 extra = 0
48 max_num = 1
49 classes = ('grp-collapse grp-open',) # style
50 inline_classes = ('grp-collapse grp-open',) # style
51 filter_horizontal = ('groups',)
52
53
54 def mark_paid(modeladmin, request, queryset):
55 queryset.update(paid=True)
56
57
58 mark_paid.short_description = "Merk som betalt"
59
60
61 def mark_not_paid(modeladmin, request, queryset):
62 queryset.update(paid=False)
63
64
65 mark_not_paid.short_description = "Merk som ikke betalt"
66
67
68 def mark_attended(modeladmin, request, queryset):
69 queryset.update(attended=True)
70
71
72 mark_attended.short_description = "Merk som møtt"
73
74
75 def mark_not_attended(modeladmin, request, queryset):
76 queryset.update(attended=False)
77
78
79 mark_not_attended.short_description = "Merk som ikke møtt"
80
81
82 class AttendeeAdmin(GuardedModelAdmin, VersionAdmin):
83 model = Attendee
84 ordering = ['-timestamp']
85 list_display = ('user', 'event', 'timestamp', 'paid', 'attended', 'note', 'extras')
86 list_filter = ('attended', 'paid', 'event__event')
87 search_fields = (
88 'event__event__title', '=event__event__id', 'user__first_name', 'user__last_name', 'user__username',
89 )
90 actions = [mark_paid, mark_attended, mark_not_paid, mark_not_attended]
91 group_owned_objects_field = 'event__event__organizer'
92 user_can_access_owned_by_group_objects_only = True
93
94 # Disable delete_selected http://bit.ly/1o4nleN
95 def get_actions(self, request):
96 actions = super(AttendeeAdmin, self).get_actions(request)
97 if 'delete_selected' in actions:
98 del actions['delete_selected']
99 return actions
100
101
102 class CompanyEventAdmin(VersionAdmin):
103 model = CompanyEvent
104 inlines = (CompanyInline,)
105
106
107 class ExtrasAdmin(VersionAdmin):
108 model = Extras
109 fk_name = 'choice'
110 # inlines = (ExtrasInline,)
111
112
113 class RuleBundleAdmin(VersionAdmin):
114 model = RuleBundle
115
116
117 class FieldOfStudyRuleAdmin(VersionAdmin):
118 model = FieldOfStudyRule
119
120
121 class GradeRuleAdmin(VersionAdmin):
122 model = GradeRule
123
124
125 class UserGroupRuleAdmin(VersionAdmin):
126 model = UserGroupRule
127
128
129 class AttendanceEventInline(admin.StackedInline):
130 model = AttendanceEvent
131 max_num = 1
132 extra = 0
133 filter_horizontal = ('rule_bundles',)
134 classes = ('grp-collapse grp-open',) # style
135 inline_classes = ('grp-collapse grp-open',) # style
136 exclude = ("marks_has_been_set",)
137
138
139 class EventAdmin(GuardedModelAdmin, VersionAdmin):
140 inlines = (AttendanceEventInline, FeedbackRelationInline, CompanyInline, GroupRestrictionInline)
141 exclude = ("author", )
142 list_display = ['__str__', 'event_type', 'organizer']
143 list_filter = ['event_type', 'organizer']
144 search_fields = ('title',)
145
146 group_owned_objects_field = 'organizer'
147 user_can_access_owned_by_group_objects_only = True
148
149 def save_model(self, request, obj, form, change):
150 if not change: # created
151 obj.author = request.user
152 obj.save()
153
154 def save_formset(self, request, form, formset, change):
155 instances = formset.save(commit=False)
156 for instance in instances:
157 instance.save()
158 formset.save_m2m()
159
160
161 class ReserveeInline(admin.TabularInline):
162 model = Reservee
163 extra = 1
164 classes = ('grp-collapse grp-open',) # style
165 inline_classes = ('grp-collapse grp-open',) # style
166
167
168 class ReservationAdmin(GuardedModelAdmin, VersionAdmin):
169 model = Reservation
170 inlines = (ReserveeInline,)
171 max_num = 1
172 extra = 0
173 list_display = ('attendance_event', '_number_of_seats_taken', 'seats', '_attendees', '_max_capacity')
174 classes = ('grp-collapse grp-open',) # style
175 inline_classes = ('grp-collapse grp-open',) # style
176 user_can_access_owned_by_group_objects_only = True
177 group_owned_objects_field = 'attendance_event__event__organizer'
178
179 def _number_of_seats_taken(self, obj):
180 return obj.number_of_seats_taken
181 _number_of_seats_taken.short_description = _("Fylte reservasjoner")
182
183 def _attendees(self, obj):
184 return obj.attendance_event.number_of_attendees
185 _attendees.short_description = _("Antall deltakere")
186
187 def _max_capacity(self, obj):
188 return obj.attendance_event.max_capacity
189 _max_capacity.short_description = _("Arrangementets maks-kapasitet")
190
191 def save_model(self, request, obj, form, change):
192 attendance_event = AttendanceEvent.objects.get(pk=obj.attendance_event.event)
193 number_of_free_seats = attendance_event.max_capacity - attendance_event.number_of_attendees
194 if number_of_free_seats < obj.seats:
195 obj.seats = number_of_free_seats
196 self.message_user(request, _(
197 "Du har valgt et antall reserverte plasser som overskrider antallet ledige plasser for dette "
198 "arrangementet. Antallet ble automatisk justert til %d (alle ledige plasser)."
199 ) % number_of_free_seats, messages.WARNING)
200 obj.save()
201
202
203 admin.site.register(Event, EventAdmin)
204 admin.site.register(Attendee, AttendeeAdmin)
205 admin.site.register(RuleBundle, RuleBundleAdmin)
206 admin.site.register(Extras, ExtrasAdmin)
207 admin.site.register(GradeRule, GradeRuleAdmin)
208 admin.site.register(UserGroupRule, UserGroupRuleAdmin)
209 admin.site.register(FieldOfStudyRule, FieldOfStudyRuleAdmin)
210 admin.site.register(Reservation, ReservationAdmin)
211
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/apps/events/admin.py b/apps/events/admin.py
--- a/apps/events/admin.py
+++ b/apps/events/admin.py
@@ -151,12 +151,6 @@
obj.author = request.user
obj.save()
- def save_formset(self, request, form, formset, change):
- instances = formset.save(commit=False)
- for instance in instances:
- instance.save()
- formset.save_m2m()
-
class ReserveeInline(admin.TabularInline):
model = Reservee
|
{"golden_diff": "diff --git a/apps/events/admin.py b/apps/events/admin.py\n--- a/apps/events/admin.py\n+++ b/apps/events/admin.py\n@@ -151,12 +151,6 @@\n obj.author = request.user\n obj.save()\n \n- def save_formset(self, request, form, formset, change):\n- instances = formset.save(commit=False)\n- for instance in instances:\n- instance.save()\n- formset.save_m2m()\n-\n \n class ReserveeInline(admin.TabularInline):\n model = Reservee\n", "issue": "It's not possible to delete event relations in django admin\n## What kind of an issue is this?\r\n\r\n- [x] Bug report\r\n\r\n\r\n## What is the expected behaviour?\r\n\r\nWhen you mark an event relation for deletion it should be deleted when you save the object.\r\n\r\n## What is the current behaviour?\r\n\r\nThe object is saved, but the relation is not removed.\r\n\r\n\r\n## How do you reproduce this problem? \r\n\r\n* Create an event\r\n* Add a relation (e.g. Company or Attendance)\r\n* Save\r\n* Go to the admin panel for the event\r\n* Mark the relation for deletion (click the delete checkbox by the relation)\r\n* Save\r\n* The relation is still there\r\n\r\n\r\n## Other information\r\n\r\nThis is most likely due to us overwriting the `save_formset()` method in EventAdmin[0] and not taking care of deleting objects marked for deleting in the formset [1].\r\n\r\nWe're not doing anything custom in the method, so there's really no reason to override it. The code could actually be removed.\r\n\r\n[0]: https://github.com/dotkom/onlineweb4/blob/develop/apps/events/admin.py#L160\r\n[1]: https://docs.djangoproject.com/en/1.11/ref/contrib/admin/#django.contrib.admin.ModelAdmin.save_formset\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n\nfrom django.contrib import admin, messages\nfrom django.utils.translation import ugettext as _\nfrom guardian.admin import GuardedModelAdmin\nfrom reversion.admin import VersionAdmin\n\nfrom apps.events.models import (AttendanceEvent, Attendee, CompanyEvent, Event, Extras,\n FieldOfStudyRule, GradeRule, GroupRestriction, Reservation,\n Reservee, RuleBundle, UserGroupRule)\nfrom apps.feedback.admin import FeedbackRelationInline\n\n\nclass AttendeeInline(admin.TabularInline):\n model = Attendee\n extra = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass CompanyInline(admin.TabularInline):\n model = CompanyEvent\n max_num = 20\n extra = 0\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass RuleBundleInline(admin.TabularInline):\n model = RuleBundle\n extra = 1\n max_num = 20\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass ExtrasInline(admin.TabularInline):\n model = Extras\n extra = 1\n max_num = 20\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass GroupRestrictionInline(admin.TabularInline):\n model = GroupRestriction\n extra = 0\n max_num = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n filter_horizontal = ('groups',)\n\n\ndef mark_paid(modeladmin, request, queryset):\n queryset.update(paid=True)\n\n\nmark_paid.short_description = \"Merk som betalt\"\n\n\ndef mark_not_paid(modeladmin, request, queryset):\n queryset.update(paid=False)\n\n\nmark_not_paid.short_description = \"Merk som ikke betalt\"\n\n\ndef mark_attended(modeladmin, request, queryset):\n queryset.update(attended=True)\n\n\nmark_attended.short_description = \"Merk som m\u00f8tt\"\n\n\ndef mark_not_attended(modeladmin, request, queryset):\n queryset.update(attended=False)\n\n\nmark_not_attended.short_description = \"Merk som ikke m\u00f8tt\"\n\n\nclass AttendeeAdmin(GuardedModelAdmin, VersionAdmin):\n model = Attendee\n ordering = ['-timestamp']\n list_display = ('user', 'event', 'timestamp', 'paid', 'attended', 'note', 'extras')\n list_filter = ('attended', 'paid', 'event__event')\n search_fields = (\n 'event__event__title', '=event__event__id', 'user__first_name', 'user__last_name', 'user__username',\n )\n actions = [mark_paid, mark_attended, mark_not_paid, mark_not_attended]\n group_owned_objects_field = 'event__event__organizer'\n user_can_access_owned_by_group_objects_only = True\n\n # Disable delete_selected http://bit.ly/1o4nleN\n def get_actions(self, request):\n actions = super(AttendeeAdmin, self).get_actions(request)\n if 'delete_selected' in actions:\n del actions['delete_selected']\n return actions\n\n\nclass CompanyEventAdmin(VersionAdmin):\n model = CompanyEvent\n inlines = (CompanyInline,)\n\n\nclass ExtrasAdmin(VersionAdmin):\n model = Extras\n fk_name = 'choice'\n # inlines = (ExtrasInline,)\n\n\nclass RuleBundleAdmin(VersionAdmin):\n model = RuleBundle\n\n\nclass FieldOfStudyRuleAdmin(VersionAdmin):\n model = FieldOfStudyRule\n\n\nclass GradeRuleAdmin(VersionAdmin):\n model = GradeRule\n\n\nclass UserGroupRuleAdmin(VersionAdmin):\n model = UserGroupRule\n\n\nclass AttendanceEventInline(admin.StackedInline):\n model = AttendanceEvent\n max_num = 1\n extra = 0\n filter_horizontal = ('rule_bundles',)\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n exclude = (\"marks_has_been_set\",)\n\n\nclass EventAdmin(GuardedModelAdmin, VersionAdmin):\n inlines = (AttendanceEventInline, FeedbackRelationInline, CompanyInline, GroupRestrictionInline)\n exclude = (\"author\", )\n list_display = ['__str__', 'event_type', 'organizer']\n list_filter = ['event_type', 'organizer']\n search_fields = ('title',)\n\n group_owned_objects_field = 'organizer'\n user_can_access_owned_by_group_objects_only = True\n\n def save_model(self, request, obj, form, change):\n if not change: # created\n obj.author = request.user\n obj.save()\n\n def save_formset(self, request, form, formset, change):\n instances = formset.save(commit=False)\n for instance in instances:\n instance.save()\n formset.save_m2m()\n\n\nclass ReserveeInline(admin.TabularInline):\n model = Reservee\n extra = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass ReservationAdmin(GuardedModelAdmin, VersionAdmin):\n model = Reservation\n inlines = (ReserveeInline,)\n max_num = 1\n extra = 0\n list_display = ('attendance_event', '_number_of_seats_taken', 'seats', '_attendees', '_max_capacity')\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n user_can_access_owned_by_group_objects_only = True\n group_owned_objects_field = 'attendance_event__event__organizer'\n\n def _number_of_seats_taken(self, obj):\n return obj.number_of_seats_taken\n _number_of_seats_taken.short_description = _(\"Fylte reservasjoner\")\n\n def _attendees(self, obj):\n return obj.attendance_event.number_of_attendees\n _attendees.short_description = _(\"Antall deltakere\")\n\n def _max_capacity(self, obj):\n return obj.attendance_event.max_capacity\n _max_capacity.short_description = _(\"Arrangementets maks-kapasitet\")\n\n def save_model(self, request, obj, form, change):\n attendance_event = AttendanceEvent.objects.get(pk=obj.attendance_event.event)\n number_of_free_seats = attendance_event.max_capacity - attendance_event.number_of_attendees\n if number_of_free_seats < obj.seats:\n obj.seats = number_of_free_seats\n self.message_user(request, _(\n \"Du har valgt et antall reserverte plasser som overskrider antallet ledige plasser for dette \"\n \"arrangementet. Antallet ble automatisk justert til %d (alle ledige plasser).\"\n ) % number_of_free_seats, messages.WARNING)\n obj.save()\n\n\nadmin.site.register(Event, EventAdmin)\nadmin.site.register(Attendee, AttendeeAdmin)\nadmin.site.register(RuleBundle, RuleBundleAdmin)\nadmin.site.register(Extras, ExtrasAdmin)\nadmin.site.register(GradeRule, GradeRuleAdmin)\nadmin.site.register(UserGroupRule, UserGroupRuleAdmin)\nadmin.site.register(FieldOfStudyRule, FieldOfStudyRuleAdmin)\nadmin.site.register(Reservation, ReservationAdmin)\n", "path": "apps/events/admin.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n\nfrom django.contrib import admin, messages\nfrom django.utils.translation import ugettext as _\nfrom guardian.admin import GuardedModelAdmin\nfrom reversion.admin import VersionAdmin\n\nfrom apps.events.models import (AttendanceEvent, Attendee, CompanyEvent, Event, Extras,\n FieldOfStudyRule, GradeRule, GroupRestriction, Reservation,\n Reservee, RuleBundle, UserGroupRule)\nfrom apps.feedback.admin import FeedbackRelationInline\n\n\nclass AttendeeInline(admin.TabularInline):\n model = Attendee\n extra = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass CompanyInline(admin.TabularInline):\n model = CompanyEvent\n max_num = 20\n extra = 0\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass RuleBundleInline(admin.TabularInline):\n model = RuleBundle\n extra = 1\n max_num = 20\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass ExtrasInline(admin.TabularInline):\n model = Extras\n extra = 1\n max_num = 20\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass GroupRestrictionInline(admin.TabularInline):\n model = GroupRestriction\n extra = 0\n max_num = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n filter_horizontal = ('groups',)\n\n\ndef mark_paid(modeladmin, request, queryset):\n queryset.update(paid=True)\n\n\nmark_paid.short_description = \"Merk som betalt\"\n\n\ndef mark_not_paid(modeladmin, request, queryset):\n queryset.update(paid=False)\n\n\nmark_not_paid.short_description = \"Merk som ikke betalt\"\n\n\ndef mark_attended(modeladmin, request, queryset):\n queryset.update(attended=True)\n\n\nmark_attended.short_description = \"Merk som m\u00f8tt\"\n\n\ndef mark_not_attended(modeladmin, request, queryset):\n queryset.update(attended=False)\n\n\nmark_not_attended.short_description = \"Merk som ikke m\u00f8tt\"\n\n\nclass AttendeeAdmin(GuardedModelAdmin, VersionAdmin):\n model = Attendee\n ordering = ['-timestamp']\n list_display = ('user', 'event', 'timestamp', 'paid', 'attended', 'note', 'extras')\n list_filter = ('attended', 'paid', 'event__event')\n search_fields = (\n 'event__event__title', '=event__event__id', 'user__first_name', 'user__last_name', 'user__username',\n )\n actions = [mark_paid, mark_attended, mark_not_paid, mark_not_attended]\n group_owned_objects_field = 'event__event__organizer'\n user_can_access_owned_by_group_objects_only = True\n\n # Disable delete_selected http://bit.ly/1o4nleN\n def get_actions(self, request):\n actions = super(AttendeeAdmin, self).get_actions(request)\n if 'delete_selected' in actions:\n del actions['delete_selected']\n return actions\n\n\nclass CompanyEventAdmin(VersionAdmin):\n model = CompanyEvent\n inlines = (CompanyInline,)\n\n\nclass ExtrasAdmin(VersionAdmin):\n model = Extras\n fk_name = 'choice'\n # inlines = (ExtrasInline,)\n\n\nclass RuleBundleAdmin(VersionAdmin):\n model = RuleBundle\n\n\nclass FieldOfStudyRuleAdmin(VersionAdmin):\n model = FieldOfStudyRule\n\n\nclass GradeRuleAdmin(VersionAdmin):\n model = GradeRule\n\n\nclass UserGroupRuleAdmin(VersionAdmin):\n model = UserGroupRule\n\n\nclass AttendanceEventInline(admin.StackedInline):\n model = AttendanceEvent\n max_num = 1\n extra = 0\n filter_horizontal = ('rule_bundles',)\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n exclude = (\"marks_has_been_set\",)\n\n\nclass EventAdmin(GuardedModelAdmin, VersionAdmin):\n inlines = (AttendanceEventInline, FeedbackRelationInline, CompanyInline, GroupRestrictionInline)\n exclude = (\"author\", )\n list_display = ['__str__', 'event_type', 'organizer']\n list_filter = ['event_type', 'organizer']\n search_fields = ('title',)\n\n group_owned_objects_field = 'organizer'\n user_can_access_owned_by_group_objects_only = True\n\n def save_model(self, request, obj, form, change):\n if not change: # created\n obj.author = request.user\n obj.save()\n\n\nclass ReserveeInline(admin.TabularInline):\n model = Reservee\n extra = 1\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n\n\nclass ReservationAdmin(GuardedModelAdmin, VersionAdmin):\n model = Reservation\n inlines = (ReserveeInline,)\n max_num = 1\n extra = 0\n list_display = ('attendance_event', '_number_of_seats_taken', 'seats', '_attendees', '_max_capacity')\n classes = ('grp-collapse grp-open',) # style\n inline_classes = ('grp-collapse grp-open',) # style\n user_can_access_owned_by_group_objects_only = True\n group_owned_objects_field = 'attendance_event__event__organizer'\n\n def _number_of_seats_taken(self, obj):\n return obj.number_of_seats_taken\n _number_of_seats_taken.short_description = _(\"Fylte reservasjoner\")\n\n def _attendees(self, obj):\n return obj.attendance_event.number_of_attendees\n _attendees.short_description = _(\"Antall deltakere\")\n\n def _max_capacity(self, obj):\n return obj.attendance_event.max_capacity\n _max_capacity.short_description = _(\"Arrangementets maks-kapasitet\")\n\n def save_model(self, request, obj, form, change):\n attendance_event = AttendanceEvent.objects.get(pk=obj.attendance_event.event)\n number_of_free_seats = attendance_event.max_capacity - attendance_event.number_of_attendees\n if number_of_free_seats < obj.seats:\n obj.seats = number_of_free_seats\n self.message_user(request, _(\n \"Du har valgt et antall reserverte plasser som overskrider antallet ledige plasser for dette \"\n \"arrangementet. Antallet ble automatisk justert til %d (alle ledige plasser).\"\n ) % number_of_free_seats, messages.WARNING)\n obj.save()\n\n\nadmin.site.register(Event, EventAdmin)\nadmin.site.register(Attendee, AttendeeAdmin)\nadmin.site.register(RuleBundle, RuleBundleAdmin)\nadmin.site.register(Extras, ExtrasAdmin)\nadmin.site.register(GradeRule, GradeRuleAdmin)\nadmin.site.register(UserGroupRule, UserGroupRuleAdmin)\nadmin.site.register(FieldOfStudyRule, FieldOfStudyRuleAdmin)\nadmin.site.register(Reservation, ReservationAdmin)\n", "path": "apps/events/admin.py"}]}
| 2,708 | 120 |
gh_patches_debug_1725
|
rasdani/github-patches
|
git_diff
|
HypothesisWorks__hypothesis-1350
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Python deployment did not remove RELEASE.rst
In c216ddca8155cdb05053bcfe2ac51814447aac4f, which blocks further merges. @DRMacIver, I think cb84c3ad4dfdecfcf947ccd24bfdee111083645b shouldn't have taken out the `git rm RELEASE.rst`...
AFAICT ea6bb819143d5c6b4e4c43aa2436a9f70d129872 would be a decent fix. (and is now in all my open PRs)
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `tooling/src/hypothesistooling/releasemanagement.py`
Content:
```
1 # coding=utf-8
2 #
3 # This file is part of Hypothesis, which may be found at
4 # https://github.com/HypothesisWorks/hypothesis-python
5 #
6 # Most of this work is copyright (C) 2013-2018 David R. MacIver
7 # ([email protected]), but it contains contributions by others. See
8 # CONTRIBUTING.rst for a full list of people who may hold copyright, and
9 # consult the git log if you need to determine who owns an individual
10 # contribution.
11 #
12 # This Source Code Form is subject to the terms of the Mozilla Public License,
13 # v. 2.0. If a copy of the MPL was not distributed with this file, You can
14 # obtain one at http://mozilla.org/MPL/2.0/.
15 #
16 # END HEADER
17
18 """Helpful common code for release management tasks that is shared across
19 multiple projects.
20
21 Note that most code in here is brittle and specific to our build and
22 probably makes all sorts of undocumented assumptions, even as it looks
23 like a nice tidy reusable set of functionality.
24 """
25
26
27 from __future__ import division, print_function, absolute_import
28
29 import re
30 from datetime import datetime, timedelta
31
32 import hypothesistooling as tools
33
34
35 def release_date_string():
36 """Returns a date string that represents what should be considered "today"
37 for the purposes of releasing. It is always measured in UTC, but if it's in
38 the last hour of the day it will actually be considered tomorrow.
39
40 The reason for counting it as the later day is that it ensures that
41 (unless our release process takes more than 23 hours) this value
42 remains consistent throughout the entire release.
43 """
44 now = datetime.utcnow()
45
46 return max([
47 d.strftime('%Y-%m-%d') for d in (now, now + timedelta(hours=1))
48 ])
49
50
51 def assignment_matcher(name):
52 """
53 Matches a single line of the form (some space)name = (some value). e.g.
54 " foo = 1".
55 The whole line up to the assigned value is the first matching group,
56 the rest of the line is the second matching group.
57 i.e. group 1 is the assignment, group 2 is the value. In the above
58 example group 1 would be " foo = " and group 2 would be "1"
59 """
60 return re.compile(r'\A(\s*%s\s*=\s*)(.+)\Z' % (re.escape(name),))
61
62
63 def extract_assignment_from_string(contents, name):
64 lines = contents.split('\n')
65
66 matcher = assignment_matcher(name)
67
68 for i, l in enumerate(lines):
69 match = matcher.match(l)
70 if match is not None:
71 return match[2].strip()
72
73 raise ValueError('Key %s not found in %s' % (
74 name, contents
75 ))
76
77
78 def extract_assignment(filename, name):
79 with open(filename) as i:
80 return extract_assignment_from_string(i.read(), name)
81
82
83 def replace_assignment_in_string(contents, name, value):
84 lines = contents.split('\n')
85
86 matcher = assignment_matcher(name)
87
88 count = 0
89
90 for i, l in enumerate(lines):
91 match = matcher.match(l)
92 if match is not None:
93 count += 1
94 lines[i] = match[1] + value
95
96 if count == 0:
97 raise ValueError('Key %s not found in %s' % (
98 name, contents
99 ))
100 if count > 1:
101 raise ValueError('Key %s found %d times in %s' % (
102 name, count, contents
103 ))
104
105 return '\n'.join(lines)
106
107
108 def replace_assignment(filename, name, value):
109 """Replaces a single assignment of the form key = value in a file with a
110 new value, attempting to preserve the existing format.
111
112 This is fairly fragile - in particular it knows nothing about
113 the file format. The existing value is simply the rest of the line after
114 the last space after the equals.
115 """
116 with open(filename) as i:
117 contents = i.read()
118 result = replace_assignment_in_string(contents, name, value)
119 with open(filename, 'w') as o:
120 o.write(result)
121
122
123 RELEASE_TYPE = re.compile(r"^RELEASE_TYPE: +(major|minor|patch)")
124
125
126 MAJOR = 'major'
127 MINOR = 'minor'
128 PATCH = 'patch'
129
130
131 VALID_RELEASE_TYPES = (MAJOR, MINOR, PATCH)
132
133
134 def parse_release_file(filename):
135 with open(filename) as i:
136 return parse_release_file_contents(i.read(), filename)
137
138
139 def parse_release_file_contents(release_contents, filename):
140 release_lines = release_contents.split('\n')
141
142 m = RELEASE_TYPE.match(release_lines[0])
143 if m is not None:
144 release_type = m.group(1)
145 if release_type not in VALID_RELEASE_TYPES:
146 raise ValueError('Unrecognised release type %r' % (release_type,))
147 del release_lines[0]
148 release_contents = '\n'.join(release_lines).strip()
149 else:
150 raise ValueError(
151 '%s does not start by specifying release type. The first '
152 'line of the file should be RELEASE_TYPE: followed by one of '
153 'major, minor, or patch, to specify the type of release that '
154 'this is (i.e. which version number to increment). Instead the '
155 'first line was %r' % (filename, release_lines[0],)
156 )
157
158 return release_type, release_contents
159
160
161 def bump_version_info(version_info, release_type):
162 new_version = list(version_info)
163 bump = VALID_RELEASE_TYPES.index(release_type)
164 new_version[bump] += 1
165 for i in range(bump + 1, len(new_version)):
166 new_version[i] = 0
167 new_version = tuple(new_version)
168 new_version_string = '.'.join(map(str, new_version))
169 return new_version_string, new_version
170
171
172 def update_markdown_changelog(changelog, name, version, entry):
173 with open(changelog) as i:
174 prev_contents = i.read()
175
176 title = '# %(name)s %(version)s (%(date)s)\n\n' % {
177 'name': name, 'version': version, 'date': release_date_string(),
178 }
179
180 with open(changelog, 'w') as o:
181 o.write(title)
182 o.write(entry.strip())
183 o.write('\n\n')
184 o.write(prev_contents)
185
186
187 def parse_version(version):
188 return tuple(map(int, version.split('.')))
189
190
191 def commit_pending_release(project):
192 """Create a commit with the new release."""
193 tools.git('add', '-u', project.BASE_DIR)
194
195 tools.git(
196 'commit', '-m',
197 'Bump %s version to %s and update changelog'
198 '\n\n[skip ci]' % (project.PACKAGE_NAME, project.current_version(),)
199 )
200
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/tooling/src/hypothesistooling/releasemanagement.py b/tooling/src/hypothesistooling/releasemanagement.py
--- a/tooling/src/hypothesistooling/releasemanagement.py
+++ b/tooling/src/hypothesistooling/releasemanagement.py
@@ -190,6 +190,7 @@
def commit_pending_release(project):
"""Create a commit with the new release."""
+ tools.git('rm', project.RELEASE_FILE)
tools.git('add', '-u', project.BASE_DIR)
tools.git(
|
{"golden_diff": "diff --git a/tooling/src/hypothesistooling/releasemanagement.py b/tooling/src/hypothesistooling/releasemanagement.py\n--- a/tooling/src/hypothesistooling/releasemanagement.py\n+++ b/tooling/src/hypothesistooling/releasemanagement.py\n@@ -190,6 +190,7 @@\n \n def commit_pending_release(project):\n \"\"\"Create a commit with the new release.\"\"\"\n+ tools.git('rm', project.RELEASE_FILE)\n tools.git('add', '-u', project.BASE_DIR)\n \n tools.git(\n", "issue": "Python deployment did not remove RELEASE.rst\nIn c216ddca8155cdb05053bcfe2ac51814447aac4f, which blocks further merges. @DRMacIver, I think cb84c3ad4dfdecfcf947ccd24bfdee111083645b shouldn't have taken out the `git rm RELEASE.rst`... \r\n\r\nAFAICT ea6bb819143d5c6b4e4c43aa2436a9f70d129872 would be a decent fix. (and is now in all my open PRs)\n", "before_files": [{"content": "# coding=utf-8\n#\n# This file is part of Hypothesis, which may be found at\n# https://github.com/HypothesisWorks/hypothesis-python\n#\n# Most of this work is copyright (C) 2013-2018 David R. MacIver\n# ([email protected]), but it contains contributions by others. See\n# CONTRIBUTING.rst for a full list of people who may hold copyright, and\n# consult the git log if you need to determine who owns an individual\n# contribution.\n#\n# This Source Code Form is subject to the terms of the Mozilla Public License,\n# v. 2.0. If a copy of the MPL was not distributed with this file, You can\n# obtain one at http://mozilla.org/MPL/2.0/.\n#\n# END HEADER\n\n\"\"\"Helpful common code for release management tasks that is shared across\nmultiple projects.\n\nNote that most code in here is brittle and specific to our build and\nprobably makes all sorts of undocumented assumptions, even as it looks\nlike a nice tidy reusable set of functionality.\n\"\"\"\n\n\nfrom __future__ import division, print_function, absolute_import\n\nimport re\nfrom datetime import datetime, timedelta\n\nimport hypothesistooling as tools\n\n\ndef release_date_string():\n \"\"\"Returns a date string that represents what should be considered \"today\"\n for the purposes of releasing. It is always measured in UTC, but if it's in\n the last hour of the day it will actually be considered tomorrow.\n\n The reason for counting it as the later day is that it ensures that\n (unless our release process takes more than 23 hours) this value\n remains consistent throughout the entire release.\n \"\"\"\n now = datetime.utcnow()\n\n return max([\n d.strftime('%Y-%m-%d') for d in (now, now + timedelta(hours=1))\n ])\n\n\ndef assignment_matcher(name):\n \"\"\"\n Matches a single line of the form (some space)name = (some value). e.g.\n \" foo = 1\".\n The whole line up to the assigned value is the first matching group,\n the rest of the line is the second matching group.\n i.e. group 1 is the assignment, group 2 is the value. In the above\n example group 1 would be \" foo = \" and group 2 would be \"1\"\n \"\"\"\n return re.compile(r'\\A(\\s*%s\\s*=\\s*)(.+)\\Z' % (re.escape(name),))\n\n\ndef extract_assignment_from_string(contents, name):\n lines = contents.split('\\n')\n\n matcher = assignment_matcher(name)\n\n for i, l in enumerate(lines):\n match = matcher.match(l)\n if match is not None:\n return match[2].strip()\n\n raise ValueError('Key %s not found in %s' % (\n name, contents\n ))\n\n\ndef extract_assignment(filename, name):\n with open(filename) as i:\n return extract_assignment_from_string(i.read(), name)\n\n\ndef replace_assignment_in_string(contents, name, value):\n lines = contents.split('\\n')\n\n matcher = assignment_matcher(name)\n\n count = 0\n\n for i, l in enumerate(lines):\n match = matcher.match(l)\n if match is not None:\n count += 1\n lines[i] = match[1] + value\n\n if count == 0:\n raise ValueError('Key %s not found in %s' % (\n name, contents\n ))\n if count > 1:\n raise ValueError('Key %s found %d times in %s' % (\n name, count, contents\n ))\n\n return '\\n'.join(lines)\n\n\ndef replace_assignment(filename, name, value):\n \"\"\"Replaces a single assignment of the form key = value in a file with a\n new value, attempting to preserve the existing format.\n\n This is fairly fragile - in particular it knows nothing about\n the file format. The existing value is simply the rest of the line after\n the last space after the equals.\n \"\"\"\n with open(filename) as i:\n contents = i.read()\n result = replace_assignment_in_string(contents, name, value)\n with open(filename, 'w') as o:\n o.write(result)\n\n\nRELEASE_TYPE = re.compile(r\"^RELEASE_TYPE: +(major|minor|patch)\")\n\n\nMAJOR = 'major'\nMINOR = 'minor'\nPATCH = 'patch'\n\n\nVALID_RELEASE_TYPES = (MAJOR, MINOR, PATCH)\n\n\ndef parse_release_file(filename):\n with open(filename) as i:\n return parse_release_file_contents(i.read(), filename)\n\n\ndef parse_release_file_contents(release_contents, filename):\n release_lines = release_contents.split('\\n')\n\n m = RELEASE_TYPE.match(release_lines[0])\n if m is not None:\n release_type = m.group(1)\n if release_type not in VALID_RELEASE_TYPES:\n raise ValueError('Unrecognised release type %r' % (release_type,))\n del release_lines[0]\n release_contents = '\\n'.join(release_lines).strip()\n else:\n raise ValueError(\n '%s does not start by specifying release type. The first '\n 'line of the file should be RELEASE_TYPE: followed by one of '\n 'major, minor, or patch, to specify the type of release that '\n 'this is (i.e. which version number to increment). Instead the '\n 'first line was %r' % (filename, release_lines[0],)\n )\n\n return release_type, release_contents\n\n\ndef bump_version_info(version_info, release_type):\n new_version = list(version_info)\n bump = VALID_RELEASE_TYPES.index(release_type)\n new_version[bump] += 1\n for i in range(bump + 1, len(new_version)):\n new_version[i] = 0\n new_version = tuple(new_version)\n new_version_string = '.'.join(map(str, new_version))\n return new_version_string, new_version\n\n\ndef update_markdown_changelog(changelog, name, version, entry):\n with open(changelog) as i:\n prev_contents = i.read()\n\n title = '# %(name)s %(version)s (%(date)s)\\n\\n' % {\n 'name': name, 'version': version, 'date': release_date_string(),\n }\n\n with open(changelog, 'w') as o:\n o.write(title)\n o.write(entry.strip())\n o.write('\\n\\n')\n o.write(prev_contents)\n\n\ndef parse_version(version):\n return tuple(map(int, version.split('.')))\n\n\ndef commit_pending_release(project):\n \"\"\"Create a commit with the new release.\"\"\"\n tools.git('add', '-u', project.BASE_DIR)\n\n tools.git(\n 'commit', '-m',\n 'Bump %s version to %s and update changelog'\n '\\n\\n[skip ci]' % (project.PACKAGE_NAME, project.current_version(),)\n )\n", "path": "tooling/src/hypothesistooling/releasemanagement.py"}], "after_files": [{"content": "# coding=utf-8\n#\n# This file is part of Hypothesis, which may be found at\n# https://github.com/HypothesisWorks/hypothesis-python\n#\n# Most of this work is copyright (C) 2013-2018 David R. MacIver\n# ([email protected]), but it contains contributions by others. See\n# CONTRIBUTING.rst for a full list of people who may hold copyright, and\n# consult the git log if you need to determine who owns an individual\n# contribution.\n#\n# This Source Code Form is subject to the terms of the Mozilla Public License,\n# v. 2.0. If a copy of the MPL was not distributed with this file, You can\n# obtain one at http://mozilla.org/MPL/2.0/.\n#\n# END HEADER\n\n\"\"\"Helpful common code for release management tasks that is shared across\nmultiple projects.\n\nNote that most code in here is brittle and specific to our build and\nprobably makes all sorts of undocumented assumptions, even as it looks\nlike a nice tidy reusable set of functionality.\n\"\"\"\n\n\nfrom __future__ import division, print_function, absolute_import\n\nimport re\nfrom datetime import datetime, timedelta\n\nimport hypothesistooling as tools\n\n\ndef release_date_string():\n \"\"\"Returns a date string that represents what should be considered \"today\"\n for the purposes of releasing. It is always measured in UTC, but if it's in\n the last hour of the day it will actually be considered tomorrow.\n\n The reason for counting it as the later day is that it ensures that\n (unless our release process takes more than 23 hours) this value\n remains consistent throughout the entire release.\n \"\"\"\n now = datetime.utcnow()\n\n return max([\n d.strftime('%Y-%m-%d') for d in (now, now + timedelta(hours=1))\n ])\n\n\ndef assignment_matcher(name):\n \"\"\"\n Matches a single line of the form (some space)name = (some value). e.g.\n \" foo = 1\".\n The whole line up to the assigned value is the first matching group,\n the rest of the line is the second matching group.\n i.e. group 1 is the assignment, group 2 is the value. In the above\n example group 1 would be \" foo = \" and group 2 would be \"1\"\n \"\"\"\n return re.compile(r'\\A(\\s*%s\\s*=\\s*)(.+)\\Z' % (re.escape(name),))\n\n\ndef extract_assignment_from_string(contents, name):\n lines = contents.split('\\n')\n\n matcher = assignment_matcher(name)\n\n for i, l in enumerate(lines):\n match = matcher.match(l)\n if match is not None:\n return match[2].strip()\n\n raise ValueError('Key %s not found in %s' % (\n name, contents\n ))\n\n\ndef extract_assignment(filename, name):\n with open(filename) as i:\n return extract_assignment_from_string(i.read(), name)\n\n\ndef replace_assignment_in_string(contents, name, value):\n lines = contents.split('\\n')\n\n matcher = assignment_matcher(name)\n\n count = 0\n\n for i, l in enumerate(lines):\n match = matcher.match(l)\n if match is not None:\n count += 1\n lines[i] = match[1] + value\n\n if count == 0:\n raise ValueError('Key %s not found in %s' % (\n name, contents\n ))\n if count > 1:\n raise ValueError('Key %s found %d times in %s' % (\n name, count, contents\n ))\n\n return '\\n'.join(lines)\n\n\ndef replace_assignment(filename, name, value):\n \"\"\"Replaces a single assignment of the form key = value in a file with a\n new value, attempting to preserve the existing format.\n\n This is fairly fragile - in particular it knows nothing about\n the file format. The existing value is simply the rest of the line after\n the last space after the equals.\n \"\"\"\n with open(filename) as i:\n contents = i.read()\n result = replace_assignment_in_string(contents, name, value)\n with open(filename, 'w') as o:\n o.write(result)\n\n\nRELEASE_TYPE = re.compile(r\"^RELEASE_TYPE: +(major|minor|patch)\")\n\n\nMAJOR = 'major'\nMINOR = 'minor'\nPATCH = 'patch'\n\n\nVALID_RELEASE_TYPES = (MAJOR, MINOR, PATCH)\n\n\ndef parse_release_file(filename):\n with open(filename) as i:\n return parse_release_file_contents(i.read(), filename)\n\n\ndef parse_release_file_contents(release_contents, filename):\n release_lines = release_contents.split('\\n')\n\n m = RELEASE_TYPE.match(release_lines[0])\n if m is not None:\n release_type = m.group(1)\n if release_type not in VALID_RELEASE_TYPES:\n raise ValueError('Unrecognised release type %r' % (release_type,))\n del release_lines[0]\n release_contents = '\\n'.join(release_lines).strip()\n else:\n raise ValueError(\n '%s does not start by specifying release type. The first '\n 'line of the file should be RELEASE_TYPE: followed by one of '\n 'major, minor, or patch, to specify the type of release that '\n 'this is (i.e. which version number to increment). Instead the '\n 'first line was %r' % (filename, release_lines[0],)\n )\n\n return release_type, release_contents\n\n\ndef bump_version_info(version_info, release_type):\n new_version = list(version_info)\n bump = VALID_RELEASE_TYPES.index(release_type)\n new_version[bump] += 1\n for i in range(bump + 1, len(new_version)):\n new_version[i] = 0\n new_version = tuple(new_version)\n new_version_string = '.'.join(map(str, new_version))\n return new_version_string, new_version\n\n\ndef update_markdown_changelog(changelog, name, version, entry):\n with open(changelog) as i:\n prev_contents = i.read()\n\n title = '# %(name)s %(version)s (%(date)s)\\n\\n' % {\n 'name': name, 'version': version, 'date': release_date_string(),\n }\n\n with open(changelog, 'w') as o:\n o.write(title)\n o.write(entry.strip())\n o.write('\\n\\n')\n o.write(prev_contents)\n\n\ndef parse_version(version):\n return tuple(map(int, version.split('.')))\n\n\ndef commit_pending_release(project):\n \"\"\"Create a commit with the new release.\"\"\"\n tools.git('rm', project.RELEASE_FILE)\n tools.git('add', '-u', project.BASE_DIR)\n\n tools.git(\n 'commit', '-m',\n 'Bump %s version to %s and update changelog'\n '\\n\\n[skip ci]' % (project.PACKAGE_NAME, project.current_version(),)\n )\n", "path": "tooling/src/hypothesistooling/releasemanagement.py"}]}
| 2,439 | 133 |
gh_patches_debug_34358
|
rasdani/github-patches
|
git_diff
|
sopel-irc__sopel-1404
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Option to specify alternate config directory on first run
When the bot is first run, it executes ```sopel/run_script.py``` at startup.
Said script is hardcoded to create ```~/.sopel``` if it doesn't already exist.
Should there be an option to allow the user to specify an alternate directory to use?
This was observed on d9cfb41 running ```python sopel.py``` without prior configuration.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `sopel/config/__init__.py`
Content:
```
1 # coding=utf-8
2 """
3 The config object provides a simplified to access Sopel's configuration file.
4 The sections of the file are attributes of the object, and the keys in the
5 section are attributes of that. So, for example, the ``eggs`` attribute in the
6 ``[spam]`` section can be accessed from ``config.spam.eggs``.
7
8 Section definitions (see "Section configuration sections" below) can be added
9 to the config object with ``define_section``. When this is done, only the
10 defined keys will be available. A section can not be given more than one
11 definition. The ``[core]`` section is defined with ``CoreSection`` when the
12 object is initialized.
13
14 .. versionadded:: 6.0.0
15 """
16 # Copyright 2012-2015, Elsie Powell, embolalia.com
17 # Copyright © 2012, Elad Alfassa <[email protected]>
18 # Licensed under the Eiffel Forum License 2.
19
20 from __future__ import unicode_literals, absolute_import, print_function, division
21
22 from sopel.tools import iteritems, stderr
23 import sopel.tools
24 from sopel.tools import get_input
25 import sopel.loader
26 import os
27 import sys
28 if sys.version_info.major < 3:
29 import ConfigParser
30 else:
31 basestring = str
32 import configparser as ConfigParser
33 import sopel.config.core_section
34 from sopel.config.types import StaticSection
35
36
37 DEFAULT_HOMEDIR = os.path.join(os.path.expanduser('~'), '.sopel')
38
39
40 class ConfigurationError(Exception):
41 """ Exception type for configuration errors """
42
43 def __init__(self, value):
44 self.value = value
45
46 def __str__(self):
47 return 'ConfigurationError: %s' % self.value
48
49
50 class Config(object):
51 def __init__(self, filename, validate=True):
52 """The bot's configuration.
53
54 The given filename will be associated with the configuration, and is
55 the file which will be written if write() is called. If load is not
56 given or True, the configuration object will load the attributes from
57 the file at filename.
58
59 A few default values will be set here if they are not defined in the
60 config file, or a config file is not loaded. They are documented below.
61 """
62 self.filename = filename
63 """The config object's associated file, as noted above."""
64 self.parser = ConfigParser.RawConfigParser(allow_no_value=True)
65 self.parser.read(self.filename)
66 self.define_section('core', sopel.config.core_section.CoreSection,
67 validate=validate)
68 self.get = self.parser.get
69
70 @property
71 def homedir(self):
72 """An alias to config.core.homedir"""
73 # Technically it's the other way around, so we can bootstrap filename
74 # attributes in the core section, but whatever.
75 configured = None
76 if self.parser.has_option('core', 'homedir'):
77 configured = self.parser.get('core', 'homedir')
78 if configured:
79 return configured
80 else:
81 return os.path.dirname(self.filename)
82
83 def save(self):
84 """Save all changes to the config file."""
85 cfgfile = open(self.filename, 'w')
86 self.parser.write(cfgfile)
87 cfgfile.flush()
88 cfgfile.close()
89
90 def add_section(self, name):
91 """Add a section to the config file.
92
93 Returns ``False`` if already exists.
94 """
95 try:
96 return self.parser.add_section(name)
97 except ConfigParser.DuplicateSectionError:
98 return False
99
100 def define_section(self, name, cls_, validate=True):
101 """Define the available settings in a section.
102
103 ``cls_`` must be a subclass of ``StaticSection``. If the section has
104 already been defined with a different class, ValueError is raised.
105
106 If ``validate`` is True, the section's values will be validated, and an
107 exception raised if they are invalid. This is desirable in a module's
108 setup function, for example, but might not be in the configure function.
109 """
110 if not issubclass(cls_, StaticSection):
111 raise ValueError("Class must be a subclass of StaticSection.")
112 current = getattr(self, name, None)
113 current_name = str(current.__class__)
114 new_name = str(cls_)
115 if (current is not None and not isinstance(current, self.ConfigSection) and
116 not current_name == new_name):
117 raise ValueError(
118 "Can not re-define class for section from {} to {}.".format(
119 current_name, new_name)
120 )
121 setattr(self, name, cls_(self, name, validate=validate))
122
123 class ConfigSection(object):
124
125 """Represents a section of the config file.
126
127 Contains all keys in thesection as attributes.
128
129 """
130
131 def __init__(self, name, items, parent):
132 object.__setattr__(self, '_name', name)
133 object.__setattr__(self, '_parent', parent)
134 for item in items:
135 value = item[1].strip()
136 if not value.lower() == 'none':
137 if value.lower() == 'false':
138 value = False
139 object.__setattr__(self, item[0], value)
140
141 def __getattr__(self, name):
142 return None
143
144 def __setattr__(self, name, value):
145 object.__setattr__(self, name, value)
146 if type(value) is list:
147 value = ','.join(value)
148 self._parent.parser.set(self._name, name, value)
149
150 def get_list(self, name):
151 value = getattr(self, name)
152 if not value:
153 return []
154 if isinstance(value, basestring):
155 value = value.split(',')
156 # Keep the split value, so we don't have to keep doing this
157 setattr(self, name, value)
158 return value
159
160 def __getattr__(self, name):
161 if name in self.parser.sections():
162 items = self.parser.items(name)
163 section = self.ConfigSection(name, items, self) # Return a section
164 setattr(self, name, section)
165 return section
166 else:
167 raise AttributeError("%r object has no attribute %r"
168 % (type(self).__name__, name))
169
170 def option(self, question, default=False):
171 """Ask "y/n" and return the corresponding boolean answer.
172
173 Show user in terminal a "y/n" prompt, and return true or false based on
174 the response. If default is passed as true, the default will be shown
175 as ``[y]``, else it will be ``[n]``. ``question`` should be phrased as
176 a question, but without a question mark at the end.
177
178 """
179 d = 'n'
180 if default:
181 d = 'y'
182 ans = get_input(question + ' (y/n)? [' + d + '] ')
183 if not ans:
184 ans = d
185 return ans.lower() == 'y'
186
187 def _modules(self):
188 home = os.getcwd()
189 modules_dir = os.path.join(home, 'modules')
190 filenames = sopel.loader.enumerate_modules(self)
191 os.sys.path.insert(0, modules_dir)
192 for name, mod_spec in iteritems(filenames):
193 path, type_ = mod_spec
194 try:
195 module, _ = sopel.loader.load_module(name, path, type_)
196 except Exception as e:
197 filename, lineno = sopel.tools.get_raising_file_and_line()
198 rel_path = os.path.relpath(filename, os.path.dirname(__file__))
199 raising_stmt = "%s:%d" % (rel_path, lineno)
200 stderr("Error loading %s: %s (%s)" % (name, e, raising_stmt))
201 else:
202 if hasattr(module, 'configure'):
203 prompt = name + ' module'
204 if module.__doc__:
205 doc = module.__doc__.split('\n', 1)[0]
206 if doc:
207 prompt = doc
208 prompt = 'Configure {} (y/n)? [n]'.format(prompt)
209 do_configure = get_input(prompt)
210 do_configure = do_configure and do_configure.lower() == 'y'
211 if do_configure:
212 module.configure(self)
213 self.save()
214
215
216 def _wizard(section, config=None):
217 dotdir = os.path.expanduser('~/.sopel')
218 configpath = os.path.join(dotdir, ((config or 'default.cfg') + ('.cfg' if config and not config.endswith('.cfg') else '')))
219 if section == 'all':
220 _create_config(configpath)
221 elif section == 'mod':
222 _check_dir(False)
223 if not os.path.isfile(configpath):
224 print("No config file found." +
225 " Please make one before configuring these options.")
226 sys.exit(1)
227 config = Config(configpath, validate=False)
228 config._modules()
229
230
231 def _check_dir(create=True):
232 dotdir = os.path.join(os.path.expanduser('~'), '.sopel')
233 if not os.path.isdir(dotdir):
234 if create:
235 print('Creating a config directory at ~/.sopel...')
236 try:
237 os.makedirs(dotdir)
238 except Exception as e:
239 print('There was a problem creating %s:' % dotdir, file=sys.stderr)
240 print('%s, %s' % (e.__class__, str(e)), file=sys.stderr)
241 print('Please fix this and then run Sopel again.', file=sys.stderr)
242 sys.exit(1)
243 else:
244 print("No config file found. Please make one before configuring these options.")
245 sys.exit(1)
246
247
248 def _create_config(configpath):
249 _check_dir()
250 print("Please answer the following questions" +
251 " to create your configuration file:\n")
252 try:
253 config = Config(configpath, validate=False)
254 sopel.config.core_section.configure(config)
255 if config.option(
256 'Would you like to see if there are any modules'
257 ' that need configuring'
258 ):
259 config._modules()
260 config.save()
261 except Exception: # TODO: Be specific
262 print("Encountered an error while writing the config file." +
263 " This shouldn't happen. Check permissions.")
264 raise
265
266 print("Config file written successfully!")
267 return config.filename
268
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/sopel/config/__init__.py b/sopel/config/__init__.py
--- a/sopel/config/__init__.py
+++ b/sopel/config/__init__.py
@@ -214,12 +214,12 @@
def _wizard(section, config=None):
- dotdir = os.path.expanduser('~/.sopel')
+ dotdir = os.path.dirname(config) if config is not None else DEFAULT_HOMEDIR
configpath = os.path.join(dotdir, ((config or 'default.cfg') + ('.cfg' if config and not config.endswith('.cfg') else '')))
if section == 'all':
_create_config(configpath)
elif section == 'mod':
- _check_dir(False)
+ _check_dir(dotdir, False)
if not os.path.isfile(configpath):
print("No config file found." +
" Please make one before configuring these options.")
@@ -228,15 +228,14 @@
config._modules()
-def _check_dir(create=True):
- dotdir = os.path.join(os.path.expanduser('~'), '.sopel')
- if not os.path.isdir(dotdir):
+def _check_dir(path=DEFAULT_HOMEDIR, create=True):
+ if not os.path.isdir(path):
if create:
- print('Creating a config directory at ~/.sopel...')
+ print('Creating a config directory at {}...'.format(path))
try:
- os.makedirs(dotdir)
+ os.makedirs(path)
except Exception as e:
- print('There was a problem creating %s:' % dotdir, file=sys.stderr)
+ print('There was a problem creating %s:' % path, file=sys.stderr)
print('%s, %s' % (e.__class__, str(e)), file=sys.stderr)
print('Please fix this and then run Sopel again.', file=sys.stderr)
sys.exit(1)
@@ -246,7 +245,7 @@
def _create_config(configpath):
- _check_dir()
+ _check_dir(os.path.dirname(configpath))
print("Please answer the following questions" +
" to create your configuration file:\n")
try:
|
{"golden_diff": "diff --git a/sopel/config/__init__.py b/sopel/config/__init__.py\n--- a/sopel/config/__init__.py\n+++ b/sopel/config/__init__.py\n@@ -214,12 +214,12 @@\n \n \n def _wizard(section, config=None):\n- dotdir = os.path.expanduser('~/.sopel')\n+ dotdir = os.path.dirname(config) if config is not None else DEFAULT_HOMEDIR\n configpath = os.path.join(dotdir, ((config or 'default.cfg') + ('.cfg' if config and not config.endswith('.cfg') else '')))\n if section == 'all':\n _create_config(configpath)\n elif section == 'mod':\n- _check_dir(False)\n+ _check_dir(dotdir, False)\n if not os.path.isfile(configpath):\n print(\"No config file found.\" +\n \" Please make one before configuring these options.\")\n@@ -228,15 +228,14 @@\n config._modules()\n \n \n-def _check_dir(create=True):\n- dotdir = os.path.join(os.path.expanduser('~'), '.sopel')\n- if not os.path.isdir(dotdir):\n+def _check_dir(path=DEFAULT_HOMEDIR, create=True):\n+ if not os.path.isdir(path):\n if create:\n- print('Creating a config directory at ~/.sopel...')\n+ print('Creating a config directory at {}...'.format(path))\n try:\n- os.makedirs(dotdir)\n+ os.makedirs(path)\n except Exception as e:\n- print('There was a problem creating %s:' % dotdir, file=sys.stderr)\n+ print('There was a problem creating %s:' % path, file=sys.stderr)\n print('%s, %s' % (e.__class__, str(e)), file=sys.stderr)\n print('Please fix this and then run Sopel again.', file=sys.stderr)\n sys.exit(1)\n@@ -246,7 +245,7 @@\n \n \n def _create_config(configpath):\n- _check_dir()\n+ _check_dir(os.path.dirname(configpath))\n print(\"Please answer the following questions\" +\n \" to create your configuration file:\\n\")\n try:\n", "issue": "Option to specify alternate config directory on first run\nWhen the bot is first run, it executes ```sopel/run_script.py``` at startup.\r\nSaid script is hardcoded to create ```~/.sopel``` if it doesn't already exist.\r\nShould there be an option to allow the user to specify an alternate directory to use?\r\n\r\nThis was observed on d9cfb41 running ```python sopel.py``` without prior configuration.\n", "before_files": [{"content": "# coding=utf-8\n\"\"\"\nThe config object provides a simplified to access Sopel's configuration file.\nThe sections of the file are attributes of the object, and the keys in the\nsection are attributes of that. So, for example, the ``eggs`` attribute in the\n``[spam]`` section can be accessed from ``config.spam.eggs``.\n\nSection definitions (see \"Section configuration sections\" below) can be added\nto the config object with ``define_section``. When this is done, only the\ndefined keys will be available. A section can not be given more than one\ndefinition. The ``[core]`` section is defined with ``CoreSection`` when the\nobject is initialized.\n\n.. versionadded:: 6.0.0\n\"\"\"\n# Copyright 2012-2015, Elsie Powell, embolalia.com\n# Copyright \u00a9 2012, Elad Alfassa <[email protected]>\n# Licensed under the Eiffel Forum License 2.\n\nfrom __future__ import unicode_literals, absolute_import, print_function, division\n\nfrom sopel.tools import iteritems, stderr\nimport sopel.tools\nfrom sopel.tools import get_input\nimport sopel.loader\nimport os\nimport sys\nif sys.version_info.major < 3:\n import ConfigParser\nelse:\n basestring = str\n import configparser as ConfigParser\nimport sopel.config.core_section\nfrom sopel.config.types import StaticSection\n\n\nDEFAULT_HOMEDIR = os.path.join(os.path.expanduser('~'), '.sopel')\n\n\nclass ConfigurationError(Exception):\n \"\"\" Exception type for configuration errors \"\"\"\n\n def __init__(self, value):\n self.value = value\n\n def __str__(self):\n return 'ConfigurationError: %s' % self.value\n\n\nclass Config(object):\n def __init__(self, filename, validate=True):\n \"\"\"The bot's configuration.\n\n The given filename will be associated with the configuration, and is\n the file which will be written if write() is called. If load is not\n given or True, the configuration object will load the attributes from\n the file at filename.\n\n A few default values will be set here if they are not defined in the\n config file, or a config file is not loaded. They are documented below.\n \"\"\"\n self.filename = filename\n \"\"\"The config object's associated file, as noted above.\"\"\"\n self.parser = ConfigParser.RawConfigParser(allow_no_value=True)\n self.parser.read(self.filename)\n self.define_section('core', sopel.config.core_section.CoreSection,\n validate=validate)\n self.get = self.parser.get\n\n @property\n def homedir(self):\n \"\"\"An alias to config.core.homedir\"\"\"\n # Technically it's the other way around, so we can bootstrap filename\n # attributes in the core section, but whatever.\n configured = None\n if self.parser.has_option('core', 'homedir'):\n configured = self.parser.get('core', 'homedir')\n if configured:\n return configured\n else:\n return os.path.dirname(self.filename)\n\n def save(self):\n \"\"\"Save all changes to the config file.\"\"\"\n cfgfile = open(self.filename, 'w')\n self.parser.write(cfgfile)\n cfgfile.flush()\n cfgfile.close()\n\n def add_section(self, name):\n \"\"\"Add a section to the config file.\n\n Returns ``False`` if already exists.\n \"\"\"\n try:\n return self.parser.add_section(name)\n except ConfigParser.DuplicateSectionError:\n return False\n\n def define_section(self, name, cls_, validate=True):\n \"\"\"Define the available settings in a section.\n\n ``cls_`` must be a subclass of ``StaticSection``. If the section has\n already been defined with a different class, ValueError is raised.\n\n If ``validate`` is True, the section's values will be validated, and an\n exception raised if they are invalid. This is desirable in a module's\n setup function, for example, but might not be in the configure function.\n \"\"\"\n if not issubclass(cls_, StaticSection):\n raise ValueError(\"Class must be a subclass of StaticSection.\")\n current = getattr(self, name, None)\n current_name = str(current.__class__)\n new_name = str(cls_)\n if (current is not None and not isinstance(current, self.ConfigSection) and\n not current_name == new_name):\n raise ValueError(\n \"Can not re-define class for section from {} to {}.\".format(\n current_name, new_name)\n )\n setattr(self, name, cls_(self, name, validate=validate))\n\n class ConfigSection(object):\n\n \"\"\"Represents a section of the config file.\n\n Contains all keys in thesection as attributes.\n\n \"\"\"\n\n def __init__(self, name, items, parent):\n object.__setattr__(self, '_name', name)\n object.__setattr__(self, '_parent', parent)\n for item in items:\n value = item[1].strip()\n if not value.lower() == 'none':\n if value.lower() == 'false':\n value = False\n object.__setattr__(self, item[0], value)\n\n def __getattr__(self, name):\n return None\n\n def __setattr__(self, name, value):\n object.__setattr__(self, name, value)\n if type(value) is list:\n value = ','.join(value)\n self._parent.parser.set(self._name, name, value)\n\n def get_list(self, name):\n value = getattr(self, name)\n if not value:\n return []\n if isinstance(value, basestring):\n value = value.split(',')\n # Keep the split value, so we don't have to keep doing this\n setattr(self, name, value)\n return value\n\n def __getattr__(self, name):\n if name in self.parser.sections():\n items = self.parser.items(name)\n section = self.ConfigSection(name, items, self) # Return a section\n setattr(self, name, section)\n return section\n else:\n raise AttributeError(\"%r object has no attribute %r\"\n % (type(self).__name__, name))\n\n def option(self, question, default=False):\n \"\"\"Ask \"y/n\" and return the corresponding boolean answer.\n\n Show user in terminal a \"y/n\" prompt, and return true or false based on\n the response. If default is passed as true, the default will be shown\n as ``[y]``, else it will be ``[n]``. ``question`` should be phrased as\n a question, but without a question mark at the end.\n\n \"\"\"\n d = 'n'\n if default:\n d = 'y'\n ans = get_input(question + ' (y/n)? [' + d + '] ')\n if not ans:\n ans = d\n return ans.lower() == 'y'\n\n def _modules(self):\n home = os.getcwd()\n modules_dir = os.path.join(home, 'modules')\n filenames = sopel.loader.enumerate_modules(self)\n os.sys.path.insert(0, modules_dir)\n for name, mod_spec in iteritems(filenames):\n path, type_ = mod_spec\n try:\n module, _ = sopel.loader.load_module(name, path, type_)\n except Exception as e:\n filename, lineno = sopel.tools.get_raising_file_and_line()\n rel_path = os.path.relpath(filename, os.path.dirname(__file__))\n raising_stmt = \"%s:%d\" % (rel_path, lineno)\n stderr(\"Error loading %s: %s (%s)\" % (name, e, raising_stmt))\n else:\n if hasattr(module, 'configure'):\n prompt = name + ' module'\n if module.__doc__:\n doc = module.__doc__.split('\\n', 1)[0]\n if doc:\n prompt = doc\n prompt = 'Configure {} (y/n)? [n]'.format(prompt)\n do_configure = get_input(prompt)\n do_configure = do_configure and do_configure.lower() == 'y'\n if do_configure:\n module.configure(self)\n self.save()\n\n\ndef _wizard(section, config=None):\n dotdir = os.path.expanduser('~/.sopel')\n configpath = os.path.join(dotdir, ((config or 'default.cfg') + ('.cfg' if config and not config.endswith('.cfg') else '')))\n if section == 'all':\n _create_config(configpath)\n elif section == 'mod':\n _check_dir(False)\n if not os.path.isfile(configpath):\n print(\"No config file found.\" +\n \" Please make one before configuring these options.\")\n sys.exit(1)\n config = Config(configpath, validate=False)\n config._modules()\n\n\ndef _check_dir(create=True):\n dotdir = os.path.join(os.path.expanduser('~'), '.sopel')\n if not os.path.isdir(dotdir):\n if create:\n print('Creating a config directory at ~/.sopel...')\n try:\n os.makedirs(dotdir)\n except Exception as e:\n print('There was a problem creating %s:' % dotdir, file=sys.stderr)\n print('%s, %s' % (e.__class__, str(e)), file=sys.stderr)\n print('Please fix this and then run Sopel again.', file=sys.stderr)\n sys.exit(1)\n else:\n print(\"No config file found. Please make one before configuring these options.\")\n sys.exit(1)\n\n\ndef _create_config(configpath):\n _check_dir()\n print(\"Please answer the following questions\" +\n \" to create your configuration file:\\n\")\n try:\n config = Config(configpath, validate=False)\n sopel.config.core_section.configure(config)\n if config.option(\n 'Would you like to see if there are any modules'\n ' that need configuring'\n ):\n config._modules()\n config.save()\n except Exception: # TODO: Be specific\n print(\"Encountered an error while writing the config file.\" +\n \" This shouldn't happen. Check permissions.\")\n raise\n\n print(\"Config file written successfully!\")\n return config.filename\n", "path": "sopel/config/__init__.py"}], "after_files": [{"content": "# coding=utf-8\n\"\"\"\nThe config object provides a simplified to access Sopel's configuration file.\nThe sections of the file are attributes of the object, and the keys in the\nsection are attributes of that. So, for example, the ``eggs`` attribute in the\n``[spam]`` section can be accessed from ``config.spam.eggs``.\n\nSection definitions (see \"Section configuration sections\" below) can be added\nto the config object with ``define_section``. When this is done, only the\ndefined keys will be available. A section can not be given more than one\ndefinition. The ``[core]`` section is defined with ``CoreSection`` when the\nobject is initialized.\n\n.. versionadded:: 6.0.0\n\"\"\"\n# Copyright 2012-2015, Elsie Powell, embolalia.com\n# Copyright \u00a9 2012, Elad Alfassa <[email protected]>\n# Licensed under the Eiffel Forum License 2.\n\nfrom __future__ import unicode_literals, absolute_import, print_function, division\n\nfrom sopel.tools import iteritems, stderr\nimport sopel.tools\nfrom sopel.tools import get_input\nimport sopel.loader\nimport os\nimport sys\nif sys.version_info.major < 3:\n import ConfigParser\nelse:\n basestring = str\n import configparser as ConfigParser\nimport sopel.config.core_section\nfrom sopel.config.types import StaticSection\n\n\nDEFAULT_HOMEDIR = os.path.join(os.path.expanduser('~'), '.sopel')\n\n\nclass ConfigurationError(Exception):\n \"\"\" Exception type for configuration errors \"\"\"\n\n def __init__(self, value):\n self.value = value\n\n def __str__(self):\n return 'ConfigurationError: %s' % self.value\n\n\nclass Config(object):\n def __init__(self, filename, validate=True):\n \"\"\"The bot's configuration.\n\n The given filename will be associated with the configuration, and is\n the file which will be written if write() is called. If load is not\n given or True, the configuration object will load the attributes from\n the file at filename.\n\n A few default values will be set here if they are not defined in the\n config file, or a config file is not loaded. They are documented below.\n \"\"\"\n self.filename = filename\n \"\"\"The config object's associated file, as noted above.\"\"\"\n self.parser = ConfigParser.RawConfigParser(allow_no_value=True)\n self.parser.read(self.filename)\n self.define_section('core', sopel.config.core_section.CoreSection,\n validate=validate)\n self.get = self.parser.get\n\n @property\n def homedir(self):\n \"\"\"An alias to config.core.homedir\"\"\"\n # Technically it's the other way around, so we can bootstrap filename\n # attributes in the core section, but whatever.\n configured = None\n if self.parser.has_option('core', 'homedir'):\n configured = self.parser.get('core', 'homedir')\n if configured:\n return configured\n else:\n return os.path.dirname(self.filename)\n\n def save(self):\n \"\"\"Save all changes to the config file.\"\"\"\n cfgfile = open(self.filename, 'w')\n self.parser.write(cfgfile)\n cfgfile.flush()\n cfgfile.close()\n\n def add_section(self, name):\n \"\"\"Add a section to the config file.\n\n Returns ``False`` if already exists.\n \"\"\"\n try:\n return self.parser.add_section(name)\n except ConfigParser.DuplicateSectionError:\n return False\n\n def define_section(self, name, cls_, validate=True):\n \"\"\"Define the available settings in a section.\n\n ``cls_`` must be a subclass of ``StaticSection``. If the section has\n already been defined with a different class, ValueError is raised.\n\n If ``validate`` is True, the section's values will be validated, and an\n exception raised if they are invalid. This is desirable in a module's\n setup function, for example, but might not be in the configure function.\n \"\"\"\n if not issubclass(cls_, StaticSection):\n raise ValueError(\"Class must be a subclass of StaticSection.\")\n current = getattr(self, name, None)\n current_name = str(current.__class__)\n new_name = str(cls_)\n if (current is not None and not isinstance(current, self.ConfigSection) and\n not current_name == new_name):\n raise ValueError(\n \"Can not re-define class for section from {} to {}.\".format(\n current_name, new_name)\n )\n setattr(self, name, cls_(self, name, validate=validate))\n\n class ConfigSection(object):\n\n \"\"\"Represents a section of the config file.\n\n Contains all keys in thesection as attributes.\n\n \"\"\"\n\n def __init__(self, name, items, parent):\n object.__setattr__(self, '_name', name)\n object.__setattr__(self, '_parent', parent)\n for item in items:\n value = item[1].strip()\n if not value.lower() == 'none':\n if value.lower() == 'false':\n value = False\n object.__setattr__(self, item[0], value)\n\n def __getattr__(self, name):\n return None\n\n def __setattr__(self, name, value):\n object.__setattr__(self, name, value)\n if type(value) is list:\n value = ','.join(value)\n self._parent.parser.set(self._name, name, value)\n\n def get_list(self, name):\n value = getattr(self, name)\n if not value:\n return []\n if isinstance(value, basestring):\n value = value.split(',')\n # Keep the split value, so we don't have to keep doing this\n setattr(self, name, value)\n return value\n\n def __getattr__(self, name):\n if name in self.parser.sections():\n items = self.parser.items(name)\n section = self.ConfigSection(name, items, self) # Return a section\n setattr(self, name, section)\n return section\n else:\n raise AttributeError(\"%r object has no attribute %r\"\n % (type(self).__name__, name))\n\n def option(self, question, default=False):\n \"\"\"Ask \"y/n\" and return the corresponding boolean answer.\n\n Show user in terminal a \"y/n\" prompt, and return true or false based on\n the response. If default is passed as true, the default will be shown\n as ``[y]``, else it will be ``[n]``. ``question`` should be phrased as\n a question, but without a question mark at the end.\n\n \"\"\"\n d = 'n'\n if default:\n d = 'y'\n ans = get_input(question + ' (y/n)? [' + d + '] ')\n if not ans:\n ans = d\n return ans.lower() == 'y'\n\n def _modules(self):\n home = os.getcwd()\n modules_dir = os.path.join(home, 'modules')\n filenames = sopel.loader.enumerate_modules(self)\n os.sys.path.insert(0, modules_dir)\n for name, mod_spec in iteritems(filenames):\n path, type_ = mod_spec\n try:\n module, _ = sopel.loader.load_module(name, path, type_)\n except Exception as e:\n filename, lineno = sopel.tools.get_raising_file_and_line()\n rel_path = os.path.relpath(filename, os.path.dirname(__file__))\n raising_stmt = \"%s:%d\" % (rel_path, lineno)\n stderr(\"Error loading %s: %s (%s)\" % (name, e, raising_stmt))\n else:\n if hasattr(module, 'configure'):\n prompt = name + ' module'\n if module.__doc__:\n doc = module.__doc__.split('\\n', 1)[0]\n if doc:\n prompt = doc\n prompt = 'Configure {} (y/n)? [n]'.format(prompt)\n do_configure = get_input(prompt)\n do_configure = do_configure and do_configure.lower() == 'y'\n if do_configure:\n module.configure(self)\n self.save()\n\n\ndef _wizard(section, config=None):\n dotdir = os.path.dirname(config) if config is not None else DEFAULT_HOMEDIR\n configpath = os.path.join(dotdir, ((config or 'default.cfg') + ('.cfg' if config and not config.endswith('.cfg') else '')))\n if section == 'all':\n _create_config(configpath)\n elif section == 'mod':\n _check_dir(dotdir, False)\n if not os.path.isfile(configpath):\n print(\"No config file found.\" +\n \" Please make one before configuring these options.\")\n sys.exit(1)\n config = Config(configpath, validate=False)\n config._modules()\n\n\ndef _check_dir(path=DEFAULT_HOMEDIR, create=True):\n if not os.path.isdir(path):\n if create:\n print('Creating a config directory at {}...'.format(path))\n try:\n os.makedirs(path)\n except Exception as e:\n print('There was a problem creating %s:' % path, file=sys.stderr)\n print('%s, %s' % (e.__class__, str(e)), file=sys.stderr)\n print('Please fix this and then run Sopel again.', file=sys.stderr)\n sys.exit(1)\n else:\n print(\"No config file found. Please make one before configuring these options.\")\n sys.exit(1)\n\n\ndef _create_config(configpath):\n _check_dir(os.path.dirname(configpath))\n print(\"Please answer the following questions\" +\n \" to create your configuration file:\\n\")\n try:\n config = Config(configpath, validate=False)\n sopel.config.core_section.configure(config)\n if config.option(\n 'Would you like to see if there are any modules'\n ' that need configuring'\n ):\n config._modules()\n config.save()\n except Exception: # TODO: Be specific\n print(\"Encountered an error while writing the config file.\" +\n \" This shouldn't happen. Check permissions.\")\n raise\n\n print(\"Config file written successfully!\")\n return config.filename\n", "path": "sopel/config/__init__.py"}]}
| 3,258 | 490 |
gh_patches_debug_40285
|
rasdani/github-patches
|
git_diff
|
optuna__optuna-3115
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
[RFC] Show progress bar based on `timeout`
<!-- Please write a clear and concise description of the feature proposal. -->
## Motivation
The current [`study.optimize`](https://optuna.readthedocs.io/en/stable/reference/generated/optuna.study.Study.html#optuna.study.Study.optimize) shows the progress bar when we specify the number of trials. We can show the progress bar based on the `timeout` value as well.
Note that before implementing this feature, we need to decide the priority of `n_trials` and `timeout`. When we specify both values as not `None` values, there are two ways to show the progress bar.
Moreover, we need to discuss the case when both `None`. I suppose not showing the progress bar makes sense when both are `None`.
## Description
<!-- Please write a detailed description of the new feature. -->
- Show progress bar using `timeout` value.
- Add test by following https://github.com/optuna/optuna/pull/2900
## Alternatives (optional)
<!-- Please write a clear and concise description of any alternative solutions or features you've considered. -->
## Additional context (optional)
We might need to support the `n_jobs!=1` case depending on https://github.com/optuna/optuna/issues/2957.
<!-- Please add any other context or screenshots about the feature request here. -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `optuna/progress_bar.py`
Content:
```
1 import logging
2 from typing import Any
3 from typing import Optional
4
5 from tqdm.auto import tqdm
6
7 from optuna import logging as optuna_logging
8 from optuna._experimental import experimental
9
10
11 _tqdm_handler: Optional["_TqdmLoggingHandler"] = None
12
13
14 # Reference: https://gist.github.com/hvy/8b80c2cedf02b15c24f85d1fa17ebe02
15 class _TqdmLoggingHandler(logging.StreamHandler):
16 def emit(self, record: Any) -> None:
17 try:
18 msg = self.format(record)
19 tqdm.write(msg)
20 self.flush()
21 except (KeyboardInterrupt, SystemExit):
22 raise
23 except Exception:
24 self.handleError(record)
25
26
27 class _ProgressBar(object):
28 """Progress Bar implementation for :func:`~optuna.study.Study.optimize` on the top of `tqdm`.
29
30 Args:
31 is_valid:
32 Whether to show progress bars in :func:`~optuna.study.Study.optimize`.
33 n_trials:
34 The number of trials.
35 timeout:
36 Stop study after the given number of second(s).
37 """
38
39 def __init__(
40 self, is_valid: bool, n_trials: Optional[int] = None, timeout: Optional[float] = None
41 ) -> None:
42 self._is_valid = is_valid
43 self._n_trials = n_trials
44 self._timeout = timeout
45
46 if self._is_valid:
47 self._init_valid()
48
49 # TODO(hvy): Remove initialization indirection via this method when the progress bar is no
50 # longer experimental.
51 @experimental("1.2.0", name="Progress bar")
52 def _init_valid(self) -> None:
53 self._progress_bar = tqdm(range(self._n_trials) if self._n_trials is not None else None)
54 global _tqdm_handler
55
56 _tqdm_handler = _TqdmLoggingHandler()
57 _tqdm_handler.setLevel(logging.INFO)
58 _tqdm_handler.setFormatter(optuna_logging.create_default_formatter())
59 optuna_logging.disable_default_handler()
60 optuna_logging._get_library_root_logger().addHandler(_tqdm_handler)
61
62 def update(self, elapsed_seconds: Optional[float]) -> None:
63 """Update the progress bars if ``is_valid`` is :obj:`True`.
64
65 Args:
66 elapsed_seconds:
67 The time past since :func:`~optuna.study.Study.optimize` started.
68 """
69 if self._is_valid:
70 self._progress_bar.update(1)
71 if self._timeout is not None and elapsed_seconds is not None:
72 self._progress_bar.set_postfix_str(
73 "{:.02f}/{} seconds".format(elapsed_seconds, self._timeout)
74 )
75
76 def close(self) -> None:
77 """Close progress bars."""
78 if self._is_valid:
79 self._progress_bar.close()
80 assert _tqdm_handler is not None
81 optuna_logging._get_library_root_logger().removeHandler(_tqdm_handler)
82 optuna_logging.enable_default_handler()
83
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/optuna/progress_bar.py b/optuna/progress_bar.py
--- a/optuna/progress_bar.py
+++ b/optuna/progress_bar.py
@@ -39,9 +39,11 @@
def __init__(
self, is_valid: bool, n_trials: Optional[int] = None, timeout: Optional[float] = None
) -> None:
- self._is_valid = is_valid
+
+ self._is_valid = is_valid and (n_trials or timeout) is not None
self._n_trials = n_trials
self._timeout = timeout
+ self._last_elapsed_seconds = 0.0
if self._is_valid:
self._init_valid()
@@ -50,7 +52,21 @@
# longer experimental.
@experimental("1.2.0", name="Progress bar")
def _init_valid(self) -> None:
- self._progress_bar = tqdm(range(self._n_trials) if self._n_trials is not None else None)
+
+ if self._n_trials is not None:
+ self._progress_bar = tqdm(total=self._n_trials)
+
+ else:
+ fmt = "{percentage:3.0f}%|{bar}| {elapsed}/{desc}"
+ self._progress_bar = tqdm(total=self._timeout, bar_format=fmt)
+
+ # Using description string instead postfix string
+ # to display formatted timeout, since postfix carries
+ # extra comma space auto-format.
+ # https://github.com/tqdm/tqdm/issues/712
+ total = tqdm.format_interval(self._timeout)
+ self._progress_bar.set_description_str(total)
+
global _tqdm_handler
_tqdm_handler = _TqdmLoggingHandler()
@@ -59,22 +75,37 @@
optuna_logging.disable_default_handler()
optuna_logging._get_library_root_logger().addHandler(_tqdm_handler)
- def update(self, elapsed_seconds: Optional[float]) -> None:
+ def update(self, elapsed_seconds: float) -> None:
"""Update the progress bars if ``is_valid`` is :obj:`True`.
Args:
elapsed_seconds:
The time past since :func:`~optuna.study.Study.optimize` started.
"""
+
if self._is_valid:
- self._progress_bar.update(1)
- if self._timeout is not None and elapsed_seconds is not None:
- self._progress_bar.set_postfix_str(
- "{:.02f}/{} seconds".format(elapsed_seconds, self._timeout)
- )
+ if self._n_trials is not None:
+ self._progress_bar.update(1)
+ if self._timeout is not None:
+ self._progress_bar.set_postfix_str(
+ "{:.02f}/{} seconds".format(elapsed_seconds, self._timeout)
+ )
+
+ elif self._timeout is not None:
+ time_diff = elapsed_seconds - self._last_elapsed_seconds
+ if elapsed_seconds > self._timeout:
+ # Clip elapsed time to avoid tqdm warnings.
+ time_diff -= elapsed_seconds - self._timeout
+
+ self._progress_bar.update(time_diff)
+ self._last_elapsed_seconds = elapsed_seconds
+
+ else:
+ assert False
def close(self) -> None:
"""Close progress bars."""
+
if self._is_valid:
self._progress_bar.close()
assert _tqdm_handler is not None
|
{"golden_diff": "diff --git a/optuna/progress_bar.py b/optuna/progress_bar.py\n--- a/optuna/progress_bar.py\n+++ b/optuna/progress_bar.py\n@@ -39,9 +39,11 @@\n def __init__(\n self, is_valid: bool, n_trials: Optional[int] = None, timeout: Optional[float] = None\n ) -> None:\n- self._is_valid = is_valid\n+\n+ self._is_valid = is_valid and (n_trials or timeout) is not None\n self._n_trials = n_trials\n self._timeout = timeout\n+ self._last_elapsed_seconds = 0.0\n \n if self._is_valid:\n self._init_valid()\n@@ -50,7 +52,21 @@\n # longer experimental.\n @experimental(\"1.2.0\", name=\"Progress bar\")\n def _init_valid(self) -> None:\n- self._progress_bar = tqdm(range(self._n_trials) if self._n_trials is not None else None)\n+\n+ if self._n_trials is not None:\n+ self._progress_bar = tqdm(total=self._n_trials)\n+\n+ else:\n+ fmt = \"{percentage:3.0f}%|{bar}| {elapsed}/{desc}\"\n+ self._progress_bar = tqdm(total=self._timeout, bar_format=fmt)\n+\n+ # Using description string instead postfix string\n+ # to display formatted timeout, since postfix carries\n+ # extra comma space auto-format.\n+ # https://github.com/tqdm/tqdm/issues/712\n+ total = tqdm.format_interval(self._timeout)\n+ self._progress_bar.set_description_str(total)\n+\n global _tqdm_handler\n \n _tqdm_handler = _TqdmLoggingHandler()\n@@ -59,22 +75,37 @@\n optuna_logging.disable_default_handler()\n optuna_logging._get_library_root_logger().addHandler(_tqdm_handler)\n \n- def update(self, elapsed_seconds: Optional[float]) -> None:\n+ def update(self, elapsed_seconds: float) -> None:\n \"\"\"Update the progress bars if ``is_valid`` is :obj:`True`.\n \n Args:\n elapsed_seconds:\n The time past since :func:`~optuna.study.Study.optimize` started.\n \"\"\"\n+\n if self._is_valid:\n- self._progress_bar.update(1)\n- if self._timeout is not None and elapsed_seconds is not None:\n- self._progress_bar.set_postfix_str(\n- \"{:.02f}/{} seconds\".format(elapsed_seconds, self._timeout)\n- )\n+ if self._n_trials is not None:\n+ self._progress_bar.update(1)\n+ if self._timeout is not None:\n+ self._progress_bar.set_postfix_str(\n+ \"{:.02f}/{} seconds\".format(elapsed_seconds, self._timeout)\n+ )\n+\n+ elif self._timeout is not None:\n+ time_diff = elapsed_seconds - self._last_elapsed_seconds\n+ if elapsed_seconds > self._timeout:\n+ # Clip elapsed time to avoid tqdm warnings.\n+ time_diff -= elapsed_seconds - self._timeout\n+\n+ self._progress_bar.update(time_diff)\n+ self._last_elapsed_seconds = elapsed_seconds\n+\n+ else:\n+ assert False\n \n def close(self) -> None:\n \"\"\"Close progress bars.\"\"\"\n+\n if self._is_valid:\n self._progress_bar.close()\n assert _tqdm_handler is not None\n", "issue": "[RFC] Show progress bar based on `timeout`\n<!-- Please write a clear and concise description of the feature proposal. -->\r\n\r\n## Motivation\r\n\r\nThe current [`study.optimize`](https://optuna.readthedocs.io/en/stable/reference/generated/optuna.study.Study.html#optuna.study.Study.optimize) shows the progress bar when we specify the number of trials. We can show the progress bar based on the `timeout` value as well.\r\n\r\nNote that before implementing this feature, we need to decide the priority of `n_trials` and `timeout`. When we specify both values as not `None` values, there are two ways to show the progress bar. \r\n\r\nMoreover, we need to discuss the case when both `None`. I suppose not showing the progress bar makes sense when both are `None`.\r\n\r\n## Description\r\n\r\n<!-- Please write a detailed description of the new feature. -->\r\n\r\n\r\n- Show progress bar using `timeout` value.\r\n- Add test by following https://github.com/optuna/optuna/pull/2900\r\n\r\n## Alternatives (optional)\r\n\r\n\r\n<!-- Please write a clear and concise description of any alternative solutions or features you've considered. -->\r\n\r\n## Additional context (optional)\r\n\r\nWe might need to support the `n_jobs!=1` case depending on https://github.com/optuna/optuna/issues/2957.\r\n\r\n<!-- Please add any other context or screenshots about the feature request here. -->\r\n\n", "before_files": [{"content": "import logging\nfrom typing import Any\nfrom typing import Optional\n\nfrom tqdm.auto import tqdm\n\nfrom optuna import logging as optuna_logging\nfrom optuna._experimental import experimental\n\n\n_tqdm_handler: Optional[\"_TqdmLoggingHandler\"] = None\n\n\n# Reference: https://gist.github.com/hvy/8b80c2cedf02b15c24f85d1fa17ebe02\nclass _TqdmLoggingHandler(logging.StreamHandler):\n def emit(self, record: Any) -> None:\n try:\n msg = self.format(record)\n tqdm.write(msg)\n self.flush()\n except (KeyboardInterrupt, SystemExit):\n raise\n except Exception:\n self.handleError(record)\n\n\nclass _ProgressBar(object):\n \"\"\"Progress Bar implementation for :func:`~optuna.study.Study.optimize` on the top of `tqdm`.\n\n Args:\n is_valid:\n Whether to show progress bars in :func:`~optuna.study.Study.optimize`.\n n_trials:\n The number of trials.\n timeout:\n Stop study after the given number of second(s).\n \"\"\"\n\n def __init__(\n self, is_valid: bool, n_trials: Optional[int] = None, timeout: Optional[float] = None\n ) -> None:\n self._is_valid = is_valid\n self._n_trials = n_trials\n self._timeout = timeout\n\n if self._is_valid:\n self._init_valid()\n\n # TODO(hvy): Remove initialization indirection via this method when the progress bar is no\n # longer experimental.\n @experimental(\"1.2.0\", name=\"Progress bar\")\n def _init_valid(self) -> None:\n self._progress_bar = tqdm(range(self._n_trials) if self._n_trials is not None else None)\n global _tqdm_handler\n\n _tqdm_handler = _TqdmLoggingHandler()\n _tqdm_handler.setLevel(logging.INFO)\n _tqdm_handler.setFormatter(optuna_logging.create_default_formatter())\n optuna_logging.disable_default_handler()\n optuna_logging._get_library_root_logger().addHandler(_tqdm_handler)\n\n def update(self, elapsed_seconds: Optional[float]) -> None:\n \"\"\"Update the progress bars if ``is_valid`` is :obj:`True`.\n\n Args:\n elapsed_seconds:\n The time past since :func:`~optuna.study.Study.optimize` started.\n \"\"\"\n if self._is_valid:\n self._progress_bar.update(1)\n if self._timeout is not None and elapsed_seconds is not None:\n self._progress_bar.set_postfix_str(\n \"{:.02f}/{} seconds\".format(elapsed_seconds, self._timeout)\n )\n\n def close(self) -> None:\n \"\"\"Close progress bars.\"\"\"\n if self._is_valid:\n self._progress_bar.close()\n assert _tqdm_handler is not None\n optuna_logging._get_library_root_logger().removeHandler(_tqdm_handler)\n optuna_logging.enable_default_handler()\n", "path": "optuna/progress_bar.py"}], "after_files": [{"content": "import logging\nfrom typing import Any\nfrom typing import Optional\n\nfrom tqdm.auto import tqdm\n\nfrom optuna import logging as optuna_logging\nfrom optuna._experimental import experimental\n\n\n_tqdm_handler: Optional[\"_TqdmLoggingHandler\"] = None\n\n\n# Reference: https://gist.github.com/hvy/8b80c2cedf02b15c24f85d1fa17ebe02\nclass _TqdmLoggingHandler(logging.StreamHandler):\n def emit(self, record: Any) -> None:\n try:\n msg = self.format(record)\n tqdm.write(msg)\n self.flush()\n except (KeyboardInterrupt, SystemExit):\n raise\n except Exception:\n self.handleError(record)\n\n\nclass _ProgressBar(object):\n \"\"\"Progress Bar implementation for :func:`~optuna.study.Study.optimize` on the top of `tqdm`.\n\n Args:\n is_valid:\n Whether to show progress bars in :func:`~optuna.study.Study.optimize`.\n n_trials:\n The number of trials.\n timeout:\n Stop study after the given number of second(s).\n \"\"\"\n\n def __init__(\n self, is_valid: bool, n_trials: Optional[int] = None, timeout: Optional[float] = None\n ) -> None:\n\n self._is_valid = is_valid and (n_trials or timeout) is not None\n self._n_trials = n_trials\n self._timeout = timeout\n self._last_elapsed_seconds = 0.0\n\n if self._is_valid:\n self._init_valid()\n\n # TODO(hvy): Remove initialization indirection via this method when the progress bar is no\n # longer experimental.\n @experimental(\"1.2.0\", name=\"Progress bar\")\n def _init_valid(self) -> None:\n\n if self._n_trials is not None:\n self._progress_bar = tqdm(total=self._n_trials)\n\n else:\n fmt = \"{percentage:3.0f}%|{bar}| {elapsed}/{desc}\"\n self._progress_bar = tqdm(total=self._timeout, bar_format=fmt)\n\n # Using description string instead postfix string\n # to display formatted timeout, since postfix carries\n # extra comma space auto-format.\n # https://github.com/tqdm/tqdm/issues/712\n total = tqdm.format_interval(self._timeout)\n self._progress_bar.set_description_str(total)\n\n global _tqdm_handler\n\n _tqdm_handler = _TqdmLoggingHandler()\n _tqdm_handler.setLevel(logging.INFO)\n _tqdm_handler.setFormatter(optuna_logging.create_default_formatter())\n optuna_logging.disable_default_handler()\n optuna_logging._get_library_root_logger().addHandler(_tqdm_handler)\n\n def update(self, elapsed_seconds: float) -> None:\n \"\"\"Update the progress bars if ``is_valid`` is :obj:`True`.\n\n Args:\n elapsed_seconds:\n The time past since :func:`~optuna.study.Study.optimize` started.\n \"\"\"\n\n if self._is_valid:\n if self._n_trials is not None:\n self._progress_bar.update(1)\n if self._timeout is not None:\n self._progress_bar.set_postfix_str(\n \"{:.02f}/{} seconds\".format(elapsed_seconds, self._timeout)\n )\n\n elif self._timeout is not None:\n time_diff = elapsed_seconds - self._last_elapsed_seconds\n if elapsed_seconds > self._timeout:\n # Clip elapsed time to avoid tqdm warnings.\n time_diff -= elapsed_seconds - self._timeout\n\n self._progress_bar.update(time_diff)\n self._last_elapsed_seconds = elapsed_seconds\n\n else:\n assert False\n\n def close(self) -> None:\n \"\"\"Close progress bars.\"\"\"\n\n if self._is_valid:\n self._progress_bar.close()\n assert _tqdm_handler is not None\n optuna_logging._get_library_root_logger().removeHandler(_tqdm_handler)\n optuna_logging.enable_default_handler()\n", "path": "optuna/progress_bar.py"}]}
| 1,363 | 768 |
gh_patches_debug_27844
|
rasdani/github-patches
|
git_diff
|
Zeroto521__my-data-toolkit-713
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
ENH: New geoaccessor for GeoSeries to return tuple of coordinates `(x, y)`
<!--
Thanks for contributing a pull request!
Please follow these standard acronyms to start the commit message:
- ENH: enhancement
- BUG: bug fix
- DOC: documentation
- TYP: type annotations
- TST: addition or modification of tests
- MAINT: maintenance commit (refactoring, typos, etc.)
- BLD: change related to building
- REL: related to releasing
- API: an (incompatible) API change
- DEP: deprecate something, or remove a deprecated object
- DEV: development tool or utility
- REV: revert an earlier commit
- PERF: performance improvement
- BOT: always commit via a bot
- CI: related to CI or CD
- CLN: Code cleanup
-->
- [ ] closes #xxxx
- [x] whatsnew entry
as title
```python
>>> import dtoolkit.geoaccessor
>>> import geopandas as gpd
>>> from shapely.geometry import Point
>>> s = geopandas.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])
>>> s.xy
0 (1.0, 1.0)
1 (2.0, 2.0)
2 (3.0, 3.0)
dtype: object
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `dtoolkit/geoaccessor/geoseries/xy.py`
Content:
```
1 import geopandas as gpd
2 import pandas as pd
3
4 from dtoolkit.geoaccessor.register import register_geoseries_method
5
6
7 @register_geoseries_method
8 def xy(s: gpd.GeoSeries, /) -> pd.Series:
9 """
10 Return the x and y location of Point geometries in a GeoSeries.
11
12 Returns
13 -------
14 Series
15 tuple of x and y coordinates.
16
17 See Also
18 --------
19 geopandas.GeoSeries.x
20 geopandas.GeoSeries.y
21
22 Examples
23 --------
24 >>> import dtoolkit.geoaccessor
25 >>> import geopandas as gpd
26 >>> from shapely.geometry import Point
27 >>> s = gpd.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])
28 >>> s
29 0 POINT (1.00000 1.00000)
30 1 POINT (2.00000 2.00000)
31 2 POINT (3.00000 3.00000)
32 dtype: geometry
33 >>> s.xy()
34 0 (1.0, 1.0)
35 1 (2.0, 2.0)
36 2 (3.0, 3.0)
37 dtype: object
38 """
39
40 return pd.concat((s.x, s.y), axis=1).apply(tuple, axis=1)
41
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/dtoolkit/geoaccessor/geoseries/xy.py b/dtoolkit/geoaccessor/geoseries/xy.py
--- a/dtoolkit/geoaccessor/geoseries/xy.py
+++ b/dtoolkit/geoaccessor/geoseries/xy.py
@@ -5,14 +5,19 @@
@register_geoseries_method
-def xy(s: gpd.GeoSeries, /) -> pd.Series:
+def xy(s: gpd.GeoSeries, /, reverse: bool = False) -> pd.Series:
"""
Return the x and y location of Point geometries in a GeoSeries.
+ Parameters
+ ----------
+ reverse : bool, default False
+ If True, return (y, x) instead.
+
Returns
-------
Series
- tuple of x and y coordinates.
+ tuple of coordinate.
See Also
--------
@@ -24,17 +29,26 @@
>>> import dtoolkit.geoaccessor
>>> import geopandas as gpd
>>> from shapely.geometry import Point
- >>> s = gpd.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])
+ >>> s = gpd.GeoSeries([Point(0, 1), Point(0, 2), Point(0, 3)])
>>> s
- 0 POINT (1.00000 1.00000)
- 1 POINT (2.00000 2.00000)
- 2 POINT (3.00000 3.00000)
+ 0 POINT (0.00000 1.00000)
+ 1 POINT (0.00000 2.00000)
+ 2 POINT (0.00000 3.00000)
dtype: geometry
>>> s.xy()
- 0 (1.0, 1.0)
- 1 (2.0, 2.0)
- 2 (3.0, 3.0)
+ 0 (0.0, 1.0)
+ 1 (0.0, 2.0)
+ 2 (0.0, 3.0)
+ dtype: object
+
+ Set ``reverse=True`` to return (y, x).
+
+ >>> s.xy(True)
+ 0 (1.0, 0.0)
+ 1 (2.0, 0.0)
+ 2 (3.0, 0.0)
dtype: object
"""
- return pd.concat((s.x, s.y), axis=1).apply(tuple, axis=1)
+ coordinates = (s.y, s.x) if reverse else (s.x, s.y)
+ return pd.concat(coordinates, axis=1).apply(tuple, axis=1)
|
{"golden_diff": "diff --git a/dtoolkit/geoaccessor/geoseries/xy.py b/dtoolkit/geoaccessor/geoseries/xy.py\n--- a/dtoolkit/geoaccessor/geoseries/xy.py\n+++ b/dtoolkit/geoaccessor/geoseries/xy.py\n@@ -5,14 +5,19 @@\n \n \n @register_geoseries_method\n-def xy(s: gpd.GeoSeries, /) -> pd.Series:\n+def xy(s: gpd.GeoSeries, /, reverse: bool = False) -> pd.Series:\n \"\"\"\n Return the x and y location of Point geometries in a GeoSeries.\n \n+ Parameters\n+ ----------\n+ reverse : bool, default False\n+ If True, return (y, x) instead.\n+\n Returns\n -------\n Series\n- tuple of x and y coordinates.\n+ tuple of coordinate.\n \n See Also\n --------\n@@ -24,17 +29,26 @@\n >>> import dtoolkit.geoaccessor\n >>> import geopandas as gpd\n >>> from shapely.geometry import Point\n- >>> s = gpd.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])\n+ >>> s = gpd.GeoSeries([Point(0, 1), Point(0, 2), Point(0, 3)])\n >>> s\n- 0 POINT (1.00000 1.00000)\n- 1 POINT (2.00000 2.00000)\n- 2 POINT (3.00000 3.00000)\n+ 0 POINT (0.00000 1.00000)\n+ 1 POINT (0.00000 2.00000)\n+ 2 POINT (0.00000 3.00000)\n dtype: geometry\n >>> s.xy()\n- 0 (1.0, 1.0)\n- 1 (2.0, 2.0)\n- 2 (3.0, 3.0)\n+ 0 (0.0, 1.0)\n+ 1 (0.0, 2.0)\n+ 2 (0.0, 3.0)\n+ dtype: object\n+\n+ Set ``reverse=True`` to return (y, x).\n+\n+ >>> s.xy(True)\n+ 0 (1.0, 0.0)\n+ 1 (2.0, 0.0)\n+ 2 (3.0, 0.0)\n dtype: object\n \"\"\"\n \n- return pd.concat((s.x, s.y), axis=1).apply(tuple, axis=1)\n+ coordinates = (s.y, s.x) if reverse else (s.x, s.y)\n+ return pd.concat(coordinates, axis=1).apply(tuple, axis=1)\n", "issue": "ENH: New geoaccessor for GeoSeries to return tuple of coordinates `(x, y)`\n<!--\r\nThanks for contributing a pull request!\r\n\r\nPlease follow these standard acronyms to start the commit message:\r\n\r\n- ENH: enhancement\r\n- BUG: bug fix\r\n- DOC: documentation\r\n- TYP: type annotations\r\n- TST: addition or modification of tests\r\n- MAINT: maintenance commit (refactoring, typos, etc.)\r\n- BLD: change related to building\r\n- REL: related to releasing\r\n- API: an (incompatible) API change\r\n- DEP: deprecate something, or remove a deprecated object\r\n- DEV: development tool or utility\r\n- REV: revert an earlier commit\r\n- PERF: performance improvement\r\n- BOT: always commit via a bot\r\n- CI: related to CI or CD\r\n- CLN: Code cleanup\r\n-->\r\n\r\n- [ ] closes #xxxx\r\n- [x] whatsnew entry\r\n\r\nas title\r\n\r\n```python\r\n >>> import dtoolkit.geoaccessor\r\n >>> import geopandas as gpd\r\n >>> from shapely.geometry import Point\r\n >>> s = geopandas.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])\r\n >>> s.xy\r\n 0 (1.0, 1.0)\r\n 1 (2.0, 2.0)\r\n 2 (3.0, 3.0)\r\n dtype: object\r\n```\n", "before_files": [{"content": "import geopandas as gpd\nimport pandas as pd\n\nfrom dtoolkit.geoaccessor.register import register_geoseries_method\n\n\n@register_geoseries_method\ndef xy(s: gpd.GeoSeries, /) -> pd.Series:\n \"\"\"\n Return the x and y location of Point geometries in a GeoSeries.\n\n Returns\n -------\n Series\n tuple of x and y coordinates.\n\n See Also\n --------\n geopandas.GeoSeries.x\n geopandas.GeoSeries.y\n\n Examples\n --------\n >>> import dtoolkit.geoaccessor\n >>> import geopandas as gpd\n >>> from shapely.geometry import Point\n >>> s = gpd.GeoSeries([Point(1, 1), Point(2, 2), Point(3, 3)])\n >>> s\n 0 POINT (1.00000 1.00000)\n 1 POINT (2.00000 2.00000)\n 2 POINT (3.00000 3.00000)\n dtype: geometry\n >>> s.xy()\n 0 (1.0, 1.0)\n 1 (2.0, 2.0)\n 2 (3.0, 3.0)\n dtype: object\n \"\"\"\n\n return pd.concat((s.x, s.y), axis=1).apply(tuple, axis=1)\n", "path": "dtoolkit/geoaccessor/geoseries/xy.py"}], "after_files": [{"content": "import geopandas as gpd\nimport pandas as pd\n\nfrom dtoolkit.geoaccessor.register import register_geoseries_method\n\n\n@register_geoseries_method\ndef xy(s: gpd.GeoSeries, /, reverse: bool = False) -> pd.Series:\n \"\"\"\n Return the x and y location of Point geometries in a GeoSeries.\n\n Parameters\n ----------\n reverse : bool, default False\n If True, return (y, x) instead.\n\n Returns\n -------\n Series\n tuple of coordinate.\n\n See Also\n --------\n geopandas.GeoSeries.x\n geopandas.GeoSeries.y\n\n Examples\n --------\n >>> import dtoolkit.geoaccessor\n >>> import geopandas as gpd\n >>> from shapely.geometry import Point\n >>> s = gpd.GeoSeries([Point(0, 1), Point(0, 2), Point(0, 3)])\n >>> s\n 0 POINT (0.00000 1.00000)\n 1 POINT (0.00000 2.00000)\n 2 POINT (0.00000 3.00000)\n dtype: geometry\n >>> s.xy()\n 0 (0.0, 1.0)\n 1 (0.0, 2.0)\n 2 (0.0, 3.0)\n dtype: object\n\n Set ``reverse=True`` to return (y, x).\n\n >>> s.xy(True)\n 0 (1.0, 0.0)\n 1 (2.0, 0.0)\n 2 (3.0, 0.0)\n dtype: object\n \"\"\"\n\n coordinates = (s.y, s.x) if reverse else (s.x, s.y)\n return pd.concat(coordinates, axis=1).apply(tuple, axis=1)\n", "path": "dtoolkit/geoaccessor/geoseries/xy.py"}]}
| 995 | 696 |
gh_patches_debug_26503
|
rasdani/github-patches
|
git_diff
|
sanic-org__sanic-2837
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Cookie totally breaks if the client sets a bare cookie
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
A cookie may be not in the `key=value` format. For example. if the JS code runs `document.cookie = "bad"`, it becomes:

I don't know how to call it. I will use the term "bare cookie" in this report. In the following requests with a bare cookie, the Cookie HTTP header becomes: `Cookie: key=value; bad`

It seems that Sanic cannot parse the header with bare cookies, and will throw all cookies (including the legimit `key=value` pair) away. See the code snippet below.
### Code snippet
```python
from sanic import Sanic
from sanic.response import html, text
app = Sanic("test")
app.config.AUTO_EXTEND = False
@app.get("/")
async def route1(request):
return html('<script>document.cookie="key=value"; document.cookie="bad"; location.href="/fire";</script>')
@app.get("/fire")
async def route2(request):
return text(f'''
headers = {request.headers.get("Cookie")}
key = {request.cookies.get("key", "none")}
''')
if __name__ == '__main__':
app.run(port=4321, debug=True)
```
Then visit `http://127.0.0.1:4321/` in Chrome. The page shows:
```
headers = key=value; bad
key = none
```
### Expected Behavior
The page should show:
```
headers = key=value; bad
key = value
```
### How do you run Sanic?
As a script (`app.run` or `Sanic.serve`)
### Operating System
Windows
### Sanic Version
22.12.0
### Additional context
I am using the latest stable Chrome (117.0.5938.150) to reproduce this.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `sanic/models/protocol_types.py`
Content:
```
1 from __future__ import annotations
2
3 import sys
4
5 from asyncio import BaseTransport
6 from typing import TYPE_CHECKING, Any, AnyStr, Optional
7
8
9 if TYPE_CHECKING:
10 from sanic.http.constants import HTTP
11 from sanic.models.asgi import ASGIScope
12
13
14 if sys.version_info < (3, 8):
15 Range = Any
16 HTMLProtocol = Any
17 else:
18 # Protocol is a 3.8+ feature
19 from typing import Protocol
20
21 class HTMLProtocol(Protocol):
22 def __html__(self) -> AnyStr:
23 ...
24
25 def _repr_html_(self) -> AnyStr:
26 ...
27
28 class Range(Protocol):
29 start: Optional[int]
30 end: Optional[int]
31 size: Optional[int]
32 total: Optional[int]
33 __slots__ = ()
34
35
36 class TransportProtocol(BaseTransport):
37 scope: ASGIScope
38 version: HTTP
39 __slots__ = ()
40
```
Path: `sanic/cookies/request.py`
Content:
```
1 import re
2
3 from typing import Any, Dict, List, Optional
4
5 from sanic.cookies.response import Cookie
6 from sanic.log import deprecation
7 from sanic.request.parameters import RequestParameters
8
9
10 COOKIE_NAME_RESERVED_CHARS = re.compile(
11 '[\x00-\x1F\x7F-\xFF()<>@,;:\\\\"/[\\]?={} \x09]'
12 )
13 OCTAL_PATTERN = re.compile(r"\\[0-3][0-7][0-7]")
14 QUOTE_PATTERN = re.compile(r"[\\].")
15
16
17 def _unquote(str): # no cov
18 if str is None or len(str) < 2:
19 return str
20 if str[0] != '"' or str[-1] != '"':
21 return str
22
23 str = str[1:-1]
24
25 i = 0
26 n = len(str)
27 res = []
28 while 0 <= i < n:
29 o_match = OCTAL_PATTERN.search(str, i)
30 q_match = QUOTE_PATTERN.search(str, i)
31 if not o_match and not q_match:
32 res.append(str[i:])
33 break
34 # else:
35 j = k = -1
36 if o_match:
37 j = o_match.start(0)
38 if q_match:
39 k = q_match.start(0)
40 if q_match and (not o_match or k < j):
41 res.append(str[i:k])
42 res.append(str[k + 1])
43 i = k + 2
44 else:
45 res.append(str[i:j])
46 res.append(chr(int(str[j + 1 : j + 4], 8))) # noqa: E203
47 i = j + 4
48 return "".join(res)
49
50
51 def parse_cookie(raw: str) -> Dict[str, List[str]]:
52 """Parses a raw cookie string into a dictionary.
53
54 The function takes a raw cookie string (usually from HTTP headers) and
55 returns a dictionary where each key is a cookie name and the value is a
56 list of values for that cookie. The function handles quoted values and
57 skips invalid cookie names.
58
59 Args:
60 raw (str): The raw cookie string to be parsed.
61
62 Returns:
63 Dict[str, List[str]]: A dictionary containing the cookie names as keys
64 and a list of values for each cookie.
65
66 Example:
67 ```python
68 raw = 'name1=value1; name2="value2"; name3=value3'
69 cookies = parse_cookie(raw)
70 # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}
71 ```
72 """ # noqa: E501
73 cookies: Dict[str, List[str]] = {}
74
75 for token in raw.split(";"):
76 name, __, value = token.partition("=")
77 name = name.strip()
78 value = value.strip()
79
80 if not name:
81 continue
82
83 if COOKIE_NAME_RESERVED_CHARS.search(name): # no cov
84 continue
85
86 if len(value) > 2 and value[0] == '"' and value[-1] == '"': # no cov
87 value = _unquote(value)
88
89 if name in cookies:
90 cookies[name].append(value)
91 else:
92 cookies[name] = [value]
93
94 return cookies
95
96
97 class CookieRequestParameters(RequestParameters):
98 """A container for accessing single and multiple cookie values.
99
100 Because the HTTP standard allows for multiple cookies with the same name,
101 a standard dictionary cannot be used to access cookie values. This class
102 provides a way to access cookie values in a way that is similar to a
103 dictionary, but also allows for accessing multiple values for a single
104 cookie name when necessary.
105
106 Args:
107 cookies (Dict[str, List[str]]): A dictionary containing the cookie
108 names as keys and a list of values for each cookie.
109
110 Example:
111 ```python
112 raw = 'name1=value1; name2="value2"; name3=value3'
113 cookies = parse_cookie(raw)
114 # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}
115
116 request_cookies = CookieRequestParameters(cookies)
117 request_cookies['name1'] # 'value1'
118 request_cookies.get('name1') # 'value1'
119 request_cookies.getlist('name1') # ['value1']
120 ```
121 """ # noqa: E501
122
123 def __getitem__(self, key: str) -> Optional[str]:
124 deprecation(
125 f"You are accessing cookie key '{key}', which is currently in "
126 "compat mode returning a single cookie value. Starting in v24.3 "
127 "accessing a cookie value like this will return a list of values. "
128 "To avoid this behavior and continue accessing a single value, "
129 f"please upgrade from request.cookies['{key}'] to "
130 f"request.cookies.get('{key}'). See more details: "
131 "https://sanic.dev/en/guide/release-notes/v23.3.html#request-cookies", # noqa
132 24.3,
133 )
134 try:
135 value = self._get_prefixed_cookie(key)
136 except KeyError:
137 value = super().__getitem__(key)
138 return value[0]
139
140 def __getattr__(self, key: str) -> str:
141 if key.startswith("_"):
142 return self.__getattribute__(key)
143 key = key.rstrip("_").replace("_", "-")
144 return str(self.get(key, ""))
145
146 def get(self, name: str, default: Optional[Any] = None) -> Optional[Any]:
147 try:
148 return self._get_prefixed_cookie(name)[0]
149 except KeyError:
150 return super().get(name, default)
151
152 def getlist(
153 self, name: str, default: Optional[Any] = None
154 ) -> Optional[Any]:
155 try:
156 return self._get_prefixed_cookie(name)
157 except KeyError:
158 return super().getlist(name, default)
159
160 def _get_prefixed_cookie(self, name: str) -> Any:
161 getitem = super().__getitem__
162 try:
163 return getitem(f"{Cookie.HOST_PREFIX}{name}")
164 except KeyError:
165 return getitem(f"{Cookie.SECURE_PREFIX}{name}")
166
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/sanic/cookies/request.py b/sanic/cookies/request.py
--- a/sanic/cookies/request.py
+++ b/sanic/cookies/request.py
@@ -73,12 +73,17 @@
cookies: Dict[str, List[str]] = {}
for token in raw.split(";"):
- name, __, value = token.partition("=")
+ name, sep, value = token.partition("=")
name = name.strip()
value = value.strip()
- if not name:
- continue
+ # Support cookies =value or plain value with no name
+ # https://github.com/httpwg/http-extensions/issues/159
+ if not sep:
+ if not name:
+ # Empty value like ;; or a cookie header with no value
+ continue
+ name, value = "", name
if COOKIE_NAME_RESERVED_CHARS.search(name): # no cov
continue
diff --git a/sanic/models/protocol_types.py b/sanic/models/protocol_types.py
--- a/sanic/models/protocol_types.py
+++ b/sanic/models/protocol_types.py
@@ -3,7 +3,7 @@
import sys
from asyncio import BaseTransport
-from typing import TYPE_CHECKING, Any, AnyStr, Optional
+from typing import TYPE_CHECKING, Any, Optional, Union
if TYPE_CHECKING:
@@ -19,10 +19,10 @@
from typing import Protocol
class HTMLProtocol(Protocol):
- def __html__(self) -> AnyStr:
+ def __html__(self) -> Union[str, bytes]:
...
- def _repr_html_(self) -> AnyStr:
+ def _repr_html_(self) -> Union[str, bytes]:
...
class Range(Protocol):
|
{"golden_diff": "diff --git a/sanic/cookies/request.py b/sanic/cookies/request.py\n--- a/sanic/cookies/request.py\n+++ b/sanic/cookies/request.py\n@@ -73,12 +73,17 @@\n cookies: Dict[str, List[str]] = {}\n \n for token in raw.split(\";\"):\n- name, __, value = token.partition(\"=\")\n+ name, sep, value = token.partition(\"=\")\n name = name.strip()\n value = value.strip()\n \n- if not name:\n- continue\n+ # Support cookies =value or plain value with no name\n+ # https://github.com/httpwg/http-extensions/issues/159\n+ if not sep:\n+ if not name:\n+ # Empty value like ;; or a cookie header with no value\n+ continue\n+ name, value = \"\", name\n \n if COOKIE_NAME_RESERVED_CHARS.search(name): # no cov\n continue\ndiff --git a/sanic/models/protocol_types.py b/sanic/models/protocol_types.py\n--- a/sanic/models/protocol_types.py\n+++ b/sanic/models/protocol_types.py\n@@ -3,7 +3,7 @@\n import sys\n \n from asyncio import BaseTransport\n-from typing import TYPE_CHECKING, Any, AnyStr, Optional\n+from typing import TYPE_CHECKING, Any, Optional, Union\n \n \n if TYPE_CHECKING:\n@@ -19,10 +19,10 @@\n from typing import Protocol\n \n class HTMLProtocol(Protocol):\n- def __html__(self) -> AnyStr:\n+ def __html__(self) -> Union[str, bytes]:\n ...\n \n- def _repr_html_(self) -> AnyStr:\n+ def _repr_html_(self) -> Union[str, bytes]:\n ...\n \n class Range(Protocol):\n", "issue": "Cookie totally breaks if the client sets a bare cookie\n### Is there an existing issue for this?\r\n\r\n- [X] I have searched the existing issues\r\n\r\n### Describe the bug\r\n\r\nA cookie may be not in the `key=value` format. For example. if the JS code runs `document.cookie = \"bad\"`, it becomes:\r\n\r\n\r\n\r\nI don't know how to call it. I will use the term \"bare cookie\" in this report. In the following requests with a bare cookie, the Cookie HTTP header becomes: `Cookie: key=value; bad`\r\n\r\n\r\n\r\nIt seems that Sanic cannot parse the header with bare cookies, and will throw all cookies (including the legimit `key=value` pair) away. See the code snippet below.\r\n\r\n### Code snippet\r\n\r\n```python\r\nfrom sanic import Sanic\r\nfrom sanic.response import html, text\r\n\r\napp = Sanic(\"test\")\r\napp.config.AUTO_EXTEND = False\r\n\r\[email protected](\"/\")\r\nasync def route1(request):\r\n return html('<script>document.cookie=\"key=value\"; document.cookie=\"bad\"; location.href=\"/fire\";</script>')\r\n \r\[email protected](\"/fire\")\r\nasync def route2(request):\r\n return text(f'''\r\n headers = {request.headers.get(\"Cookie\")}\r\n key = {request.cookies.get(\"key\", \"none\")}\r\n ''')\r\n\r\nif __name__ == '__main__':\r\n app.run(port=4321, debug=True)\r\n```\r\n\r\nThen visit `http://127.0.0.1:4321/` in Chrome. The page shows:\r\n\r\n```\r\nheaders = key=value; bad\r\nkey = none\r\n```\r\n\r\n### Expected Behavior\r\n\r\nThe page should show:\r\n\r\n```\r\nheaders = key=value; bad\r\nkey = value\r\n```\r\n\r\n### How do you run Sanic?\r\n\r\nAs a script (`app.run` or `Sanic.serve`)\r\n\r\n### Operating System\r\n\r\nWindows\r\n\r\n### Sanic Version\r\n\r\n22.12.0\r\n\r\n### Additional context\r\n\r\nI am using the latest stable Chrome (117.0.5938.150) to reproduce this.\r\n\r\n\r\n\n", "before_files": [{"content": "from __future__ import annotations\n\nimport sys\n\nfrom asyncio import BaseTransport\nfrom typing import TYPE_CHECKING, Any, AnyStr, Optional\n\n\nif TYPE_CHECKING:\n from sanic.http.constants import HTTP\n from sanic.models.asgi import ASGIScope\n\n\nif sys.version_info < (3, 8):\n Range = Any\n HTMLProtocol = Any\nelse:\n # Protocol is a 3.8+ feature\n from typing import Protocol\n\n class HTMLProtocol(Protocol):\n def __html__(self) -> AnyStr:\n ...\n\n def _repr_html_(self) -> AnyStr:\n ...\n\n class Range(Protocol):\n start: Optional[int]\n end: Optional[int]\n size: Optional[int]\n total: Optional[int]\n __slots__ = ()\n\n\nclass TransportProtocol(BaseTransport):\n scope: ASGIScope\n version: HTTP\n __slots__ = ()\n", "path": "sanic/models/protocol_types.py"}, {"content": "import re\n\nfrom typing import Any, Dict, List, Optional\n\nfrom sanic.cookies.response import Cookie\nfrom sanic.log import deprecation\nfrom sanic.request.parameters import RequestParameters\n\n\nCOOKIE_NAME_RESERVED_CHARS = re.compile(\n '[\\x00-\\x1F\\x7F-\\xFF()<>@,;:\\\\\\\\\"/[\\\\]?={} \\x09]'\n)\nOCTAL_PATTERN = re.compile(r\"\\\\[0-3][0-7][0-7]\")\nQUOTE_PATTERN = re.compile(r\"[\\\\].\")\n\n\ndef _unquote(str): # no cov\n if str is None or len(str) < 2:\n return str\n if str[0] != '\"' or str[-1] != '\"':\n return str\n\n str = str[1:-1]\n\n i = 0\n n = len(str)\n res = []\n while 0 <= i < n:\n o_match = OCTAL_PATTERN.search(str, i)\n q_match = QUOTE_PATTERN.search(str, i)\n if not o_match and not q_match:\n res.append(str[i:])\n break\n # else:\n j = k = -1\n if o_match:\n j = o_match.start(0)\n if q_match:\n k = q_match.start(0)\n if q_match and (not o_match or k < j):\n res.append(str[i:k])\n res.append(str[k + 1])\n i = k + 2\n else:\n res.append(str[i:j])\n res.append(chr(int(str[j + 1 : j + 4], 8))) # noqa: E203\n i = j + 4\n return \"\".join(res)\n\n\ndef parse_cookie(raw: str) -> Dict[str, List[str]]:\n \"\"\"Parses a raw cookie string into a dictionary.\n\n The function takes a raw cookie string (usually from HTTP headers) and\n returns a dictionary where each key is a cookie name and the value is a\n list of values for that cookie. The function handles quoted values and\n skips invalid cookie names.\n\n Args:\n raw (str): The raw cookie string to be parsed.\n\n Returns:\n Dict[str, List[str]]: A dictionary containing the cookie names as keys\n and a list of values for each cookie.\n\n Example:\n ```python\n raw = 'name1=value1; name2=\"value2\"; name3=value3'\n cookies = parse_cookie(raw)\n # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}\n ```\n \"\"\" # noqa: E501\n cookies: Dict[str, List[str]] = {}\n\n for token in raw.split(\";\"):\n name, __, value = token.partition(\"=\")\n name = name.strip()\n value = value.strip()\n\n if not name:\n continue\n\n if COOKIE_NAME_RESERVED_CHARS.search(name): # no cov\n continue\n\n if len(value) > 2 and value[0] == '\"' and value[-1] == '\"': # no cov\n value = _unquote(value)\n\n if name in cookies:\n cookies[name].append(value)\n else:\n cookies[name] = [value]\n\n return cookies\n\n\nclass CookieRequestParameters(RequestParameters):\n \"\"\"A container for accessing single and multiple cookie values.\n\n Because the HTTP standard allows for multiple cookies with the same name,\n a standard dictionary cannot be used to access cookie values. This class\n provides a way to access cookie values in a way that is similar to a\n dictionary, but also allows for accessing multiple values for a single\n cookie name when necessary.\n\n Args:\n cookies (Dict[str, List[str]]): A dictionary containing the cookie\n names as keys and a list of values for each cookie.\n\n Example:\n ```python\n raw = 'name1=value1; name2=\"value2\"; name3=value3'\n cookies = parse_cookie(raw)\n # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}\n\n request_cookies = CookieRequestParameters(cookies)\n request_cookies['name1'] # 'value1'\n request_cookies.get('name1') # 'value1'\n request_cookies.getlist('name1') # ['value1']\n ```\n \"\"\" # noqa: E501\n\n def __getitem__(self, key: str) -> Optional[str]:\n deprecation(\n f\"You are accessing cookie key '{key}', which is currently in \"\n \"compat mode returning a single cookie value. Starting in v24.3 \"\n \"accessing a cookie value like this will return a list of values. \"\n \"To avoid this behavior and continue accessing a single value, \"\n f\"please upgrade from request.cookies['{key}'] to \"\n f\"request.cookies.get('{key}'). See more details: \"\n \"https://sanic.dev/en/guide/release-notes/v23.3.html#request-cookies\", # noqa\n 24.3,\n )\n try:\n value = self._get_prefixed_cookie(key)\n except KeyError:\n value = super().__getitem__(key)\n return value[0]\n\n def __getattr__(self, key: str) -> str:\n if key.startswith(\"_\"):\n return self.__getattribute__(key)\n key = key.rstrip(\"_\").replace(\"_\", \"-\")\n return str(self.get(key, \"\"))\n\n def get(self, name: str, default: Optional[Any] = None) -> Optional[Any]:\n try:\n return self._get_prefixed_cookie(name)[0]\n except KeyError:\n return super().get(name, default)\n\n def getlist(\n self, name: str, default: Optional[Any] = None\n ) -> Optional[Any]:\n try:\n return self._get_prefixed_cookie(name)\n except KeyError:\n return super().getlist(name, default)\n\n def _get_prefixed_cookie(self, name: str) -> Any:\n getitem = super().__getitem__\n try:\n return getitem(f\"{Cookie.HOST_PREFIX}{name}\")\n except KeyError:\n return getitem(f\"{Cookie.SECURE_PREFIX}{name}\")\n", "path": "sanic/cookies/request.py"}], "after_files": [{"content": "from __future__ import annotations\n\nimport sys\n\nfrom asyncio import BaseTransport\nfrom typing import TYPE_CHECKING, Any, Optional, Union\n\n\nif TYPE_CHECKING:\n from sanic.http.constants import HTTP\n from sanic.models.asgi import ASGIScope\n\n\nif sys.version_info < (3, 8):\n Range = Any\n HTMLProtocol = Any\nelse:\n # Protocol is a 3.8+ feature\n from typing import Protocol\n\n class HTMLProtocol(Protocol):\n def __html__(self) -> Union[str, bytes]:\n ...\n\n def _repr_html_(self) -> Union[str, bytes]:\n ...\n\n class Range(Protocol):\n start: Optional[int]\n end: Optional[int]\n size: Optional[int]\n total: Optional[int]\n __slots__ = ()\n\n\nclass TransportProtocol(BaseTransport):\n scope: ASGIScope\n version: HTTP\n __slots__ = ()\n", "path": "sanic/models/protocol_types.py"}, {"content": "import re\n\nfrom typing import Any, Dict, List, Optional\n\nfrom sanic.cookies.response import Cookie\nfrom sanic.log import deprecation\nfrom sanic.request.parameters import RequestParameters\n\n\nCOOKIE_NAME_RESERVED_CHARS = re.compile(\n '[\\x00-\\x1F\\x7F-\\xFF()<>@,;:\\\\\\\\\"/[\\\\]?={} \\x09]'\n)\nOCTAL_PATTERN = re.compile(r\"\\\\[0-3][0-7][0-7]\")\nQUOTE_PATTERN = re.compile(r\"[\\\\].\")\n\n\ndef _unquote(str): # no cov\n if str is None or len(str) < 2:\n return str\n if str[0] != '\"' or str[-1] != '\"':\n return str\n\n str = str[1:-1]\n\n i = 0\n n = len(str)\n res = []\n while 0 <= i < n:\n o_match = OCTAL_PATTERN.search(str, i)\n q_match = QUOTE_PATTERN.search(str, i)\n if not o_match and not q_match:\n res.append(str[i:])\n break\n # else:\n j = k = -1\n if o_match:\n j = o_match.start(0)\n if q_match:\n k = q_match.start(0)\n if q_match and (not o_match or k < j):\n res.append(str[i:k])\n res.append(str[k + 1])\n i = k + 2\n else:\n res.append(str[i:j])\n res.append(chr(int(str[j + 1 : j + 4], 8))) # noqa: E203\n i = j + 4\n return \"\".join(res)\n\n\ndef parse_cookie(raw: str) -> Dict[str, List[str]]:\n \"\"\"Parses a raw cookie string into a dictionary.\n\n The function takes a raw cookie string (usually from HTTP headers) and\n returns a dictionary where each key is a cookie name and the value is a\n list of values for that cookie. The function handles quoted values and\n skips invalid cookie names.\n\n Args:\n raw (str): The raw cookie string to be parsed.\n\n Returns:\n Dict[str, List[str]]: A dictionary containing the cookie names as keys\n and a list of values for each cookie.\n\n Example:\n ```python\n raw = 'name1=value1; name2=\"value2\"; name3=value3'\n cookies = parse_cookie(raw)\n # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}\n ```\n \"\"\" # noqa: E501\n cookies: Dict[str, List[str]] = {}\n\n for token in raw.split(\";\"):\n name, sep, value = token.partition(\"=\")\n name = name.strip()\n value = value.strip()\n\n # Support cookies =value or plain value with no name\n # https://github.com/httpwg/http-extensions/issues/159\n if not sep:\n if not name:\n # Empty value like ;; or a cookie header with no value\n continue\n name, value = \"\", name\n\n if COOKIE_NAME_RESERVED_CHARS.search(name): # no cov\n continue\n\n if len(value) > 2 and value[0] == '\"' and value[-1] == '\"': # no cov\n value = _unquote(value)\n\n if name in cookies:\n cookies[name].append(value)\n else:\n cookies[name] = [value]\n\n return cookies\n\n\nclass CookieRequestParameters(RequestParameters):\n \"\"\"A container for accessing single and multiple cookie values.\n\n Because the HTTP standard allows for multiple cookies with the same name,\n a standard dictionary cannot be used to access cookie values. This class\n provides a way to access cookie values in a way that is similar to a\n dictionary, but also allows for accessing multiple values for a single\n cookie name when necessary.\n\n Args:\n cookies (Dict[str, List[str]]): A dictionary containing the cookie\n names as keys and a list of values for each cookie.\n\n Example:\n ```python\n raw = 'name1=value1; name2=\"value2\"; name3=value3'\n cookies = parse_cookie(raw)\n # cookies will be {'name1': ['value1'], 'name2': ['value2'], 'name3': ['value3']}\n\n request_cookies = CookieRequestParameters(cookies)\n request_cookies['name1'] # 'value1'\n request_cookies.get('name1') # 'value1'\n request_cookies.getlist('name1') # ['value1']\n ```\n \"\"\" # noqa: E501\n\n def __getitem__(self, key: str) -> Optional[str]:\n deprecation(\n f\"You are accessing cookie key '{key}', which is currently in \"\n \"compat mode returning a single cookie value. Starting in v24.3 \"\n \"accessing a cookie value like this will return a list of values. \"\n \"To avoid this behavior and continue accessing a single value, \"\n f\"please upgrade from request.cookies['{key}'] to \"\n f\"request.cookies.get('{key}'). See more details: \"\n \"https://sanic.dev/en/guide/release-notes/v23.3.html#request-cookies\", # noqa\n 24.3,\n )\n try:\n value = self._get_prefixed_cookie(key)\n except KeyError:\n value = super().__getitem__(key)\n return value[0]\n\n def __getattr__(self, key: str) -> str:\n if key.startswith(\"_\"):\n return self.__getattribute__(key)\n key = key.rstrip(\"_\").replace(\"_\", \"-\")\n return str(self.get(key, \"\"))\n\n def get(self, name: str, default: Optional[Any] = None) -> Optional[Any]:\n try:\n return self._get_prefixed_cookie(name)[0]\n except KeyError:\n return super().get(name, default)\n\n def getlist(\n self, name: str, default: Optional[Any] = None\n ) -> Optional[Any]:\n try:\n return self._get_prefixed_cookie(name)\n except KeyError:\n return super().getlist(name, default)\n\n def _get_prefixed_cookie(self, name: str) -> Any:\n getitem = super().__getitem__\n try:\n return getitem(f\"{Cookie.HOST_PREFIX}{name}\")\n except KeyError:\n return getitem(f\"{Cookie.SECURE_PREFIX}{name}\")\n", "path": "sanic/cookies/request.py"}]}
| 2,857 | 392 |
gh_patches_debug_26594
|
rasdani/github-patches
|
git_diff
|
python-discord__bot-842
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
IndexError when re-evaluating snekbox code
Sentry Issue: [BOT-2C](https://sentry.io/organizations/python-discord/issues/1567544008/?referrer=github_integration)
```
IndexError: list index out of range
File "discord/ext/commands/core.py", line 83, in wrapped
ret = await coro(*args, **kwargs)
File "bot/cogs/snekbox.py", line 281, in eval_command
code = await self.continue_eval(ctx, response)
File "bot/cogs/snekbox.py", line 235, in continue_eval
code = new_message.content.split(' ', maxsplit=1)[1]
```
Using `get_context` seems to be a more robust option than parsing the message ourselves. Either way, if it can't parse it, it should just assume the entire content is the code.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `bot/cogs/snekbox.py`
Content:
```
1 import asyncio
2 import contextlib
3 import datetime
4 import logging
5 import re
6 import textwrap
7 from functools import partial
8 from signal import Signals
9 from typing import Optional, Tuple
10
11 from discord import HTTPException, Message, NotFound, Reaction, User
12 from discord.ext.commands import Cog, Context, command, guild_only
13
14 from bot.bot import Bot
15 from bot.constants import Channels, Roles, URLs
16 from bot.decorators import in_channel
17 from bot.utils.messages import wait_for_deletion
18
19 log = logging.getLogger(__name__)
20
21 ESCAPE_REGEX = re.compile("[`\u202E\u200B]{3,}")
22 FORMATTED_CODE_REGEX = re.compile(
23 r"^\s*" # any leading whitespace from the beginning of the string
24 r"(?P<delim>(?P<block>```)|``?)" # code delimiter: 1-3 backticks; (?P=block) only matches if it's a block
25 r"(?(block)(?:(?P<lang>[a-z]+)\n)?)" # if we're in a block, match optional language (only letters plus newline)
26 r"(?:[ \t]*\n)*" # any blank (empty or tabs/spaces only) lines before the code
27 r"(?P<code>.*?)" # extract all code inside the markup
28 r"\s*" # any more whitespace before the end of the code markup
29 r"(?P=delim)" # match the exact same delimiter from the start again
30 r"\s*$", # any trailing whitespace until the end of the string
31 re.DOTALL | re.IGNORECASE # "." also matches newlines, case insensitive
32 )
33 RAW_CODE_REGEX = re.compile(
34 r"^(?:[ \t]*\n)*" # any blank (empty or tabs/spaces only) lines before the code
35 r"(?P<code>.*?)" # extract all the rest as code
36 r"\s*$", # any trailing whitespace until the end of the string
37 re.DOTALL # "." also matches newlines
38 )
39
40 MAX_PASTE_LEN = 1000
41 EVAL_ROLES = (Roles.helpers, Roles.moderators, Roles.admins, Roles.owners, Roles.python_community, Roles.partners)
42
43 SIGKILL = 9
44
45 REEVAL_EMOJI = '\U0001f501' # :repeat:
46
47
48 class Snekbox(Cog):
49 """Safe evaluation of Python code using Snekbox."""
50
51 def __init__(self, bot: Bot):
52 self.bot = bot
53 self.jobs = {}
54
55 async def post_eval(self, code: str) -> dict:
56 """Send a POST request to the Snekbox API to evaluate code and return the results."""
57 url = URLs.snekbox_eval_api
58 data = {"input": code}
59 async with self.bot.http_session.post(url, json=data, raise_for_status=True) as resp:
60 return await resp.json()
61
62 async def upload_output(self, output: str) -> Optional[str]:
63 """Upload the eval output to a paste service and return a URL to it if successful."""
64 log.trace("Uploading full output to paste service...")
65
66 if len(output) > MAX_PASTE_LEN:
67 log.info("Full output is too long to upload")
68 return "too long to upload"
69
70 url = URLs.paste_service.format(key="documents")
71 try:
72 async with self.bot.http_session.post(url, data=output, raise_for_status=True) as resp:
73 data = await resp.json()
74
75 if "key" in data:
76 return URLs.paste_service.format(key=data["key"])
77 except Exception:
78 # 400 (Bad Request) means there are too many characters
79 log.exception("Failed to upload full output to paste service!")
80
81 @staticmethod
82 def prepare_input(code: str) -> str:
83 """Extract code from the Markdown, format it, and insert it into the code template."""
84 match = FORMATTED_CODE_REGEX.fullmatch(code)
85 if match:
86 code, block, lang, delim = match.group("code", "block", "lang", "delim")
87 code = textwrap.dedent(code)
88 if block:
89 info = (f"'{lang}' highlighted" if lang else "plain") + " code block"
90 else:
91 info = f"{delim}-enclosed inline code"
92 log.trace(f"Extracted {info} for evaluation:\n{code}")
93 else:
94 code = textwrap.dedent(RAW_CODE_REGEX.fullmatch(code).group("code"))
95 log.trace(
96 f"Eval message contains unformatted or badly formatted code, "
97 f"stripping whitespace only:\n{code}"
98 )
99
100 return code
101
102 @staticmethod
103 def get_results_message(results: dict) -> Tuple[str, str]:
104 """Return a user-friendly message and error corresponding to the process's return code."""
105 stdout, returncode = results["stdout"], results["returncode"]
106 msg = f"Your eval job has completed with return code {returncode}"
107 error = ""
108
109 if returncode is None:
110 msg = "Your eval job has failed"
111 error = stdout.strip()
112 elif returncode == 128 + SIGKILL:
113 msg = "Your eval job timed out or ran out of memory"
114 elif returncode == 255:
115 msg = "Your eval job has failed"
116 error = "A fatal NsJail error occurred"
117 else:
118 # Try to append signal's name if one exists
119 try:
120 name = Signals(returncode - 128).name
121 msg = f"{msg} ({name})"
122 except ValueError:
123 pass
124
125 return msg, error
126
127 @staticmethod
128 def get_status_emoji(results: dict) -> str:
129 """Return an emoji corresponding to the status code or lack of output in result."""
130 if not results["stdout"].strip(): # No output
131 return ":warning:"
132 elif results["returncode"] == 0: # No error
133 return ":white_check_mark:"
134 else: # Exception
135 return ":x:"
136
137 async def format_output(self, output: str) -> Tuple[str, Optional[str]]:
138 """
139 Format the output and return a tuple of the formatted output and a URL to the full output.
140
141 Prepend each line with a line number. Truncate if there are over 10 lines or 1000 characters
142 and upload the full output to a paste service.
143 """
144 log.trace("Formatting output...")
145
146 output = output.rstrip("\n")
147 original_output = output # To be uploaded to a pasting service if needed
148 paste_link = None
149
150 if "<@" in output:
151 output = output.replace("<@", "<@\u200B") # Zero-width space
152
153 if "<!@" in output:
154 output = output.replace("<!@", "<!@\u200B") # Zero-width space
155
156 if ESCAPE_REGEX.findall(output):
157 return "Code block escape attempt detected; will not output result", paste_link
158
159 truncated = False
160 lines = output.count("\n")
161
162 if lines > 0:
163 output = [f"{i:03d} | {line}" for i, line in enumerate(output.split('\n'), 1)]
164 output = output[:11] # Limiting to only 11 lines
165 output = "\n".join(output)
166
167 if lines > 10:
168 truncated = True
169 if len(output) >= 1000:
170 output = f"{output[:1000]}\n... (truncated - too long, too many lines)"
171 else:
172 output = f"{output}\n... (truncated - too many lines)"
173 elif len(output) >= 1000:
174 truncated = True
175 output = f"{output[:1000]}\n... (truncated - too long)"
176
177 if truncated:
178 paste_link = await self.upload_output(original_output)
179
180 output = output or "[No output]"
181
182 return output, paste_link
183
184 async def send_eval(self, ctx: Context, code: str) -> Message:
185 """
186 Evaluate code, format it, and send the output to the corresponding channel.
187
188 Return the bot response.
189 """
190 async with ctx.typing():
191 results = await self.post_eval(code)
192 msg, error = self.get_results_message(results)
193
194 if error:
195 output, paste_link = error, None
196 else:
197 output, paste_link = await self.format_output(results["stdout"])
198
199 icon = self.get_status_emoji(results)
200 msg = f"{ctx.author.mention} {icon} {msg}.\n\n```py\n{output}\n```"
201 if paste_link:
202 msg = f"{msg}\nFull output: {paste_link}"
203
204 response = await ctx.send(msg)
205 self.bot.loop.create_task(
206 wait_for_deletion(response, user_ids=(ctx.author.id,), client=ctx.bot)
207 )
208
209 log.info(f"{ctx.author}'s job had a return code of {results['returncode']}")
210 return response
211
212 async def continue_eval(self, ctx: Context, response: Message) -> Optional[str]:
213 """
214 Check if the eval session should continue.
215
216 Return the new code to evaluate or None if the eval session should be terminated.
217 """
218 _predicate_eval_message_edit = partial(predicate_eval_message_edit, ctx)
219 _predicate_emoji_reaction = partial(predicate_eval_emoji_reaction, ctx)
220
221 with contextlib.suppress(NotFound):
222 try:
223 _, new_message = await self.bot.wait_for(
224 'message_edit',
225 check=_predicate_eval_message_edit,
226 timeout=10
227 )
228 await ctx.message.add_reaction(REEVAL_EMOJI)
229 await self.bot.wait_for(
230 'reaction_add',
231 check=_predicate_emoji_reaction,
232 timeout=10
233 )
234
235 code = new_message.content.split(' ', maxsplit=1)[1]
236 await ctx.message.clear_reactions()
237 with contextlib.suppress(HTTPException):
238 await response.delete()
239
240 except asyncio.TimeoutError:
241 await ctx.message.clear_reactions()
242 return None
243
244 return code
245
246 @command(name="eval", aliases=("e",))
247 @guild_only()
248 @in_channel(Channels.bot_commands, hidden_channels=(Channels.esoteric,), bypass_roles=EVAL_ROLES)
249 async def eval_command(self, ctx: Context, *, code: str = None) -> None:
250 """
251 Run Python code and get the results.
252
253 This command supports multiple lines of code, including code wrapped inside a formatted code
254 block. Code can be re-evaluated by editing the original message within 10 seconds and
255 clicking the reaction that subsequently appears.
256
257 We've done our best to make this sandboxed, but do let us know if you manage to find an
258 issue with it!
259 """
260 if ctx.author.id in self.jobs:
261 await ctx.send(
262 f"{ctx.author.mention} You've already got a job running - "
263 "please wait for it to finish!"
264 )
265 return
266
267 if not code: # None or empty string
268 await ctx.invoke(self.bot.get_command("help"), "eval")
269 return
270
271 log.info(f"Received code from {ctx.author} for evaluation:\n{code}")
272
273 while True:
274 self.jobs[ctx.author.id] = datetime.datetime.now()
275 code = self.prepare_input(code)
276 try:
277 response = await self.send_eval(ctx, code)
278 finally:
279 del self.jobs[ctx.author.id]
280
281 code = await self.continue_eval(ctx, response)
282 if not code:
283 break
284 log.info(f"Re-evaluating message {ctx.message.id}")
285
286
287 def predicate_eval_message_edit(ctx: Context, old_msg: Message, new_msg: Message) -> bool:
288 """Return True if the edited message is the context message and the content was indeed modified."""
289 return new_msg.id == ctx.message.id and old_msg.content != new_msg.content
290
291
292 def predicate_eval_emoji_reaction(ctx: Context, reaction: Reaction, user: User) -> bool:
293 """Return True if the reaction REEVAL_EMOJI was added by the context message author on this message."""
294 return reaction.message.id == ctx.message.id and user.id == ctx.author.id and str(reaction) == REEVAL_EMOJI
295
296
297 def setup(bot: Bot) -> None:
298 """Load the Snekbox cog."""
299 bot.add_cog(Snekbox(bot))
300
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/bot/cogs/snekbox.py b/bot/cogs/snekbox.py
--- a/bot/cogs/snekbox.py
+++ b/bot/cogs/snekbox.py
@@ -232,7 +232,7 @@
timeout=10
)
- code = new_message.content.split(' ', maxsplit=1)[1]
+ code = await self.get_code(new_message)
await ctx.message.clear_reactions()
with contextlib.suppress(HTTPException):
await response.delete()
@@ -243,6 +243,26 @@
return code
+ async def get_code(self, message: Message) -> Optional[str]:
+ """
+ Return the code from `message` to be evaluated.
+
+ If the message is an invocation of the eval command, return the first argument or None if it
+ doesn't exist. Otherwise, return the full content of the message.
+ """
+ log.trace(f"Getting context for message {message.id}.")
+ new_ctx = await self.bot.get_context(message)
+
+ if new_ctx.command is self.eval_command:
+ log.trace(f"Message {message.id} invokes eval command.")
+ split = message.content.split(maxsplit=1)
+ code = split[1] if len(split) > 1 else None
+ else:
+ log.trace(f"Message {message.id} does not invoke eval command.")
+ code = message.content
+
+ return code
+
@command(name="eval", aliases=("e",))
@guild_only()
@in_channel(Channels.bot_commands, hidden_channels=(Channels.esoteric,), bypass_roles=EVAL_ROLES)
|
{"golden_diff": "diff --git a/bot/cogs/snekbox.py b/bot/cogs/snekbox.py\n--- a/bot/cogs/snekbox.py\n+++ b/bot/cogs/snekbox.py\n@@ -232,7 +232,7 @@\n timeout=10\n )\n \n- code = new_message.content.split(' ', maxsplit=1)[1]\n+ code = await self.get_code(new_message)\n await ctx.message.clear_reactions()\n with contextlib.suppress(HTTPException):\n await response.delete()\n@@ -243,6 +243,26 @@\n \n return code\n \n+ async def get_code(self, message: Message) -> Optional[str]:\n+ \"\"\"\n+ Return the code from `message` to be evaluated.\n+\n+ If the message is an invocation of the eval command, return the first argument or None if it\n+ doesn't exist. Otherwise, return the full content of the message.\n+ \"\"\"\n+ log.trace(f\"Getting context for message {message.id}.\")\n+ new_ctx = await self.bot.get_context(message)\n+\n+ if new_ctx.command is self.eval_command:\n+ log.trace(f\"Message {message.id} invokes eval command.\")\n+ split = message.content.split(maxsplit=1)\n+ code = split[1] if len(split) > 1 else None\n+ else:\n+ log.trace(f\"Message {message.id} does not invoke eval command.\")\n+ code = message.content\n+\n+ return code\n+\n @command(name=\"eval\", aliases=(\"e\",))\n @guild_only()\n @in_channel(Channels.bot_commands, hidden_channels=(Channels.esoteric,), bypass_roles=EVAL_ROLES)\n", "issue": "IndexError when re-evaluating snekbox code\nSentry Issue: [BOT-2C](https://sentry.io/organizations/python-discord/issues/1567544008/?referrer=github_integration)\n\n```\nIndexError: list index out of range\n File \"discord/ext/commands/core.py\", line 83, in wrapped\n ret = await coro(*args, **kwargs)\n File \"bot/cogs/snekbox.py\", line 281, in eval_command\n code = await self.continue_eval(ctx, response)\n File \"bot/cogs/snekbox.py\", line 235, in continue_eval\n code = new_message.content.split(' ', maxsplit=1)[1]\n```\n\nUsing `get_context` seems to be a more robust option than parsing the message ourselves. Either way, if it can't parse it, it should just assume the entire content is the code.\n", "before_files": [{"content": "import asyncio\nimport contextlib\nimport datetime\nimport logging\nimport re\nimport textwrap\nfrom functools import partial\nfrom signal import Signals\nfrom typing import Optional, Tuple\n\nfrom discord import HTTPException, Message, NotFound, Reaction, User\nfrom discord.ext.commands import Cog, Context, command, guild_only\n\nfrom bot.bot import Bot\nfrom bot.constants import Channels, Roles, URLs\nfrom bot.decorators import in_channel\nfrom bot.utils.messages import wait_for_deletion\n\nlog = logging.getLogger(__name__)\n\nESCAPE_REGEX = re.compile(\"[`\\u202E\\u200B]{3,}\")\nFORMATTED_CODE_REGEX = re.compile(\n r\"^\\s*\" # any leading whitespace from the beginning of the string\n r\"(?P<delim>(?P<block>```)|``?)\" # code delimiter: 1-3 backticks; (?P=block) only matches if it's a block\n r\"(?(block)(?:(?P<lang>[a-z]+)\\n)?)\" # if we're in a block, match optional language (only letters plus newline)\n r\"(?:[ \\t]*\\n)*\" # any blank (empty or tabs/spaces only) lines before the code\n r\"(?P<code>.*?)\" # extract all code inside the markup\n r\"\\s*\" # any more whitespace before the end of the code markup\n r\"(?P=delim)\" # match the exact same delimiter from the start again\n r\"\\s*$\", # any trailing whitespace until the end of the string\n re.DOTALL | re.IGNORECASE # \".\" also matches newlines, case insensitive\n)\nRAW_CODE_REGEX = re.compile(\n r\"^(?:[ \\t]*\\n)*\" # any blank (empty or tabs/spaces only) lines before the code\n r\"(?P<code>.*?)\" # extract all the rest as code\n r\"\\s*$\", # any trailing whitespace until the end of the string\n re.DOTALL # \".\" also matches newlines\n)\n\nMAX_PASTE_LEN = 1000\nEVAL_ROLES = (Roles.helpers, Roles.moderators, Roles.admins, Roles.owners, Roles.python_community, Roles.partners)\n\nSIGKILL = 9\n\nREEVAL_EMOJI = '\\U0001f501' # :repeat:\n\n\nclass Snekbox(Cog):\n \"\"\"Safe evaluation of Python code using Snekbox.\"\"\"\n\n def __init__(self, bot: Bot):\n self.bot = bot\n self.jobs = {}\n\n async def post_eval(self, code: str) -> dict:\n \"\"\"Send a POST request to the Snekbox API to evaluate code and return the results.\"\"\"\n url = URLs.snekbox_eval_api\n data = {\"input\": code}\n async with self.bot.http_session.post(url, json=data, raise_for_status=True) as resp:\n return await resp.json()\n\n async def upload_output(self, output: str) -> Optional[str]:\n \"\"\"Upload the eval output to a paste service and return a URL to it if successful.\"\"\"\n log.trace(\"Uploading full output to paste service...\")\n\n if len(output) > MAX_PASTE_LEN:\n log.info(\"Full output is too long to upload\")\n return \"too long to upload\"\n\n url = URLs.paste_service.format(key=\"documents\")\n try:\n async with self.bot.http_session.post(url, data=output, raise_for_status=True) as resp:\n data = await resp.json()\n\n if \"key\" in data:\n return URLs.paste_service.format(key=data[\"key\"])\n except Exception:\n # 400 (Bad Request) means there are too many characters\n log.exception(\"Failed to upload full output to paste service!\")\n\n @staticmethod\n def prepare_input(code: str) -> str:\n \"\"\"Extract code from the Markdown, format it, and insert it into the code template.\"\"\"\n match = FORMATTED_CODE_REGEX.fullmatch(code)\n if match:\n code, block, lang, delim = match.group(\"code\", \"block\", \"lang\", \"delim\")\n code = textwrap.dedent(code)\n if block:\n info = (f\"'{lang}' highlighted\" if lang else \"plain\") + \" code block\"\n else:\n info = f\"{delim}-enclosed inline code\"\n log.trace(f\"Extracted {info} for evaluation:\\n{code}\")\n else:\n code = textwrap.dedent(RAW_CODE_REGEX.fullmatch(code).group(\"code\"))\n log.trace(\n f\"Eval message contains unformatted or badly formatted code, \"\n f\"stripping whitespace only:\\n{code}\"\n )\n\n return code\n\n @staticmethod\n def get_results_message(results: dict) -> Tuple[str, str]:\n \"\"\"Return a user-friendly message and error corresponding to the process's return code.\"\"\"\n stdout, returncode = results[\"stdout\"], results[\"returncode\"]\n msg = f\"Your eval job has completed with return code {returncode}\"\n error = \"\"\n\n if returncode is None:\n msg = \"Your eval job has failed\"\n error = stdout.strip()\n elif returncode == 128 + SIGKILL:\n msg = \"Your eval job timed out or ran out of memory\"\n elif returncode == 255:\n msg = \"Your eval job has failed\"\n error = \"A fatal NsJail error occurred\"\n else:\n # Try to append signal's name if one exists\n try:\n name = Signals(returncode - 128).name\n msg = f\"{msg} ({name})\"\n except ValueError:\n pass\n\n return msg, error\n\n @staticmethod\n def get_status_emoji(results: dict) -> str:\n \"\"\"Return an emoji corresponding to the status code or lack of output in result.\"\"\"\n if not results[\"stdout\"].strip(): # No output\n return \":warning:\"\n elif results[\"returncode\"] == 0: # No error\n return \":white_check_mark:\"\n else: # Exception\n return \":x:\"\n\n async def format_output(self, output: str) -> Tuple[str, Optional[str]]:\n \"\"\"\n Format the output and return a tuple of the formatted output and a URL to the full output.\n\n Prepend each line with a line number. Truncate if there are over 10 lines or 1000 characters\n and upload the full output to a paste service.\n \"\"\"\n log.trace(\"Formatting output...\")\n\n output = output.rstrip(\"\\n\")\n original_output = output # To be uploaded to a pasting service if needed\n paste_link = None\n\n if \"<@\" in output:\n output = output.replace(\"<@\", \"<@\\u200B\") # Zero-width space\n\n if \"<!@\" in output:\n output = output.replace(\"<!@\", \"<!@\\u200B\") # Zero-width space\n\n if ESCAPE_REGEX.findall(output):\n return \"Code block escape attempt detected; will not output result\", paste_link\n\n truncated = False\n lines = output.count(\"\\n\")\n\n if lines > 0:\n output = [f\"{i:03d} | {line}\" for i, line in enumerate(output.split('\\n'), 1)]\n output = output[:11] # Limiting to only 11 lines\n output = \"\\n\".join(output)\n\n if lines > 10:\n truncated = True\n if len(output) >= 1000:\n output = f\"{output[:1000]}\\n... (truncated - too long, too many lines)\"\n else:\n output = f\"{output}\\n... (truncated - too many lines)\"\n elif len(output) >= 1000:\n truncated = True\n output = f\"{output[:1000]}\\n... (truncated - too long)\"\n\n if truncated:\n paste_link = await self.upload_output(original_output)\n\n output = output or \"[No output]\"\n\n return output, paste_link\n\n async def send_eval(self, ctx: Context, code: str) -> Message:\n \"\"\"\n Evaluate code, format it, and send the output to the corresponding channel.\n\n Return the bot response.\n \"\"\"\n async with ctx.typing():\n results = await self.post_eval(code)\n msg, error = self.get_results_message(results)\n\n if error:\n output, paste_link = error, None\n else:\n output, paste_link = await self.format_output(results[\"stdout\"])\n\n icon = self.get_status_emoji(results)\n msg = f\"{ctx.author.mention} {icon} {msg}.\\n\\n```py\\n{output}\\n```\"\n if paste_link:\n msg = f\"{msg}\\nFull output: {paste_link}\"\n\n response = await ctx.send(msg)\n self.bot.loop.create_task(\n wait_for_deletion(response, user_ids=(ctx.author.id,), client=ctx.bot)\n )\n\n log.info(f\"{ctx.author}'s job had a return code of {results['returncode']}\")\n return response\n\n async def continue_eval(self, ctx: Context, response: Message) -> Optional[str]:\n \"\"\"\n Check if the eval session should continue.\n\n Return the new code to evaluate or None if the eval session should be terminated.\n \"\"\"\n _predicate_eval_message_edit = partial(predicate_eval_message_edit, ctx)\n _predicate_emoji_reaction = partial(predicate_eval_emoji_reaction, ctx)\n\n with contextlib.suppress(NotFound):\n try:\n _, new_message = await self.bot.wait_for(\n 'message_edit',\n check=_predicate_eval_message_edit,\n timeout=10\n )\n await ctx.message.add_reaction(REEVAL_EMOJI)\n await self.bot.wait_for(\n 'reaction_add',\n check=_predicate_emoji_reaction,\n timeout=10\n )\n\n code = new_message.content.split(' ', maxsplit=1)[1]\n await ctx.message.clear_reactions()\n with contextlib.suppress(HTTPException):\n await response.delete()\n\n except asyncio.TimeoutError:\n await ctx.message.clear_reactions()\n return None\n\n return code\n\n @command(name=\"eval\", aliases=(\"e\",))\n @guild_only()\n @in_channel(Channels.bot_commands, hidden_channels=(Channels.esoteric,), bypass_roles=EVAL_ROLES)\n async def eval_command(self, ctx: Context, *, code: str = None) -> None:\n \"\"\"\n Run Python code and get the results.\n\n This command supports multiple lines of code, including code wrapped inside a formatted code\n block. Code can be re-evaluated by editing the original message within 10 seconds and\n clicking the reaction that subsequently appears.\n\n We've done our best to make this sandboxed, but do let us know if you manage to find an\n issue with it!\n \"\"\"\n if ctx.author.id in self.jobs:\n await ctx.send(\n f\"{ctx.author.mention} You've already got a job running - \"\n \"please wait for it to finish!\"\n )\n return\n\n if not code: # None or empty string\n await ctx.invoke(self.bot.get_command(\"help\"), \"eval\")\n return\n\n log.info(f\"Received code from {ctx.author} for evaluation:\\n{code}\")\n\n while True:\n self.jobs[ctx.author.id] = datetime.datetime.now()\n code = self.prepare_input(code)\n try:\n response = await self.send_eval(ctx, code)\n finally:\n del self.jobs[ctx.author.id]\n\n code = await self.continue_eval(ctx, response)\n if not code:\n break\n log.info(f\"Re-evaluating message {ctx.message.id}\")\n\n\ndef predicate_eval_message_edit(ctx: Context, old_msg: Message, new_msg: Message) -> bool:\n \"\"\"Return True if the edited message is the context message and the content was indeed modified.\"\"\"\n return new_msg.id == ctx.message.id and old_msg.content != new_msg.content\n\n\ndef predicate_eval_emoji_reaction(ctx: Context, reaction: Reaction, user: User) -> bool:\n \"\"\"Return True if the reaction REEVAL_EMOJI was added by the context message author on this message.\"\"\"\n return reaction.message.id == ctx.message.id and user.id == ctx.author.id and str(reaction) == REEVAL_EMOJI\n\n\ndef setup(bot: Bot) -> None:\n \"\"\"Load the Snekbox cog.\"\"\"\n bot.add_cog(Snekbox(bot))\n", "path": "bot/cogs/snekbox.py"}], "after_files": [{"content": "import asyncio\nimport contextlib\nimport datetime\nimport logging\nimport re\nimport textwrap\nfrom functools import partial\nfrom signal import Signals\nfrom typing import Optional, Tuple\n\nfrom discord import HTTPException, Message, NotFound, Reaction, User\nfrom discord.ext.commands import Cog, Context, command, guild_only\n\nfrom bot.bot import Bot\nfrom bot.constants import Channels, Roles, URLs\nfrom bot.decorators import in_channel\nfrom bot.utils.messages import wait_for_deletion\n\nlog = logging.getLogger(__name__)\n\nESCAPE_REGEX = re.compile(\"[`\\u202E\\u200B]{3,}\")\nFORMATTED_CODE_REGEX = re.compile(\n r\"^\\s*\" # any leading whitespace from the beginning of the string\n r\"(?P<delim>(?P<block>```)|``?)\" # code delimiter: 1-3 backticks; (?P=block) only matches if it's a block\n r\"(?(block)(?:(?P<lang>[a-z]+)\\n)?)\" # if we're in a block, match optional language (only letters plus newline)\n r\"(?:[ \\t]*\\n)*\" # any blank (empty or tabs/spaces only) lines before the code\n r\"(?P<code>.*?)\" # extract all code inside the markup\n r\"\\s*\" # any more whitespace before the end of the code markup\n r\"(?P=delim)\" # match the exact same delimiter from the start again\n r\"\\s*$\", # any trailing whitespace until the end of the string\n re.DOTALL | re.IGNORECASE # \".\" also matches newlines, case insensitive\n)\nRAW_CODE_REGEX = re.compile(\n r\"^(?:[ \\t]*\\n)*\" # any blank (empty or tabs/spaces only) lines before the code\n r\"(?P<code>.*?)\" # extract all the rest as code\n r\"\\s*$\", # any trailing whitespace until the end of the string\n re.DOTALL # \".\" also matches newlines\n)\n\nMAX_PASTE_LEN = 1000\nEVAL_ROLES = (Roles.helpers, Roles.moderators, Roles.admins, Roles.owners, Roles.python_community, Roles.partners)\n\nSIGKILL = 9\n\nREEVAL_EMOJI = '\\U0001f501' # :repeat:\n\n\nclass Snekbox(Cog):\n \"\"\"Safe evaluation of Python code using Snekbox.\"\"\"\n\n def __init__(self, bot: Bot):\n self.bot = bot\n self.jobs = {}\n\n async def post_eval(self, code: str) -> dict:\n \"\"\"Send a POST request to the Snekbox API to evaluate code and return the results.\"\"\"\n url = URLs.snekbox_eval_api\n data = {\"input\": code}\n async with self.bot.http_session.post(url, json=data, raise_for_status=True) as resp:\n return await resp.json()\n\n async def upload_output(self, output: str) -> Optional[str]:\n \"\"\"Upload the eval output to a paste service and return a URL to it if successful.\"\"\"\n log.trace(\"Uploading full output to paste service...\")\n\n if len(output) > MAX_PASTE_LEN:\n log.info(\"Full output is too long to upload\")\n return \"too long to upload\"\n\n url = URLs.paste_service.format(key=\"documents\")\n try:\n async with self.bot.http_session.post(url, data=output, raise_for_status=True) as resp:\n data = await resp.json()\n\n if \"key\" in data:\n return URLs.paste_service.format(key=data[\"key\"])\n except Exception:\n # 400 (Bad Request) means there are too many characters\n log.exception(\"Failed to upload full output to paste service!\")\n\n @staticmethod\n def prepare_input(code: str) -> str:\n \"\"\"Extract code from the Markdown, format it, and insert it into the code template.\"\"\"\n match = FORMATTED_CODE_REGEX.fullmatch(code)\n if match:\n code, block, lang, delim = match.group(\"code\", \"block\", \"lang\", \"delim\")\n code = textwrap.dedent(code)\n if block:\n info = (f\"'{lang}' highlighted\" if lang else \"plain\") + \" code block\"\n else:\n info = f\"{delim}-enclosed inline code\"\n log.trace(f\"Extracted {info} for evaluation:\\n{code}\")\n else:\n code = textwrap.dedent(RAW_CODE_REGEX.fullmatch(code).group(\"code\"))\n log.trace(\n f\"Eval message contains unformatted or badly formatted code, \"\n f\"stripping whitespace only:\\n{code}\"\n )\n\n return code\n\n @staticmethod\n def get_results_message(results: dict) -> Tuple[str, str]:\n \"\"\"Return a user-friendly message and error corresponding to the process's return code.\"\"\"\n stdout, returncode = results[\"stdout\"], results[\"returncode\"]\n msg = f\"Your eval job has completed with return code {returncode}\"\n error = \"\"\n\n if returncode is None:\n msg = \"Your eval job has failed\"\n error = stdout.strip()\n elif returncode == 128 + SIGKILL:\n msg = \"Your eval job timed out or ran out of memory\"\n elif returncode == 255:\n msg = \"Your eval job has failed\"\n error = \"A fatal NsJail error occurred\"\n else:\n # Try to append signal's name if one exists\n try:\n name = Signals(returncode - 128).name\n msg = f\"{msg} ({name})\"\n except ValueError:\n pass\n\n return msg, error\n\n @staticmethod\n def get_status_emoji(results: dict) -> str:\n \"\"\"Return an emoji corresponding to the status code or lack of output in result.\"\"\"\n if not results[\"stdout\"].strip(): # No output\n return \":warning:\"\n elif results[\"returncode\"] == 0: # No error\n return \":white_check_mark:\"\n else: # Exception\n return \":x:\"\n\n async def format_output(self, output: str) -> Tuple[str, Optional[str]]:\n \"\"\"\n Format the output and return a tuple of the formatted output and a URL to the full output.\n\n Prepend each line with a line number. Truncate if there are over 10 lines or 1000 characters\n and upload the full output to a paste service.\n \"\"\"\n log.trace(\"Formatting output...\")\n\n output = output.rstrip(\"\\n\")\n original_output = output # To be uploaded to a pasting service if needed\n paste_link = None\n\n if \"<@\" in output:\n output = output.replace(\"<@\", \"<@\\u200B\") # Zero-width space\n\n if \"<!@\" in output:\n output = output.replace(\"<!@\", \"<!@\\u200B\") # Zero-width space\n\n if ESCAPE_REGEX.findall(output):\n return \"Code block escape attempt detected; will not output result\", paste_link\n\n truncated = False\n lines = output.count(\"\\n\")\n\n if lines > 0:\n output = [f\"{i:03d} | {line}\" for i, line in enumerate(output.split('\\n'), 1)]\n output = output[:11] # Limiting to only 11 lines\n output = \"\\n\".join(output)\n\n if lines > 10:\n truncated = True\n if len(output) >= 1000:\n output = f\"{output[:1000]}\\n... (truncated - too long, too many lines)\"\n else:\n output = f\"{output}\\n... (truncated - too many lines)\"\n elif len(output) >= 1000:\n truncated = True\n output = f\"{output[:1000]}\\n... (truncated - too long)\"\n\n if truncated:\n paste_link = await self.upload_output(original_output)\n\n output = output or \"[No output]\"\n\n return output, paste_link\n\n async def send_eval(self, ctx: Context, code: str) -> Message:\n \"\"\"\n Evaluate code, format it, and send the output to the corresponding channel.\n\n Return the bot response.\n \"\"\"\n async with ctx.typing():\n results = await self.post_eval(code)\n msg, error = self.get_results_message(results)\n\n if error:\n output, paste_link = error, None\n else:\n output, paste_link = await self.format_output(results[\"stdout\"])\n\n icon = self.get_status_emoji(results)\n msg = f\"{ctx.author.mention} {icon} {msg}.\\n\\n```py\\n{output}\\n```\"\n if paste_link:\n msg = f\"{msg}\\nFull output: {paste_link}\"\n\n response = await ctx.send(msg)\n self.bot.loop.create_task(\n wait_for_deletion(response, user_ids=(ctx.author.id,), client=ctx.bot)\n )\n\n log.info(f\"{ctx.author}'s job had a return code of {results['returncode']}\")\n return response\n\n async def continue_eval(self, ctx: Context, response: Message) -> Optional[str]:\n \"\"\"\n Check if the eval session should continue.\n\n Return the new code to evaluate or None if the eval session should be terminated.\n \"\"\"\n _predicate_eval_message_edit = partial(predicate_eval_message_edit, ctx)\n _predicate_emoji_reaction = partial(predicate_eval_emoji_reaction, ctx)\n\n with contextlib.suppress(NotFound):\n try:\n _, new_message = await self.bot.wait_for(\n 'message_edit',\n check=_predicate_eval_message_edit,\n timeout=10\n )\n await ctx.message.add_reaction(REEVAL_EMOJI)\n await self.bot.wait_for(\n 'reaction_add',\n check=_predicate_emoji_reaction,\n timeout=10\n )\n\n code = await self.get_code(new_message)\n await ctx.message.clear_reactions()\n with contextlib.suppress(HTTPException):\n await response.delete()\n\n except asyncio.TimeoutError:\n await ctx.message.clear_reactions()\n return None\n\n return code\n\n async def get_code(self, message: Message) -> Optional[str]:\n \"\"\"\n Return the code from `message` to be evaluated.\n\n If the message is an invocation of the eval command, return the first argument or None if it\n doesn't exist. Otherwise, return the full content of the message.\n \"\"\"\n log.trace(f\"Getting context for message {message.id}.\")\n new_ctx = await self.bot.get_context(message)\n\n if new_ctx.command is self.eval_command:\n log.trace(f\"Message {message.id} invokes eval command.\")\n split = message.content.split(maxsplit=1)\n code = split[1] if len(split) > 1 else None\n else:\n log.trace(f\"Message {message.id} does not invoke eval command.\")\n code = message.content\n\n return code\n\n @command(name=\"eval\", aliases=(\"e\",))\n @guild_only()\n @in_channel(Channels.bot_commands, hidden_channels=(Channels.esoteric,), bypass_roles=EVAL_ROLES)\n async def eval_command(self, ctx: Context, *, code: str = None) -> None:\n \"\"\"\n Run Python code and get the results.\n\n This command supports multiple lines of code, including code wrapped inside a formatted code\n block. Code can be re-evaluated by editing the original message within 10 seconds and\n clicking the reaction that subsequently appears.\n\n We've done our best to make this sandboxed, but do let us know if you manage to find an\n issue with it!\n \"\"\"\n if ctx.author.id in self.jobs:\n await ctx.send(\n f\"{ctx.author.mention} You've already got a job running - \"\n \"please wait for it to finish!\"\n )\n return\n\n if not code: # None or empty string\n await ctx.invoke(self.bot.get_command(\"help\"), \"eval\")\n return\n\n log.info(f\"Received code from {ctx.author} for evaluation:\\n{code}\")\n\n while True:\n self.jobs[ctx.author.id] = datetime.datetime.now()\n code = self.prepare_input(code)\n try:\n response = await self.send_eval(ctx, code)\n finally:\n del self.jobs[ctx.author.id]\n\n code = await self.continue_eval(ctx, response)\n if not code:\n break\n log.info(f\"Re-evaluating message {ctx.message.id}\")\n\n\ndef predicate_eval_message_edit(ctx: Context, old_msg: Message, new_msg: Message) -> bool:\n \"\"\"Return True if the edited message is the context message and the content was indeed modified.\"\"\"\n return new_msg.id == ctx.message.id and old_msg.content != new_msg.content\n\n\ndef predicate_eval_emoji_reaction(ctx: Context, reaction: Reaction, user: User) -> bool:\n \"\"\"Return True if the reaction REEVAL_EMOJI was added by the context message author on this message.\"\"\"\n return reaction.message.id == ctx.message.id and user.id == ctx.author.id and str(reaction) == REEVAL_EMOJI\n\n\ndef setup(bot: Bot) -> None:\n \"\"\"Load the Snekbox cog.\"\"\"\n bot.add_cog(Snekbox(bot))\n", "path": "bot/cogs/snekbox.py"}]}
| 3,992 | 370 |
gh_patches_debug_18990
|
rasdani/github-patches
|
git_diff
|
qutebrowser__qutebrowser-5916
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Port completion.models.listcategory to QRegularExpression
In `qutebrowser/completion/models/listcategory.py`, we use `QRegExp`, which is going to be removed in Qt 6:
```python
rx = QRegExp(val, Qt.CaseInsensitive)
self.setFilterRegExp(rx)
```
We should instead [port](https://doc-snapshots.qt.io/qt6-dev/qtcore-changes-qt6.html#regular-expression-classes) this to use [QRegularExpression](https://doc.qt.io/qt-5/qregularexpression.html) (or, perhaps, Python's `re`?) instead.
~~(Also, we should probably call `qtutils.ensure_valid(rx)`)~~ (done in `master` already)
cc @rcorre
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `qutebrowser/completion/models/listcategory.py`
Content:
```
1 # vim: ft=python fileencoding=utf-8 sts=4 sw=4 et:
2
3 # Copyright 2017-2020 Ryan Roden-Corrent (rcorre) <[email protected]>
4 #
5 # This file is part of qutebrowser.
6 #
7 # qutebrowser is free software: you can redistribute it and/or modify
8 # it under the terms of the GNU General Public License as published by
9 # the Free Software Foundation, either version 3 of the License, or
10 # (at your option) any later version.
11 #
12 # qutebrowser is distributed in the hope that it will be useful,
13 # but WITHOUT ANY WARRANTY; without even the implied warranty of
14 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 # GNU General Public License for more details.
16 #
17 # You should have received a copy of the GNU General Public License
18 # along with qutebrowser. If not, see <http://www.gnu.org/licenses/>.
19
20 """Completion category that uses a list of tuples as a data source."""
21
22 import re
23 from typing import Iterable, Tuple
24
25 from PyQt5.QtCore import Qt, QSortFilterProxyModel, QRegExp
26 from PyQt5.QtGui import QStandardItem, QStandardItemModel
27 from PyQt5.QtWidgets import QWidget
28
29 from qutebrowser.completion.models import util
30 from qutebrowser.utils import qtutils, log
31
32
33 class ListCategory(QSortFilterProxyModel):
34
35 """Expose a list of items as a category for the CompletionModel."""
36
37 def __init__(self,
38 name: str,
39 items: Iterable[Tuple[str, ...]],
40 sort: bool = True,
41 delete_func: util.DeleteFuncType = None,
42 parent: QWidget = None):
43 super().__init__(parent)
44 self.name = name
45 self.srcmodel = QStandardItemModel(parent=self)
46 self._pattern = ''
47 # ListCategory filters all columns
48 self.columns_to_filter = [0, 1, 2]
49 self.setFilterKeyColumn(-1)
50 for item in items:
51 self.srcmodel.appendRow([QStandardItem(x) for x in item])
52 self.setSourceModel(self.srcmodel)
53 self.delete_func = delete_func
54 self._sort = sort
55
56 def set_pattern(self, val):
57 """Setter for pattern.
58
59 Args:
60 val: The value to set.
61 """
62 self._pattern = val
63 val = re.sub(r' +', r' ', val) # See #1919
64 val = re.escape(val)
65 val = val.replace(r'\ ', '.*')
66 rx = QRegExp(val, Qt.CaseInsensitive)
67 qtutils.ensure_valid(rx)
68 self.setFilterRegExp(rx)
69 self.invalidate()
70 sortcol = 0
71 self.sort(sortcol)
72
73 def lessThan(self, lindex, rindex):
74 """Custom sorting implementation.
75
76 Prefers all items which start with self._pattern. Other than that, uses
77 normal Python string sorting.
78
79 Args:
80 lindex: The QModelIndex of the left item (*left* < right)
81 rindex: The QModelIndex of the right item (left < *right*)
82
83 Return:
84 True if left < right, else False
85 """
86 qtutils.ensure_valid(lindex)
87 qtutils.ensure_valid(rindex)
88
89 left = self.srcmodel.data(lindex)
90 right = self.srcmodel.data(rindex)
91
92 if left is None or right is None: # pragma: no cover
93 log.completion.warning("Got unexpected None value, "
94 "left={!r} right={!r} "
95 "lindex={!r} rindex={!r}"
96 .format(left, right, lindex, rindex))
97 return False
98
99 leftstart = left.startswith(self._pattern)
100 rightstart = right.startswith(self._pattern)
101
102 if leftstart and not rightstart:
103 return True
104 elif rightstart and not leftstart:
105 return False
106 elif self._sort:
107 return left < right
108 else:
109 return False
110
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/qutebrowser/completion/models/listcategory.py b/qutebrowser/completion/models/listcategory.py
--- a/qutebrowser/completion/models/listcategory.py
+++ b/qutebrowser/completion/models/listcategory.py
@@ -22,7 +22,7 @@
import re
from typing import Iterable, Tuple
-from PyQt5.QtCore import Qt, QSortFilterProxyModel, QRegExp
+from PyQt5.QtCore import QSortFilterProxyModel, QRegularExpression
from PyQt5.QtGui import QStandardItem, QStandardItemModel
from PyQt5.QtWidgets import QWidget
@@ -63,9 +63,9 @@
val = re.sub(r' +', r' ', val) # See #1919
val = re.escape(val)
val = val.replace(r'\ ', '.*')
- rx = QRegExp(val, Qt.CaseInsensitive)
+ rx = QRegularExpression(val, QRegularExpression.CaseInsensitiveOption)
qtutils.ensure_valid(rx)
- self.setFilterRegExp(rx)
+ self.setFilterRegularExpression(rx)
self.invalidate()
sortcol = 0
self.sort(sortcol)
|
{"golden_diff": "diff --git a/qutebrowser/completion/models/listcategory.py b/qutebrowser/completion/models/listcategory.py\n--- a/qutebrowser/completion/models/listcategory.py\n+++ b/qutebrowser/completion/models/listcategory.py\n@@ -22,7 +22,7 @@\n import re\n from typing import Iterable, Tuple\n \n-from PyQt5.QtCore import Qt, QSortFilterProxyModel, QRegExp\n+from PyQt5.QtCore import QSortFilterProxyModel, QRegularExpression\n from PyQt5.QtGui import QStandardItem, QStandardItemModel\n from PyQt5.QtWidgets import QWidget\n \n@@ -63,9 +63,9 @@\n val = re.sub(r' +', r' ', val) # See #1919\n val = re.escape(val)\n val = val.replace(r'\\ ', '.*')\n- rx = QRegExp(val, Qt.CaseInsensitive)\n+ rx = QRegularExpression(val, QRegularExpression.CaseInsensitiveOption)\n qtutils.ensure_valid(rx)\n- self.setFilterRegExp(rx)\n+ self.setFilterRegularExpression(rx)\n self.invalidate()\n sortcol = 0\n self.sort(sortcol)\n", "issue": "Port completion.models.listcategory to QRegularExpression\nIn `qutebrowser/completion/models/listcategory.py`, we use `QRegExp`, which is going to be removed in Qt 6:\r\n\r\n```python\r\n rx = QRegExp(val, Qt.CaseInsensitive)\r\n self.setFilterRegExp(rx)\r\n```\r\n\r\nWe should instead [port](https://doc-snapshots.qt.io/qt6-dev/qtcore-changes-qt6.html#regular-expression-classes) this to use [QRegularExpression](https://doc.qt.io/qt-5/qregularexpression.html) (or, perhaps, Python's `re`?) instead.\r\n\r\n~~(Also, we should probably call `qtutils.ensure_valid(rx)`)~~ (done in `master` already)\r\n\r\ncc @rcorre \n", "before_files": [{"content": "# vim: ft=python fileencoding=utf-8 sts=4 sw=4 et:\n\n# Copyright 2017-2020 Ryan Roden-Corrent (rcorre) <[email protected]>\n#\n# This file is part of qutebrowser.\n#\n# qutebrowser is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# qutebrowser is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with qutebrowser. If not, see <http://www.gnu.org/licenses/>.\n\n\"\"\"Completion category that uses a list of tuples as a data source.\"\"\"\n\nimport re\nfrom typing import Iterable, Tuple\n\nfrom PyQt5.QtCore import Qt, QSortFilterProxyModel, QRegExp\nfrom PyQt5.QtGui import QStandardItem, QStandardItemModel\nfrom PyQt5.QtWidgets import QWidget\n\nfrom qutebrowser.completion.models import util\nfrom qutebrowser.utils import qtutils, log\n\n\nclass ListCategory(QSortFilterProxyModel):\n\n \"\"\"Expose a list of items as a category for the CompletionModel.\"\"\"\n\n def __init__(self,\n name: str,\n items: Iterable[Tuple[str, ...]],\n sort: bool = True,\n delete_func: util.DeleteFuncType = None,\n parent: QWidget = None):\n super().__init__(parent)\n self.name = name\n self.srcmodel = QStandardItemModel(parent=self)\n self._pattern = ''\n # ListCategory filters all columns\n self.columns_to_filter = [0, 1, 2]\n self.setFilterKeyColumn(-1)\n for item in items:\n self.srcmodel.appendRow([QStandardItem(x) for x in item])\n self.setSourceModel(self.srcmodel)\n self.delete_func = delete_func\n self._sort = sort\n\n def set_pattern(self, val):\n \"\"\"Setter for pattern.\n\n Args:\n val: The value to set.\n \"\"\"\n self._pattern = val\n val = re.sub(r' +', r' ', val) # See #1919\n val = re.escape(val)\n val = val.replace(r'\\ ', '.*')\n rx = QRegExp(val, Qt.CaseInsensitive)\n qtutils.ensure_valid(rx)\n self.setFilterRegExp(rx)\n self.invalidate()\n sortcol = 0\n self.sort(sortcol)\n\n def lessThan(self, lindex, rindex):\n \"\"\"Custom sorting implementation.\n\n Prefers all items which start with self._pattern. Other than that, uses\n normal Python string sorting.\n\n Args:\n lindex: The QModelIndex of the left item (*left* < right)\n rindex: The QModelIndex of the right item (left < *right*)\n\n Return:\n True if left < right, else False\n \"\"\"\n qtutils.ensure_valid(lindex)\n qtutils.ensure_valid(rindex)\n\n left = self.srcmodel.data(lindex)\n right = self.srcmodel.data(rindex)\n\n if left is None or right is None: # pragma: no cover\n log.completion.warning(\"Got unexpected None value, \"\n \"left={!r} right={!r} \"\n \"lindex={!r} rindex={!r}\"\n .format(left, right, lindex, rindex))\n return False\n\n leftstart = left.startswith(self._pattern)\n rightstart = right.startswith(self._pattern)\n\n if leftstart and not rightstart:\n return True\n elif rightstart and not leftstart:\n return False\n elif self._sort:\n return left < right\n else:\n return False\n", "path": "qutebrowser/completion/models/listcategory.py"}], "after_files": [{"content": "# vim: ft=python fileencoding=utf-8 sts=4 sw=4 et:\n\n# Copyright 2017-2020 Ryan Roden-Corrent (rcorre) <[email protected]>\n#\n# This file is part of qutebrowser.\n#\n# qutebrowser is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# qutebrowser is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with qutebrowser. If not, see <http://www.gnu.org/licenses/>.\n\n\"\"\"Completion category that uses a list of tuples as a data source.\"\"\"\n\nimport re\nfrom typing import Iterable, Tuple\n\nfrom PyQt5.QtCore import QSortFilterProxyModel, QRegularExpression\nfrom PyQt5.QtGui import QStandardItem, QStandardItemModel\nfrom PyQt5.QtWidgets import QWidget\n\nfrom qutebrowser.completion.models import util\nfrom qutebrowser.utils import qtutils, log\n\n\nclass ListCategory(QSortFilterProxyModel):\n\n \"\"\"Expose a list of items as a category for the CompletionModel.\"\"\"\n\n def __init__(self,\n name: str,\n items: Iterable[Tuple[str, ...]],\n sort: bool = True,\n delete_func: util.DeleteFuncType = None,\n parent: QWidget = None):\n super().__init__(parent)\n self.name = name\n self.srcmodel = QStandardItemModel(parent=self)\n self._pattern = ''\n # ListCategory filters all columns\n self.columns_to_filter = [0, 1, 2]\n self.setFilterKeyColumn(-1)\n for item in items:\n self.srcmodel.appendRow([QStandardItem(x) for x in item])\n self.setSourceModel(self.srcmodel)\n self.delete_func = delete_func\n self._sort = sort\n\n def set_pattern(self, val):\n \"\"\"Setter for pattern.\n\n Args:\n val: The value to set.\n \"\"\"\n self._pattern = val\n val = re.sub(r' +', r' ', val) # See #1919\n val = re.escape(val)\n val = val.replace(r'\\ ', '.*')\n rx = QRegularExpression(val, QRegularExpression.CaseInsensitiveOption)\n qtutils.ensure_valid(rx)\n self.setFilterRegularExpression(rx)\n self.invalidate()\n sortcol = 0\n self.sort(sortcol)\n\n def lessThan(self, lindex, rindex):\n \"\"\"Custom sorting implementation.\n\n Prefers all items which start with self._pattern. Other than that, uses\n normal Python string sorting.\n\n Args:\n lindex: The QModelIndex of the left item (*left* < right)\n rindex: The QModelIndex of the right item (left < *right*)\n\n Return:\n True if left < right, else False\n \"\"\"\n qtutils.ensure_valid(lindex)\n qtutils.ensure_valid(rindex)\n\n left = self.srcmodel.data(lindex)\n right = self.srcmodel.data(rindex)\n\n if left is None or right is None: # pragma: no cover\n log.completion.warning(\"Got unexpected None value, \"\n \"left={!r} right={!r} \"\n \"lindex={!r} rindex={!r}\"\n .format(left, right, lindex, rindex))\n return False\n\n leftstart = left.startswith(self._pattern)\n rightstart = right.startswith(self._pattern)\n\n if leftstart and not rightstart:\n return True\n elif rightstart and not leftstart:\n return False\n elif self._sort:\n return left < right\n else:\n return False\n", "path": "qutebrowser/completion/models/listcategory.py"}]}
| 1,507 | 245 |
gh_patches_debug_1842
|
rasdani/github-patches
|
git_diff
|
DistrictDataLabs__yellowbrick-1162
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
pytest-runner is deprecated
pytest-runner is deprecated: https://github.com/pytest-dev/pytest-runner/#deprecation-notice
If I find time, then I can make a PR, but I thought I'd let you know in the meantime.
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `setup.py`
Content:
```
1 #!/usr/bin/env python
2 # setup
3 # Setup script for installing yellowbrick
4 #
5 # Author: Benjamin Bengfort
6 # Created: Wed May 18 14:33:26 2016 -0400
7 #
8 # Copyright (C) 2016 The scikit-yb developers
9 # For license information, see LICENSE.txt and NOTICE.md
10 #
11 # ID: setup.py [c4f3ba7] [email protected] $
12
13 """
14 Setup script for installing yellowbrick.
15 See http://bbengfort.github.io/programmer/2016/01/20/packaging-with-pypi.html
16 """
17
18 ##########################################################################
19 ## Imports
20 ##########################################################################
21
22 import os
23 import codecs
24
25 from setuptools import setup
26 from setuptools import find_packages
27
28 ##########################################################################
29 ## Package Information
30 ##########################################################################
31
32 ## Basic information
33 ## Basic information
34 NAME = "yellowbrick"
35 DESCRIPTION = "A suite of visual analysis and diagnostic tools for machine learning."
36 AUTHOR = "The scikit-yb developers"
37 EMAIL = "[email protected]"
38 MAINTAINER = "The scikit-yb developers"
39 LICENSE = "Apache 2"
40 REPOSITORY = "https://github.com/DistrictDataLabs/yellowbrick"
41 PACKAGE = "yellowbrick"
42 URL = "http://scikit-yb.org/"
43
44 ## Define the keywords
45 KEYWORDS = (
46 "visualization",
47 "machine learning",
48 "scikit-learn",
49 "matplotlib",
50 "data science",
51 )
52
53 ## Define the classifiers
54 ## See https://pypi.python.org/pypi?%3Aaction=list_classifiers
55 CLASSIFIERS = (
56 "Development Status :: 5 - Production/Stable",
57 "Intended Audience :: Developers",
58 "Intended Audience :: Science/Research",
59 "License :: OSI Approved :: Apache Software License",
60 "Natural Language :: English",
61 "Operating System :: OS Independent",
62 "Programming Language :: Python",
63 "Programming Language :: Python :: 3.5",
64 "Programming Language :: Python :: 3.6",
65 "Programming Language :: Python :: 3.7",
66 "Programming Language :: Python :: 3.8",
67 "Topic :: Software Development",
68 "Topic :: Software Development :: Libraries :: Python Modules",
69 "Topic :: Scientific/Engineering :: Visualization",
70 )
71
72 ## Important Paths
73 PROJECT = os.path.abspath(os.path.dirname(__file__))
74 REQUIRE_PATH = "requirements.txt"
75 VERSION_PATH = os.path.join(PACKAGE, "version.py")
76 PKG_DESCRIBE = "DESCRIPTION.md"
77
78 ## Directories to ignore in find_packages
79 EXCLUDES = (
80 "tests", "tests.*",
81 "bin",
82 "docs", "docs.*",
83 "fixtures",
84 "register",
85 "notebooks", "notebooks.*",
86 "examples", "examples.*",
87 "binder", "binder.*",
88 "paper",
89 )
90
91 ##########################################################################
92 ## Helper Functions
93 ##########################################################################
94
95
96 def read(*parts):
97 """
98 Assume UTF-8 encoding and return the contents of the file located at the
99 absolute path from the REPOSITORY joined with *parts.
100 """
101 with codecs.open(os.path.join(PROJECT, *parts), "rb", "utf-8") as f:
102 return f.read()
103
104
105 def get_version(path=VERSION_PATH):
106 """
107 Reads the python file defined in the VERSION_PATH to find the get_version
108 function, and executes it to ensure that it is loaded correctly. Separating
109 the version in this way ensures no additional code is executed.
110 """
111 namespace = {}
112 exec(read(path), namespace)
113 return namespace["get_version"](short=True)
114
115
116 def get_requires(path=REQUIRE_PATH):
117 """
118 Yields a generator of requirements as defined by the REQUIRE_PATH which
119 should point to a requirements.txt output by `pip freeze`.
120 """
121 for line in read(path).splitlines():
122 line = line.strip()
123 if line and not line.startswith("#"):
124 yield line
125
126
127 def get_description_type(path=PKG_DESCRIBE):
128 """
129 Returns the long_description_content_type based on the extension of the
130 package describe path (e.g. .txt, .rst, or .md).
131 """
132 _, ext = os.path.splitext(path)
133 return {".rst": "text/x-rst", ".txt": "text/plain", ".md": "text/markdown"}[ext]
134
135
136 ##########################################################################
137 ## Define the configuration
138 ##########################################################################
139
140 config = {
141 "name": NAME,
142 "version": get_version(),
143 "description": DESCRIPTION,
144 "long_description": read(PKG_DESCRIBE),
145 "long_description_content_type": get_description_type(PKG_DESCRIBE),
146 "classifiers": CLASSIFIERS,
147 "keywords": KEYWORDS,
148 "license": LICENSE,
149 "author": AUTHOR,
150 "author_email": EMAIL,
151 "url": URL,
152 "maintainer": MAINTAINER,
153 "maintainer_email": EMAIL,
154 "project_urls": {
155 "Documentation": URL,
156 "Download": "{}/tarball/v{}".format(REPOSITORY, get_version()),
157 "Source": REPOSITORY,
158 "Tracker": "{}/issues".format(REPOSITORY),
159 },
160 "download_url": "{}/tarball/v{}".format(REPOSITORY, get_version()),
161 "packages": find_packages(where=PROJECT, exclude=EXCLUDES),
162 "package_data": {"yellowbrick": ["datasets/manifest.json"]},
163 "zip_safe": False,
164 "entry_points": {"console_scripts": []},
165 "install_requires": list(get_requires()),
166 "python_requires": ">=3.4, <4",
167 "setup_requires": ["pytest-runner"],
168 "tests_require": ["pytest"],
169 }
170
171
172 ##########################################################################
173 ## Run setup script
174 ##########################################################################
175
176 if __name__ == "__main__":
177 setup(**config)
178
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/setup.py b/setup.py
--- a/setup.py
+++ b/setup.py
@@ -163,9 +163,7 @@
"zip_safe": False,
"entry_points": {"console_scripts": []},
"install_requires": list(get_requires()),
- "python_requires": ">=3.4, <4",
- "setup_requires": ["pytest-runner"],
- "tests_require": ["pytest"],
+ "python_requires": ">=3.4, <4"
}
|
{"golden_diff": "diff --git a/setup.py b/setup.py\n--- a/setup.py\n+++ b/setup.py\n@@ -163,9 +163,7 @@\n \"zip_safe\": False,\n \"entry_points\": {\"console_scripts\": []},\n \"install_requires\": list(get_requires()),\n- \"python_requires\": \">=3.4, <4\",\n- \"setup_requires\": [\"pytest-runner\"],\n- \"tests_require\": [\"pytest\"],\n+ \"python_requires\": \">=3.4, <4\"\n }\n", "issue": "pytest-runner is deprecated\npytest-runner is deprecated: https://github.com/pytest-dev/pytest-runner/#deprecation-notice\r\n\r\nIf I find time, then I can make a PR, but I thought I'd let you know in the meantime.\n", "before_files": [{"content": "#!/usr/bin/env python\n# setup\n# Setup script for installing yellowbrick\n#\n# Author: Benjamin Bengfort\n# Created: Wed May 18 14:33:26 2016 -0400\n#\n# Copyright (C) 2016 The scikit-yb developers\n# For license information, see LICENSE.txt and NOTICE.md\n#\n# ID: setup.py [c4f3ba7] [email protected] $\n\n\"\"\"\nSetup script for installing yellowbrick.\nSee http://bbengfort.github.io/programmer/2016/01/20/packaging-with-pypi.html\n\"\"\"\n\n##########################################################################\n## Imports\n##########################################################################\n\nimport os\nimport codecs\n\nfrom setuptools import setup\nfrom setuptools import find_packages\n\n##########################################################################\n## Package Information\n##########################################################################\n\n## Basic information\n## Basic information\nNAME = \"yellowbrick\"\nDESCRIPTION = \"A suite of visual analysis and diagnostic tools for machine learning.\"\nAUTHOR = \"The scikit-yb developers\"\nEMAIL = \"[email protected]\"\nMAINTAINER = \"The scikit-yb developers\"\nLICENSE = \"Apache 2\"\nREPOSITORY = \"https://github.com/DistrictDataLabs/yellowbrick\"\nPACKAGE = \"yellowbrick\"\nURL = \"http://scikit-yb.org/\"\n\n## Define the keywords\nKEYWORDS = (\n \"visualization\",\n \"machine learning\",\n \"scikit-learn\",\n \"matplotlib\",\n \"data science\",\n)\n\n## Define the classifiers\n## See https://pypi.python.org/pypi?%3Aaction=list_classifiers\nCLASSIFIERS = (\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Topic :: Software Development\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n \"Topic :: Scientific/Engineering :: Visualization\",\n)\n\n## Important Paths\nPROJECT = os.path.abspath(os.path.dirname(__file__))\nREQUIRE_PATH = \"requirements.txt\"\nVERSION_PATH = os.path.join(PACKAGE, \"version.py\")\nPKG_DESCRIBE = \"DESCRIPTION.md\"\n\n## Directories to ignore in find_packages\nEXCLUDES = (\n \"tests\", \"tests.*\",\n \"bin\",\n \"docs\", \"docs.*\",\n \"fixtures\",\n \"register\",\n \"notebooks\", \"notebooks.*\",\n \"examples\", \"examples.*\",\n \"binder\", \"binder.*\",\n \"paper\",\n)\n\n##########################################################################\n## Helper Functions\n##########################################################################\n\n\ndef read(*parts):\n \"\"\"\n Assume UTF-8 encoding and return the contents of the file located at the\n absolute path from the REPOSITORY joined with *parts.\n \"\"\"\n with codecs.open(os.path.join(PROJECT, *parts), \"rb\", \"utf-8\") as f:\n return f.read()\n\n\ndef get_version(path=VERSION_PATH):\n \"\"\"\n Reads the python file defined in the VERSION_PATH to find the get_version\n function, and executes it to ensure that it is loaded correctly. Separating\n the version in this way ensures no additional code is executed.\n \"\"\"\n namespace = {}\n exec(read(path), namespace)\n return namespace[\"get_version\"](short=True)\n\n\ndef get_requires(path=REQUIRE_PATH):\n \"\"\"\n Yields a generator of requirements as defined by the REQUIRE_PATH which\n should point to a requirements.txt output by `pip freeze`.\n \"\"\"\n for line in read(path).splitlines():\n line = line.strip()\n if line and not line.startswith(\"#\"):\n yield line\n\n\ndef get_description_type(path=PKG_DESCRIBE):\n \"\"\"\n Returns the long_description_content_type based on the extension of the\n package describe path (e.g. .txt, .rst, or .md).\n \"\"\"\n _, ext = os.path.splitext(path)\n return {\".rst\": \"text/x-rst\", \".txt\": \"text/plain\", \".md\": \"text/markdown\"}[ext]\n\n\n##########################################################################\n## Define the configuration\n##########################################################################\n\nconfig = {\n \"name\": NAME,\n \"version\": get_version(),\n \"description\": DESCRIPTION,\n \"long_description\": read(PKG_DESCRIBE),\n \"long_description_content_type\": get_description_type(PKG_DESCRIBE),\n \"classifiers\": CLASSIFIERS,\n \"keywords\": KEYWORDS,\n \"license\": LICENSE,\n \"author\": AUTHOR,\n \"author_email\": EMAIL,\n \"url\": URL,\n \"maintainer\": MAINTAINER,\n \"maintainer_email\": EMAIL,\n \"project_urls\": {\n \"Documentation\": URL,\n \"Download\": \"{}/tarball/v{}\".format(REPOSITORY, get_version()),\n \"Source\": REPOSITORY,\n \"Tracker\": \"{}/issues\".format(REPOSITORY),\n },\n \"download_url\": \"{}/tarball/v{}\".format(REPOSITORY, get_version()),\n \"packages\": find_packages(where=PROJECT, exclude=EXCLUDES),\n \"package_data\": {\"yellowbrick\": [\"datasets/manifest.json\"]},\n \"zip_safe\": False,\n \"entry_points\": {\"console_scripts\": []},\n \"install_requires\": list(get_requires()),\n \"python_requires\": \">=3.4, <4\",\n \"setup_requires\": [\"pytest-runner\"],\n \"tests_require\": [\"pytest\"],\n}\n\n\n##########################################################################\n## Run setup script\n##########################################################################\n\nif __name__ == \"__main__\":\n setup(**config)\n", "path": "setup.py"}], "after_files": [{"content": "#!/usr/bin/env python\n# setup\n# Setup script for installing yellowbrick\n#\n# Author: Benjamin Bengfort\n# Created: Wed May 18 14:33:26 2016 -0400\n#\n# Copyright (C) 2016 The scikit-yb developers\n# For license information, see LICENSE.txt and NOTICE.md\n#\n# ID: setup.py [c4f3ba7] [email protected] $\n\n\"\"\"\nSetup script for installing yellowbrick.\nSee http://bbengfort.github.io/programmer/2016/01/20/packaging-with-pypi.html\n\"\"\"\n\n##########################################################################\n## Imports\n##########################################################################\n\nimport os\nimport codecs\n\nfrom setuptools import setup\nfrom setuptools import find_packages\n\n##########################################################################\n## Package Information\n##########################################################################\n\n## Basic information\n## Basic information\nNAME = \"yellowbrick\"\nDESCRIPTION = \"A suite of visual analysis and diagnostic tools for machine learning.\"\nAUTHOR = \"The scikit-yb developers\"\nEMAIL = \"[email protected]\"\nMAINTAINER = \"The scikit-yb developers\"\nLICENSE = \"Apache 2\"\nREPOSITORY = \"https://github.com/DistrictDataLabs/yellowbrick\"\nPACKAGE = \"yellowbrick\"\nURL = \"http://scikit-yb.org/\"\n\n## Define the keywords\nKEYWORDS = (\n \"visualization\",\n \"machine learning\",\n \"scikit-learn\",\n \"matplotlib\",\n \"data science\",\n)\n\n## Define the classifiers\n## See https://pypi.python.org/pypi?%3Aaction=list_classifiers\nCLASSIFIERS = (\n \"Development Status :: 5 - Production/Stable\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science/Research\",\n \"License :: OSI Approved :: Apache Software License\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3.5\",\n \"Programming Language :: Python :: 3.6\",\n \"Programming Language :: Python :: 3.7\",\n \"Programming Language :: Python :: 3.8\",\n \"Topic :: Software Development\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n \"Topic :: Scientific/Engineering :: Visualization\",\n)\n\n## Important Paths\nPROJECT = os.path.abspath(os.path.dirname(__file__))\nREQUIRE_PATH = \"requirements.txt\"\nVERSION_PATH = os.path.join(PACKAGE, \"version.py\")\nPKG_DESCRIBE = \"DESCRIPTION.md\"\n\n## Directories to ignore in find_packages\nEXCLUDES = (\n \"tests\", \"tests.*\",\n \"bin\",\n \"docs\", \"docs.*\",\n \"fixtures\",\n \"register\",\n \"notebooks\", \"notebooks.*\",\n \"examples\", \"examples.*\",\n \"binder\", \"binder.*\",\n \"paper\",\n)\n\n##########################################################################\n## Helper Functions\n##########################################################################\n\n\ndef read(*parts):\n \"\"\"\n Assume UTF-8 encoding and return the contents of the file located at the\n absolute path from the REPOSITORY joined with *parts.\n \"\"\"\n with codecs.open(os.path.join(PROJECT, *parts), \"rb\", \"utf-8\") as f:\n return f.read()\n\n\ndef get_version(path=VERSION_PATH):\n \"\"\"\n Reads the python file defined in the VERSION_PATH to find the get_version\n function, and executes it to ensure that it is loaded correctly. Separating\n the version in this way ensures no additional code is executed.\n \"\"\"\n namespace = {}\n exec(read(path), namespace)\n return namespace[\"get_version\"](short=True)\n\n\ndef get_requires(path=REQUIRE_PATH):\n \"\"\"\n Yields a generator of requirements as defined by the REQUIRE_PATH which\n should point to a requirements.txt output by `pip freeze`.\n \"\"\"\n for line in read(path).splitlines():\n line = line.strip()\n if line and not line.startswith(\"#\"):\n yield line\n\n\ndef get_description_type(path=PKG_DESCRIBE):\n \"\"\"\n Returns the long_description_content_type based on the extension of the\n package describe path (e.g. .txt, .rst, or .md).\n \"\"\"\n _, ext = os.path.splitext(path)\n return {\".rst\": \"text/x-rst\", \".txt\": \"text/plain\", \".md\": \"text/markdown\"}[ext]\n\n\n##########################################################################\n## Define the configuration\n##########################################################################\n\nconfig = {\n \"name\": NAME,\n \"version\": get_version(),\n \"description\": DESCRIPTION,\n \"long_description\": read(PKG_DESCRIBE),\n \"long_description_content_type\": get_description_type(PKG_DESCRIBE),\n \"classifiers\": CLASSIFIERS,\n \"keywords\": KEYWORDS,\n \"license\": LICENSE,\n \"author\": AUTHOR,\n \"author_email\": EMAIL,\n \"url\": URL,\n \"maintainer\": MAINTAINER,\n \"maintainer_email\": EMAIL,\n \"project_urls\": {\n \"Documentation\": URL,\n \"Download\": \"{}/tarball/v{}\".format(REPOSITORY, get_version()),\n \"Source\": REPOSITORY,\n \"Tracker\": \"{}/issues\".format(REPOSITORY),\n },\n \"download_url\": \"{}/tarball/v{}\".format(REPOSITORY, get_version()),\n \"packages\": find_packages(where=PROJECT, exclude=EXCLUDES),\n \"package_data\": {\"yellowbrick\": [\"datasets/manifest.json\"]},\n \"zip_safe\": False,\n \"entry_points\": {\"console_scripts\": []},\n \"install_requires\": list(get_requires()),\n \"python_requires\": \">=3.4, <4\"\n}\n\n\n##########################################################################\n## Run setup script\n##########################################################################\n\nif __name__ == \"__main__\":\n setup(**config)\n", "path": "setup.py"}]}
| 1,983 | 111 |
gh_patches_debug_36242
|
rasdani/github-patches
|
git_diff
|
pytorch__ignite-1756
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Improve `ignite.contirb.metrics` implementation
For these metrics in `ignite.contrib.metrics` :
- [Average Precision](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/average_precision.py)
- [Precision Recall Curve](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/precision_recall_curve.py)
- [Roc Auc](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/roc_auc.py)
We need to make the implementation simpler. Similar to the one here #1690. New arg `device` will be added, and we need to add the necessary distributed tests as well. See here for the tests https://github.com/pytorch/ignite/blob/master/tests/ignite/contrib/metrics/regression/test_canberra_metric.py#L99
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `ignite/contrib/metrics/average_precision.py`
Content:
```
1 from typing import Callable
2
3 import torch
4
5 from ignite.metrics import EpochMetric
6
7
8 def average_precision_compute_fn(y_preds: torch.Tensor, y_targets: torch.Tensor) -> float:
9 try:
10 from sklearn.metrics import average_precision_score
11 except ImportError:
12 raise RuntimeError("This contrib module requires sklearn to be installed.")
13
14 y_true = y_targets.numpy()
15 y_pred = y_preds.numpy()
16 return average_precision_score(y_true, y_pred)
17
18
19 class AveragePrecision(EpochMetric):
20 """Computes Average Precision accumulating predictions and the ground-truth during an epoch
21 and applying `sklearn.metrics.average_precision_score <https://scikit-learn.org/stable/modules/generated/
22 sklearn.metrics.average_precision_score.html#sklearn.metrics.average_precision_score>`_ .
23
24 Args:
25 output_transform: a callable that is used to transform the
26 :class:`~ignite.engine.engine.Engine`'s ``process_function``'s output into the
27 form expected by the metric. This can be useful if, for example, you have a multi-output model and
28 you want to compute the metric with respect to one of the outputs.
29 check_compute_fn: Default False. If True, `average_precision_score
30 <https://scikit-learn.org/stable/modules/generated/sklearn.metrics.average_precision_score.html
31 #sklearn.metrics.average_precision_score>`_ is run on the first batch of data to ensure there are
32 no issues. User will be warned in case there are any issues computing the function.
33
34 AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or
35 confidence values. To apply an activation to y_pred, use output_transform as shown below:
36
37 .. code-block:: python
38
39 def activated_output_transform(output):
40 y_pred, y = output
41 y_pred = torch.softmax(y_pred, dim=1)
42 return y_pred, y
43
44 avg_precision = AveragePrecision(activated_output_transform)
45
46 """
47
48 def __init__(self, output_transform: Callable = lambda x: x, check_compute_fn: bool = False) -> None:
49 super(AveragePrecision, self).__init__(
50 average_precision_compute_fn, output_transform=output_transform, check_compute_fn=check_compute_fn
51 )
52
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/ignite/contrib/metrics/average_precision.py b/ignite/contrib/metrics/average_precision.py
--- a/ignite/contrib/metrics/average_precision.py
+++ b/ignite/contrib/metrics/average_precision.py
@@ -1,4 +1,4 @@
-from typing import Callable
+from typing import Callable, Optional, Union
import torch
@@ -6,13 +6,10 @@
def average_precision_compute_fn(y_preds: torch.Tensor, y_targets: torch.Tensor) -> float:
- try:
- from sklearn.metrics import average_precision_score
- except ImportError:
- raise RuntimeError("This contrib module requires sklearn to be installed.")
+ from sklearn.metrics import average_precision_score
- y_true = y_targets.numpy()
- y_pred = y_preds.numpy()
+ y_true = y_targets.cpu().numpy()
+ y_pred = y_preds.cpu().numpy()
return average_precision_score(y_true, y_pred)
@@ -30,6 +27,7 @@
<https://scikit-learn.org/stable/modules/generated/sklearn.metrics.average_precision_score.html
#sklearn.metrics.average_precision_score>`_ is run on the first batch of data to ensure there are
no issues. User will be warned in case there are any issues computing the function.
+ device: optional device specification for internal storage.
AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or
confidence values. To apply an activation to y_pred, use output_transform as shown below:
@@ -45,7 +43,21 @@
"""
- def __init__(self, output_transform: Callable = lambda x: x, check_compute_fn: bool = False) -> None:
+ def __init__(
+ self,
+ output_transform: Callable = lambda x: x,
+ check_compute_fn: bool = False,
+ device: Union[str, torch.device] = torch.device("cpu"),
+ ):
+
+ try:
+ from sklearn.metrics import average_precision_score
+ except ImportError:
+ raise RuntimeError("This contrib module requires sklearn to be installed.")
+
super(AveragePrecision, self).__init__(
- average_precision_compute_fn, output_transform=output_transform, check_compute_fn=check_compute_fn
+ average_precision_compute_fn,
+ output_transform=output_transform,
+ check_compute_fn=check_compute_fn,
+ device=device,
)
|
{"golden_diff": "diff --git a/ignite/contrib/metrics/average_precision.py b/ignite/contrib/metrics/average_precision.py\n--- a/ignite/contrib/metrics/average_precision.py\n+++ b/ignite/contrib/metrics/average_precision.py\n@@ -1,4 +1,4 @@\n-from typing import Callable\n+from typing import Callable, Optional, Union\n \n import torch\n \n@@ -6,13 +6,10 @@\n \n \n def average_precision_compute_fn(y_preds: torch.Tensor, y_targets: torch.Tensor) -> float:\n- try:\n- from sklearn.metrics import average_precision_score\n- except ImportError:\n- raise RuntimeError(\"This contrib module requires sklearn to be installed.\")\n+ from sklearn.metrics import average_precision_score\n \n- y_true = y_targets.numpy()\n- y_pred = y_preds.numpy()\n+ y_true = y_targets.cpu().numpy()\n+ y_pred = y_preds.cpu().numpy()\n return average_precision_score(y_true, y_pred)\n \n \n@@ -30,6 +27,7 @@\n <https://scikit-learn.org/stable/modules/generated/sklearn.metrics.average_precision_score.html\n #sklearn.metrics.average_precision_score>`_ is run on the first batch of data to ensure there are\n no issues. User will be warned in case there are any issues computing the function.\n+ device: optional device specification for internal storage.\n \n AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or\n confidence values. To apply an activation to y_pred, use output_transform as shown below:\n@@ -45,7 +43,21 @@\n \n \"\"\"\n \n- def __init__(self, output_transform: Callable = lambda x: x, check_compute_fn: bool = False) -> None:\n+ def __init__(\n+ self,\n+ output_transform: Callable = lambda x: x,\n+ check_compute_fn: bool = False,\n+ device: Union[str, torch.device] = torch.device(\"cpu\"),\n+ ):\n+\n+ try:\n+ from sklearn.metrics import average_precision_score\n+ except ImportError:\n+ raise RuntimeError(\"This contrib module requires sklearn to be installed.\")\n+\n super(AveragePrecision, self).__init__(\n- average_precision_compute_fn, output_transform=output_transform, check_compute_fn=check_compute_fn\n+ average_precision_compute_fn,\n+ output_transform=output_transform,\n+ check_compute_fn=check_compute_fn,\n+ device=device,\n )\n", "issue": "Improve `ignite.contirb.metrics` implementation \nFor these metrics in `ignite.contrib.metrics` :\r\n- [Average Precision](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/average_precision.py)\r\n- [Precision Recall Curve](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/precision_recall_curve.py)\r\n- [Roc Auc](https://github.com/pytorch/ignite/blob/master/ignite/contrib/metrics/roc_auc.py) \r\nWe need to make the implementation simpler. Similar to the one here #1690. New arg `device` will be added, and we need to add the necessary distributed tests as well. See here for the tests https://github.com/pytorch/ignite/blob/master/tests/ignite/contrib/metrics/regression/test_canberra_metric.py#L99\r\n\n", "before_files": [{"content": "from typing import Callable\n\nimport torch\n\nfrom ignite.metrics import EpochMetric\n\n\ndef average_precision_compute_fn(y_preds: torch.Tensor, y_targets: torch.Tensor) -> float:\n try:\n from sklearn.metrics import average_precision_score\n except ImportError:\n raise RuntimeError(\"This contrib module requires sklearn to be installed.\")\n\n y_true = y_targets.numpy()\n y_pred = y_preds.numpy()\n return average_precision_score(y_true, y_pred)\n\n\nclass AveragePrecision(EpochMetric):\n \"\"\"Computes Average Precision accumulating predictions and the ground-truth during an epoch\n and applying `sklearn.metrics.average_precision_score <https://scikit-learn.org/stable/modules/generated/\n sklearn.metrics.average_precision_score.html#sklearn.metrics.average_precision_score>`_ .\n\n Args:\n output_transform: a callable that is used to transform the\n :class:`~ignite.engine.engine.Engine`'s ``process_function``'s output into the\n form expected by the metric. This can be useful if, for example, you have a multi-output model and\n you want to compute the metric with respect to one of the outputs.\n check_compute_fn: Default False. If True, `average_precision_score\n <https://scikit-learn.org/stable/modules/generated/sklearn.metrics.average_precision_score.html\n #sklearn.metrics.average_precision_score>`_ is run on the first batch of data to ensure there are\n no issues. User will be warned in case there are any issues computing the function.\n\n AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or\n confidence values. To apply an activation to y_pred, use output_transform as shown below:\n\n .. code-block:: python\n\n def activated_output_transform(output):\n y_pred, y = output\n y_pred = torch.softmax(y_pred, dim=1)\n return y_pred, y\n\n avg_precision = AveragePrecision(activated_output_transform)\n\n \"\"\"\n\n def __init__(self, output_transform: Callable = lambda x: x, check_compute_fn: bool = False) -> None:\n super(AveragePrecision, self).__init__(\n average_precision_compute_fn, output_transform=output_transform, check_compute_fn=check_compute_fn\n )\n", "path": "ignite/contrib/metrics/average_precision.py"}], "after_files": [{"content": "from typing import Callable, Optional, Union\n\nimport torch\n\nfrom ignite.metrics import EpochMetric\n\n\ndef average_precision_compute_fn(y_preds: torch.Tensor, y_targets: torch.Tensor) -> float:\n from sklearn.metrics import average_precision_score\n\n y_true = y_targets.cpu().numpy()\n y_pred = y_preds.cpu().numpy()\n return average_precision_score(y_true, y_pred)\n\n\nclass AveragePrecision(EpochMetric):\n \"\"\"Computes Average Precision accumulating predictions and the ground-truth during an epoch\n and applying `sklearn.metrics.average_precision_score <https://scikit-learn.org/stable/modules/generated/\n sklearn.metrics.average_precision_score.html#sklearn.metrics.average_precision_score>`_ .\n\n Args:\n output_transform: a callable that is used to transform the\n :class:`~ignite.engine.engine.Engine`'s ``process_function``'s output into the\n form expected by the metric. This can be useful if, for example, you have a multi-output model and\n you want to compute the metric with respect to one of the outputs.\n check_compute_fn: Default False. If True, `average_precision_score\n <https://scikit-learn.org/stable/modules/generated/sklearn.metrics.average_precision_score.html\n #sklearn.metrics.average_precision_score>`_ is run on the first batch of data to ensure there are\n no issues. User will be warned in case there are any issues computing the function.\n device: optional device specification for internal storage.\n\n AveragePrecision expects y to be comprised of 0's and 1's. y_pred must either be probability estimates or\n confidence values. To apply an activation to y_pred, use output_transform as shown below:\n\n .. code-block:: python\n\n def activated_output_transform(output):\n y_pred, y = output\n y_pred = torch.softmax(y_pred, dim=1)\n return y_pred, y\n\n avg_precision = AveragePrecision(activated_output_transform)\n\n \"\"\"\n\n def __init__(\n self,\n output_transform: Callable = lambda x: x,\n check_compute_fn: bool = False,\n device: Union[str, torch.device] = torch.device(\"cpu\"),\n ):\n\n try:\n from sklearn.metrics import average_precision_score\n except ImportError:\n raise RuntimeError(\"This contrib module requires sklearn to be installed.\")\n\n super(AveragePrecision, self).__init__(\n average_precision_compute_fn,\n output_transform=output_transform,\n check_compute_fn=check_compute_fn,\n device=device,\n )\n", "path": "ignite/contrib/metrics/average_precision.py"}]}
| 1,021 | 533 |
gh_patches_debug_24410
|
rasdani/github-patches
|
git_diff
|
pre-commit__pre-commit-977
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
`minimum_pre_commit_version` at the top level in configuration
Similar to the one that's currently allowed on individual hooks, allow this at the top level of the `.pre-commit-config.yaml`
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `pre_commit/clientlib.py`
Content:
```
1 from __future__ import absolute_import
2 from __future__ import unicode_literals
3
4 import argparse
5 import functools
6 import pipes
7 import sys
8
9 import cfgv
10 from aspy.yaml import ordered_load
11 from identify.identify import ALL_TAGS
12
13 import pre_commit.constants as C
14 from pre_commit.error_handler import FatalError
15 from pre_commit.languages.all import all_languages
16
17
18 def check_type_tag(tag):
19 if tag not in ALL_TAGS:
20 raise cfgv.ValidationError(
21 'Type tag {!r} is not recognized. '
22 'Try upgrading identify and pre-commit?'.format(tag),
23 )
24
25
26 def _make_argparser(filenames_help):
27 parser = argparse.ArgumentParser()
28 parser.add_argument('filenames', nargs='*', help=filenames_help)
29 parser.add_argument('-V', '--version', action='version', version=C.VERSION)
30 return parser
31
32
33 MANIFEST_HOOK_DICT = cfgv.Map(
34 'Hook', 'id',
35
36 cfgv.Required('id', cfgv.check_string),
37 cfgv.Required('name', cfgv.check_string),
38 cfgv.Required('entry', cfgv.check_string),
39 cfgv.Required('language', cfgv.check_one_of(all_languages)),
40 cfgv.Optional('alias', cfgv.check_string, ''),
41
42 cfgv.Optional(
43 'files', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '',
44 ),
45 cfgv.Optional(
46 'exclude', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '^$',
47 ),
48 cfgv.Optional('types', cfgv.check_array(check_type_tag), ['file']),
49 cfgv.Optional('exclude_types', cfgv.check_array(check_type_tag), []),
50
51 cfgv.Optional(
52 'additional_dependencies', cfgv.check_array(cfgv.check_string), [],
53 ),
54 cfgv.Optional('args', cfgv.check_array(cfgv.check_string), []),
55 cfgv.Optional('always_run', cfgv.check_bool, False),
56 cfgv.Optional('pass_filenames', cfgv.check_bool, True),
57 cfgv.Optional('description', cfgv.check_string, ''),
58 cfgv.Optional('language_version', cfgv.check_string, C.DEFAULT),
59 cfgv.Optional('log_file', cfgv.check_string, ''),
60 cfgv.Optional('minimum_pre_commit_version', cfgv.check_string, '0'),
61 cfgv.Optional('require_serial', cfgv.check_bool, False),
62 cfgv.Optional('stages', cfgv.check_array(cfgv.check_one_of(C.STAGES)), []),
63 cfgv.Optional('verbose', cfgv.check_bool, False),
64 )
65 MANIFEST_SCHEMA = cfgv.Array(MANIFEST_HOOK_DICT)
66
67
68 class InvalidManifestError(FatalError):
69 pass
70
71
72 load_manifest = functools.partial(
73 cfgv.load_from_filename,
74 schema=MANIFEST_SCHEMA,
75 load_strategy=ordered_load,
76 exc_tp=InvalidManifestError,
77 )
78
79
80 def validate_manifest_main(argv=None):
81 parser = _make_argparser('Manifest filenames.')
82 args = parser.parse_args(argv)
83 ret = 0
84 for filename in args.filenames:
85 try:
86 load_manifest(filename)
87 except InvalidManifestError as e:
88 print(e)
89 ret = 1
90 return ret
91
92
93 LOCAL = 'local'
94 META = 'meta'
95
96
97 class MigrateShaToRev(object):
98 @staticmethod
99 def _cond(key):
100 return cfgv.Conditional(
101 key, cfgv.check_string,
102 condition_key='repo',
103 condition_value=cfgv.NotIn(LOCAL, META),
104 ensure_absent=True,
105 )
106
107 def check(self, dct):
108 if dct.get('repo') in {LOCAL, META}:
109 self._cond('rev').check(dct)
110 self._cond('sha').check(dct)
111 elif 'sha' in dct and 'rev' in dct:
112 raise cfgv.ValidationError('Cannot specify both sha and rev')
113 elif 'sha' in dct:
114 self._cond('sha').check(dct)
115 else:
116 self._cond('rev').check(dct)
117
118 def apply_default(self, dct):
119 if 'sha' in dct:
120 dct['rev'] = dct.pop('sha')
121
122 def remove_default(self, dct):
123 pass
124
125
126 def _entry(modname):
127 """the hook `entry` is passed through `shlex.split()` by the command
128 runner, so to prevent issues with spaces and backslashes (on Windows)
129 it must be quoted here.
130 """
131 return '{} -m pre_commit.meta_hooks.{}'.format(
132 pipes.quote(sys.executable), modname,
133 )
134
135
136 _meta = (
137 (
138 'check-hooks-apply', (
139 ('name', 'Check hooks apply to the repository'),
140 ('files', C.CONFIG_FILE),
141 ('entry', _entry('check_hooks_apply')),
142 ),
143 ),
144 (
145 'check-useless-excludes', (
146 ('name', 'Check for useless excludes'),
147 ('files', C.CONFIG_FILE),
148 ('entry', _entry('check_useless_excludes')),
149 ),
150 ),
151 (
152 'identity', (
153 ('name', 'identity'),
154 ('verbose', True),
155 ('entry', _entry('identity')),
156 ),
157 ),
158 )
159
160 META_HOOK_DICT = cfgv.Map(
161 'Hook', 'id',
162 cfgv.Required('id', cfgv.check_string),
163 cfgv.Required('id', cfgv.check_one_of(tuple(k for k, _ in _meta))),
164 # language must be system
165 cfgv.Optional('language', cfgv.check_one_of({'system'}), 'system'),
166 *([
167 # default to the hook definition for the meta hooks
168 cfgv.ConditionalOptional(key, cfgv.check_any, value, 'id', hook_id)
169 for hook_id, values in _meta
170 for key, value in values
171 ] + [
172 # default to the "manifest" parsing
173 cfgv.OptionalNoDefault(item.key, item.check_fn)
174 # these will always be defaulted above
175 if item.key in {'name', 'language', 'entry'} else
176 item
177 for item in MANIFEST_HOOK_DICT.items
178 ])
179 )
180 CONFIG_HOOK_DICT = cfgv.Map(
181 'Hook', 'id',
182
183 cfgv.Required('id', cfgv.check_string),
184
185 # All keys in manifest hook dict are valid in a config hook dict, but
186 # are optional.
187 # No defaults are provided here as the config is merged on top of the
188 # manifest.
189 *[
190 cfgv.OptionalNoDefault(item.key, item.check_fn)
191 for item in MANIFEST_HOOK_DICT.items
192 if item.key != 'id'
193 ]
194 )
195 CONFIG_REPO_DICT = cfgv.Map(
196 'Repository', 'repo',
197
198 cfgv.Required('repo', cfgv.check_string),
199
200 cfgv.ConditionalRecurse(
201 'hooks', cfgv.Array(CONFIG_HOOK_DICT),
202 'repo', cfgv.NotIn(LOCAL, META),
203 ),
204 cfgv.ConditionalRecurse(
205 'hooks', cfgv.Array(MANIFEST_HOOK_DICT),
206 'repo', LOCAL,
207 ),
208 cfgv.ConditionalRecurse(
209 'hooks', cfgv.Array(META_HOOK_DICT),
210 'repo', META,
211 ),
212
213 MigrateShaToRev(),
214 )
215 DEFAULT_LANGUAGE_VERSION = cfgv.Map(
216 'DefaultLanguageVersion', None,
217 cfgv.NoAdditionalKeys(all_languages),
218 *[cfgv.Optional(x, cfgv.check_string, C.DEFAULT) for x in all_languages]
219 )
220 CONFIG_SCHEMA = cfgv.Map(
221 'Config', None,
222
223 cfgv.RequiredRecurse('repos', cfgv.Array(CONFIG_REPO_DICT)),
224 cfgv.OptionalRecurse(
225 'default_language_version', DEFAULT_LANGUAGE_VERSION, {},
226 ),
227 cfgv.Optional(
228 'default_stages',
229 cfgv.check_array(cfgv.check_one_of(C.STAGES)),
230 C.STAGES,
231 ),
232 cfgv.Optional('exclude', cfgv.check_regex, '^$'),
233 cfgv.Optional('fail_fast', cfgv.check_bool, False),
234 )
235
236
237 class InvalidConfigError(FatalError):
238 pass
239
240
241 def ordered_load_normalize_legacy_config(contents):
242 data = ordered_load(contents)
243 if isinstance(data, list):
244 # TODO: Once happy, issue a deprecation warning and instructions
245 return {'repos': data}
246 else:
247 return data
248
249
250 load_config = functools.partial(
251 cfgv.load_from_filename,
252 schema=CONFIG_SCHEMA,
253 load_strategy=ordered_load_normalize_legacy_config,
254 exc_tp=InvalidConfigError,
255 )
256
257
258 def validate_config_main(argv=None):
259 parser = _make_argparser('Config filenames.')
260 args = parser.parse_args(argv)
261 ret = 0
262 for filename in args.filenames:
263 try:
264 load_config(filename)
265 except InvalidConfigError as e:
266 print(e)
267 ret = 1
268 return ret
269
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/pre_commit/clientlib.py b/pre_commit/clientlib.py
--- a/pre_commit/clientlib.py
+++ b/pre_commit/clientlib.py
@@ -13,6 +13,7 @@
import pre_commit.constants as C
from pre_commit.error_handler import FatalError
from pre_commit.languages.all import all_languages
+from pre_commit.util import parse_version
def check_type_tag(tag):
@@ -23,6 +24,16 @@
)
+def check_min_version(version):
+ if parse_version(version) > parse_version(C.VERSION):
+ raise cfgv.ValidationError(
+ 'pre-commit version {} is required but version {} is installed. '
+ 'Perhaps run `pip install --upgrade pre-commit`.'.format(
+ version, C.VERSION,
+ ),
+ )
+
+
def _make_argparser(filenames_help):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help=filenames_help)
@@ -231,6 +242,11 @@
),
cfgv.Optional('exclude', cfgv.check_regex, '^$'),
cfgv.Optional('fail_fast', cfgv.check_bool, False),
+ cfgv.Optional(
+ 'minimum_pre_commit_version',
+ cfgv.check_and(cfgv.check_string, check_min_version),
+ '0',
+ ),
)
|
{"golden_diff": "diff --git a/pre_commit/clientlib.py b/pre_commit/clientlib.py\n--- a/pre_commit/clientlib.py\n+++ b/pre_commit/clientlib.py\n@@ -13,6 +13,7 @@\n import pre_commit.constants as C\n from pre_commit.error_handler import FatalError\n from pre_commit.languages.all import all_languages\n+from pre_commit.util import parse_version\n \n \n def check_type_tag(tag):\n@@ -23,6 +24,16 @@\n )\n \n \n+def check_min_version(version):\n+ if parse_version(version) > parse_version(C.VERSION):\n+ raise cfgv.ValidationError(\n+ 'pre-commit version {} is required but version {} is installed. '\n+ 'Perhaps run `pip install --upgrade pre-commit`.'.format(\n+ version, C.VERSION,\n+ ),\n+ )\n+\n+\n def _make_argparser(filenames_help):\n parser = argparse.ArgumentParser()\n parser.add_argument('filenames', nargs='*', help=filenames_help)\n@@ -231,6 +242,11 @@\n ),\n cfgv.Optional('exclude', cfgv.check_regex, '^$'),\n cfgv.Optional('fail_fast', cfgv.check_bool, False),\n+ cfgv.Optional(\n+ 'minimum_pre_commit_version',\n+ cfgv.check_and(cfgv.check_string, check_min_version),\n+ '0',\n+ ),\n )\n", "issue": "`minimum_pre_commit_version` at the top level in configuration\nSimilar to the one that's currently allowed on individual hooks, allow this at the top level of the `.pre-commit-config.yaml`\n", "before_files": [{"content": "from __future__ import absolute_import\nfrom __future__ import unicode_literals\n\nimport argparse\nimport functools\nimport pipes\nimport sys\n\nimport cfgv\nfrom aspy.yaml import ordered_load\nfrom identify.identify import ALL_TAGS\n\nimport pre_commit.constants as C\nfrom pre_commit.error_handler import FatalError\nfrom pre_commit.languages.all import all_languages\n\n\ndef check_type_tag(tag):\n if tag not in ALL_TAGS:\n raise cfgv.ValidationError(\n 'Type tag {!r} is not recognized. '\n 'Try upgrading identify and pre-commit?'.format(tag),\n )\n\n\ndef _make_argparser(filenames_help):\n parser = argparse.ArgumentParser()\n parser.add_argument('filenames', nargs='*', help=filenames_help)\n parser.add_argument('-V', '--version', action='version', version=C.VERSION)\n return parser\n\n\nMANIFEST_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n\n cfgv.Required('id', cfgv.check_string),\n cfgv.Required('name', cfgv.check_string),\n cfgv.Required('entry', cfgv.check_string),\n cfgv.Required('language', cfgv.check_one_of(all_languages)),\n cfgv.Optional('alias', cfgv.check_string, ''),\n\n cfgv.Optional(\n 'files', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '',\n ),\n cfgv.Optional(\n 'exclude', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '^$',\n ),\n cfgv.Optional('types', cfgv.check_array(check_type_tag), ['file']),\n cfgv.Optional('exclude_types', cfgv.check_array(check_type_tag), []),\n\n cfgv.Optional(\n 'additional_dependencies', cfgv.check_array(cfgv.check_string), [],\n ),\n cfgv.Optional('args', cfgv.check_array(cfgv.check_string), []),\n cfgv.Optional('always_run', cfgv.check_bool, False),\n cfgv.Optional('pass_filenames', cfgv.check_bool, True),\n cfgv.Optional('description', cfgv.check_string, ''),\n cfgv.Optional('language_version', cfgv.check_string, C.DEFAULT),\n cfgv.Optional('log_file', cfgv.check_string, ''),\n cfgv.Optional('minimum_pre_commit_version', cfgv.check_string, '0'),\n cfgv.Optional('require_serial', cfgv.check_bool, False),\n cfgv.Optional('stages', cfgv.check_array(cfgv.check_one_of(C.STAGES)), []),\n cfgv.Optional('verbose', cfgv.check_bool, False),\n)\nMANIFEST_SCHEMA = cfgv.Array(MANIFEST_HOOK_DICT)\n\n\nclass InvalidManifestError(FatalError):\n pass\n\n\nload_manifest = functools.partial(\n cfgv.load_from_filename,\n schema=MANIFEST_SCHEMA,\n load_strategy=ordered_load,\n exc_tp=InvalidManifestError,\n)\n\n\ndef validate_manifest_main(argv=None):\n parser = _make_argparser('Manifest filenames.')\n args = parser.parse_args(argv)\n ret = 0\n for filename in args.filenames:\n try:\n load_manifest(filename)\n except InvalidManifestError as e:\n print(e)\n ret = 1\n return ret\n\n\nLOCAL = 'local'\nMETA = 'meta'\n\n\nclass MigrateShaToRev(object):\n @staticmethod\n def _cond(key):\n return cfgv.Conditional(\n key, cfgv.check_string,\n condition_key='repo',\n condition_value=cfgv.NotIn(LOCAL, META),\n ensure_absent=True,\n )\n\n def check(self, dct):\n if dct.get('repo') in {LOCAL, META}:\n self._cond('rev').check(dct)\n self._cond('sha').check(dct)\n elif 'sha' in dct and 'rev' in dct:\n raise cfgv.ValidationError('Cannot specify both sha and rev')\n elif 'sha' in dct:\n self._cond('sha').check(dct)\n else:\n self._cond('rev').check(dct)\n\n def apply_default(self, dct):\n if 'sha' in dct:\n dct['rev'] = dct.pop('sha')\n\n def remove_default(self, dct):\n pass\n\n\ndef _entry(modname):\n \"\"\"the hook `entry` is passed through `shlex.split()` by the command\n runner, so to prevent issues with spaces and backslashes (on Windows)\n it must be quoted here.\n \"\"\"\n return '{} -m pre_commit.meta_hooks.{}'.format(\n pipes.quote(sys.executable), modname,\n )\n\n\n_meta = (\n (\n 'check-hooks-apply', (\n ('name', 'Check hooks apply to the repository'),\n ('files', C.CONFIG_FILE),\n ('entry', _entry('check_hooks_apply')),\n ),\n ),\n (\n 'check-useless-excludes', (\n ('name', 'Check for useless excludes'),\n ('files', C.CONFIG_FILE),\n ('entry', _entry('check_useless_excludes')),\n ),\n ),\n (\n 'identity', (\n ('name', 'identity'),\n ('verbose', True),\n ('entry', _entry('identity')),\n ),\n ),\n)\n\nMETA_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n cfgv.Required('id', cfgv.check_string),\n cfgv.Required('id', cfgv.check_one_of(tuple(k for k, _ in _meta))),\n # language must be system\n cfgv.Optional('language', cfgv.check_one_of({'system'}), 'system'),\n *([\n # default to the hook definition for the meta hooks\n cfgv.ConditionalOptional(key, cfgv.check_any, value, 'id', hook_id)\n for hook_id, values in _meta\n for key, value in values\n ] + [\n # default to the \"manifest\" parsing\n cfgv.OptionalNoDefault(item.key, item.check_fn)\n # these will always be defaulted above\n if item.key in {'name', 'language', 'entry'} else\n item\n for item in MANIFEST_HOOK_DICT.items\n ])\n)\nCONFIG_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n\n cfgv.Required('id', cfgv.check_string),\n\n # All keys in manifest hook dict are valid in a config hook dict, but\n # are optional.\n # No defaults are provided here as the config is merged on top of the\n # manifest.\n *[\n cfgv.OptionalNoDefault(item.key, item.check_fn)\n for item in MANIFEST_HOOK_DICT.items\n if item.key != 'id'\n ]\n)\nCONFIG_REPO_DICT = cfgv.Map(\n 'Repository', 'repo',\n\n cfgv.Required('repo', cfgv.check_string),\n\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(CONFIG_HOOK_DICT),\n 'repo', cfgv.NotIn(LOCAL, META),\n ),\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(MANIFEST_HOOK_DICT),\n 'repo', LOCAL,\n ),\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(META_HOOK_DICT),\n 'repo', META,\n ),\n\n MigrateShaToRev(),\n)\nDEFAULT_LANGUAGE_VERSION = cfgv.Map(\n 'DefaultLanguageVersion', None,\n cfgv.NoAdditionalKeys(all_languages),\n *[cfgv.Optional(x, cfgv.check_string, C.DEFAULT) for x in all_languages]\n)\nCONFIG_SCHEMA = cfgv.Map(\n 'Config', None,\n\n cfgv.RequiredRecurse('repos', cfgv.Array(CONFIG_REPO_DICT)),\n cfgv.OptionalRecurse(\n 'default_language_version', DEFAULT_LANGUAGE_VERSION, {},\n ),\n cfgv.Optional(\n 'default_stages',\n cfgv.check_array(cfgv.check_one_of(C.STAGES)),\n C.STAGES,\n ),\n cfgv.Optional('exclude', cfgv.check_regex, '^$'),\n cfgv.Optional('fail_fast', cfgv.check_bool, False),\n)\n\n\nclass InvalidConfigError(FatalError):\n pass\n\n\ndef ordered_load_normalize_legacy_config(contents):\n data = ordered_load(contents)\n if isinstance(data, list):\n # TODO: Once happy, issue a deprecation warning and instructions\n return {'repos': data}\n else:\n return data\n\n\nload_config = functools.partial(\n cfgv.load_from_filename,\n schema=CONFIG_SCHEMA,\n load_strategy=ordered_load_normalize_legacy_config,\n exc_tp=InvalidConfigError,\n)\n\n\ndef validate_config_main(argv=None):\n parser = _make_argparser('Config filenames.')\n args = parser.parse_args(argv)\n ret = 0\n for filename in args.filenames:\n try:\n load_config(filename)\n except InvalidConfigError as e:\n print(e)\n ret = 1\n return ret\n", "path": "pre_commit/clientlib.py"}], "after_files": [{"content": "from __future__ import absolute_import\nfrom __future__ import unicode_literals\n\nimport argparse\nimport functools\nimport pipes\nimport sys\n\nimport cfgv\nfrom aspy.yaml import ordered_load\nfrom identify.identify import ALL_TAGS\n\nimport pre_commit.constants as C\nfrom pre_commit.error_handler import FatalError\nfrom pre_commit.languages.all import all_languages\nfrom pre_commit.util import parse_version\n\n\ndef check_type_tag(tag):\n if tag not in ALL_TAGS:\n raise cfgv.ValidationError(\n 'Type tag {!r} is not recognized. '\n 'Try upgrading identify and pre-commit?'.format(tag),\n )\n\n\ndef check_min_version(version):\n if parse_version(version) > parse_version(C.VERSION):\n raise cfgv.ValidationError(\n 'pre-commit version {} is required but version {} is installed. '\n 'Perhaps run `pip install --upgrade pre-commit`.'.format(\n version, C.VERSION,\n ),\n )\n\n\ndef _make_argparser(filenames_help):\n parser = argparse.ArgumentParser()\n parser.add_argument('filenames', nargs='*', help=filenames_help)\n parser.add_argument('-V', '--version', action='version', version=C.VERSION)\n return parser\n\n\nMANIFEST_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n\n cfgv.Required('id', cfgv.check_string),\n cfgv.Required('name', cfgv.check_string),\n cfgv.Required('entry', cfgv.check_string),\n cfgv.Required('language', cfgv.check_one_of(all_languages)),\n cfgv.Optional('alias', cfgv.check_string, ''),\n\n cfgv.Optional(\n 'files', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '',\n ),\n cfgv.Optional(\n 'exclude', cfgv.check_and(cfgv.check_string, cfgv.check_regex), '^$',\n ),\n cfgv.Optional('types', cfgv.check_array(check_type_tag), ['file']),\n cfgv.Optional('exclude_types', cfgv.check_array(check_type_tag), []),\n\n cfgv.Optional(\n 'additional_dependencies', cfgv.check_array(cfgv.check_string), [],\n ),\n cfgv.Optional('args', cfgv.check_array(cfgv.check_string), []),\n cfgv.Optional('always_run', cfgv.check_bool, False),\n cfgv.Optional('pass_filenames', cfgv.check_bool, True),\n cfgv.Optional('description', cfgv.check_string, ''),\n cfgv.Optional('language_version', cfgv.check_string, C.DEFAULT),\n cfgv.Optional('log_file', cfgv.check_string, ''),\n cfgv.Optional('minimum_pre_commit_version', cfgv.check_string, '0'),\n cfgv.Optional('require_serial', cfgv.check_bool, False),\n cfgv.Optional('stages', cfgv.check_array(cfgv.check_one_of(C.STAGES)), []),\n cfgv.Optional('verbose', cfgv.check_bool, False),\n)\nMANIFEST_SCHEMA = cfgv.Array(MANIFEST_HOOK_DICT)\n\n\nclass InvalidManifestError(FatalError):\n pass\n\n\nload_manifest = functools.partial(\n cfgv.load_from_filename,\n schema=MANIFEST_SCHEMA,\n load_strategy=ordered_load,\n exc_tp=InvalidManifestError,\n)\n\n\ndef validate_manifest_main(argv=None):\n parser = _make_argparser('Manifest filenames.')\n args = parser.parse_args(argv)\n ret = 0\n for filename in args.filenames:\n try:\n load_manifest(filename)\n except InvalidManifestError as e:\n print(e)\n ret = 1\n return ret\n\n\nLOCAL = 'local'\nMETA = 'meta'\n\n\nclass MigrateShaToRev(object):\n @staticmethod\n def _cond(key):\n return cfgv.Conditional(\n key, cfgv.check_string,\n condition_key='repo',\n condition_value=cfgv.NotIn(LOCAL, META),\n ensure_absent=True,\n )\n\n def check(self, dct):\n if dct.get('repo') in {LOCAL, META}:\n self._cond('rev').check(dct)\n self._cond('sha').check(dct)\n elif 'sha' in dct and 'rev' in dct:\n raise cfgv.ValidationError('Cannot specify both sha and rev')\n elif 'sha' in dct:\n self._cond('sha').check(dct)\n else:\n self._cond('rev').check(dct)\n\n def apply_default(self, dct):\n if 'sha' in dct:\n dct['rev'] = dct.pop('sha')\n\n def remove_default(self, dct):\n pass\n\n\ndef _entry(modname):\n \"\"\"the hook `entry` is passed through `shlex.split()` by the command\n runner, so to prevent issues with spaces and backslashes (on Windows)\n it must be quoted here.\n \"\"\"\n return '{} -m pre_commit.meta_hooks.{}'.format(\n pipes.quote(sys.executable), modname,\n )\n\n\n_meta = (\n (\n 'check-hooks-apply', (\n ('name', 'Check hooks apply to the repository'),\n ('files', C.CONFIG_FILE),\n ('entry', _entry('check_hooks_apply')),\n ),\n ),\n (\n 'check-useless-excludes', (\n ('name', 'Check for useless excludes'),\n ('files', C.CONFIG_FILE),\n ('entry', _entry('check_useless_excludes')),\n ),\n ),\n (\n 'identity', (\n ('name', 'identity'),\n ('verbose', True),\n ('entry', _entry('identity')),\n ),\n ),\n)\n\nMETA_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n cfgv.Required('id', cfgv.check_string),\n cfgv.Required('id', cfgv.check_one_of(tuple(k for k, _ in _meta))),\n # language must be system\n cfgv.Optional('language', cfgv.check_one_of({'system'}), 'system'),\n *([\n # default to the hook definition for the meta hooks\n cfgv.ConditionalOptional(key, cfgv.check_any, value, 'id', hook_id)\n for hook_id, values in _meta\n for key, value in values\n ] + [\n # default to the \"manifest\" parsing\n cfgv.OptionalNoDefault(item.key, item.check_fn)\n # these will always be defaulted above\n if item.key in {'name', 'language', 'entry'} else\n item\n for item in MANIFEST_HOOK_DICT.items\n ])\n)\nCONFIG_HOOK_DICT = cfgv.Map(\n 'Hook', 'id',\n\n cfgv.Required('id', cfgv.check_string),\n\n # All keys in manifest hook dict are valid in a config hook dict, but\n # are optional.\n # No defaults are provided here as the config is merged on top of the\n # manifest.\n *[\n cfgv.OptionalNoDefault(item.key, item.check_fn)\n for item in MANIFEST_HOOK_DICT.items\n if item.key != 'id'\n ]\n)\nCONFIG_REPO_DICT = cfgv.Map(\n 'Repository', 'repo',\n\n cfgv.Required('repo', cfgv.check_string),\n\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(CONFIG_HOOK_DICT),\n 'repo', cfgv.NotIn(LOCAL, META),\n ),\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(MANIFEST_HOOK_DICT),\n 'repo', LOCAL,\n ),\n cfgv.ConditionalRecurse(\n 'hooks', cfgv.Array(META_HOOK_DICT),\n 'repo', META,\n ),\n\n MigrateShaToRev(),\n)\nDEFAULT_LANGUAGE_VERSION = cfgv.Map(\n 'DefaultLanguageVersion', None,\n cfgv.NoAdditionalKeys(all_languages),\n *[cfgv.Optional(x, cfgv.check_string, C.DEFAULT) for x in all_languages]\n)\nCONFIG_SCHEMA = cfgv.Map(\n 'Config', None,\n\n cfgv.RequiredRecurse('repos', cfgv.Array(CONFIG_REPO_DICT)),\n cfgv.OptionalRecurse(\n 'default_language_version', DEFAULT_LANGUAGE_VERSION, {},\n ),\n cfgv.Optional(\n 'default_stages',\n cfgv.check_array(cfgv.check_one_of(C.STAGES)),\n C.STAGES,\n ),\n cfgv.Optional('exclude', cfgv.check_regex, '^$'),\n cfgv.Optional('fail_fast', cfgv.check_bool, False),\n cfgv.Optional(\n 'minimum_pre_commit_version',\n cfgv.check_and(cfgv.check_string, check_min_version),\n '0',\n ),\n)\n\n\nclass InvalidConfigError(FatalError):\n pass\n\n\ndef ordered_load_normalize_legacy_config(contents):\n data = ordered_load(contents)\n if isinstance(data, list):\n # TODO: Once happy, issue a deprecation warning and instructions\n return {'repos': data}\n else:\n return data\n\n\nload_config = functools.partial(\n cfgv.load_from_filename,\n schema=CONFIG_SCHEMA,\n load_strategy=ordered_load_normalize_legacy_config,\n exc_tp=InvalidConfigError,\n)\n\n\ndef validate_config_main(argv=None):\n parser = _make_argparser('Config filenames.')\n args = parser.parse_args(argv)\n ret = 0\n for filename in args.filenames:\n try:\n load_config(filename)\n except InvalidConfigError as e:\n print(e)\n ret = 1\n return ret\n", "path": "pre_commit/clientlib.py"}]}
| 2,877 | 293 |
gh_patches_debug_37409
|
rasdani/github-patches
|
git_diff
|
alltheplaces__alltheplaces-2875
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Spider tacocabana is broken
During the global build at 2021-05-26-14-42-23, spider **tacocabana** failed with **0 features** and **1 errors**.
Here's [the log](https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/logs/tacocabana.log) and [the output](https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/output/tacocabana.geojson) ([on a map](https://data.alltheplaces.xyz/map.html?show=https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/output/tacocabana.geojson))
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `locations/spiders/tacocabana.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 import scrapy
3 import json
4 import re
5
6 from locations.items import GeojsonPointItem
7
8 class TacocabanaSpider(scrapy.Spider):
9 name = "tacocabana"
10 item_attributes = { 'brand': "Taco Cabana" }
11 allowed_domains = ["www.tacocabana.com"]
12 start_urls = (
13 "http://www.tacocabana.com/wp-admin/admin-ajax.php?action=get_ajax_processor&processor=get-locations&queryType=&postID=816",
14 )
15
16 def parse(self, response):
17 data = json.loads(re.sub(r"\s<.*?>.*<.*?>\s", "", response.body_as_unicode()))
18
19 for store in data:
20 properties = {
21 "phone" : store["phone_number"],
22 "ref" : str(store["locator_store_number"]),
23 "name" : store["post_title"],
24 "opening_hours" : store["hours"],
25 "website" : store["permalink"],
26 "lat" : store["x_coordinate"],
27 "lon" : store["y_coordinate"],
28 "street" : store["street_address_1"] + store["street_address_2"],
29 "city" : store["city"],
30 "state" : store["state"],
31 "postcode" : store["zip_code"]
32 }
33
34 yield GeojsonPointItem(**properties)
35
36 else:
37 self.logger.info("No results")
38
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/locations/spiders/tacocabana.py b/locations/spiders/tacocabana.py
--- a/locations/spiders/tacocabana.py
+++ b/locations/spiders/tacocabana.py
@@ -1,37 +1,55 @@
# -*- coding: utf-8 -*-
import scrapy
-import json
-import re
from locations.items import GeojsonPointItem
+
class TacocabanaSpider(scrapy.Spider):
name = "tacocabana"
- item_attributes = { 'brand': "Taco Cabana" }
- allowed_domains = ["www.tacocabana.com"]
- start_urls = (
- "http://www.tacocabana.com/wp-admin/admin-ajax.php?action=get_ajax_processor&processor=get-locations&queryType=&postID=816",
- )
-
+ item_attributes = {"brand": "Taco Cabana"}
+ allowed_domains = ["api.koala.fuzzhq.com"]
+
+ def start_requests(self):
+ yield scrapy.http.JsonRequest(
+ "https://api.koala.fuzzhq.com/oauth/access_token",
+ data={
+ "client_id": "3nA4STkGif0fZGApqxMlVewy3h8HN6Fsy7jVOACP",
+ "client_secret": "8oBU5gWiNg04zYzz61hN3ETrTIzvmbGyeLCX0F1s",
+ "grant_type": "ordering_app_credentials",
+ "scope": "group:ordering_app",
+ },
+ callback=self.fetch_locations,
+ )
+
+ def fetch_locations(self, response):
+ self.access_token = response.json()["access_token"]
+ yield self.request(
+ "https://api.koala.fuzzhq.com/v1/ordering/store-locations/?include[]=operating_hours&include[]=attributes&per_page=50"
+ )
+
+ def request(self, url):
+ return scrapy.Request(
+ url, headers={"Authorization": f"Bearer {self.access_token}"}
+ )
+
def parse(self, response):
- data = json.loads(re.sub(r"\s<.*?>.*<.*?>\s", "", response.body_as_unicode()))
+ data = response.json()
- for store in data:
+ for store in data["data"]:
properties = {
- "phone" : store["phone_number"],
- "ref" : str(store["locator_store_number"]),
- "name" : store["post_title"],
- "opening_hours" : store["hours"],
- "website" : store["permalink"],
- "lat" : store["x_coordinate"],
- "lon" : store["y_coordinate"],
- "street" : store["street_address_1"] + store["street_address_2"],
- "city" : store["city"],
- "state" : store["state"],
- "postcode" : store["zip_code"]
+ "website": f'https://olo.tacocabana.com/menu/{store["slug"]}?showInfoModal=true',
+ "ref": store["brand_id"],
+ "lat": store["latitude"],
+ "lon": store["longitude"],
+ "addr_full": store["street_address"],
+ "city": store["city"],
+ "state": store["cached_data"]["state"],
+ "country": store["country"],
+ "postcode": store["zip_code"],
+ "phone": store["phone_number"],
}
-
yield GeojsonPointItem(**properties)
-
- else:
- self.logger.info("No results")
+
+ next_url = data["meta"]["pagination"]["links"]["next"]
+ if next_url:
+ yield self.request(next_url)
|
{"golden_diff": "diff --git a/locations/spiders/tacocabana.py b/locations/spiders/tacocabana.py\n--- a/locations/spiders/tacocabana.py\n+++ b/locations/spiders/tacocabana.py\n@@ -1,37 +1,55 @@\n # -*- coding: utf-8 -*-\n import scrapy\n-import json\n-import re\n \n from locations.items import GeojsonPointItem\n \n+\n class TacocabanaSpider(scrapy.Spider):\n name = \"tacocabana\"\n- item_attributes = { 'brand': \"Taco Cabana\" }\n- allowed_domains = [\"www.tacocabana.com\"]\n- start_urls = (\n- \"http://www.tacocabana.com/wp-admin/admin-ajax.php?action=get_ajax_processor&processor=get-locations&queryType=&postID=816\",\n- )\n- \n+ item_attributes = {\"brand\": \"Taco Cabana\"}\n+ allowed_domains = [\"api.koala.fuzzhq.com\"]\n+\n+ def start_requests(self):\n+ yield scrapy.http.JsonRequest(\n+ \"https://api.koala.fuzzhq.com/oauth/access_token\",\n+ data={\n+ \"client_id\": \"3nA4STkGif0fZGApqxMlVewy3h8HN6Fsy7jVOACP\",\n+ \"client_secret\": \"8oBU5gWiNg04zYzz61hN3ETrTIzvmbGyeLCX0F1s\",\n+ \"grant_type\": \"ordering_app_credentials\",\n+ \"scope\": \"group:ordering_app\",\n+ },\n+ callback=self.fetch_locations,\n+ )\n+\n+ def fetch_locations(self, response):\n+ self.access_token = response.json()[\"access_token\"]\n+ yield self.request(\n+ \"https://api.koala.fuzzhq.com/v1/ordering/store-locations/?include[]=operating_hours&include[]=attributes&per_page=50\"\n+ )\n+\n+ def request(self, url):\n+ return scrapy.Request(\n+ url, headers={\"Authorization\": f\"Bearer {self.access_token}\"}\n+ )\n+\n def parse(self, response):\n- data = json.loads(re.sub(r\"\\s<.*?>.*<.*?>\\s\", \"\", response.body_as_unicode()))\n+ data = response.json()\n \n- for store in data:\n+ for store in data[\"data\"]:\n properties = {\n- \"phone\" : store[\"phone_number\"],\n- \"ref\" : str(store[\"locator_store_number\"]),\n- \"name\" : store[\"post_title\"],\n- \"opening_hours\" : store[\"hours\"],\n- \"website\" : store[\"permalink\"],\n- \"lat\" : store[\"x_coordinate\"],\n- \"lon\" : store[\"y_coordinate\"],\n- \"street\" : store[\"street_address_1\"] + store[\"street_address_2\"],\n- \"city\" : store[\"city\"],\n- \"state\" : store[\"state\"],\n- \"postcode\" : store[\"zip_code\"]\n+ \"website\": f'https://olo.tacocabana.com/menu/{store[\"slug\"]}?showInfoModal=true',\n+ \"ref\": store[\"brand_id\"],\n+ \"lat\": store[\"latitude\"],\n+ \"lon\": store[\"longitude\"],\n+ \"addr_full\": store[\"street_address\"],\n+ \"city\": store[\"city\"],\n+ \"state\": store[\"cached_data\"][\"state\"],\n+ \"country\": store[\"country\"],\n+ \"postcode\": store[\"zip_code\"],\n+ \"phone\": store[\"phone_number\"],\n }\n- \n yield GeojsonPointItem(**properties)\n- \n- else:\n- self.logger.info(\"No results\")\n+\n+ next_url = data[\"meta\"][\"pagination\"][\"links\"][\"next\"]\n+ if next_url:\n+ yield self.request(next_url)\n", "issue": "Spider tacocabana is broken\nDuring the global build at 2021-05-26-14-42-23, spider **tacocabana** failed with **0 features** and **1 errors**.\n\nHere's [the log](https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/logs/tacocabana.log) and [the output](https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/output/tacocabana.geojson) ([on a map](https://data.alltheplaces.xyz/map.html?show=https://data.alltheplaces.xyz/runs/2021-05-26-14-42-23/output/tacocabana.geojson))\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\nimport scrapy\nimport json\nimport re\n\nfrom locations.items import GeojsonPointItem\n\nclass TacocabanaSpider(scrapy.Spider):\n name = \"tacocabana\"\n item_attributes = { 'brand': \"Taco Cabana\" }\n allowed_domains = [\"www.tacocabana.com\"]\n start_urls = (\n \"http://www.tacocabana.com/wp-admin/admin-ajax.php?action=get_ajax_processor&processor=get-locations&queryType=&postID=816\",\n )\n \n def parse(self, response):\n data = json.loads(re.sub(r\"\\s<.*?>.*<.*?>\\s\", \"\", response.body_as_unicode()))\n\n for store in data:\n properties = {\n \"phone\" : store[\"phone_number\"],\n \"ref\" : str(store[\"locator_store_number\"]),\n \"name\" : store[\"post_title\"],\n \"opening_hours\" : store[\"hours\"],\n \"website\" : store[\"permalink\"],\n \"lat\" : store[\"x_coordinate\"],\n \"lon\" : store[\"y_coordinate\"],\n \"street\" : store[\"street_address_1\"] + store[\"street_address_2\"],\n \"city\" : store[\"city\"],\n \"state\" : store[\"state\"],\n \"postcode\" : store[\"zip_code\"]\n }\n \n yield GeojsonPointItem(**properties)\n \n else:\n self.logger.info(\"No results\")\n", "path": "locations/spiders/tacocabana.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\nimport scrapy\n\nfrom locations.items import GeojsonPointItem\n\n\nclass TacocabanaSpider(scrapy.Spider):\n name = \"tacocabana\"\n item_attributes = {\"brand\": \"Taco Cabana\"}\n allowed_domains = [\"api.koala.fuzzhq.com\"]\n\n def start_requests(self):\n yield scrapy.http.JsonRequest(\n \"https://api.koala.fuzzhq.com/oauth/access_token\",\n data={\n \"client_id\": \"3nA4STkGif0fZGApqxMlVewy3h8HN6Fsy7jVOACP\",\n \"client_secret\": \"8oBU5gWiNg04zYzz61hN3ETrTIzvmbGyeLCX0F1s\",\n \"grant_type\": \"ordering_app_credentials\",\n \"scope\": \"group:ordering_app\",\n },\n callback=self.fetch_locations,\n )\n\n def fetch_locations(self, response):\n self.access_token = response.json()[\"access_token\"]\n yield self.request(\n \"https://api.koala.fuzzhq.com/v1/ordering/store-locations/?include[]=operating_hours&include[]=attributes&per_page=50\"\n )\n\n def request(self, url):\n return scrapy.Request(\n url, headers={\"Authorization\": f\"Bearer {self.access_token}\"}\n )\n\n def parse(self, response):\n data = response.json()\n\n for store in data[\"data\"]:\n properties = {\n \"website\": f'https://olo.tacocabana.com/menu/{store[\"slug\"]}?showInfoModal=true',\n \"ref\": store[\"brand_id\"],\n \"lat\": store[\"latitude\"],\n \"lon\": store[\"longitude\"],\n \"addr_full\": store[\"street_address\"],\n \"city\": store[\"city\"],\n \"state\": store[\"cached_data\"][\"state\"],\n \"country\": store[\"country\"],\n \"postcode\": store[\"zip_code\"],\n \"phone\": store[\"phone_number\"],\n }\n yield GeojsonPointItem(**properties)\n\n next_url = data[\"meta\"][\"pagination\"][\"links\"][\"next\"]\n if next_url:\n yield self.request(next_url)\n", "path": "locations/spiders/tacocabana.py"}]}
| 832 | 839 |
gh_patches_debug_11219
|
rasdani/github-patches
|
git_diff
|
spyder-ide__spyder-6523
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
AttributeError in jedi_plugin.py:102
I am getting the following error with the current master:
```
Response error:Traceback (most recent call last):
File "plugin_server.py", line 95, in run
response['result'] = func(*args, **kwargs)
File "/home/jitse/misc/spyder/spyder/utils/introspection/jedi_plugin.py", line 102, in get_info
elif call_def.doc and '(' in call_def.doc.splitlines()[0]:
AttributeError: 'Definition' object has no attribute 'doc'
```
This is almost certainly due to a mistake on my part in #6497 that was not found by the automatic tests, so I will take care of it shortly.
To trigger, press `Ctrl+I` when the cursor is on `end_callbacks` on line 119 of spyder/widgets/editor.py .
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `spyder/utils/introspection/jedi_plugin.py`
Content:
```
1 # -*- coding: utf-8 -*-
2 #
3 # Copyright © Spyder Project Contributors
4 # Licensed under the terms of the MIT License
5 # (see spyder/__init__.py for details)
6
7 """
8 Jedi Introspection Plugin
9 """
10 import re
11 import os.path as osp
12 import sys
13 import time
14
15 from spyder.config.base import debug_print
16 from spyder.utils import programs
17 from spyder.utils.debug import log_last_error, log_dt
18 from spyder.utils.dochelpers import getsignaturefromtext
19 from spyder.utils.introspection.manager import (
20 DEBUG_EDITOR, LOG_FILENAME, IntrospectionPlugin)
21 from spyder.utils.introspection.utils import (default_info_response,
22 get_parent_until)
23 from spyder.utils.introspection.manager import JEDI_REQVER
24
25 try:
26 import jedi
27 except ImportError:
28 jedi = None
29
30
31 class JediPlugin(IntrospectionPlugin):
32 """
33 Jedi based introspection plugin for jedi
34
35 Experimental Editor's code completion, go-to-definition and help
36 """
37
38 # ---- IntrospectionPlugin API --------------------------------------------
39 name = 'jedi'
40
41 def load_plugin(self):
42 """Load the Jedi introspection plugin"""
43 if not programs.is_module_installed('jedi', JEDI_REQVER):
44 raise ImportError('Requires Jedi %s' % JEDI_REQVER)
45 jedi.settings.case_insensitive_completion = False
46 for lib in ['numpy', 'matplotlib']:
47 jedi.preload_module(lib)
48
49 def get_completions(self, info):
50 """Return a list of (completion, type) tuples"""
51 completions = self.get_jedi_object('completions', info)
52 if DEBUG_EDITOR:
53 log_last_error(LOG_FILENAME, str("comp: " + str(completions)[:100]))
54 if completions is not None:
55 completions = [(c.name, c.type) for c in completions]
56 else:
57 completions = []
58 debug_print(str(completions)[:100])
59 return completions
60
61 def get_info(self, info):
62 """
63 Find the calltip and docs
64
65 Returns a dict like the following:
66 {'note': 'Function of numpy.core.numeric...',
67 'argspec': "(shape, dtype=None, order='C')'
68 'docstring': 'Return an array of given...'
69 'name': 'ones',
70 'calltip': 'ones(shape, dtype=None, order='C')'}
71 """
72 call_def = self.get_jedi_object('goto_definitions', info)
73
74 for cd in call_def:
75 docstring = cd.docstring()
76 if docstring and not docstring.rstrip().endswith(')'):
77 call_def = cd
78 break
79 else:
80 try:
81 call_def = call_def[0]
82 docstring = call_def.docstring()
83 except IndexError:
84 return default_info_response()
85
86 name = call_def.name
87 if name is None:
88 return default_info_response()
89
90 if call_def.module_path:
91 mod_name = get_parent_until(call_def.module_path)
92 else:
93 mod_name = None
94
95 if not mod_name:
96 mod_name = call_def.module_name
97
98 if docstring.startswith(name + '('):
99 calltip = getsignaturefromtext(docstring, name)
100 argspec = calltip[calltip.find('('):]
101 docstring = docstring[docstring.find(')') + 3:]
102 elif call_def.doc and '(' in call_def.doc.splitlines()[0]:
103 calltip = docstring.splitlines()[0]
104 name = docstring.split('(')[0]
105 docstring = docstring[docstring.find(')') + 3:]
106 argspec = calltip[calltip.find('('):]
107 else:
108 calltip = name + '(...)'
109 argspec = '()'
110
111 if call_def.type == 'module':
112 note = 'Module %s' % mod_name
113 argspec = ''
114 calltip = name
115 elif call_def.type == 'class':
116 note = 'Class in %s module' % mod_name
117 elif docstring.startswith('%s(self' % name):
118 class_name = call_def.full_name.split('.')[-2]
119 note = 'Method of %s class in %s module' % (
120 class_name.capitalize(), mod_name)
121 else:
122 note = '%s in %s module' % (call_def.type.capitalize(),
123 mod_name)
124
125 argspec = argspec.replace(' = ', '=')
126 calltip = calltip.replace(' = ', '=')
127 debug_print(call_def.name)
128
129 doc_info = dict(name=name, argspec=argspec,
130 note=note, docstring=docstring, calltip=calltip)
131 return doc_info
132
133 def get_definition(self, info):
134 """
135 Find a definition location using Jedi
136
137 Follows gotos until a definition is found, or it reaches a builtin
138 module. Falls back on token lookup if it is in an enaml file or does
139 not find a match
140 """
141 line, filename = info['line_num'], info['filename']
142 def_info, module_path, line_nr = None, None, None
143 gotos = self.get_jedi_object('goto_assignments', info)
144
145 if gotos:
146 def_info = self.get_definition_info(gotos[0])
147 if def_info and def_info['goto_next']:
148 defns = self.get_jedi_object('goto_definitions', info)
149 new_info = None
150 if defns:
151 new_info = self.get_definition_info(defns[0])
152 if new_info and not new_info['in_builtin']:
153 def_info = new_info
154 elif not def_info:
155 return
156
157 # handle builtins -> try and find the module
158 if def_info and def_info['in_builtin']:
159 module_path, line_nr = self.find_in_builtin(def_info)
160 elif def_info:
161 module_path = def_info['module_path']
162 line_nr = def_info['line_nr']
163
164 # Handle failures to find module_path and line_nr
165 if module_path == filename and line_nr == line:
166 return
167 elif module_path is None:
168 return
169
170 return module_path, line_nr
171
172 # ---- Private API -------------------------------------------------------
173
174 def get_jedi_object(self, func_name, info, use_filename=True):
175 """Call a desired function on a Jedi Script and return the result"""
176 if not jedi:
177 return
178 if DEBUG_EDITOR:
179 t0 = time.time()
180 # override IPython qt_loaders ImportDenier behavior
181 metas = sys.meta_path
182 for meta in metas:
183 if (meta.__class__.__name__ == 'ImportDenier'
184 and hasattr(meta, 'forbid')):
185 sys.meta_path.remove(meta)
186
187 if use_filename:
188 filename = info['filename']
189 else:
190 filename = None
191
192 try:
193 script = jedi.api.Script(info['source_code'], info['line_num'],
194 info['column'], filename,
195 sys_path=info['sys_path'])
196 func = getattr(script, func_name)
197 val = func()
198 except Exception as e:
199 val = None
200 debug_print('Jedi error (%s)' % func_name)
201 debug_print(str(e))
202 if DEBUG_EDITOR:
203 log_last_error(LOG_FILENAME, str(e))
204 if DEBUG_EDITOR:
205 log_dt(LOG_FILENAME, func_name, t0)
206 if not val and filename:
207 return self.get_jedi_object(func_name, info, False)
208 else:
209 return val
210
211 @staticmethod
212 def get_definition_info(defn):
213 """Extract definition information from the Jedi definition object"""
214 try:
215 module_path = defn.module_path
216 name = defn.name
217 if hasattr(defn, 'line_nr'):
218 line_nr = defn.line_nr
219 else:
220 line_nr = defn.line
221 description = defn.description
222 in_builtin = defn.in_builtin_module()
223 except Exception as e:
224 if DEBUG_EDITOR:
225 log_last_error(LOG_FILENAME, 'Get Defintion: %s' % e)
226 return None
227 pattern = r'class\s+{0}|def\s+{0}|self.{0}\s*=|{0}\s*='.format(name)
228 if not re.match(pattern, description):
229 goto_next = True
230 else:
231 goto_next = False
232 return dict(module_path=module_path, line_nr=line_nr,
233 description=description, name=name, in_builtin=in_builtin,
234 goto_next=goto_next)
235
236 def find_in_builtin(self, info):
237 """Find a definition in a builtin file"""
238 module_path = info['module_path']
239 line_nr = info['line_nr']
240
241 if module_path is None:
242 return None, None
243
244 ext = osp.splitext(info['module_path'])[1]
245 desc = info['description']
246 name = info['name']
247 if ext in self.python_like_exts() and (
248 desc.startswith('import ') or desc.startswith('from ')):
249 path = self.python_like_mod_finder(desc,
250 osp.dirname(module_path), name)
251 if path:
252 info['module_path'] = module_path = path
253 info['line_nr'] = line_nr = 1
254
255 if ext in self.all_editable_exts():
256 pattern = r'from.*\W{0}\W?.*c?import|import.*\W{0}'
257 if not re.match(pattern.format(info['name']), desc):
258 line_nr = self.get_definition_from_file(module_path, name,
259 line_nr)
260 if not line_nr:
261 module_path = None
262
263 if not ext in self.all_editable_exts():
264 line_nr = None
265
266 return module_path, line_nr
267
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/spyder/utils/introspection/jedi_plugin.py b/spyder/utils/introspection/jedi_plugin.py
--- a/spyder/utils/introspection/jedi_plugin.py
+++ b/spyder/utils/introspection/jedi_plugin.py
@@ -99,7 +99,7 @@
calltip = getsignaturefromtext(docstring, name)
argspec = calltip[calltip.find('('):]
docstring = docstring[docstring.find(')') + 3:]
- elif call_def.doc and '(' in call_def.doc.splitlines()[0]:
+ elif docstring and '(' in docstring.splitlines()[0]:
calltip = docstring.splitlines()[0]
name = docstring.split('(')[0]
docstring = docstring[docstring.find(')') + 3:]
|
{"golden_diff": "diff --git a/spyder/utils/introspection/jedi_plugin.py b/spyder/utils/introspection/jedi_plugin.py\n--- a/spyder/utils/introspection/jedi_plugin.py\n+++ b/spyder/utils/introspection/jedi_plugin.py\n@@ -99,7 +99,7 @@\n calltip = getsignaturefromtext(docstring, name)\n argspec = calltip[calltip.find('('):]\n docstring = docstring[docstring.find(')') + 3:]\n- elif call_def.doc and '(' in call_def.doc.splitlines()[0]:\n+ elif docstring and '(' in docstring.splitlines()[0]:\n calltip = docstring.splitlines()[0]\n name = docstring.split('(')[0]\n docstring = docstring[docstring.find(')') + 3:]\n", "issue": "AttributeError in jedi_plugin.py:102\nI am getting the following error with the current master:\r\n```\r\nResponse error:Traceback (most recent call last):\r\n File \"plugin_server.py\", line 95, in run\r\n response['result'] = func(*args, **kwargs)\r\n File \"/home/jitse/misc/spyder/spyder/utils/introspection/jedi_plugin.py\", line 102, in get_info\r\n elif call_def.doc and '(' in call_def.doc.splitlines()[0]:\r\nAttributeError: 'Definition' object has no attribute 'doc'\r\n```\r\nThis is almost certainly due to a mistake on my part in #6497 that was not found by the automatic tests, so I will take care of it shortly.\r\n\r\nTo trigger, press `Ctrl+I` when the cursor is on `end_callbacks` on line 119 of spyder/widgets/editor.py .\n", "before_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Copyright \u00a9 Spyder Project Contributors\n# Licensed under the terms of the MIT License\n# (see spyder/__init__.py for details)\n\n\"\"\"\nJedi Introspection Plugin\n\"\"\"\nimport re\nimport os.path as osp\nimport sys\nimport time\n\nfrom spyder.config.base import debug_print\nfrom spyder.utils import programs\nfrom spyder.utils.debug import log_last_error, log_dt\nfrom spyder.utils.dochelpers import getsignaturefromtext\nfrom spyder.utils.introspection.manager import (\n DEBUG_EDITOR, LOG_FILENAME, IntrospectionPlugin)\nfrom spyder.utils.introspection.utils import (default_info_response,\n get_parent_until)\nfrom spyder.utils.introspection.manager import JEDI_REQVER\n\ntry:\n import jedi\nexcept ImportError:\n jedi = None\n\n\nclass JediPlugin(IntrospectionPlugin):\n \"\"\"\n Jedi based introspection plugin for jedi\n\n Experimental Editor's code completion, go-to-definition and help\n \"\"\"\n\n # ---- IntrospectionPlugin API --------------------------------------------\n name = 'jedi'\n\n def load_plugin(self):\n \"\"\"Load the Jedi introspection plugin\"\"\"\n if not programs.is_module_installed('jedi', JEDI_REQVER):\n raise ImportError('Requires Jedi %s' % JEDI_REQVER)\n jedi.settings.case_insensitive_completion = False\n for lib in ['numpy', 'matplotlib']:\n jedi.preload_module(lib)\n\n def get_completions(self, info):\n \"\"\"Return a list of (completion, type) tuples\"\"\"\n completions = self.get_jedi_object('completions', info)\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, str(\"comp: \" + str(completions)[:100]))\n if completions is not None:\n completions = [(c.name, c.type) for c in completions]\n else:\n completions = []\n debug_print(str(completions)[:100])\n return completions\n\n def get_info(self, info):\n \"\"\"\n Find the calltip and docs\n\n Returns a dict like the following:\n {'note': 'Function of numpy.core.numeric...',\n 'argspec': \"(shape, dtype=None, order='C')'\n 'docstring': 'Return an array of given...'\n 'name': 'ones',\n 'calltip': 'ones(shape, dtype=None, order='C')'}\n \"\"\"\n call_def = self.get_jedi_object('goto_definitions', info)\n\n for cd in call_def:\n docstring = cd.docstring()\n if docstring and not docstring.rstrip().endswith(')'):\n call_def = cd\n break\n else:\n try:\n call_def = call_def[0]\n docstring = call_def.docstring()\n except IndexError:\n return default_info_response()\n\n name = call_def.name\n if name is None:\n return default_info_response()\n\n if call_def.module_path:\n mod_name = get_parent_until(call_def.module_path)\n else:\n mod_name = None\n\n if not mod_name:\n mod_name = call_def.module_name\n\n if docstring.startswith(name + '('):\n calltip = getsignaturefromtext(docstring, name)\n argspec = calltip[calltip.find('('):]\n docstring = docstring[docstring.find(')') + 3:]\n elif call_def.doc and '(' in call_def.doc.splitlines()[0]:\n calltip = docstring.splitlines()[0]\n name = docstring.split('(')[0]\n docstring = docstring[docstring.find(')') + 3:]\n argspec = calltip[calltip.find('('):]\n else:\n calltip = name + '(...)'\n argspec = '()'\n\n if call_def.type == 'module':\n note = 'Module %s' % mod_name\n argspec = ''\n calltip = name\n elif call_def.type == 'class':\n note = 'Class in %s module' % mod_name\n elif docstring.startswith('%s(self' % name):\n class_name = call_def.full_name.split('.')[-2]\n note = 'Method of %s class in %s module' % (\n class_name.capitalize(), mod_name)\n else:\n note = '%s in %s module' % (call_def.type.capitalize(),\n mod_name)\n\n argspec = argspec.replace(' = ', '=')\n calltip = calltip.replace(' = ', '=')\n debug_print(call_def.name)\n\n doc_info = dict(name=name, argspec=argspec,\n note=note, docstring=docstring, calltip=calltip)\n return doc_info\n\n def get_definition(self, info):\n \"\"\"\n Find a definition location using Jedi\n\n Follows gotos until a definition is found, or it reaches a builtin\n module. Falls back on token lookup if it is in an enaml file or does\n not find a match\n \"\"\"\n line, filename = info['line_num'], info['filename']\n def_info, module_path, line_nr = None, None, None\n gotos = self.get_jedi_object('goto_assignments', info)\n\n if gotos:\n def_info = self.get_definition_info(gotos[0])\n if def_info and def_info['goto_next']:\n defns = self.get_jedi_object('goto_definitions', info)\n new_info = None\n if defns:\n new_info = self.get_definition_info(defns[0])\n if new_info and not new_info['in_builtin']:\n def_info = new_info\n elif not def_info:\n return\n\n # handle builtins -> try and find the module\n if def_info and def_info['in_builtin']:\n module_path, line_nr = self.find_in_builtin(def_info)\n elif def_info:\n module_path = def_info['module_path']\n line_nr = def_info['line_nr']\n\n # Handle failures to find module_path and line_nr\n if module_path == filename and line_nr == line:\n return\n elif module_path is None:\n return\n\n return module_path, line_nr\n\n # ---- Private API -------------------------------------------------------\n\n def get_jedi_object(self, func_name, info, use_filename=True):\n \"\"\"Call a desired function on a Jedi Script and return the result\"\"\"\n if not jedi:\n return\n if DEBUG_EDITOR:\n t0 = time.time()\n # override IPython qt_loaders ImportDenier behavior\n metas = sys.meta_path\n for meta in metas:\n if (meta.__class__.__name__ == 'ImportDenier'\n and hasattr(meta, 'forbid')):\n sys.meta_path.remove(meta)\n\n if use_filename:\n filename = info['filename']\n else:\n filename = None\n\n try:\n script = jedi.api.Script(info['source_code'], info['line_num'],\n info['column'], filename,\n sys_path=info['sys_path'])\n func = getattr(script, func_name)\n val = func()\n except Exception as e:\n val = None\n debug_print('Jedi error (%s)' % func_name)\n debug_print(str(e))\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, str(e))\n if DEBUG_EDITOR:\n log_dt(LOG_FILENAME, func_name, t0)\n if not val and filename:\n return self.get_jedi_object(func_name, info, False)\n else:\n return val\n\n @staticmethod\n def get_definition_info(defn):\n \"\"\"Extract definition information from the Jedi definition object\"\"\"\n try:\n module_path = defn.module_path\n name = defn.name\n if hasattr(defn, 'line_nr'):\n line_nr = defn.line_nr\n else:\n line_nr = defn.line\n description = defn.description\n in_builtin = defn.in_builtin_module()\n except Exception as e:\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, 'Get Defintion: %s' % e)\n return None\n pattern = r'class\\s+{0}|def\\s+{0}|self.{0}\\s*=|{0}\\s*='.format(name)\n if not re.match(pattern, description):\n goto_next = True\n else:\n goto_next = False\n return dict(module_path=module_path, line_nr=line_nr,\n description=description, name=name, in_builtin=in_builtin,\n goto_next=goto_next)\n\n def find_in_builtin(self, info):\n \"\"\"Find a definition in a builtin file\"\"\"\n module_path = info['module_path']\n line_nr = info['line_nr']\n\n if module_path is None:\n return None, None\n\n ext = osp.splitext(info['module_path'])[1]\n desc = info['description']\n name = info['name']\n if ext in self.python_like_exts() and (\n desc.startswith('import ') or desc.startswith('from ')):\n path = self.python_like_mod_finder(desc,\n osp.dirname(module_path), name)\n if path:\n info['module_path'] = module_path = path\n info['line_nr'] = line_nr = 1\n\n if ext in self.all_editable_exts():\n pattern = r'from.*\\W{0}\\W?.*c?import|import.*\\W{0}'\n if not re.match(pattern.format(info['name']), desc):\n line_nr = self.get_definition_from_file(module_path, name,\n line_nr)\n if not line_nr:\n module_path = None\n\n if not ext in self.all_editable_exts():\n line_nr = None\n\n return module_path, line_nr\n", "path": "spyder/utils/introspection/jedi_plugin.py"}], "after_files": [{"content": "# -*- coding: utf-8 -*-\n#\n# Copyright \u00a9 Spyder Project Contributors\n# Licensed under the terms of the MIT License\n# (see spyder/__init__.py for details)\n\n\"\"\"\nJedi Introspection Plugin\n\"\"\"\nimport re\nimport os.path as osp\nimport sys\nimport time\n\nfrom spyder.config.base import debug_print\nfrom spyder.utils import programs\nfrom spyder.utils.debug import log_last_error, log_dt\nfrom spyder.utils.dochelpers import getsignaturefromtext\nfrom spyder.utils.introspection.manager import (\n DEBUG_EDITOR, LOG_FILENAME, IntrospectionPlugin)\nfrom spyder.utils.introspection.utils import (default_info_response,\n get_parent_until)\nfrom spyder.utils.introspection.manager import JEDI_REQVER\n\ntry:\n import jedi\nexcept ImportError:\n jedi = None\n\n\nclass JediPlugin(IntrospectionPlugin):\n \"\"\"\n Jedi based introspection plugin for jedi\n\n Experimental Editor's code completion, go-to-definition and help\n \"\"\"\n\n # ---- IntrospectionPlugin API --------------------------------------------\n name = 'jedi'\n\n def load_plugin(self):\n \"\"\"Load the Jedi introspection plugin\"\"\"\n if not programs.is_module_installed('jedi', JEDI_REQVER):\n raise ImportError('Requires Jedi %s' % JEDI_REQVER)\n jedi.settings.case_insensitive_completion = False\n for lib in ['numpy', 'matplotlib']:\n jedi.preload_module(lib)\n\n def get_completions(self, info):\n \"\"\"Return a list of (completion, type) tuples\"\"\"\n completions = self.get_jedi_object('completions', info)\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, str(\"comp: \" + str(completions)[:100]))\n if completions is not None:\n completions = [(c.name, c.type) for c in completions]\n else:\n completions = []\n debug_print(str(completions)[:100])\n return completions\n\n def get_info(self, info):\n \"\"\"\n Find the calltip and docs\n\n Returns a dict like the following:\n {'note': 'Function of numpy.core.numeric...',\n 'argspec': \"(shape, dtype=None, order='C')'\n 'docstring': 'Return an array of given...'\n 'name': 'ones',\n 'calltip': 'ones(shape, dtype=None, order='C')'}\n \"\"\"\n call_def = self.get_jedi_object('goto_definitions', info)\n\n for cd in call_def:\n docstring = cd.docstring()\n if docstring and not docstring.rstrip().endswith(')'):\n call_def = cd\n break\n else:\n try:\n call_def = call_def[0]\n docstring = call_def.docstring()\n except IndexError:\n return default_info_response()\n\n name = call_def.name\n if name is None:\n return default_info_response()\n\n if call_def.module_path:\n mod_name = get_parent_until(call_def.module_path)\n else:\n mod_name = None\n\n if not mod_name:\n mod_name = call_def.module_name\n\n if docstring.startswith(name + '('):\n calltip = getsignaturefromtext(docstring, name)\n argspec = calltip[calltip.find('('):]\n docstring = docstring[docstring.find(')') + 3:]\n elif docstring and '(' in docstring.splitlines()[0]:\n calltip = docstring.splitlines()[0]\n name = docstring.split('(')[0]\n docstring = docstring[docstring.find(')') + 3:]\n argspec = calltip[calltip.find('('):]\n else:\n calltip = name + '(...)'\n argspec = '()'\n\n if call_def.type == 'module':\n note = 'Module %s' % mod_name\n argspec = ''\n calltip = name\n elif call_def.type == 'class':\n note = 'Class in %s module' % mod_name\n elif docstring.startswith('%s(self' % name):\n class_name = call_def.full_name.split('.')[-2]\n note = 'Method of %s class in %s module' % (\n class_name.capitalize(), mod_name)\n else:\n note = '%s in %s module' % (call_def.type.capitalize(),\n mod_name)\n\n argspec = argspec.replace(' = ', '=')\n calltip = calltip.replace(' = ', '=')\n debug_print(call_def.name)\n\n doc_info = dict(name=name, argspec=argspec,\n note=note, docstring=docstring, calltip=calltip)\n return doc_info\n\n def get_definition(self, info):\n \"\"\"\n Find a definition location using Jedi\n\n Follows gotos until a definition is found, or it reaches a builtin\n module. Falls back on token lookup if it is in an enaml file or does\n not find a match\n \"\"\"\n line, filename = info['line_num'], info['filename']\n def_info, module_path, line_nr = None, None, None\n gotos = self.get_jedi_object('goto_assignments', info)\n\n if gotos:\n def_info = self.get_definition_info(gotos[0])\n if def_info and def_info['goto_next']:\n defns = self.get_jedi_object('goto_definitions', info)\n new_info = None\n if defns:\n new_info = self.get_definition_info(defns[0])\n if new_info and not new_info['in_builtin']:\n def_info = new_info\n elif not def_info:\n return\n\n # handle builtins -> try and find the module\n if def_info and def_info['in_builtin']:\n module_path, line_nr = self.find_in_builtin(def_info)\n elif def_info:\n module_path = def_info['module_path']\n line_nr = def_info['line_nr']\n\n # Handle failures to find module_path and line_nr\n if module_path == filename and line_nr == line:\n return\n elif module_path is None:\n return\n\n return module_path, line_nr\n\n # ---- Private API -------------------------------------------------------\n\n def get_jedi_object(self, func_name, info, use_filename=True):\n \"\"\"Call a desired function on a Jedi Script and return the result\"\"\"\n if not jedi:\n return\n if DEBUG_EDITOR:\n t0 = time.time()\n # override IPython qt_loaders ImportDenier behavior\n metas = sys.meta_path\n for meta in metas:\n if (meta.__class__.__name__ == 'ImportDenier'\n and hasattr(meta, 'forbid')):\n sys.meta_path.remove(meta)\n\n if use_filename:\n filename = info['filename']\n else:\n filename = None\n\n try:\n script = jedi.api.Script(info['source_code'], info['line_num'],\n info['column'], filename,\n sys_path=info['sys_path'])\n func = getattr(script, func_name)\n val = func()\n except Exception as e:\n val = None\n debug_print('Jedi error (%s)' % func_name)\n debug_print(str(e))\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, str(e))\n if DEBUG_EDITOR:\n log_dt(LOG_FILENAME, func_name, t0)\n if not val and filename:\n return self.get_jedi_object(func_name, info, False)\n else:\n return val\n\n @staticmethod\n def get_definition_info(defn):\n \"\"\"Extract definition information from the Jedi definition object\"\"\"\n try:\n module_path = defn.module_path\n name = defn.name\n if hasattr(defn, 'line_nr'):\n line_nr = defn.line_nr\n else:\n line_nr = defn.line\n description = defn.description\n in_builtin = defn.in_builtin_module()\n except Exception as e:\n if DEBUG_EDITOR:\n log_last_error(LOG_FILENAME, 'Get Defintion: %s' % e)\n return None\n pattern = r'class\\s+{0}|def\\s+{0}|self.{0}\\s*=|{0}\\s*='.format(name)\n if not re.match(pattern, description):\n goto_next = True\n else:\n goto_next = False\n return dict(module_path=module_path, line_nr=line_nr,\n description=description, name=name, in_builtin=in_builtin,\n goto_next=goto_next)\n\n def find_in_builtin(self, info):\n \"\"\"Find a definition in a builtin file\"\"\"\n module_path = info['module_path']\n line_nr = info['line_nr']\n\n if module_path is None:\n return None, None\n\n ext = osp.splitext(info['module_path'])[1]\n desc = info['description']\n name = info['name']\n if ext in self.python_like_exts() and (\n desc.startswith('import ') or desc.startswith('from ')):\n path = self.python_like_mod_finder(desc,\n osp.dirname(module_path), name)\n if path:\n info['module_path'] = module_path = path\n info['line_nr'] = line_nr = 1\n\n if ext in self.all_editable_exts():\n pattern = r'from.*\\W{0}\\W?.*c?import|import.*\\W{0}'\n if not re.match(pattern.format(info['name']), desc):\n line_nr = self.get_definition_from_file(module_path, name,\n line_nr)\n if not line_nr:\n module_path = None\n\n if not ext in self.all_editable_exts():\n line_nr = None\n\n return module_path, line_nr\n", "path": "spyder/utils/introspection/jedi_plugin.py"}]}
| 3,240 | 182 |
gh_patches_debug_2630
|
rasdani/github-patches
|
git_diff
|
joke2k__faker-826
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
pt_BR email not returning valid email addresses
When creating a fake Factory with the pt_BR it is not returning valid email addresses.
Example:
```
melocauã@bol.com.br
joã[email protected]
laví[email protected]
vitó[email protected]
```
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `faker/providers/internet/pt_BR/__init__.py`
Content:
```
1 # coding=utf-8
2 from __future__ import unicode_literals
3 from .. import Provider as InternetProvider
4
5
6 class Provider(InternetProvider):
7 safe_email_tlds = ('com', 'net', 'br', 'br')
8 free_email_domains = (
9 'gmail.com',
10 'hotmail.com',
11 'yahoo.com.br',
12 'uol.com.br',
13 'bol.com.br',
14 'ig.com.br')
15 tlds = ('com', 'com', 'com', 'net', 'org', 'br', 'br', 'br')
16
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/faker/providers/internet/pt_BR/__init__.py b/faker/providers/internet/pt_BR/__init__.py
--- a/faker/providers/internet/pt_BR/__init__.py
+++ b/faker/providers/internet/pt_BR/__init__.py
@@ -13,3 +13,11 @@
'bol.com.br',
'ig.com.br')
tlds = ('com', 'com', 'com', 'net', 'org', 'br', 'br', 'br')
+ replacements = (
+ ('à', 'a'), ('â', 'a'), ('ã', 'a'),
+ ('ç', 'c'),
+ ('é', 'e'), ('ê', 'e'),
+ ('í', 'i'),
+ ('ô', 'o'), ('ö', 'o'), ('õ', 'o'),
+ ('ú', 'u'),
+ )
|
{"golden_diff": "diff --git a/faker/providers/internet/pt_BR/__init__.py b/faker/providers/internet/pt_BR/__init__.py\n--- a/faker/providers/internet/pt_BR/__init__.py\n+++ b/faker/providers/internet/pt_BR/__init__.py\n@@ -13,3 +13,11 @@\n 'bol.com.br',\n 'ig.com.br')\n tlds = ('com', 'com', 'com', 'net', 'org', 'br', 'br', 'br')\n+ replacements = (\n+ ('\u00e0', 'a'), ('\u00e2', 'a'), ('\u00e3', 'a'),\n+ ('\u00e7', 'c'),\n+ ('\u00e9', 'e'), ('\u00ea', 'e'),\n+ ('\u00ed', 'i'),\n+ ('\u00f4', 'o'), ('\u00f6', 'o'), ('\u00f5', 'o'),\n+ ('\u00fa', 'u'),\n+ )\n", "issue": "pt_BR email not returning valid email addresses\nWhen creating a fake Factory with the pt_BR it is not returning valid email addresses.\r\nExample:\r\n```\r\nmelocau\[email protected]\r\njo\[email protected]\r\nlav\[email protected]\r\nvit\[email protected]\r\n```\n", "before_files": [{"content": "# coding=utf-8\nfrom __future__ import unicode_literals\nfrom .. import Provider as InternetProvider\n\n\nclass Provider(InternetProvider):\n safe_email_tlds = ('com', 'net', 'br', 'br')\n free_email_domains = (\n 'gmail.com',\n 'hotmail.com',\n 'yahoo.com.br',\n 'uol.com.br',\n 'bol.com.br',\n 'ig.com.br')\n tlds = ('com', 'com', 'com', 'net', 'org', 'br', 'br', 'br')\n", "path": "faker/providers/internet/pt_BR/__init__.py"}], "after_files": [{"content": "# coding=utf-8\nfrom __future__ import unicode_literals\nfrom .. import Provider as InternetProvider\n\n\nclass Provider(InternetProvider):\n safe_email_tlds = ('com', 'net', 'br', 'br')\n free_email_domains = (\n 'gmail.com',\n 'hotmail.com',\n 'yahoo.com.br',\n 'uol.com.br',\n 'bol.com.br',\n 'ig.com.br')\n tlds = ('com', 'com', 'com', 'net', 'org', 'br', 'br', 'br')\n replacements = (\n ('\u00e0', 'a'), ('\u00e2', 'a'), ('\u00e3', 'a'),\n ('\u00e7', 'c'),\n ('\u00e9', 'e'), ('\u00ea', 'e'),\n ('\u00ed', 'i'),\n ('\u00f4', 'o'), ('\u00f6', 'o'), ('\u00f5', 'o'),\n ('\u00fa', 'u'),\n )\n", "path": "faker/providers/internet/pt_BR/__init__.py"}]}
| 477 | 199 |
gh_patches_debug_19253
|
rasdani/github-patches
|
git_diff
|
internetarchive__openlibrary-9114
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Update Github Action "Recent Comment Slack Bot" to mark issues as Needs: Response
<!-- IMPORTANT: Before posting, be sure to redact or remove sensitive data, such as passwords, secret keys, session cookies, etc. -->
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
```
def add_label_to_issues(ids):
owner = 'internetarchive'
repo = 'openlibrary'
labels_url = f"https://api.github.com/repos/{owner}/{repo}/issues/labels"
payload = {"labels": ["Needs: Response"]}
for issue_number in ids:
issue_labels_url = f"https://api.github.com/repos/{owner}/{repo}/issues/{issue_number}/labels"
response = requests.post(issue_labels_url, json=payload, headers=github_headers)
if __name__ == "__main__":
add_label_to_issues([9054, 9065])
```
to https://github.com/internetarchive/openlibrary/blob/master/scripts/gh_scripts/issue_comment_bot.py#L255
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->
### Additional context
<!-- Add any other context or screenshots about the feature request here. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `scripts/gh_scripts/issue_comment_bot.py`
Content:
```
1 #!/usr/bin/env python
2 """
3 Fetches Open Library GitHub issues that have been commented on
4 within some amount of time, in hours.
5
6 Writes links to each issue to given Slack channel.
7 """
8 import argparse
9 import errno
10 import os
11 import sys
12 import time
13
14 from datetime import datetime, timedelta
15 from typing import Any
16
17 import requests
18
19 # Maps lead label to GitHub username
20 lead_label_to_username = {
21 'Lead: @mekarpeles': 'mekarpeles',
22 'Lead: @cdrini': 'cdrini',
23 'Lead: @scottbarnes': 'scottbarnes',
24 'Lead: @seabelis': 'seabelis',
25 'Lead: @jimchamp': 'jimchamp',
26 }
27
28 # Maps GitHub username to Slack ID
29 username_to_slack_id = {
30 'mekarpeles': '<@mek>',
31 'cdrini': '<@cdrini>',
32 'scottbarnes': '<@U03MNR6T7FH>',
33 'seabelis': '<@UAHQ39ACT>',
34 'jimchamp': '<@U01ARTHG9EV>',
35 'hornc': '<@U0EUS8DV0>',
36 }
37
38 github_headers = {
39 'X-GitHub-Api-Version': '2022-11-28',
40 'Accept': 'application/vnd.github+json',
41 }
42
43
44 def fetch_issues(updated_since: str):
45 """
46 Fetches all GitHub issues that have been updated since the given date string and have at least one comment.
47
48 GitHub results are paginated. This functions appends each result to a list, and does so for all pages.
49 To keep API calls to a minimum, we request the maximum number of results per request (100 per page, as of writing).
50
51 Important: Updated issues need not have a recent comment. Update events include many other things, such as adding a
52 label to an issue, or moving an issue to a milestone. Issues returned by this function will require additional
53 processing in order to determine if they have recent comments.
54 """
55 # Make initial query for updated issues:
56 query = f'repo:internetarchive/openlibrary is:open is:issue comments:>0 updated:>{updated_since}'
57 p: dict[str, str | int] = {
58 'q': query,
59 'per_page': 100,
60 }
61 response = requests.get(
62 'https://api.github.com/search/issues', params=p, headers=github_headers
63 )
64 d = response.json()
65 results = d['items']
66
67 # Fetch additional updated issues, if any exist
68 def get_next_page(url: str):
69 """Returns list of issues and optional url for next page"""
70 resp = requests.get(url, headers=github_headers)
71 # Get issues
72 d = resp.json()
73 issues = d['items']
74 # Prepare url for next page
75 next = resp.links.get('next', {})
76 next_url = next.get('url', '')
77
78 return issues, next_url
79
80 links = response.links
81 next = links.get('next', {})
82 next_url = next.get('url', '')
83 while next_url:
84 # Make call with next link
85 issues, next_url = get_next_page(next_url)
86 results = results + issues
87
88 return results
89
90
91 def filter_issues(issues: list, since: datetime):
92 """
93 Returns list of issues that were not last responded to by staff.
94 Requires fetching the most recent comments for the given issues.
95 """
96 results = []
97
98 for i in issues:
99 # Fetch comments using URL from previous GitHub search results
100 comments_url = i.get('comments_url')
101 resp = requests.get(
102 comments_url, params={'per_page': 100}, headers=github_headers
103 )
104
105 # Ensure that we have the last page of comments
106 links = resp.links
107 last = links.get('last', {})
108 last_url = last.get('url', '')
109
110 if last_url:
111 resp = requests.get(last_url, headers=github_headers)
112
113 # Get last comment
114 comments = resp.json()
115 last_comment = comments[-1]
116
117 # Determine if last comment meets our criteria for Slack notifications
118 # First step: Ensure that the last comment was left after the given `since` datetime
119 created = datetime.fromisoformat(last_comment['created_at'])
120 # Removing timezone info to avoid TypeErrors, which occur when
121 # comparing a timezone-aware datetime with a timezone-naive datetime
122 created = created.replace(tzinfo=None)
123 if created > since:
124 # Next step: Determine if the last commenter is a staff member
125 last_commenter = last_comment['user']['login']
126 if last_commenter not in username_to_slack_id:
127 lead_label = find_lead_label(i.get('labels', []))
128 results.append(
129 {
130 'comment_url': last_comment['html_url'],
131 'commenter': last_commenter,
132 'issue_title': i['title'],
133 'lead_label': lead_label,
134 }
135 )
136
137 return results
138
139
140 def find_lead_label(labels: list[dict[str, Any]]) -> str:
141 """
142 Finds and returns the name of the first lead label found in the given list of GitHub labels.
143
144 Returns an empty string if no lead label is found
145 """
146 result = ''
147 for label in labels:
148 if label['name'].startswith('Lead:'):
149 result = label['name']
150 break
151
152 return result
153
154
155 def publish_digest(
156 issues: list[dict[str, str]],
157 slack_channel: str,
158 slack_token: str,
159 hours_passed: int,
160 ):
161 """
162 Creates a threaded Slack messaged containing a digest of recently commented GitHub issues.
163
164 Parent Slack message will say how many comments were left, and the timeframe. Each reply
165 will include a link to the comment, as well as additional information.
166 """
167 # Create the parent message
168 parent_thread_msg = (
169 f'{len(issues)} new GitHub comment(s) since {hours_passed} hour(s) ago'
170 )
171
172 response = requests.post(
173 'https://slack.com/api/chat.postMessage',
174 headers={
175 'Authorization': f"Bearer {slack_token}",
176 'Content-Type': 'application/json; charset=utf-8',
177 },
178 json={
179 'channel': slack_channel,
180 'text': parent_thread_msg,
181 },
182 )
183
184 if response.status_code != 200:
185 # XXX : Log this
186 print(f'Failed to send message to Slack. Status code: {response.status_code}')
187 # XXX : Add retry logic?
188 sys.exit(errno.ECOMM)
189
190 d = response.json()
191 # Store timestamp, which, along with the channel, uniquely identifies the parent thread
192 ts = d.get('ts')
193
194 def comment_on_thread(message: str):
195 """
196 Posts the given message as a reply to the parent message.
197 """
198 response = requests.post(
199 'https://slack.com/api/chat.postMessage',
200 headers={
201 'Authorization': f"Bearer {slack_token}",
202 'Content-Type': 'application/json; charset=utf-8',
203 },
204 json={
205 'channel': slack_channel,
206 'text': message,
207 'thread_ts': ts,
208 },
209 )
210 if response.status_code != 200:
211 # XXX : Check "ok" field for errors
212 # XXX : Log this
213 print(
214 f'Failed to POST slack message\n Status code: {response.status_code}\n Message: {message}'
215 )
216 # XXX : Retry logic?
217
218 for i in issues:
219 # Slack rate limit is roughly 1 request per second
220 time.sleep(1)
221
222 comment_url = i['comment_url']
223 issue_title = i['issue_title']
224 commenter = i['commenter']
225 message = f'<{comment_url}|Latest comment for: *{issue_title}*>\n'
226
227 username = lead_label_to_username.get(i['lead_label'], '')
228 slack_id = username_to_slack_id.get(username, '')
229 if slack_id:
230 message += f'Lead: {slack_id}\n'
231 elif i['lead_label']:
232 message += f'{i["lead_label"]}\n'
233 else:
234 message += 'Lead: N/A\n'
235
236 message += f'Commenter: *{commenter}*'
237 comment_on_thread(message)
238
239
240 def time_since(hours):
241 """Returns datetime and string representations of the current time, minus the given hour"""
242 now = datetime.now()
243 # XXX : Add a minute or two to the delta (to avoid dropping issues)?
244 since = now - timedelta(hours=hours)
245 return since, since.strftime('%Y-%m-%dT%H:%M:%S')
246
247
248 def start_job(args: argparse.Namespace):
249 """
250 Starts the new comment digest job.
251 """
252 since, date_string = time_since(args.hours)
253 issues = fetch_issues(date_string)
254
255 filtered_issues = filter_issues(issues, since)
256 publish_digest(filtered_issues, args.channel, args.slack_token, args.hours)
257 print('Digest posted to Slack.')
258
259
260 def _get_parser() -> argparse.ArgumentParser:
261 """
262 Creates and returns an ArgumentParser containing default values which were
263 read from the config file.
264 """
265 parser = argparse.ArgumentParser(description=__doc__)
266 parser.add_argument(
267 'hours',
268 help='Fetch issues that have been updated since this many hours ago',
269 type=int,
270 )
271 parser.add_argument(
272 'channel',
273 help="Issues will be published to this Slack channel",
274 type=str,
275 )
276 parser.add_argument(
277 'slack_token',
278 metavar='slack-token',
279 help='Slack auth token',
280 type=str,
281 )
282
283 return parser
284
285
286 if __name__ == '__main__':
287 # Process command-line arguments and starts the notification job
288 parser = _get_parser()
289 args = parser.parse_args()
290
291 # If found, add token to GitHub request headers:
292 github_token = os.environ.get('GITHUB_TOKEN', '')
293 if github_token:
294 github_headers['Authorization'] = f'Bearer {github_token}'
295 start_job(args)
296
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/scripts/gh_scripts/issue_comment_bot.py b/scripts/gh_scripts/issue_comment_bot.py
--- a/scripts/gh_scripts/issue_comment_bot.py
+++ b/scripts/gh_scripts/issue_comment_bot.py
@@ -245,6 +245,16 @@
return since, since.strftime('%Y-%m-%dT%H:%M:%S')
+def add_label_to_issues(issues):
+ for issue in issues:
+ issue_labels_url = f"https://api.github.com/repos/internetarchive/openlibrary/issues/{issue['number']}/labels"
+ response = requests.post(
+ issue_labels_url,
+ json={"labels": ["Needs: Response"]},
+ headers=github_headers,
+ )
+
+
def start_job(args: argparse.Namespace):
"""
Starts the new comment digest job.
@@ -253,6 +263,7 @@
issues = fetch_issues(date_string)
filtered_issues = filter_issues(issues, since)
+ add_label_to_issues(filtered_issues)
publish_digest(filtered_issues, args.channel, args.slack_token, args.hours)
print('Digest posted to Slack.')
|
{"golden_diff": "diff --git a/scripts/gh_scripts/issue_comment_bot.py b/scripts/gh_scripts/issue_comment_bot.py\n--- a/scripts/gh_scripts/issue_comment_bot.py\n+++ b/scripts/gh_scripts/issue_comment_bot.py\n@@ -245,6 +245,16 @@\n return since, since.strftime('%Y-%m-%dT%H:%M:%S')\n \n \n+def add_label_to_issues(issues):\n+ for issue in issues:\n+ issue_labels_url = f\"https://api.github.com/repos/internetarchive/openlibrary/issues/{issue['number']}/labels\"\n+ response = requests.post(\n+ issue_labels_url,\n+ json={\"labels\": [\"Needs: Response\"]},\n+ headers=github_headers,\n+ )\n+\n+\n def start_job(args: argparse.Namespace):\n \"\"\"\n Starts the new comment digest job.\n@@ -253,6 +263,7 @@\n issues = fetch_issues(date_string)\n \n filtered_issues = filter_issues(issues, since)\n+ add_label_to_issues(filtered_issues)\n publish_digest(filtered_issues, args.channel, args.slack_token, args.hours)\n print('Digest posted to Slack.')\n", "issue": "Update Github Action \"Recent Comment Slack Bot\" to mark issues as Needs: Response \n<!-- IMPORTANT: Before posting, be sure to redact or remove sensitive data, such as passwords, secret keys, session cookies, etc. -->\r\n\r\n<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->\r\n\r\n### Describe the problem that you'd like solved\r\n<!-- A clear and concise description of what you want to happen. -->\r\n\r\n```\r\ndef add_label_to_issues(ids):\r\n owner = 'internetarchive'\r\n repo = 'openlibrary'\r\n labels_url = f\"https://api.github.com/repos/{owner}/{repo}/issues/labels\"\r\n payload = {\"labels\": [\"Needs: Response\"]}\r\n for issue_number in ids:\r\n issue_labels_url = f\"https://api.github.com/repos/{owner}/{repo}/issues/{issue_number}/labels\"\r\n response = requests.post(issue_labels_url, json=payload, headers=github_headers)\r\n\r\nif __name__ == \"__main__\":\r\n add_label_to_issues([9054, 9065])\r\n```\r\n\r\nto https://github.com/internetarchive/openlibrary/blob/master/scripts/gh_scripts/issue_comment_bot.py#L255\r\n\r\n### Proposal & Constraints\r\n<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->\r\n\r\n<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->\r\n\r\n### Additional context\r\n<!-- Add any other context or screenshots about the feature request here. -->\r\n\r\n### Stakeholders\r\n<!-- @ tag stakeholders of this bug -->\r\n\r\n\r\n\n", "before_files": [{"content": "#!/usr/bin/env python\n\"\"\"\nFetches Open Library GitHub issues that have been commented on\nwithin some amount of time, in hours.\n\nWrites links to each issue to given Slack channel.\n\"\"\"\nimport argparse\nimport errno\nimport os\nimport sys\nimport time\n\nfrom datetime import datetime, timedelta\nfrom typing import Any\n\nimport requests\n\n# Maps lead label to GitHub username\nlead_label_to_username = {\n 'Lead: @mekarpeles': 'mekarpeles',\n 'Lead: @cdrini': 'cdrini',\n 'Lead: @scottbarnes': 'scottbarnes',\n 'Lead: @seabelis': 'seabelis',\n 'Lead: @jimchamp': 'jimchamp',\n}\n\n# Maps GitHub username to Slack ID\nusername_to_slack_id = {\n 'mekarpeles': '<@mek>',\n 'cdrini': '<@cdrini>',\n 'scottbarnes': '<@U03MNR6T7FH>',\n 'seabelis': '<@UAHQ39ACT>',\n 'jimchamp': '<@U01ARTHG9EV>',\n 'hornc': '<@U0EUS8DV0>',\n}\n\ngithub_headers = {\n 'X-GitHub-Api-Version': '2022-11-28',\n 'Accept': 'application/vnd.github+json',\n}\n\n\ndef fetch_issues(updated_since: str):\n \"\"\"\n Fetches all GitHub issues that have been updated since the given date string and have at least one comment.\n\n GitHub results are paginated. This functions appends each result to a list, and does so for all pages.\n To keep API calls to a minimum, we request the maximum number of results per request (100 per page, as of writing).\n\n Important: Updated issues need not have a recent comment. Update events include many other things, such as adding a\n label to an issue, or moving an issue to a milestone. Issues returned by this function will require additional\n processing in order to determine if they have recent comments.\n \"\"\"\n # Make initial query for updated issues:\n query = f'repo:internetarchive/openlibrary is:open is:issue comments:>0 updated:>{updated_since}'\n p: dict[str, str | int] = {\n 'q': query,\n 'per_page': 100,\n }\n response = requests.get(\n 'https://api.github.com/search/issues', params=p, headers=github_headers\n )\n d = response.json()\n results = d['items']\n\n # Fetch additional updated issues, if any exist\n def get_next_page(url: str):\n \"\"\"Returns list of issues and optional url for next page\"\"\"\n resp = requests.get(url, headers=github_headers)\n # Get issues\n d = resp.json()\n issues = d['items']\n # Prepare url for next page\n next = resp.links.get('next', {})\n next_url = next.get('url', '')\n\n return issues, next_url\n\n links = response.links\n next = links.get('next', {})\n next_url = next.get('url', '')\n while next_url:\n # Make call with next link\n issues, next_url = get_next_page(next_url)\n results = results + issues\n\n return results\n\n\ndef filter_issues(issues: list, since: datetime):\n \"\"\"\n Returns list of issues that were not last responded to by staff.\n Requires fetching the most recent comments for the given issues.\n \"\"\"\n results = []\n\n for i in issues:\n # Fetch comments using URL from previous GitHub search results\n comments_url = i.get('comments_url')\n resp = requests.get(\n comments_url, params={'per_page': 100}, headers=github_headers\n )\n\n # Ensure that we have the last page of comments\n links = resp.links\n last = links.get('last', {})\n last_url = last.get('url', '')\n\n if last_url:\n resp = requests.get(last_url, headers=github_headers)\n\n # Get last comment\n comments = resp.json()\n last_comment = comments[-1]\n\n # Determine if last comment meets our criteria for Slack notifications\n # First step: Ensure that the last comment was left after the given `since` datetime\n created = datetime.fromisoformat(last_comment['created_at'])\n # Removing timezone info to avoid TypeErrors, which occur when\n # comparing a timezone-aware datetime with a timezone-naive datetime\n created = created.replace(tzinfo=None)\n if created > since:\n # Next step: Determine if the last commenter is a staff member\n last_commenter = last_comment['user']['login']\n if last_commenter not in username_to_slack_id:\n lead_label = find_lead_label(i.get('labels', []))\n results.append(\n {\n 'comment_url': last_comment['html_url'],\n 'commenter': last_commenter,\n 'issue_title': i['title'],\n 'lead_label': lead_label,\n }\n )\n\n return results\n\n\ndef find_lead_label(labels: list[dict[str, Any]]) -> str:\n \"\"\"\n Finds and returns the name of the first lead label found in the given list of GitHub labels.\n\n Returns an empty string if no lead label is found\n \"\"\"\n result = ''\n for label in labels:\n if label['name'].startswith('Lead:'):\n result = label['name']\n break\n\n return result\n\n\ndef publish_digest(\n issues: list[dict[str, str]],\n slack_channel: str,\n slack_token: str,\n hours_passed: int,\n):\n \"\"\"\n Creates a threaded Slack messaged containing a digest of recently commented GitHub issues.\n\n Parent Slack message will say how many comments were left, and the timeframe. Each reply\n will include a link to the comment, as well as additional information.\n \"\"\"\n # Create the parent message\n parent_thread_msg = (\n f'{len(issues)} new GitHub comment(s) since {hours_passed} hour(s) ago'\n )\n\n response = requests.post(\n 'https://slack.com/api/chat.postMessage',\n headers={\n 'Authorization': f\"Bearer {slack_token}\",\n 'Content-Type': 'application/json; charset=utf-8',\n },\n json={\n 'channel': slack_channel,\n 'text': parent_thread_msg,\n },\n )\n\n if response.status_code != 200:\n # XXX : Log this\n print(f'Failed to send message to Slack. Status code: {response.status_code}')\n # XXX : Add retry logic?\n sys.exit(errno.ECOMM)\n\n d = response.json()\n # Store timestamp, which, along with the channel, uniquely identifies the parent thread\n ts = d.get('ts')\n\n def comment_on_thread(message: str):\n \"\"\"\n Posts the given message as a reply to the parent message.\n \"\"\"\n response = requests.post(\n 'https://slack.com/api/chat.postMessage',\n headers={\n 'Authorization': f\"Bearer {slack_token}\",\n 'Content-Type': 'application/json; charset=utf-8',\n },\n json={\n 'channel': slack_channel,\n 'text': message,\n 'thread_ts': ts,\n },\n )\n if response.status_code != 200:\n # XXX : Check \"ok\" field for errors\n # XXX : Log this\n print(\n f'Failed to POST slack message\\n Status code: {response.status_code}\\n Message: {message}'\n )\n # XXX : Retry logic?\n\n for i in issues:\n # Slack rate limit is roughly 1 request per second\n time.sleep(1)\n\n comment_url = i['comment_url']\n issue_title = i['issue_title']\n commenter = i['commenter']\n message = f'<{comment_url}|Latest comment for: *{issue_title}*>\\n'\n\n username = lead_label_to_username.get(i['lead_label'], '')\n slack_id = username_to_slack_id.get(username, '')\n if slack_id:\n message += f'Lead: {slack_id}\\n'\n elif i['lead_label']:\n message += f'{i[\"lead_label\"]}\\n'\n else:\n message += 'Lead: N/A\\n'\n\n message += f'Commenter: *{commenter}*'\n comment_on_thread(message)\n\n\ndef time_since(hours):\n \"\"\"Returns datetime and string representations of the current time, minus the given hour\"\"\"\n now = datetime.now()\n # XXX : Add a minute or two to the delta (to avoid dropping issues)?\n since = now - timedelta(hours=hours)\n return since, since.strftime('%Y-%m-%dT%H:%M:%S')\n\n\ndef start_job(args: argparse.Namespace):\n \"\"\"\n Starts the new comment digest job.\n \"\"\"\n since, date_string = time_since(args.hours)\n issues = fetch_issues(date_string)\n\n filtered_issues = filter_issues(issues, since)\n publish_digest(filtered_issues, args.channel, args.slack_token, args.hours)\n print('Digest posted to Slack.')\n\n\ndef _get_parser() -> argparse.ArgumentParser:\n \"\"\"\n Creates and returns an ArgumentParser containing default values which were\n read from the config file.\n \"\"\"\n parser = argparse.ArgumentParser(description=__doc__)\n parser.add_argument(\n 'hours',\n help='Fetch issues that have been updated since this many hours ago',\n type=int,\n )\n parser.add_argument(\n 'channel',\n help=\"Issues will be published to this Slack channel\",\n type=str,\n )\n parser.add_argument(\n 'slack_token',\n metavar='slack-token',\n help='Slack auth token',\n type=str,\n )\n\n return parser\n\n\nif __name__ == '__main__':\n # Process command-line arguments and starts the notification job\n parser = _get_parser()\n args = parser.parse_args()\n\n # If found, add token to GitHub request headers:\n github_token = os.environ.get('GITHUB_TOKEN', '')\n if github_token:\n github_headers['Authorization'] = f'Bearer {github_token}'\n start_job(args)\n", "path": "scripts/gh_scripts/issue_comment_bot.py"}], "after_files": [{"content": "#!/usr/bin/env python\n\"\"\"\nFetches Open Library GitHub issues that have been commented on\nwithin some amount of time, in hours.\n\nWrites links to each issue to given Slack channel.\n\"\"\"\nimport argparse\nimport errno\nimport os\nimport sys\nimport time\n\nfrom datetime import datetime, timedelta\nfrom typing import Any\n\nimport requests\n\n# Maps lead label to GitHub username\nlead_label_to_username = {\n 'Lead: @mekarpeles': 'mekarpeles',\n 'Lead: @cdrini': 'cdrini',\n 'Lead: @scottbarnes': 'scottbarnes',\n 'Lead: @seabelis': 'seabelis',\n 'Lead: @jimchamp': 'jimchamp',\n}\n\n# Maps GitHub username to Slack ID\nusername_to_slack_id = {\n 'mekarpeles': '<@mek>',\n 'cdrini': '<@cdrini>',\n 'scottbarnes': '<@U03MNR6T7FH>',\n 'seabelis': '<@UAHQ39ACT>',\n 'jimchamp': '<@U01ARTHG9EV>',\n 'hornc': '<@U0EUS8DV0>',\n}\n\ngithub_headers = {\n 'X-GitHub-Api-Version': '2022-11-28',\n 'Accept': 'application/vnd.github+json',\n}\n\n\ndef fetch_issues(updated_since: str):\n \"\"\"\n Fetches all GitHub issues that have been updated since the given date string and have at least one comment.\n\n GitHub results are paginated. This functions appends each result to a list, and does so for all pages.\n To keep API calls to a minimum, we request the maximum number of results per request (100 per page, as of writing).\n\n Important: Updated issues need not have a recent comment. Update events include many other things, such as adding a\n label to an issue, or moving an issue to a milestone. Issues returned by this function will require additional\n processing in order to determine if they have recent comments.\n \"\"\"\n # Make initial query for updated issues:\n query = f'repo:internetarchive/openlibrary is:open is:issue comments:>0 updated:>{updated_since}'\n p: dict[str, str | int] = {\n 'q': query,\n 'per_page': 100,\n }\n response = requests.get(\n 'https://api.github.com/search/issues', params=p, headers=github_headers\n )\n d = response.json()\n results = d['items']\n\n # Fetch additional updated issues, if any exist\n def get_next_page(url: str):\n \"\"\"Returns list of issues and optional url for next page\"\"\"\n resp = requests.get(url, headers=github_headers)\n # Get issues\n d = resp.json()\n issues = d['items']\n # Prepare url for next page\n next = resp.links.get('next', {})\n next_url = next.get('url', '')\n\n return issues, next_url\n\n links = response.links\n next = links.get('next', {})\n next_url = next.get('url', '')\n while next_url:\n # Make call with next link\n issues, next_url = get_next_page(next_url)\n results = results + issues\n\n return results\n\n\ndef filter_issues(issues: list, since: datetime):\n \"\"\"\n Returns list of issues that were not last responded to by staff.\n Requires fetching the most recent comments for the given issues.\n \"\"\"\n results = []\n\n for i in issues:\n # Fetch comments using URL from previous GitHub search results\n comments_url = i.get('comments_url')\n resp = requests.get(\n comments_url, params={'per_page': 100}, headers=github_headers\n )\n\n # Ensure that we have the last page of comments\n links = resp.links\n last = links.get('last', {})\n last_url = last.get('url', '')\n\n if last_url:\n resp = requests.get(last_url, headers=github_headers)\n\n # Get last comment\n comments = resp.json()\n last_comment = comments[-1]\n\n # Determine if last comment meets our criteria for Slack notifications\n # First step: Ensure that the last comment was left after the given `since` datetime\n created = datetime.fromisoformat(last_comment['created_at'])\n # Removing timezone info to avoid TypeErrors, which occur when\n # comparing a timezone-aware datetime with a timezone-naive datetime\n created = created.replace(tzinfo=None)\n if created > since:\n # Next step: Determine if the last commenter is a staff member\n last_commenter = last_comment['user']['login']\n if last_commenter not in username_to_slack_id:\n lead_label = find_lead_label(i.get('labels', []))\n results.append(\n {\n 'comment_url': last_comment['html_url'],\n 'commenter': last_commenter,\n 'issue_title': i['title'],\n 'lead_label': lead_label,\n }\n )\n\n return results\n\n\ndef find_lead_label(labels: list[dict[str, Any]]) -> str:\n \"\"\"\n Finds and returns the name of the first lead label found in the given list of GitHub labels.\n\n Returns an empty string if no lead label is found\n \"\"\"\n result = ''\n for label in labels:\n if label['name'].startswith('Lead:'):\n result = label['name']\n break\n\n return result\n\n\ndef publish_digest(\n issues: list[dict[str, str]],\n slack_channel: str,\n slack_token: str,\n hours_passed: int,\n):\n \"\"\"\n Creates a threaded Slack messaged containing a digest of recently commented GitHub issues.\n\n Parent Slack message will say how many comments were left, and the timeframe. Each reply\n will include a link to the comment, as well as additional information.\n \"\"\"\n # Create the parent message\n parent_thread_msg = (\n f'{len(issues)} new GitHub comment(s) since {hours_passed} hour(s) ago'\n )\n\n response = requests.post(\n 'https://slack.com/api/chat.postMessage',\n headers={\n 'Authorization': f\"Bearer {slack_token}\",\n 'Content-Type': 'application/json; charset=utf-8',\n },\n json={\n 'channel': slack_channel,\n 'text': parent_thread_msg,\n },\n )\n\n if response.status_code != 200:\n # XXX : Log this\n print(f'Failed to send message to Slack. Status code: {response.status_code}')\n # XXX : Add retry logic?\n sys.exit(errno.ECOMM)\n\n d = response.json()\n # Store timestamp, which, along with the channel, uniquely identifies the parent thread\n ts = d.get('ts')\n\n def comment_on_thread(message: str):\n \"\"\"\n Posts the given message as a reply to the parent message.\n \"\"\"\n response = requests.post(\n 'https://slack.com/api/chat.postMessage',\n headers={\n 'Authorization': f\"Bearer {slack_token}\",\n 'Content-Type': 'application/json; charset=utf-8',\n },\n json={\n 'channel': slack_channel,\n 'text': message,\n 'thread_ts': ts,\n },\n )\n if response.status_code != 200:\n # XXX : Check \"ok\" field for errors\n # XXX : Log this\n print(\n f'Failed to POST slack message\\n Status code: {response.status_code}\\n Message: {message}'\n )\n # XXX : Retry logic?\n\n for i in issues:\n # Slack rate limit is roughly 1 request per second\n time.sleep(1)\n\n comment_url = i['comment_url']\n issue_title = i['issue_title']\n commenter = i['commenter']\n message = f'<{comment_url}|Latest comment for: *{issue_title}*>\\n'\n\n username = lead_label_to_username.get(i['lead_label'], '')\n slack_id = username_to_slack_id.get(username, '')\n if slack_id:\n message += f'Lead: {slack_id}\\n'\n elif i['lead_label']:\n message += f'{i[\"lead_label\"]}\\n'\n else:\n message += 'Lead: N/A\\n'\n\n message += f'Commenter: *{commenter}*'\n comment_on_thread(message)\n\n\ndef time_since(hours):\n \"\"\"Returns datetime and string representations of the current time, minus the given hour\"\"\"\n now = datetime.now()\n # XXX : Add a minute or two to the delta (to avoid dropping issues)?\n since = now - timedelta(hours=hours)\n return since, since.strftime('%Y-%m-%dT%H:%M:%S')\n\n\ndef add_label_to_issues(issues):\n for issue in issues:\n issue_labels_url = f\"https://api.github.com/repos/internetarchive/openlibrary/issues/{issue['number']}/labels\"\n response = requests.post(\n issue_labels_url,\n json={\"labels\": [\"Needs: Response\"]},\n headers=github_headers,\n )\n\n\ndef start_job(args: argparse.Namespace):\n \"\"\"\n Starts the new comment digest job.\n \"\"\"\n since, date_string = time_since(args.hours)\n issues = fetch_issues(date_string)\n\n filtered_issues = filter_issues(issues, since)\n add_label_to_issues(filtered_issues)\n publish_digest(filtered_issues, args.channel, args.slack_token, args.hours)\n print('Digest posted to Slack.')\n\n\ndef _get_parser() -> argparse.ArgumentParser:\n \"\"\"\n Creates and returns an ArgumentParser containing default values which were\n read from the config file.\n \"\"\"\n parser = argparse.ArgumentParser(description=__doc__)\n parser.add_argument(\n 'hours',\n help='Fetch issues that have been updated since this many hours ago',\n type=int,\n )\n parser.add_argument(\n 'channel',\n help=\"Issues will be published to this Slack channel\",\n type=str,\n )\n parser.add_argument(\n 'slack_token',\n metavar='slack-token',\n help='Slack auth token',\n type=str,\n )\n\n return parser\n\n\nif __name__ == '__main__':\n # Process command-line arguments and starts the notification job\n parser = _get_parser()\n args = parser.parse_args()\n\n # If found, add token to GitHub request headers:\n github_token = os.environ.get('GITHUB_TOKEN', '')\n if github_token:\n github_headers['Authorization'] = f'Bearer {github_token}'\n start_job(args)\n", "path": "scripts/gh_scripts/issue_comment_bot.py"}]}
| 3,568 | 248 |
gh_patches_debug_14522
|
rasdani/github-patches
|
git_diff
|
cisagov__manage.get.gov-1302
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
Unable to access Manage Domain for a transition domain
### Current Behavior
Logged in as a full access user I manually created a domain that just had domain and domain info, but no application. When going to its Manage Domain page via the django admin domain object I got a 500 error and the logs showed an Exception was thrown. This was to simulate a transition domain and looks like this bug would affect analysts too.
### Expected Behavior
Expected as I am full access to not see a 500 error going to the manage domain page. Nothing should have crashed.
### Steps to Reproduce
1. Log in as full access or admin user
2. Create a domain with a domain information object
3. click the manage domain button on the Domain object
4. see a 500 error
### Environment
_No response_
### Additional Context

This line "if requested_domain.domain_application.status not in valid_domain_statuses" in the error message above is the cause of the problem. it should not be checking domain application and needs to check domain information instead.
### Issue Links
_No response_
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/registrar/views/utility/mixins.py`
Content:
```
1 """Permissions-related mixin classes."""
2
3 from django.contrib.auth.mixins import PermissionRequiredMixin
4
5 from registrar.models import (
6 DomainApplication,
7 DomainInvitation,
8 DomainInformation,
9 UserDomainRole,
10 )
11 import logging
12
13
14 logger = logging.getLogger(__name__)
15
16
17 class PermissionsLoginMixin(PermissionRequiredMixin):
18
19 """Mixin that redirects to login page if not logged in, otherwise 403."""
20
21 def handle_no_permission(self):
22 self.raise_exception = self.request.user.is_authenticated
23 return super().handle_no_permission()
24
25
26 class DomainPermission(PermissionsLoginMixin):
27
28 """Does the logged-in user have access to this domain?"""
29
30 def has_permission(self):
31 """Check if this user has access to this domain.
32
33 The user is in self.request.user and the domain needs to be looked
34 up from the domain's primary key in self.kwargs["pk"]
35 """
36
37 if not self.request.user.is_authenticated:
38 return False
39
40 if self.request.user.is_restricted():
41 return False
42
43 pk = self.kwargs["pk"]
44 # If pk is none then something went very wrong...
45 if pk is None:
46 raise ValueError("Primary key is None")
47
48 if self.can_access_other_user_domains(pk):
49 return True
50
51 # user needs to have a role on the domain
52 if not UserDomainRole.objects.filter(
53 user=self.request.user, domain__id=pk
54 ).exists():
55 return False
56
57 # if we need to check more about the nature of role, do it here.
58 return True
59
60 def can_access_other_user_domains(self, pk):
61 """Checks to see if an authorized user (staff or superuser)
62 can access a domain that they did not create or was invited to.
63 """
64
65 # Check if the user is permissioned...
66 user_is_analyst_or_superuser = self.request.user.has_perm(
67 "registrar.analyst_access_permission"
68 ) or self.request.user.has_perm("registrar.full_access_permission")
69
70 if not user_is_analyst_or_superuser:
71 return False
72
73 # Check if the user is attempting a valid edit action.
74 # In other words, if the analyst/admin did not click
75 # the 'Manage Domain' button in /admin,
76 # then they cannot access this page.
77 session = self.request.session
78 can_do_action = (
79 "analyst_action" in session
80 and "analyst_action_location" in session
81 and session["analyst_action_location"] == pk
82 )
83
84 if not can_do_action:
85 return False
86
87 # Analysts may manage domains, when they are in these statuses:
88 valid_domain_statuses = [
89 DomainApplication.APPROVED,
90 DomainApplication.IN_REVIEW,
91 DomainApplication.REJECTED,
92 DomainApplication.ACTION_NEEDED,
93 # Edge case - some domains do not have
94 # a status or DomainInformation... aka a status of 'None'.
95 # It is necessary to access those to correct errors.
96 None,
97 ]
98
99 requested_domain = None
100 if DomainInformation.objects.filter(id=pk).exists():
101 requested_domain = DomainInformation.objects.get(id=pk)
102
103 if requested_domain.domain_application.status not in valid_domain_statuses:
104 return False
105
106 # Valid session keys exist,
107 # the user is permissioned,
108 # and it is in a valid status
109 return True
110
111
112 class DomainApplicationPermission(PermissionsLoginMixin):
113
114 """Does the logged-in user have access to this domain application?"""
115
116 def has_permission(self):
117 """Check if this user has access to this domain application.
118
119 The user is in self.request.user and the domain needs to be looked
120 up from the domain's primary key in self.kwargs["pk"]
121 """
122 if not self.request.user.is_authenticated:
123 return False
124
125 # user needs to be the creator of the application
126 # this query is empty if there isn't a domain application with this
127 # id and this user as creator
128 if not DomainApplication.objects.filter(
129 creator=self.request.user, id=self.kwargs["pk"]
130 ).exists():
131 return False
132
133 return True
134
135
136 class ApplicationWizardPermission(PermissionsLoginMixin):
137
138 """Does the logged-in user have permission to start or edit an application?"""
139
140 def has_permission(self):
141 """Check if this user has permission to start or edit an application.
142
143 The user is in self.request.user
144 """
145
146 # The user has an ineligible flag
147 if self.request.user.is_restricted():
148 return False
149
150 return True
151
152
153 class DomainInvitationPermission(PermissionsLoginMixin):
154
155 """Does the logged-in user have access to this domain invitation?
156
157 A user has access to a domain invitation if they have a role on the
158 associated domain.
159 """
160
161 def has_permission(self):
162 """Check if this user has a role on the domain of this invitation."""
163 if not self.request.user.is_authenticated:
164 return False
165
166 if not DomainInvitation.objects.filter(
167 id=self.kwargs["pk"], domain__permissions__user=self.request.user
168 ).exists():
169 return False
170
171 return True
172
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/registrar/views/utility/mixins.py b/src/registrar/views/utility/mixins.py
--- a/src/registrar/views/utility/mixins.py
+++ b/src/registrar/views/utility/mixins.py
@@ -100,7 +100,15 @@
if DomainInformation.objects.filter(id=pk).exists():
requested_domain = DomainInformation.objects.get(id=pk)
- if requested_domain.domain_application.status not in valid_domain_statuses:
+ # if no domain information or application exist, the user
+ # should be able to manage the domain; however, if domain information
+ # and domain application exist, and application is not in valid status,
+ # user should not be able to manage domain
+ if (
+ requested_domain
+ and requested_domain.domain_application
+ and requested_domain.domain_application.status not in valid_domain_statuses
+ ):
return False
# Valid session keys exist,
|
{"golden_diff": "diff --git a/src/registrar/views/utility/mixins.py b/src/registrar/views/utility/mixins.py\n--- a/src/registrar/views/utility/mixins.py\n+++ b/src/registrar/views/utility/mixins.py\n@@ -100,7 +100,15 @@\n if DomainInformation.objects.filter(id=pk).exists():\n requested_domain = DomainInformation.objects.get(id=pk)\n \n- if requested_domain.domain_application.status not in valid_domain_statuses:\n+ # if no domain information or application exist, the user\n+ # should be able to manage the domain; however, if domain information\n+ # and domain application exist, and application is not in valid status,\n+ # user should not be able to manage domain\n+ if (\n+ requested_domain\n+ and requested_domain.domain_application\n+ and requested_domain.domain_application.status not in valid_domain_statuses\n+ ):\n return False\n \n # Valid session keys exist,\n", "issue": "Unable to access Manage Domain for a transition domain\n### Current Behavior\n\nLogged in as a full access user I manually created a domain that just had domain and domain info, but no application. When going to its Manage Domain page via the django admin domain object I got a 500 error and the logs showed an Exception was thrown. This was to simulate a transition domain and looks like this bug would affect analysts too. \n\n### Expected Behavior\n\nExpected as I am full access to not see a 500 error going to the manage domain page. Nothing should have crashed.\n\n### Steps to Reproduce\n\n1. Log in as full access or admin user\r\n2. Create a domain with a domain information object \r\n3. click the manage domain button on the Domain object\r\n4. see a 500 error\n\n### Environment\n\n_No response_\n\n### Additional Context\n\n\r\nThis line \"if requested_domain.domain_application.status not in valid_domain_statuses\" in the error message above is the cause of the problem. it should not be checking domain application and needs to check domain information instead. \n\n### Issue Links\n\n_No response_\n", "before_files": [{"content": "\"\"\"Permissions-related mixin classes.\"\"\"\n\nfrom django.contrib.auth.mixins import PermissionRequiredMixin\n\nfrom registrar.models import (\n DomainApplication,\n DomainInvitation,\n DomainInformation,\n UserDomainRole,\n)\nimport logging\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass PermissionsLoginMixin(PermissionRequiredMixin):\n\n \"\"\"Mixin that redirects to login page if not logged in, otherwise 403.\"\"\"\n\n def handle_no_permission(self):\n self.raise_exception = self.request.user.is_authenticated\n return super().handle_no_permission()\n\n\nclass DomainPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has access to this domain.\n\n The user is in self.request.user and the domain needs to be looked\n up from the domain's primary key in self.kwargs[\"pk\"]\n \"\"\"\n\n if not self.request.user.is_authenticated:\n return False\n\n if self.request.user.is_restricted():\n return False\n\n pk = self.kwargs[\"pk\"]\n # If pk is none then something went very wrong...\n if pk is None:\n raise ValueError(\"Primary key is None\")\n\n if self.can_access_other_user_domains(pk):\n return True\n\n # user needs to have a role on the domain\n if not UserDomainRole.objects.filter(\n user=self.request.user, domain__id=pk\n ).exists():\n return False\n\n # if we need to check more about the nature of role, do it here.\n return True\n\n def can_access_other_user_domains(self, pk):\n \"\"\"Checks to see if an authorized user (staff or superuser)\n can access a domain that they did not create or was invited to.\n \"\"\"\n\n # Check if the user is permissioned...\n user_is_analyst_or_superuser = self.request.user.has_perm(\n \"registrar.analyst_access_permission\"\n ) or self.request.user.has_perm(\"registrar.full_access_permission\")\n\n if not user_is_analyst_or_superuser:\n return False\n\n # Check if the user is attempting a valid edit action.\n # In other words, if the analyst/admin did not click\n # the 'Manage Domain' button in /admin,\n # then they cannot access this page.\n session = self.request.session\n can_do_action = (\n \"analyst_action\" in session\n and \"analyst_action_location\" in session\n and session[\"analyst_action_location\"] == pk\n )\n\n if not can_do_action:\n return False\n\n # Analysts may manage domains, when they are in these statuses:\n valid_domain_statuses = [\n DomainApplication.APPROVED,\n DomainApplication.IN_REVIEW,\n DomainApplication.REJECTED,\n DomainApplication.ACTION_NEEDED,\n # Edge case - some domains do not have\n # a status or DomainInformation... aka a status of 'None'.\n # It is necessary to access those to correct errors.\n None,\n ]\n\n requested_domain = None\n if DomainInformation.objects.filter(id=pk).exists():\n requested_domain = DomainInformation.objects.get(id=pk)\n\n if requested_domain.domain_application.status not in valid_domain_statuses:\n return False\n\n # Valid session keys exist,\n # the user is permissioned,\n # and it is in a valid status\n return True\n\n\nclass DomainApplicationPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain application?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has access to this domain application.\n\n The user is in self.request.user and the domain needs to be looked\n up from the domain's primary key in self.kwargs[\"pk\"]\n \"\"\"\n if not self.request.user.is_authenticated:\n return False\n\n # user needs to be the creator of the application\n # this query is empty if there isn't a domain application with this\n # id and this user as creator\n if not DomainApplication.objects.filter(\n creator=self.request.user, id=self.kwargs[\"pk\"]\n ).exists():\n return False\n\n return True\n\n\nclass ApplicationWizardPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have permission to start or edit an application?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has permission to start or edit an application.\n\n The user is in self.request.user\n \"\"\"\n\n # The user has an ineligible flag\n if self.request.user.is_restricted():\n return False\n\n return True\n\n\nclass DomainInvitationPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain invitation?\n\n A user has access to a domain invitation if they have a role on the\n associated domain.\n \"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has a role on the domain of this invitation.\"\"\"\n if not self.request.user.is_authenticated:\n return False\n\n if not DomainInvitation.objects.filter(\n id=self.kwargs[\"pk\"], domain__permissions__user=self.request.user\n ).exists():\n return False\n\n return True\n", "path": "src/registrar/views/utility/mixins.py"}], "after_files": [{"content": "\"\"\"Permissions-related mixin classes.\"\"\"\n\nfrom django.contrib.auth.mixins import PermissionRequiredMixin\n\nfrom registrar.models import (\n DomainApplication,\n DomainInvitation,\n DomainInformation,\n UserDomainRole,\n)\nimport logging\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass PermissionsLoginMixin(PermissionRequiredMixin):\n\n \"\"\"Mixin that redirects to login page if not logged in, otherwise 403.\"\"\"\n\n def handle_no_permission(self):\n self.raise_exception = self.request.user.is_authenticated\n return super().handle_no_permission()\n\n\nclass DomainPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has access to this domain.\n\n The user is in self.request.user and the domain needs to be looked\n up from the domain's primary key in self.kwargs[\"pk\"]\n \"\"\"\n\n if not self.request.user.is_authenticated:\n return False\n\n if self.request.user.is_restricted():\n return False\n\n pk = self.kwargs[\"pk\"]\n # If pk is none then something went very wrong...\n if pk is None:\n raise ValueError(\"Primary key is None\")\n\n if self.can_access_other_user_domains(pk):\n return True\n\n # user needs to have a role on the domain\n if not UserDomainRole.objects.filter(\n user=self.request.user, domain__id=pk\n ).exists():\n return False\n\n # if we need to check more about the nature of role, do it here.\n return True\n\n def can_access_other_user_domains(self, pk):\n \"\"\"Checks to see if an authorized user (staff or superuser)\n can access a domain that they did not create or was invited to.\n \"\"\"\n\n # Check if the user is permissioned...\n user_is_analyst_or_superuser = self.request.user.has_perm(\n \"registrar.analyst_access_permission\"\n ) or self.request.user.has_perm(\"registrar.full_access_permission\")\n\n if not user_is_analyst_or_superuser:\n return False\n\n # Check if the user is attempting a valid edit action.\n # In other words, if the analyst/admin did not click\n # the 'Manage Domain' button in /admin,\n # then they cannot access this page.\n session = self.request.session\n can_do_action = (\n \"analyst_action\" in session\n and \"analyst_action_location\" in session\n and session[\"analyst_action_location\"] == pk\n )\n\n if not can_do_action:\n return False\n\n # Analysts may manage domains, when they are in these statuses:\n valid_domain_statuses = [\n DomainApplication.APPROVED,\n DomainApplication.IN_REVIEW,\n DomainApplication.REJECTED,\n DomainApplication.ACTION_NEEDED,\n # Edge case - some domains do not have\n # a status or DomainInformation... aka a status of 'None'.\n # It is necessary to access those to correct errors.\n None,\n ]\n\n requested_domain = None\n if DomainInformation.objects.filter(id=pk).exists():\n requested_domain = DomainInformation.objects.get(id=pk)\n\n # if no domain information or application exist, the user\n # should be able to manage the domain; however, if domain information\n # and domain application exist, and application is not in valid status,\n # user should not be able to manage domain\n if (\n requested_domain\n and requested_domain.domain_application\n and requested_domain.domain_application.status not in valid_domain_statuses\n ):\n return False\n\n # Valid session keys exist,\n # the user is permissioned,\n # and it is in a valid status\n return True\n\n\nclass DomainApplicationPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain application?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has access to this domain application.\n\n The user is in self.request.user and the domain needs to be looked\n up from the domain's primary key in self.kwargs[\"pk\"]\n \"\"\"\n if not self.request.user.is_authenticated:\n return False\n\n # user needs to be the creator of the application\n # this query is empty if there isn't a domain application with this\n # id and this user as creator\n if not DomainApplication.objects.filter(\n creator=self.request.user, id=self.kwargs[\"pk\"]\n ).exists():\n return False\n\n return True\n\n\nclass ApplicationWizardPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have permission to start or edit an application?\"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has permission to start or edit an application.\n\n The user is in self.request.user\n \"\"\"\n\n # The user has an ineligible flag\n if self.request.user.is_restricted():\n return False\n\n return True\n\n\nclass DomainInvitationPermission(PermissionsLoginMixin):\n\n \"\"\"Does the logged-in user have access to this domain invitation?\n\n A user has access to a domain invitation if they have a role on the\n associated domain.\n \"\"\"\n\n def has_permission(self):\n \"\"\"Check if this user has a role on the domain of this invitation.\"\"\"\n if not self.request.user.is_authenticated:\n return False\n\n if not DomainInvitation.objects.filter(\n id=self.kwargs[\"pk\"], domain__permissions__user=self.request.user\n ).exists():\n return False\n\n return True\n", "path": "src/registrar/views/utility/mixins.py"}]}
| 2,057 | 204 |
gh_patches_debug_8784
|
rasdani/github-patches
|
git_diff
|
digitalfabrik__integreat-cms-632
|
We are currently solving the following issue within our repository. Here is the issue text:
--- BEGIN ISSUE ---
get_last_root_page template tag not working
### Describe the Bug
<!-- A clear and concise description of what the bug is. -->
The `get_last_root_page` template tag is not working:
https://github.com/Integreat/integreat-cms/blob/develop/src/cms/templatetags/page_filters.py#L10-L20
This is because in the page QuerySet aggregation, the difference() function is used:
https://github.com/Integreat/integreat-cms/blob/a285cb3c74b2a2c501147076338e2b2a70c89bd6/src/cms/models/regions/region.py#L177
After difference(), it is not possible to use filter(), see https://docs.djangoproject.com/en/2.2/ref/models/querysets/#union
This leads to the last drop-region in the page tree being incorrectly tied to the last page in the list, not necessarily the last root page.
### Steps to Reproduce
1. Go to Page tree
2. Drag & drop page to most bottom line
### Expected Behavior
<!-- A clear and concise description of what you expected to happen. -->
The page should be placed next to the last root page
### Actual Behavior
<!-- A clear and concise description of what actually happened. -->
The page is placed next to the last page in the list
### Additional Information
<!-- Add any other context (e.g. logs, screenshots, etc.) about the problem here. -->
--- END ISSUE ---
Below are some code segments, each from a relevant file. One or more of these files may contain bugs.
--- BEGIN FILES ---
Path: `src/cms/templatetags/page_filters.py`
Content:
```
1 """
2 This is a collection of tags and filters for :class:`~cms.models.pages.page.Page` objects.
3 """
4 from django import template
5
6 register = template.Library()
7
8
9 @register.simple_tag
10 def get_last_root_page(pages):
11 """
12 This tag returns the last page on the root level.
13
14 :param pages: The requested page tree
15 :type pages: ~mptt.querysets.TreeQuerySet [ ~cms.models.pages.page.Page ]
16
17 :return: The last root page of the given :class:`~django.db.models.query.QuerySet`
18 :rtype: ~cms.models.pages.page.Page
19 """
20 return pages.filter(parent=None).last()
21
```
--- END FILES ---
Please first localize the bug based on the issue statement, and then generate a patch according to the `git diff` format fenced by three backticks.
Here is an example:
```diff
diff --git a/examples/server_async.py b/examples/server_async.py
--- a/examples/server_async.py
+++ b/examples/server_async.py
@@ -313,4 +313,4 @@
if __name__ == "__main__":
- asyncio.run(run_async_server("."), debug=True)
+ asyncio.run(run_async_server(), debug=True)
diff --git a/examples/server_sync.py b/examples/server_sync.py
--- a/examples/server_sync.py
+++ b/examples/server_sync.py
@@ -313,5 +313,5 @@
if __name__ == "__main__":
- server = run_sync_server(".")
+ server = run_sync_server()
server.shutdown()
```
|
diff --git a/src/cms/templatetags/page_filters.py b/src/cms/templatetags/page_filters.py
--- a/src/cms/templatetags/page_filters.py
+++ b/src/cms/templatetags/page_filters.py
@@ -12,9 +12,9 @@
This tag returns the last page on the root level.
:param pages: The requested page tree
- :type pages: ~mptt.querysets.TreeQuerySet [ ~cms.models.pages.page.Page ]
+ :type pages: list [ ~cms.models.pages.page.Page ]
- :return: The last root page of the given :class:`~django.db.models.query.QuerySet`
+ :return: The last root page of the given page list
:rtype: ~cms.models.pages.page.Page
"""
- return pages.filter(parent=None).last()
+ return list(filter(lambda p: not p.parent, pages))[-1]
|
{"golden_diff": "diff --git a/src/cms/templatetags/page_filters.py b/src/cms/templatetags/page_filters.py\n--- a/src/cms/templatetags/page_filters.py\n+++ b/src/cms/templatetags/page_filters.py\n@@ -12,9 +12,9 @@\n This tag returns the last page on the root level.\n \n :param pages: The requested page tree\n- :type pages: ~mptt.querysets.TreeQuerySet [ ~cms.models.pages.page.Page ]\n+ :type pages: list [ ~cms.models.pages.page.Page ]\n \n- :return: The last root page of the given :class:`~django.db.models.query.QuerySet`\n+ :return: The last root page of the given page list\n :rtype: ~cms.models.pages.page.Page\n \"\"\"\n- return pages.filter(parent=None).last()\n+ return list(filter(lambda p: not p.parent, pages))[-1]\n", "issue": "get_last_root_page template tag not working\n### Describe the Bug\r\n<!-- A clear and concise description of what the bug is. -->\r\nThe `get_last_root_page` template tag is not working:\r\nhttps://github.com/Integreat/integreat-cms/blob/develop/src/cms/templatetags/page_filters.py#L10-L20\r\n\r\nThis is because in the page QuerySet aggregation, the difference() function is used:\r\nhttps://github.com/Integreat/integreat-cms/blob/a285cb3c74b2a2c501147076338e2b2a70c89bd6/src/cms/models/regions/region.py#L177\r\n\r\nAfter difference(), it is not possible to use filter(), see https://docs.djangoproject.com/en/2.2/ref/models/querysets/#union\r\n\r\nThis leads to the last drop-region in the page tree being incorrectly tied to the last page in the list, not necessarily the last root page.\r\n\r\n### Steps to Reproduce\r\n\r\n1. Go to Page tree\r\n2. Drag & drop page to most bottom line\r\n\r\n### Expected Behavior\r\n<!-- A clear and concise description of what you expected to happen. -->\r\nThe page should be placed next to the last root page\r\n\r\n### Actual Behavior\r\n<!-- A clear and concise description of what actually happened. -->\r\nThe page is placed next to the last page in the list\r\n\r\n### Additional Information\r\n<!-- Add any other context (e.g. logs, screenshots, etc.) about the problem here. -->\r\n\r\n\n", "before_files": [{"content": "\"\"\"\nThis is a collection of tags and filters for :class:`~cms.models.pages.page.Page` objects.\n\"\"\"\nfrom django import template\n\nregister = template.Library()\n\n\[email protected]_tag\ndef get_last_root_page(pages):\n \"\"\"\n This tag returns the last page on the root level.\n\n :param pages: The requested page tree\n :type pages: ~mptt.querysets.TreeQuerySet [ ~cms.models.pages.page.Page ]\n\n :return: The last root page of the given :class:`~django.db.models.query.QuerySet`\n :rtype: ~cms.models.pages.page.Page\n \"\"\"\n return pages.filter(parent=None).last()\n", "path": "src/cms/templatetags/page_filters.py"}], "after_files": [{"content": "\"\"\"\nThis is a collection of tags and filters for :class:`~cms.models.pages.page.Page` objects.\n\"\"\"\nfrom django import template\n\nregister = template.Library()\n\n\[email protected]_tag\ndef get_last_root_page(pages):\n \"\"\"\n This tag returns the last page on the root level.\n\n :param pages: The requested page tree\n :type pages: list [ ~cms.models.pages.page.Page ]\n\n :return: The last root page of the given page list\n :rtype: ~cms.models.pages.page.Page\n \"\"\"\n return list(filter(lambda p: not p.parent, pages))[-1]\n", "path": "src/cms/templatetags/page_filters.py"}]}
| 760 | 203 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.