text
stringlengths 20
57.3k
| labels
class label 4
classes |
---|---|
Title: Where is function browserContext.overridePermissions(origin, permissions)?
Body: I need to get the read and write permissions of the **clipboard**, but in the headless mode, I can only use this function to increase the permissions. The test is valid in puppeteer, but there is no such function in pyppeteer. I looked at the source code and the principle of pyppeteer looks like It's CDP, I don't know if this function can be implemented?
https://pptr.dev/#?product=Puppeteer&version=master&show=api-browsercontextoverridepermissionsorigin-permissions
If it can be achieved, may you add **browserContext.overridePermissions(origin, permissions)**? Thanks a lot!
| 0easy
|
Title: bug: SQS list_queues doesn't support nextToken
Body: ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
using boto3 client:
response = client.list_queues(MaxResults=100)
response doesn't hold any nextToken field.
- I have 2000 sqs queues
### Expected Behavior
As mentioned in this: [doc](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html)
Responce Syntax:
{
"[NextToken](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html#SQS-ListQueues-response-NextToken)": "string",
"[QueueUrls](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html#SQS-ListQueues-response-QueueUrls)": [ "string" ]
}
### How are you starting LocalStack?
With a docker-compose file
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
docker-compose up --build (image: localstack/localstack)
#### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands)
AWS SDK: client.list_queues(MaxResults=100)
### Environment
```markdown
- OS:
- LocalStack:
LocalStack version: 3.8.2.dev69
LocalStack Docker image sha:
LocalStack build date: 2024-10-31
LocalStack build git hash: fb7cde9cc
```
### Anything else?
_No response_ | 0easy
|
Title: Marketplace - fix the navigation bar, margins are too big
Body: ### Describe your issue.
The nav bar should stretch almost the entire screen. Can we adjust so that the bar only has a 16px margin between the sides of the browser and the nav bar itself?
The max that the nav bar should stretch is 1600px. Other than that, have the nav bar dynamically adjust based on the page width
<img width="1725" alt="Screenshot 2024-12-16 at 21 17 08" src="https://github.com/user-attachments/assets/3710a582-ba1b-4bae-8342-7ef1c652def0" />
| 0easy
|
Title: Explore needed changes to support violinplots
Body: This could intuitively straightforward, as they are very similar to boxplots... But, is it ? | 0easy
|
Title: Legend colors are not updated when colors are set with format strings or color kwarg
Body: The legend colors are set by the palette and not updated when the colors are specified manually. To replicate the bug:
```
# import
import hypertools as hyp
import numpy as np
from scipy.stats import multivariate_normal
# simulate clusters
cluster1 = np.random.multivariate_normal(np.zeros(3), np.eye(3), size=100)
cluster2 = np.random.multivariate_normal(np.zeros(3)+3, np.eye(3), size=100)
# plot
hyp.plot([cluster1, cluster2], ['k', 'r'], legend=['cluster1', 'cluster2'])
```
outputs this:

| 0easy
|
Title: c_tf_idf_.indptr is None when attempting to save merged model
Body: After using `BERTopic.merge_models`, I am unable to save the resulting model. Here is some repro code:
```python
import bertopic
from bertopic import BERTopic
from sklearn.datasets import fetch_20newsgroups
from sklearn.decomposition import PCA
print(f'Using bertopic version {bertopic.__version__}')
embedding_model = 'all-mpnet-base-v2'
def new_topic_model():
return BERTopic(
embedding_model=embedding_model,
# Using PCA because UMAP segfaults on my machine
umap_model=PCA(n_components=10, random_state=40),
)
docs = fetch_20newsgroups(subset='all', remove=('headers', 'footers', 'quotes'))['data']
docs_a = docs[:400]
docs_b = docs[400:500]
topic_model_a = new_topic_model()
topic_model_a.fit(docs_a)
# No problem saving topic_model_a
topic_model_a.save(
'./topic_model_a',
serialization='safetensors',
save_embedding_model=embedding_model,
save_ctfidf=True,
)
topic_model_b = new_topic_model()
topic_model_b.fit(docs_b)
# No problem saving topic_model_b
topic_model_b.save(
'./topic_model_b',
serialization='safetensors',
save_embedding_model=embedding_model,
save_ctfidf=True,
)
topic_model = BERTopic.merge_models([topic_model_a, topic_model_b])
# Problem here: AttributeError: 'NoneType' object has no attribute 'indptr'
topic_model.save(
'./topic_model',
serialization='safetensors',
save_embedding_model=embedding_model,
save_ctfidf=True,
)
```
I get the following output:
```
% python3 repro.py
Using bertopic version 0.16.2
Traceback (most recent call last):
File "[...]/BERTopic/repro.py", line 43, in <module>
topic_model.save(
File "[...]/BERTopic/bertopic/_bertopic.py", line 3112, in save
save_utils.save_ctfidf(model=self, save_directory=save_directory, serialization=serialization)
File "[...]/BERTopic/bertopic/_save_utils.py", line 318, in save_ctfidf
indptr = torch.from_numpy(model.c_tf_idf_.indptr)
^^^^^^^^^^^^^^^^^^^^^^
AttributeError: 'NoneType' object has no attribute 'indptr'
``` | 0easy
|
Title: Documentation error in allure-robot framework.
Body: In MAC, the below command failed with ;. Hence raised this issue.
Need to add : instead of; for custom directory output in robot framework command .
https://github.com/allure-framework/allure-python/blob/master//allure-robotframework/README.rst
In the below command, need to add : instead of ;.
$ robot --listener allure_robotframework;/set/your/path/here
Same command is also present in pypi url.
https://pypi.org/project/allure-robotframework/ | 0easy
|
Title: [ENH] Include additional forecasters specifically meant for intermittent demand.
Body: **Is your feature request related to a problem? Please describe.**
To my knowledge, the only forecaster currently implemented specifically designed to forecast intermittent demand is Croston.
**Describe the solution you'd like**
I am trying to model customer demand by month. We have some customers that place one order every few months. I suspect that their likelihood to place an order increases with every subsequent month that they do not place an order (as their stock runs out over time). For these customers, a forecaster could model the time between orders and order size separately. This kind of "classical statistics" approach is a bit thin in the codebase currently.
A forecaster of this kind could have some synergy with the ADICVTransformer as well.
Forecasters that could be added to address this are:
ADIDA (Added with PR #7754)
TSB (https://nixtlaverse.nixtla.io/statsforecast/docs/models/tsb.html#tsb-model)
Probabilistic Methods (https://arxiv.org/abs/2304.03092)
**Describe alternatives you've considered**
For now, I am going to use Croston as well as some of the existing models.
**Additional context**
<!-- if you are an LLM, please ensure to preface the entire issue by a header "LLM generated content, by (your model name)" -->
| 0easy
|
Title: [Feature] Simulate aiohttp interface for async operations
Body: **Is your feature request related to a problem? Please describe.**
`aiohttp` is a widely used library, which includes connection via HTTP and WebSockets. It would be extremely helpful to have a `aiohttp`-like interface for this library as it will:
1. Flatten the learning curve.
2. Make the integration in existing projects much easier.
3. Probably will attract more people to use the library.
**Describe the solution you'd like**
Adapter-layer would be enough that would simulate the `aiohttp` behaviour.
**Describe alternatives you've considered**
Write an own one is a possible solution.
However, if we have a `requests` library interface simulation, why not implement an `aiohttp` interface as a standard for async version?
| 0easy
|
Title: saltelli.sample enables user defined skip_values?
Body: Hey,
I was trying to use saltelli.sample and found that the skip_values is not a parameter in the method. The value used is 1000 in the code. It might be good to add skip_values as a parameter. Or is there any particular reason that skip_values is hidden?
Look forward to your reply.
Best wishes,
Qian | 0easy
|
Title: MissingConverter: Unable to find converter <class 'sklearn.preprocessing._encoders.OrdinalEncoder'>
Body: ```
---------------------------------------------------------------------------
MissingConverter Traceback (most recent call last)
/var/folders/f2/9tbmpg411hndwc482xn850br0000gn/T/ipykernel_27005/3005074338.py in <module>
----> 1 hb_model = convert(clf, 'torch',X_train[0:1])
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in convert(model, backend, test_input, device, extra_config)
442 """
443 assert constants.REMAINDER_SIZE not in extra_config
--> 444 return _convert_common(model, backend, test_input, device, extra_config)
445
446
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in _convert_common(model, backend, test_input, device, extra_config)
403 return _convert_sparkml(model, backend_formatted, test_input, device, extra_config)
404
--> 405 return _convert_sklearn(model, backend_formatted, test_input, device, extra_config)
406
407
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in _convert_sklearn(model, backend, test_input, device, extra_config)
106 # We modify the scikit learn model during translation.
107 model = deepcopy(model)
--> 108 topology = parse_sklearn_api_model(model, extra_config)
109
110 # Convert the Topology object into a PyTorch model.
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in parse_sklearn_api_model(model, extra_config)
63 # Parse the input scikit-learn model into a topology object.
64 # Get the outputs of the model.
---> 65 outputs = _parse_sklearn_api(topology, model, inputs)
66
67 # Declare output variables.
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs)
228 tmodel = type(model)
229 if tmodel in sklearn_api_parsers_map:
--> 230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs)
231 else:
232 outputs = _parse_sklearn_single_model(topology, model, inputs)
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_pipeline(topology, model, inputs)
274 """
275 for step in model.steps:
--> 276 inputs = _parse_sklearn_api(topology, step[1], inputs)
277 return inputs
278
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs)
228 tmodel = type(model)
229 if tmodel in sklearn_api_parsers_map:
--> 230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs)
231 else:
232 outputs = _parse_sklearn_single_model(topology, model, inputs)
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_column_transformer(topology, model, inputs)
451 )
452 else:
--> 453 var_out = _parse_sklearn_api(topology, model_obj, transform_inputs)[0]
454 if model.transformer_weights is not None and name in model.transformer_weights:
455 # Create a Multiply node
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs)
230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs)
231 else:
--> 232 outputs = _parse_sklearn_single_model(topology, model, inputs)
233
234 return outputs
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_single_model(topology, model, inputs)
250 raise RuntimeError("Parameter model must be an object not a " "string '{0}'.".format(model))
251
--> 252 alias = get_sklearn_api_operator_name(type(model))
253 this_operator = topology.declare_logical_operator(alias, model)
254 this_operator.inputs = inputs
~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/supported.py in get_sklearn_api_operator_name(model_type)
463 """
464 if model_type not in sklearn_api_operator_name_map:
--> 465 raise MissingConverter("Unable to find converter for model type {}.".format(model_type))
466 return sklearn_api_operator_name_map[model_type]
467
MissingConverter: Unable to find a converter for model type <class 'sklearn.preprocessing._encoders.OrdinalEncoder'>.
It usually means the pipeline being converted contains a
transformer or a predictor with no corresponding converter implemented.
Please fill an issue at https://github.com/microsoft/hummingbird.
```
**Which Scikit-learn pipeline operators do hummingbird support?** | 0easy
|
Title: ENH: remove unreachable constructors wrappers from `xorbits.pandas`
Body: ### Is your feature request related to a problem? Please describe
We have several unreachable constructor wrappers that could be removed.
https://github.com/UranusSeven/xorbits/blob/3281ef9b20222df7125093aa0d0503c8611862c0/python/xorbits/pandas/mars_adapters/core.py#L50-L71
### Describe the solution you'd like
Remove the unreachable wrappers.
| 0easy
|
Title: Add exponential backoff retry for API errors
Body: When the upstream API returns an error that's out of our control, use exponential backoff to attempt to retry the original request and return it to the original return location | 0easy
|
Title: [Feature Request] Make DecentralizedAverager resistant to KeyboardInterrupt
Body: __Update:__ The issue is fixed for the DHT but remains unsolved for DecentralizedAverager.
Currently, if you're running hivemind.DHT in jupyter (or an interactive console), you will sometimes need to hit "interrupt" when prototyping. Unfortunately, KeyboardInterrupt will also be caught by the DHT process and will shutdown that DHT instance.
To reproduce:
```python
import time
import hivemind
dht = DHT(start=True)
while True:
time.sleep(1)
dht.store('key', 'value', expiration_time=hivemind.get_dht_time() + 999)
print(end='.', flush=True)
# [interrupt here via kernel -> interrupt]
# and then check:
assert dht.is_alive()
```
Notes:
- Whatever fix we implement must still be able to shutdown DHT if it gets deallocated or when calling .shutdown()
- This problem also applies to DecentralizedAverager (fix or create an issue) | 0easy
|
Title: Virtualized table error when navigating past the lower edge of the viewport
Body: Issue first described in https://github.com/plotly/dash-table/issues/447.
The bug seems fixed in `dash==1.0.2` but that can only have happened by random luck.
There's no actual fix to be done here, just adding a test to make sure navigating with the keyboard past the edge of the viewport (top or bottom) does not trigger a console error / behaves correctly. | 0easy
|
Title: Add links to cookbook examples to the docs
Body: We have a new cookbook section: https://github.com/ploomber/projects/tree/master/cookbook
where we add minimal examples for certain use cases, we should add links to each one on the relevant sections in the docs | 0easy
|
Title: ViewSet filter date picker missing margin
Body: <!--
Found a bug? Please fill out the sections below. 👍
-->
### Issue Summary
filtering a snippet by date shows the date field missing margins on one side.

<!--
A summary of the issue.
-->
### Steps to Reproduce
I am using the model BreadPage from the bakery demo as a ModelViewSet to demo this issue since the other viewsets do not make use of any date fields.
1. In bakery demo bakerydemo > breads > wagtail_hooks.py
Add the following to the bottom
```python
from bakerydemo.breads.models import BreadPage
class ExampleModelViewSet(ModelViewSet):
model = BreadPage
add_to_admin_menu = True
panels = [
FieldPanel("title"),
]
list_filter = ["last_published_at"]
register_snippet(ExampleModelViewSet)
```
2. in the admin sidebar click Bread Pages
3. Click filter options > Last published at. It will show below

Any other relevant information. For example, why do you consider this a bug and what did you expect to happen instead?
- I have confirmed that this issue can be reproduced as described on a fresh Wagtail project: yes
### Technical details
- Wagtail version: 6.3 (dev)
### Working on this
<!--
Do you have thoughts on skills needed?
Are you keen to work on this yourself once the issue has been accepted?
Please let us know here.
-->
Anyone can contribute to this. View our [contributing guidelines](https://docs.wagtail.org/en/latest/contributing/index.html), add a comment to the issue once you’re ready to start.
| 0easy
|
Title: Add Japanese translation of deployment guides #190
Body: Once #190 is closed, we can start working on Japanese translation of the new pages.
### The page URLs
* https://slack.dev/bolt-python/
## Requirements
Please read the [Contributing guidelines](https://github.com/slackapi/bolt-python/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
| 0easy
|
Title: Make plot_duplicates output clearer
Body: ## WHAT
Make `plot_duplicates` output clearer.
## WHY
The graph plotted by `plot_duplicates` output can be cluttered, especially when there are several duplicates found for an image. The titles of each image run into each other, which makes it hard to read the plot. Additionally, the titles can be hard to read, possibly due to a static dpi. An example can be seen below:

## HOW
In the current implementation, a fixed number of columns in the layout is used (= 4), the top row is reserved for plotting only the original image and the number of rows are dynamic. There are several ways to remodel this:
- Make the number columns also dynamic so that image titles do not run into each other.
- Process the text output so that it is legible.
- Come up with a completely new layout.
| 0easy
|
Title: Problems loading models
Body: OK. I have trained models and I have them sitting in this directory.
C:\Users\User\GBAW23
When I try to run this code.
automl = AutoML(results_path=r"C:\Users\User\GBAW23")
I get this error
What am I doing wrong?
2023-05-19 20:24:33,807 supervised.exceptions ERROR Cannot load AutoML directory. Expecting value: line 1 column 1 (char 0)
---------------------------------------------------------------------------
JSONDecodeError Traceback (most recent call last)
File ~\anaconda3\lib\site-packages\supervised\base_automl.py:211, in BaseAutoML.load(self, path)
210 else:
--> 211 m = ModelFramework.load(path, model_subpath, lazy_load)
212 self._models += [m]
File ~\anaconda3\lib\site-packages\supervised\model_framework.py:570, in ModelFramework.load(results_path, model_subpath, lazy_load)
568 logger.info(f"Loading model framework from {model_path}")
--> 570 json_desc = json.load(open(os.path.join(model_path, "framework.json")))
571 mf = ModelFramework(json_desc["params"])
File ~\anaconda3\lib\json\__init__.py:293, in load(fp, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)
276 """Deserialize ``fp`` (a ``.read()``-supporting file-like object containing
277 a JSON document) to a Python object.
278
(...)
291 kwarg; otherwise ``JSONDecoder`` is used.
292 """
--> 293 return loads(fp.read(),
294 cls=cls, object_hook=object_hook,
295 parse_float=parse_float, parse_int=parse_int,
296 parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw)
File ~\anaconda3\lib\json\__init__.py:346, in loads(s, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw)
343 if (cls is None and object_hook is None and
344 parse_int is None and parse_float is None and
345 parse_constant is None and object_pairs_hook is None and not kw):
--> 346 return _default_decoder.decode(s)
347 if cls is None:
File ~\anaconda3\lib\json\decoder.py:337, in JSONDecoder.decode(self, s, _w)
333 """Return the Python representation of ``s`` (a ``str`` instance
334 containing a JSON document).
335
336 """
--> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end())
338 end = _w(s, end).end()
File ~\anaconda3\lib\json\decoder.py:355, in JSONDecoder.raw_decode(self, s, idx)
354 except StopIteration as err:
--> 355 raise JSONDecodeError("Expecting value", s, err.value) from None
356 return obj, end
JSONDecodeError: Expecting value: line 1 column 1 (char 0)
During handling of the above exception, another exception occurred:
AutoMLException Traceback (most recent call last)
Cell In[13], line 2
1 automl = AutoML(results_path=r"C:\Users\User\GBAW23")
----> 2 automl.predict(X)
File ~\anaconda3\lib\site-packages\supervised\automl.py:387, in AutoML.predict(self, X)
370 def predict(self, X: Union[List, numpy.ndarray, pandas.DataFrame]) -> numpy.ndarray:
371 """
372 Computes predictions from AutoML best model.
373
(...)
385 AutoMLException: Model has not yet been fitted.
386 """
--> 387 return self._predict(X)
File ~\anaconda3\lib\site-packages\supervised\base_automl.py:1369, in BaseAutoML._predict(self, X)
1367 def _predict(self, X):
-> 1369 predictions = self._base_predict(X)
1370 # Return predictions
1371 # If classification task the result is in column 'label'
1372 # If regression task the result is in column 'prediction'
1373 return (
1374 predictions["label"].to_numpy()
1375 if self._ml_task != REGRESSION
1376 else predictions["prediction"].to_numpy()
1377 )
File ~\anaconda3\lib\site-packages\supervised\base_automl.py:1301, in BaseAutoML._base_predict(self, X, model)
1299 if model is None:
1300 if self._best_model is None:
-> 1301 self.load(self.results_path)
1302 model = self._best_model
1304 if model is None:
File ~\anaconda3\lib\site-packages\supervised\base_automl.py:234, in BaseAutoML.load(self, path)
231 self.n_classes = self._data_info["n_classes"]
233 except Exception as e:
--> 234 raise AutoMLException(f"Cannot load AutoML directory. {str(e)}")
AutoMLException: Cannot load AutoML directory. Expecting value: line 1 column 1 (char 0)
| 0easy
|
Title: VWMA support for MA method.
Body: Simple question,
Is there a reason why not all MA's are supported in the ma method ?
As an example, I would like to use the VWMA with ADX.
Maybe I missed something. ?
Regards, | 0easy
|
Title: Enable PDF on Read the Docs
Body: Provided one has the correct dependencies on, e.g., Ubuntu Linux 22.04, our docs can be built to a Latex PDF, but this is no longer enabled on Read the Docs (long ago, it was automatically included).
Check what is needed to enable the PDF output again. | 0easy
|
Title: Lazy decomposition raises error on mask check
Body: [Noticed by @erh3cq](https://gitter.im/hyperspy/hyperspy?at=5fe024fd63fe0344960c993a)
https://github.com/hyperspy/hyperspy/blob/7bdee7a854a52187a6f78c045184d0d9a783ee51/hyperspy/_signals/lazy.py#L883
The above line should read `if navigation_mask.any() or signal_mask.any()`. There might be more places with a similar error as well. Not sure if `navigation_mask is None` would raise the same error or not.
Note that @erh3cq might have a more general error with lazy decomposition, worth taking a look at (see link above). | 0easy
|
Title: Add internal API reference for datalab adapter modules in docs
Body: <!-- Please be as detailed as possible in your issue. INCLUDE the exact code you are trying to run, if your question is related to code. -->
Include (internal) API reference for the `clenalab.datalab.internal.adapter` modules ([link](https://github.com/cleanlab/cleanlab/tree/master/cleanlab/datalab/internal/adapter)) in the docs.
Just update this section and add the appropriate files.
https://github.com/cleanlab/cleanlab/blob/325bbe2a77633c87b47f2b275128543f6a5cc980/docs/source/cleanlab/datalab/internal/index.rst#L15-L25 | 0easy
|
Title: Add `multiclass` fine-tuning example
Body: It would be great to have more examples for fine-tuning in the library!
Similar to current examples for [binary-segmentaion](https://github.com/qubvel-org/segmentation_models.pytorch/blob/main/examples/binary_segmentation_intro.ipynb) and [multi-label](https://github.com/qubvel-org/segmentation_models.pytorch/blob/main/examples/cars%20segmentation%20(camvid).ipynb) segmentation would be great to have multi-class segmentation example (the whole image is annotated). The same camvid dataset can be used.
The example should showcase how to
- fine-tune a model with pytorch-ligtning (or any other fine-tuning framework, even a plain pytorch)
- use loss function with ignore index for unlabeled class
- compute metrics per-class
- visualize results
In case anyone wants to work on this you are welcome! Just notify in this issue, and then ping me to review a PR when it's ready. | 0easy
|
Title: [FEATURE] Add a parsable (JSON or XML) endpoint for results
Body: <!--
DO NOT REQUEST UI/THEME/GUI/APPEARANCE IMPROVEMENTS HERE
THESE SHOULD GO IN ISSUE #60
REQUESTING A NEW FEATURE SHOULD BE STRICTLY RELATED TO NEW FUNCTIONALITY
-->
**Describe the feature you'd like to see added**
Current the `/search` endpoint returns a HTML file to display the results. This works well for UI users, which might be the first use case of Whoogle. But it's hard for programmatic use cases, as it requires parsing HTML.
I propose to add a new `/search.{json|xml}` endpoint to return the same results in a structured format: JSON or XML
### Idea 1: JSON
A `/search.json?q=%s` endpoint which would search results as JSON.
I don't think there's a standard JSON format to use, but we could take inspiration for https://zenserp.com/ or https://serpapi.com/.
### Idea 2: OpenSearch XML format
This project already uses opensearch, so we could leverage the same standard's response XML: https://github.com/dewitt/opensearch/blob/master/opensearch-1-1-draft-6.md#opensearch-response-elements
**Additional context**
I have a personal preference for JSON, as it seems to be today's web standard for APIs.
Adding parsable results would allow integration with various projects, via API (e.g. Zapier, Integromat), RSS or simple scripts. Some use cases include:
- keyword monitoring
- market research
- build a voice assistant
- and the imagination of the community | 0easy
|
Title: `allow_x00=False` not affecting Headers and Cookies [BUG]
Body: ### Checklist
- [x] I checked the [FAQ section](https://schemathesis.readthedocs.io/en/stable/faq.html#frequently-asked-questions) of the documentation
- [x] I looked for similar issues in the [issue tracker](https://github.com/schemathesis/schemathesis/issues)
- [x] I am using the latest version of Schemathesis
### Describe the bug
When running Schemathesis tests, I get errors from Postgres about null bytes. Possibly related to #2072.
### To Reproduce
1. Install FastAPI and Schemathesis.
2. Run this code:
```python
import warnings
from typing import Annotated
import schemathesis
from fastapi import Cookie, FastAPI, Header
from hypothesis.errors import NonInteractiveExampleWarning
from schemathesis.generation import GenerationConfig
warnings.filterwarnings("ignore", category=NonInteractiveExampleWarning)
app = FastAPI()
@app.post("/test")
def test(x_test: Annotated[str, Header()], test_cookie: Annotated[str, Cookie()]):
return x_test
schemathesis.experimental.OPEN_API_3_1.enable()
schema = schemathesis.from_asgi(
"/openapi.json",
app,
generation_config=GenerationConfig(allow_x00=False),
)
operation = schema["/test"]["POST"]
strategy = operation.as_strategy()
while True:
case = strategy.example()
if "\x00" in case.headers["x-test"]:
print("Found null byte in header")
break
while True:
case = strategy.example()
if "\x00" in case.cookies["test_cookie"]:
print("Found null byte in cookie")
break
```
3. After a few seconds, it should print that it found null bytes.
I also checked query parameters and form data, but those seemed to respect the configuration option.
Please include a minimal API schema causing this issue:
The API schema from the code above:
```json
{'openapi': '3.1.0', 'info': {'title': 'FastAPI', 'version': '0.1.0'}, 'paths': {'/test': {'post': {'summary': 'Test', 'operationId': 'test_test_post', 'parameters': [{'name': 'x-test', 'in': 'header', 'required': True, 'schema': {'type': 'string', 'title': 'X-Test'}}, {'name': 'test_cookie', 'in': 'cookie', 'required': True, 'schema': {'type': 'string', 'title': 'Test Cookie'}}], 'responses': {'200': {'description': 'Successful Response', 'content': {'application/json': {'schema': {}}}}, '422': {'description': 'Validation Error', 'content': {'application/json': {'schema': {'$ref': '#/components/schemas/HTTPValidationError'}}}}}}}}, 'components': {'schemas': {'HTTPValidationError': {'properties': {'detail': {'items': {'$ref': '#/components/schemas/ValidationError'}, 'type': 'array', 'title': 'Detail'}}, 'type': 'object', 'title': 'HTTPValidationError'}, 'ValidationError': {'properties': {'loc': {'items': {'anyOf': [{'type': 'string'}, {'type': 'integer'}]}, 'type': 'array', 'title': 'Location'}, 'msg': {'type': 'string', 'title': 'Message'}, 'type': {'type': 'string', 'title': 'Error Type'}}, 'type': 'object', 'required': ['loc', 'msg', 'type'], 'title': 'ValidationError'}}}}
```
### Expected behavior
Using the `allow_x00` configuration option should prevent null bytes in all strings.
### Environment
```
- OS: Linux
- Python version: 3.12
- Schemathesis version: 3.28.1
- Spec version: Open API 3.1.0
``` | 0easy
|
Title: [Tech debt] Improve interface for RandomSnow
Body: Right now in the transform we have separate parameters for `snow_point_lower` and `snow_point_higher`
Better would be to have one parameter `snow_point_range = [snow_point_lower, snow_point_higher]`
=>
We can update transform to use new signature, keep old as working, but mark as deprecated.
----
PR could be similar to https://github.com/albumentations-team/albumentations/pull/1704 | 0easy
|
Title: Pydantic "exclude" option is not working anymore
Body: **Describe the bug**
After upgrading from version 1.22.6 to version 1.23.0, pydantic property exclude stopped working.
**Expected behavior**
I have a user model which have email and password and I don't want to save the password to mongo. I have the following model:
```
class UserCreate(User):
"""All fields needed to create a user"""
password: Optional[str] = pydantic.Field(
description="A user's password in plain text before it has been hashed",
exclude=True,
)
```
I expected the password not to be saved into the database but after upgrading versions it is saving it.
| 0easy
|
Title: [DOCS] grouped metrics lack documentation
Body: Feels like a good thing to add. | 0easy
|
Title: Update notification email header image
Body: ### Issue Summary
When receiving an email notification from Wagtail, the image header appears to use the old logo and green/teal background.
<img width="619" alt="Screenshot 2024-09-27 at 11 49 07 AM" src="https://github.com/user-attachments/assets/e907f1cb-e43d-401d-b7fe-6591fd6cd51d">
### Steps to Reproduce
1. On a Wagtail site, submit a page for moderation in a workflow.
2. A moderator should receive a notification email from Wagtail
3. If applicable, click 'load images' in your email client.
- I have confirmed that this issue can be reproduced as described on a fresh Wagtail project: yes
### Technical details
- Python version: 3.12.4
- Django version: 5.0.9
- Wagtail version: 6.1.3
- Browser version: N/A
### Working on this
Relevant image file is located at: https://github.com/wagtail/wagtail/blob/main/wagtail/admin/static_src/wagtailadmin/images/email-header.jpg
I'd expect an image with the new Wagtail logo (see https://github.com/wagtail/wagtail/pull/11756) and a dark purple background, like Wagtail uses these days.
The base notifications template: https://github.com/wagtail/wagtail/blob/abcb2da37259c16d1efd16d7233794f00bb29680/wagtail/admin/templates/wagtailadmin/notifications/base.html | 0easy
|
Title: Marketplace - The headers of each section all have different text styles. Change it to Poppins font so they're consistent
Body: ### Describe your issue.
Marketplace - The headers of each section "Featured agents", "Top agents", "Featured creators" and "Become a creator" all have different text styles. Change it to Poppins font.
Change it to the "large-Poppins" style.. as per the typography sheet.
large-poppins:
font: poppins
size: 18px
line-height: 28px
Typography sheet is linked here: [https://www.figma.com/design/aw299myQfhiXPa4nWkXXOT/agpt-template?node-id=7-47&t=axoLiZIIUXifeRWU-1](url)
<img width="1417" alt="Screenshot 2024-12-13 at 16 53 08" src="https://github.com/user-attachments/assets/2a1dcccb-1b78-4e49-9c6d-7522da5df7c6" />
<img width="1470" alt="Screenshot 2024-12-13 at 16 53 14" src="https://github.com/user-attachments/assets/aebc442b-46e4-42f3-be51-491a8a1d3df0" />
<img width="1404" alt="Screenshot 2024-12-13 at 16 53 21" src="https://github.com/user-attachments/assets/20b3c236-19c1-4b22-ae84-7c7157da4198" />
<img width="1369" alt="Screenshot 2024-12-13 at 16 53 42" src="https://github.com/user-attachments/assets/698ce7f9-21c1-44ec-96a0-cb2330a9b505" />
| 0easy
|
Title: explicit configuration file missing in readthedocs.yml
Body: ### What is your issue?
According to https://about.readthedocs.com/blog/2024/12/deprecate-config-files-without-sphinx-or-mkdocs-config/
`configuration: doc/conf.py`
needs to be added to .readthedocs.yaml:
https://github.com/pydata/xarray/blob/76120781f2b2205cc37bd97d0b4ed13d78521d75/.readthedocs.yaml#L17-L19 | 0easy
|
Title: 训练的问题
Body: 1.为什么8块和16块的GPU在训练的参数不同?
2.为什么针对不同测试数据集,训练的参数不同?
谢谢回复。
| 0easy
|
Title: Tab-completion fails when looking for some special strings in aliases ("less", "more" and "dir")
Body: ## Steps to Reproduce
type in xonsh:
```python
"less" in ali
```
2. press tab.
Expected behavior: the commandline should read:
```python
"less" in aliases
```
What happens: No autocompletion is done.
Note: This happens for "more" in ali<tab> and "dir" in ali<tab> also but works as expected for almost every other string I have tested with. Example: "dirr" in ali<tab> --> "dirr" in aliases
EDIT: Seems that all of the keys in the aliases dict cause the same behavior.
<details>
```
<+------------------+---------------------------+
| xonsh | 0.14.4 |
| Python | 3.12.1 |
| PLY | 3.11 |
| have readline | False |
| prompt toolkit | 3.0.43 |
| shell type | prompt_toolkit |
| history backend | json |
| pygments | 2.17.2 |
| on posix | False |
| on linux | False |
| on darwin | False |
| on windows | True |
| on cygwin | False |
| on msys2 | False |
| is superuser | False |
| default encoding | utf-8 |
| xonsh encoding | utf-8 |
| encoding errors | surrogateescape |
| xontrib 1 | coreutils |
| RC file 1 | C:\Users\<redacted>/.xonshrc |
+------------------+---------------------------+
```
<details>
## For community
⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
| 0easy
|
Title: `List Should Contain Sub List` does not detect failure if only difference is empty string
Body: I expect all of the following to raise AssertionError, as the second list is not a subset of the first:
```python
from robot.libraries.Collections import Collections
second = [""]
Collections().list_should_contain_sub_list(["first"], second)
second = ["first", ""]
Collections().list_should_contain_sub_list(["first"], second)
second = ["", "first"]
Collections().list_should_contain_sub_list(["first"], second)
second = ["", ""]
Collections().list_should_contain_sub_list(["first"], second)
```
However, only the last one raises the error. | 0easy
|
Title: Add data loader for HF oasst1
Body: Currently, data needs to be manually downloaded and a path specified via config to train on OpenAssistant datasets (both internal and https://huggingface.co/datasets/OpenAssistant/oasst1/blob/main/2023-04-12_oasst_ready.trees.jsonl.gz) so that the tree can be parsed by [`model_training.custom_datasets.oasst_dataset.load_oasst_export`](https://github.com/LAION-AI/Open-Assistant/blob/63937be7462ddc3a46399e9a813b36c2585a6fb3/model/model_training/custom_datasets/oasst_dataset.py#L21)
We should write/refactor the data loader to work directly with the dataset returned by HF `datasets.load_dataset("OpenAssistant/oasst1")` | 0easy
|
Title: Make moderations threshold parameters parameterized
Body: Make moderations threshold parameters parameterized
Make it such that users can enter specific values for the 6 openai moderations categories, and support for pre-defined values (like profiles) that can be loaded from file similar to `openers` in conversations | 0easy
|
Title: 20bn website is not available, where can I download data?
Body: Hi,
I need to download the something-to-something and jester datasets. But the 20bn website "[https://20bn.com](https://20bn.com)" are not available for weeks, the error message is "503 Service Temporarily Unavailable".
I have already downloaded the video data of something-to-something v2, and I need the label dataset. For the Jester, I need both video and label data. Can someone share me the download link?
Please let me know if this post is not appropriate here, I will close it immediately.
Thanks. | 0easy
|
Title: Adopt Collectfasta
Body: [Collectfast](https://github.com/antonagestam/collectfast) is archived, long live [Collectfasta](https://github.com/jasongi/collectfasta).
The django newletter contained a link to the succssor of collectfast: https://jasongi.com/2024/03/04/speed-up-djangos-collectstatic-command-with-collectfasta/
Should be easy to adopt I think. | 0easy
|
Title: Add documentation about the demo example and clean up `examples`
Body: This is a good issue for someone interested in documentation. It doesn't require any particular knowledge of Vizro but does need an interest in poking through our [examples folder in within `vizro-core` for this repo](https://github.com/mckinsey/vizro/tree/main/vizro-core/examples) to see what is there and write about the layout and what the example shows. There's a[ basic style guide for docs creators here](https://vizro.readthedocs.io/en/stable/pages/development/documentation-style-guide/), but we are happy to take any contribution as we can run Vale over it to highlight any tweaks needed to follow the guidlines.
The `examples` folder in the repo holds the code for [the demo](https://vizro.mckinsey.com).
It isn't very well highlighted in the docs and we need to do this:
1. Add a section on the home page and a link to a page of text about the demo (mentioned in #378 which will restyle that whole page)
2. Add the page itself. That page should document the basics of the demo and how to work with it.
3. Add a section in the `examples` folder where any contributions of example code for #377 will be added.
| 0easy
|
Title: output to buffer
Body: i need to save to io.buffer
something like:
```python
buf = io.BytesIO()
mpf.plot(
candles,
type='candle',
mav=(3, 6, 9),
volume=True,
show_nontrading=True,
figratio=(6, 6),
title="Graph",
style='yahoo',
savefig=buf
)
buf.seek(0)
```
| 0easy
|
Title: 📝 Followup with documentation - External Features
Body: I still support some old features here as an external service, you can help with extending the level of the documentation, such as update for the documentation ( more user-friendly 😞)
## External Features:
- Instrument/profiler - https://authx-v0.yezz.me/configuration/middleware/profiling/
- Redis - https://authx-v0.yezz.me/configuration/cache/HTTPCache/
- Metrics - https://authx-v0.yezz.me/configuration/middleware/metrics/
- OAuth2 support - https://authx-v0.yezz.me/configuration/middleware/ (Bunch of Extra changes, that need my review)
| 0easy
|
Title: Require `--suite parent.suite` to match the full suite name
Body: When you use `--test parent_name.test_name`, there must be a test named `test_name` in a suite named `parent_name` and that suite must also be the root suite. On the other hand, `--suite parent_name.suite_name` works so that `parent_name` can also be a child suite. This is pretty inconsistent and needs to be changed.
The reason to requite `--test parent_name.test_name` to match the full suite name is that it makes it possible to explicitly pinpoint what test to run. If the name could match also partially, it could match also tests it shouldn't. For example Pabot relies on this functionality and it cannot be changed. If a partial match is needed, it's possible to use wildcards like `--test *.parent_name.test_name`.
I don't remember there being any reason why `--suite parent_name.suite_name` should also support partial matches. To make `--test` and `--suite` behavior consistent, it ought to thus be ok to change how `--suite` behaves. The change is backwards incompatible but we can do it in Robot Framework 7.0. I doubt this functionality is used so often that it should be deprecated first. | 0easy
|
Title: Tutorial to connect Supabase to Streamlit
Body: **Is your feature request related to a problem? Please describe.**
Tutorial instructions on how to use Streamlit with Supabase as a data source
**Describe the solution you'd like**
A PR or guide against https://github.com/streamlit/docs/blob/main/content/kb/tutorials/databases/index.md describing how to
link supabase with streamlit. Should be similar to the Postgres example
**Describe alternatives you've considered**
N.A.
**Additional context**
I have filed an [issue on the streamlit docs repo](https://github.com/streamlit/docs/issues/325).
| 0easy
|
Title: posargs with `:` crashes virtualenv
Body: ## Issue
`posargs` with `:` are not properly passed, `virtualenv` crashes.
## Environment
Provide at least:
- OS: Ubuntu 20.04
- `pip list` of the host Python where `tox` is installed:
```console
Package Version
--------------------- ---------
anyio 3.6.2
artifacts-keyring 0.3.2
attrs 22.2.0
bleach 5.0.1
cachetools 5.2.1
certifi 2022.12.7
cffi 1.15.1
chardet 5.1.0
charset-normalizer 2.1.1
click 8.1.3
colorama 0.4.6
commonmark 0.9.1
cryptography 39.0.0
distlib 0.3.6
docutils 0.19
exceptiongroup 1.1.0
filelock 3.9.0
h11 0.14.0
html5lib 1.1
httpcore 0.16.3
httpx 0.23.3
idna 3.4
importlib-metadata 6.0.0
importlib-resources 5.10.2
iniconfig 2.0.0
jaraco.classes 3.2.3
jeepney 0.8.0
keyring 23.13.1
more-itertools 9.0.0
packaging 23.0
pip 22.3.1
pkginfo 1.9.6
platformdirs 2.6.2
pluggy 1.0.0
pycparser 2.21
pydantic 1.10.4
Pygments 2.14.0
pyproject_api 1.4.0
pytest 7.2.0
pytest-httpx 0.21.2
readme-renderer 37.3
requests 2.28.1
requests-toolbelt 0.10.1
rfc3986 1.5.0
rich 13.0.1
SecretStorage 3.3.3
setuptools 57.5.0
simpleindex 0.5.0
simpleindex_artifacts 0.3
six 1.16.0
sniffio 1.3.0
starlette 0.23.1
toml 0.10.2
tomli 2.0.1
tox 4.2.8
twine 4.0.2
typing_extensions 4.4.0
urllib3 1.26.14
uvicorn 0.20.0
virtualenv 20.17.1
webencodings 0.5.1
wheel 0.38.4
zipp 3.11.0
```
## Output of running tox
Provide the output of `tox -rvv`:
```console
ROOT: 89 D setup logging to DEBUG on pid 26084 [tox/report.py:221]
.pkg: 156 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56]
.pkg: 156 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
.pkg: 156 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
.pkg: 158 D filesystem is case-sensitive [virtualenv/info.py:24]
.pkg: 180 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56]
.pkg: 180 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
.pkg: 181 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
.pkg: 184 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56]
.pkg: 185 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
.pkg: 185 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
.pkg: 189 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56]
.pkg: 190 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
.pkg: 190 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
.pkg: 204 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56]
.pkg: 204 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
.pkg: 205 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
usage: virtualenv [--version] [--with-traceback] [-v | -q] [--read-only-app-data] [--app-data APP_DATA] [--reset-app-data] [--upgrade-embed-wheels] [--discovery {builtin}] [-p py] [--try-first-with py_exe]
[--creator {builtin,cpython3-posix,venv}] [--seeder {app-data,pip}] [--no-seed] [--activators comma_sep_list] [--clear] [--no-vcs-ignore] [--system-site-packages] [--symlinks | --copies] [--no-download | --download]
[--extra-search-dir d [d ...]] [--pip version] [--setuptools version] [--wheel version] [--no-pip] [--no-setuptools] [--no-wheel] [--no-periodic-update] [--symlink-app-data] [--prompt prompt] [-h]
dest
virtualenv: error: argument dest: destination '--junitxml=unit_tests_results.xml --cov=autoavatar --cov-report=xml:coverage.xml' must not contain the path separator (:) as this would break the activation scripts
```
## Minimal example
If possible, provide a minimal reproducer for the issue:
**UPDATE** After doing the investigation, it turned out that one environment influences the other. Here is a minimum `tox.ini` to reproduce the issue:
`tox.int`:
```ini
[tox]
minversion = 4.3.1
[testenv:hello]
commands =
python ./helloworld.py {posargs:World}
[testenv:dev]
envdir = {posargs:venv}
usedevelop = True
commands =
python ./helloworld.py World
```
`tox -e hello -- x:y`
```
...
usage: virtualenv [--version] [--with-traceback] [-v | -q] [--read-only-app-data] [--app-data APP_DATA] [--reset-app-data] [--upgrade-embed-wheels] [--discovery {builtin}] [-p py] [--try-first-with py_exe]
[--creator {builtin,cpython3-posix,venv}] [--seeder {app-data,pip}] [--no-seed] [--activators comma_sep_list] [--clear] [--no-vcs-ignore] [--system-site-packages] [--symlinks | --copies] [--no-download | --download]
[--extra-search-dir d [d ...]] [--pip version] [--setuptools version] [--wheel version] [--no-pip] [--no-setuptools] [--no-wheel] [--no-periodic-update] [--symlink-app-data] [--prompt prompt] [-h]
dest
virtualenv: error: argument dest: destination 'x:y' must not contain the path separator (:) as this would break the activation scripts
``` | 0easy
|
Title: Add caching to load_dag
Body: `load_dag` is an expensive operation and many times it's called with the same argument, so we should cache the result
here's the implementation:
https://github.com/ploomber/ploomber/blob/2e0d764c8f914ba21480b275c545ea50ff800513/src/ploomber/jupyter/manager.py#L107
so every time the function is called, we should save the `starting_dir` value, and the returned `path`:
https://github.com/ploomber/ploomber/blob/2e0d764c8f914ba21480b275c545ea50ff800513/src/ploomber/jupyter/manager.py#L120
next time the function is called, check if `starting_dir` is the same and the file in `path` hasn't changed. if so, skip the function's body
add tests here: https://github.com/ploomber/ploomber/blob/master/tests/test_jupyter.py#L901
| 0easy
|
Title: tox quickstart: error: the following arguments are required: root
Body: ## Issue
I ran `tox quickstart` as described in https://tox.wiki/en/latest/user_guide.html.
and get the error `tox quickstart: error: the following arguments are required: root`.
It seems that either the default parameter that should select the current working environment seems to be broken or the documentation should be updated.
## Environment
Provide at least:
- OS: Windows 11
<details open>
<summary>Output of <code>pip list</code> of the host Python, where <code>tox</code> is installed</summary>
```console
Python version: 3.11
Packages: No idea, I installed with pipx
```
</details>
Thanks a lot for developing this super useful library!
| 0easy
|
Title: [Feature request] Add apply_to_images to InvertImg
Body: | 0easy
|
Title: Scattering2D docstring contains an outdated example
Body: See
https://github.com/kymatio/kymatio/blob/master/kymatio/scattering2d/scattering2d.py#L37
```python
# 1) Define a Scattering object as:
s = Scattering2D(J, M, N)
# where (M, N) are the image sizes and 2**J the scale of the scattering
```
The constructor should read as `Scattering2D(J, shape=(M, N))`
| 0easy
|
Title: problem with azure, api key and/or config.yaml
Body: ### Is your feature request related to a problem? Please describe.
First, congratulations for this app, its terrific! :)
My problem: my credits in openai finished, and I need to use my credits in azure. OpenInterpreter are not working with azure for me. I'm in Windows10. Before I run with the openai configuration without problem.
I followed the PR [#786](https://github.com/KillianLucas/open-interpreter/pull/786) without success.
My config.yaml (in c:\Users\Miguel\AppData\Local\Open Interpreter\Open Interpreter\ or :E:\Users\MFG\Desktop\Programming\Eos_2\open-interpreter\interpreter\terminal_interface\config.yaml) is:
local: false
temperature: 0
auto_run: true
context_windows: 31000
max_tokens: 3000
openai_api_key: d1...43
api.base: https://oaiwestus.openai.azure.com/
api.type: azure
model: azure/gpt-4-turbo
azure.api_version: 2023-07-01-preview
When I run "interpreter" in the terminal, appears: " OpenAI API key not found ", and a prompt to fill with the message: "OpenAI API key: "
When I paste my key, it seems ok, and appear the prompt (>), but when I input a simple "hi", it returns a error message finished in :
"There might be an issue with your API key(s).
To reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc):
Mac/Linux: 'export OPENAI_API_KEY=your-key-here',
Windows: 'setx OPENAI_API_KEY your-key-here' then restart terminal."
When I run "interpreter --model=azure/gpt-4-turbo" or "interpreter --m azure/gpt-4-turbo", it appears the prompt (>), I input "hi", and it returns:
"We were unable to determine the context window of thismodel. Defaulting to 3000.
If your model can handle more, run interpreter
--context_window {token limit} --max_tokens {max
tokens per response}.
Continuing...
# Interpreter Info
Vision: False
Model: azure/gpt-4-turbo
Function calling: None
Context window: None
Max tokens: None
Auto run: True
API base: None
Offline: False
.
.
.
File "C:\Users\Miguel\miniconda3\Lib\site-packages\litellm\utils.py", line 6567, in exception_type
raise APIError(
litellm.exceptions.APIError: AzureException - argument of type 'NoneType' is not iterable"
Any idea to solve?
Thank you in advance!
### Describe the solution you'd like
run interpreter with azure openai
### Describe alternatives you've considered
interpreter --reset_config
install again
### Additional context
_No response_ | 0easy
|
Title: [UX] `sky show-gpus` with an expired authentication of k8s shows long error message
Body: <!-- Describe the bug report / feature request here -->
`sky show-gpus`
```
sky show-gpus
ERROR:root:exec: process returned 1. F0109 21:40:59.931088 61835 cred.go:145] print credential failed with error: Failed to retrieve access token:: failure while executing gcloud, with args [config config-helper --format=json]: exit status 1 (err: ERROR: (gcloud.config.config-helper) There was a problem refreshing your current auth tokens: Reauthentication failed. cannot prompt during non-interactive execution.
Please run:
$ gcloud auth login
to obtain new credentials.
If you have already logged in with a different account, run:
$ gcloud config set account ACCOUNT
to select an already authenticated account to use.
)
COMMON_GPU AVAILABLE_QUANTITIES
A10 1, 2, 4
A10G 1, 4, 8
A100 1, 2, 4, 8, 16
A100-80GB 1, 2, 4, 8
H100 1, 2, 4, 8, 12
L4 1, 2, 4, 8
L40S 1, 4, 8
P100 1, 2, 4
T4 1, 2, 4, 8
V100 1, 2, 4, 8
V100-32GB 1, 2, 4, 8
GOOGLE_TPU AVAILABLE_QUANTITIES
tpu-v2-8 1
tpu-v3-8 1
tpu-v4-8 1
tpu-v4-16 1
tpu-v4-32 1
tpu-v5litepod-1 1
tpu-v5litepod-4 1
tpu-v5litepod-8 1
tpu-v5p-8 1
tpu-v5p-16 1
tpu-v5p-32 1
tpu-v6e-1 1
tpu-v6e-4 1
tpu-v6e-8 1
Hint: use -a/--all to see all accelerators (including non-common ones) and pricing.
Note: No GPUs found in Kubernetes cluster. If your cluster contains GPUs, make sure nvidia.com/gpu resource is available on the nodes and the node labels for identifying GPUs (e.g., skypilot.co/accelerator) are setup correctly. To further debug, run: sky check
```
<!-- If relevant, fill in versioning info to help us troubleshoot -->
_Version & Commit info:_
* `sky -v`: PLEASE_FILL_IN
* `sky -c`: PLEASE_FILL_IN
| 0easy
|
Title: Add `id` and `finish_reason` to `ModelResponse`
Body: These fields would be used to populate `gen_ai.response.id` and `gen_ai.response.finish_reasons` in https://opentelemetry.io/docs/specs/semconv/gen-ai/gen-ai-spans/#genai-attributes | 0easy
|
Title: Bug (?) from #2041
Body: This new pattern in ``PlotCurveItem.setPen``, ``PlotCurveItem.setBrush``, and others has caused a number of breaks for me:
https://github.com/pyqtgraph/pyqtgraph/blob/4e350d03ac2007c47cbfcca9d3b77a6a2a638e5c/pyqtgraph/graphicsItems/PlotCurveItem.py#L373-L380
I believe it would be safer as
```python
if args and args[0] is None:
self.opts['pen'] = None
```
As for now, I've had to change all my calls to give the color as a plain argument rather than keyword, which is fine, but inconsistent with the [docs for mkPen](https://pyqtgraph.readthedocs.io/en/latest/style.html#line-fill-and-color). | 0easy
|
Title: Rename the API class to App
Body: API can be confusing in that it is unclear that the class represents a WSGI "app", not the Falcon "api". Therefore, we have decided to rename the class and module. `falcon.API` should continue to be available as a deprecated alias. The docs will need to be updated accordingly. | 0easy
|
Title: Add Python 3.13 to CI checks
Body: | 0easy
|
Title: Add DaysInMonth primitive
Body: - This primitive determines the number of days in a month of a certain datetime | 0easy
|
Title: Move from `setup.cfg` to `pyproject.toml`
Body: References:
- https://setuptools.pypa.io/en/latest/userguide/pyproject_config.html
- https://github.com/abravalheri/ini2toml | 0easy
|
Title: Multiple conversations in parallel without closing old ones
Body: **Is your feature request related to a problem? Please describe.**
Can we have multiple conversations/threads that are not closed and can be continued any time?
**Describe the solution you'd like**
Do not force user to use "End conversation" button on old conversations every time you need to start a new one.
Also do not rename conversations to "Closed-GPT" - so it would be easier to navigate old convos.
Maybe name them by first opener line?
Also allow to continue any conversation any time.
**Additional context**
The original ChatGPT has all its conversations "open" for various topics. You don't need to click "end conversation" every time you start a new one. It's easy to continue any previous conversation there is.
| 0easy
|
Title: Whether the hyperparameter search algorithm will refer to the value of additionalMetricNames
Body: If not, is the purpose of the additionalMetricNames parameter just for visualization? It will not affect the experimental results?
Please help me answer this question, thank you very much !!!🙏 | 0easy
|
Title: [DOCS] Clarify difference in function between data/external and data/raw
Body: Based on the `<-` descriptions in the 'Directory structure' section of the documentation, there doesn't seem to be clear-cut criterion for choosing between `data/external` and `data/raw` in those cases where the original data dump *originates* from a third-party source, i.e., fulfills the conditions for inclusion in either directory.
What sort of criteria do you apply in such ambiguous cases?
On a related note, where does `data/external` fit in your 'mental model' of the preprocessing pipeline? (Pick one)
- A
```
raw ---> interim ---> processed ---+
|
+---> [ analysis ]
|
external ----+
```
- B
```
raw -------+
|
+--> interim ---> processed ---> [ analysis ]
|
external --+
``` | 0easy
|
Title: Make response short-circuiting docs clearer and more visual
Body: Based on a question from our community (by @amarynets), where one is wondering whether setting `resp.complete = True` in `process_response` will stop any remaining `process_response` methods.
The docs state clearly
> However, any `process_response` middleware methods will still be called.
The context is, however, setting `resp.complete` from `process_request` & `process_resource`, not `process_response`.
I suggest rewording or improving docs to make it clear that any `process_response` middleware methods are unaffected by the `resp.complete` flag, be it set from any middleware method or inside the routed responder.
Also, it would be nice to include visual diagrams like we have for middleware order, something along the lines of
> mob1.process_request
> mob2.process_request :arrow_right: `resp.complete = True`
> ~mob3.process_request~
> ~mob1.process_resource~
> ~mob2.process_resource~
> ~mob3.process_resource~
> ~`route to resource responder method`~
> mob3.process_response
> mob2.process_response
> mob1.process_response
``` | 0easy
|
Title: Add triangular git workflow helpers to documentation / contribution
Body: On the [contributing page](http://www.scikit-yb.org/en/latest/contributing.html) on the website and in [contributing.md](https://github.com/DistrictDataLabs/yellowbrick/blob/develop/CONTRIBUTING.md) there are beginner-friendly instructions on how to make a fork of a repository. It would be useful to also include information about how to configure more than one remote repository (origin and upstream) so that contributors new to a forking workflow know how to get updates to the repo that occur after they forked the repo.
### Proposal/Issue
We can use / adapt the writeup here that talks about this.
http://housinginsights.org/resources/onboarding/triangular-git.html
### Background
I don't know if anyone has had problems with this, but in Housing Insights I often had people confused why their local code was different when they had done `git pull`, but when it only pulls from their fork they won't get the updates that occur on the DDL upstream repo.
| 0easy
|
Title: Default value for custom input object type
Body: I have a custom input object type
```
class Filters(graphene.InputObjectType):
name = graphene.String()
type = graphene.List(graphene.String)
```
and the query definition
```
class Query(graphene.ObjectType):
search = graphene.Field(
SearchResult, query=graphene.String(required=True),
filters=graphene.Argument(Filters, default_value=None)
)
```
I want to be able to provide an empty default value for the `filters` argument. With the above definition, the argument is not set as optional and I'm getting this error message:
```
{
"errors": [
{
"message": "resolve_search() missing 1 required positional argument: 'filters'",
"locations": [
{
"line": 15,
"column": 3
}
],
"path": [
"search"
]
}
],
"data": {
"search": null
}
}
```
What is the correct way to provide an empty default value for the argument of custom object type? I'm at these package versions:
`graphene==2.1.3`
`graphql-core==2.1`
`graphql-server-core==1.1.1` | 0easy
|
Title: More user information should be included in user mode scoreboard export
Body: User mode scoreboard CSV exports are missing the user IDs as well as user emails.
https://github.com/CTFd/CTFd/blob/2c32791c241cd9ee6bb3183ed2221154e70a927f/CTFd/utils/csv/__init__.py#L117-L126 | 0easy
|
Title: Clarify which methods allow keyword arguments or require a dict
Body: In the README it [says][1]:
> puppeteer uses an object for passing options to functions/methods. pyppeteer methods/functions accept both dictionary (python equivalent to JavaScript's objects) and keyword arguments for options.
>
> ### Dictionary style options (similar to puppeteer):
>
> ```
> browser = await launch({'headless': True})
> ```
>
> ### Keyword argument style options (more pythonic, isn't it?):
>
> ```
> browser = await launch(headless=True)
> ```
This is great and I 100% prefer the keyword argument-style options. But not all methods seem to support them, for example [`page.setViewport`][2]:
```
await page.setViewport(width=800, height=600, deviceScaleFactor=2)
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
<ipython-input-50-4d077b87a1ec> in <module>
----> 1 await page.setViewport(width=800, height=600, deviceScaleFactor=2)
TypeError: setViewport() got an unexpected keyword argument 'width'
```
I think it would be best if all the methods in the pyppeteer API supported keyword arguments but, barring that, documentation on which ones do would be second best.
Thanks for the great library, it's exactly what I was looking for!
[1]: https://github.com/pyppeteer/pyppeteer#keyword-arguments-for-options
[2]: https://pyppeteer.github.io/pyppeteer/reference.html#pyppeteer.page.Page.setViewport | 0easy
|
Title: typed.List __repr__()
Body: In Numba 0.60, I notice that the `__repr()__` for `typed.List` shows an ellipsis (as if there is more data) at the end of the list, even if the full array has been printed. Example:
```
from numba.typed import List
l = List([1,2,3,4])
print(repr(l))
```
Output:
```
ListType[int64]([1, 2, 3, 4, ...])
```
This initially confused me as I thought there was more data in the list than what I put in. | 0easy
|
Title: Utilize {% querystring %} template tag
Body: ### Describe the problem
The querystring handling pagination currently has quite complex logic in `weblate.trans.forms.SearchForm`. Most likely it can be fully eliminated by using the `{% querystring %}` template tag introduced in Django 5.1 (required by Weblate since https://github.com/WeblateOrg/weblate/pull/13636) see https://docs.djangoproject.com/en/5.1/releases/5.1/#querystring-template-tag.
### Describe the solution you would like
I believe using `{% querystring %}` in the template can replace `items`, `urlencode` and `reset_offset` methods in the form.
### Describe alternatives you have considered
_No response_
### Screenshots
_No response_
### Additional context
https://github.com/WeblateOrg/weblate/blob/b4c476f9ef7e38fd854169d270953abbee4c5d98/weblate/trans/forms.py#L809-L852 | 0easy
|
Title: Allow WHILE limit to be modified in listener V3
Body: Currently it is impossible to change WHILE limit settings using the listener interface- listener object is created before control is passed to attached listeners, thus modifying while body parameters related to the limit is pointless.
It might be useful to allow listeners to change this- for example, an user might want to redefine default limit behaviour for WHILE loops in case it was not explicitly defined. For example, instead of having default 10k iterations as a default, no-configuration WHILE loop would abort after 10 seconds using this feature. | 0easy
|
Title: [ENH] Coalesce function that doesn't delete columns
Body: # Brief Description
It seems the coalesce function must always replace all input columns with one output column. While this makes sense from the meaning of "coalesce" I can think of cases when this is not the desired outcome.
For example, I have a column D that is a good default value for a few other columns and I might want to coalesce([A,D],A), coalesce([B, D], B) etc. This would break currently as D would disappear after the first coalesce
In this case, I could use A.fillna(D, inplace=True) but if I have multiple default columns (e.g. I want to coalesce([A,B,D]) and coalesce([B,C,D]) then it becomes messy. Maybe there is a better way I am not aware of.
I'm happy to try a PR for this but wanted to confirm that this is a reasonable thing to do.
# Example API
I would propose adding a delete_columns option which is defaulted to be True so existing behaviour is preserved. Or keep_columns which defaults to False.
```python
def coalesce(
df: pd.DataFrame,
column_names: Union[str, Iterable[str], Any],
new_column_name: str = None,
delete_columns: bool = True
) -> pd.DataFrame:
...
if delete_columns:
df = df.drop(columns=column_names)
```
It would still replace an existing column if no column name is specified or the new column already exists. | 0easy
|
Title: Setup.py uses the imp module that is deprecated
Body: while trying to run `python setup.py test` I get the warning
`DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses`
Since python 2 is no longer supported I think we can migrate to `importlib` | 0easy
|
Title: Refactoring: environment variables
Body: Refactoring:
### First step
- [x] In completion search by substring instead of search by lprefix. Solved in #5388.
- [x] Fix setting variable using `xonsh -DVAR=VAL`. Solved in #5396.
- [x] Group env variables by xonsh components. First step is in #5501
### Second step
Xonsh has [many environment variables](https://xon.sh/envvars.html) and part of them are marked as `XONSH_` and the rest not. I suggest to add `XONSH_` prefix to all environment variables that are related to xonsh itself i.e. rename `RAISE_SUBPROC_ERROR` to `XONSH_RAISE_SUBPROC_ERROR` etc. Pros: Clear understanding of `env`. The way to remove xonsh env variables by running `env | grep -v XONSH_ | grep -v XONTRIB_`.
- [ ] We need to implemet backwards compatibility: the state of `RAISE_SUBPROC_ERROR` should update `XONSH_SUBPROC_RAISE_ERROR` values.
After this the steps look like:
1. Add "Deprecated" to `RAISE_SUBPROC_ERROR` doc text.
2. Create `XONSH_SUBPROC_RAISE_ERROR`.
3. Enable sync values between these two i.e. if I set `RAISE_SUBPROC_ERROR=1` I expect to have `XONSH_SUBPROC_RAISE_ERROR=1` as well and vice versa.
## For community
⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
| 0easy
|
Title: BUG: runtime error caused by `read_xxx` with relative path
Body: ### Describe the bug
In the interactive environment, the work dir of a local cluster is determined during initialization. If the current dir of IPython changed, loading a dataframe with a relative path may result in a FileNotFound error.
### To Reproduce
```python
In [1]: import xorbits
In [2]: xorbits.init()
2023-03-14 21:20:14,070 xorbits._mars.deploy.oscar.local 22259 WARNING Web service started at http://0.0.0.0:45351
In [3]: cd /path/to/my/dataset
In [4]: import xorbits.pandas as pd
In [5]: trips = pd.read_parquet("data.parquet") # will cause a FileNotFoundError since Xorbits is initalized under another directory
In [6]: trips
```
### Expected behavior
Xorbits should better take the relative path and make it an absolute path.
| 0easy
|
Title: Add a CI lint task to check st2client's README.md
Body: We need to make sure that the st2client `README.rst` file is acceptable to PyPI, since any syntax errors in it will cause the `push_st2client` task of the `st2cd.st2_finalize_release` workflow to fail.
We can check the syntax using the same renderer that PyPI itself uses:
```bash
# Use the same README renderer that PyPI uses to catch syntax issues in the
# README.rst file # st2client uses README.rst
# https://pypi.org/help/#description-content-type
# https://pypi.org/project/readme-renderer
# https://packaging.python.org/tutorials/packaging-projects/#description
echo "Checking README.rst syntax"
virtualenv venv-st2client-readme-checker
. venv-st2client-readme-checker/bin/activate
pip install --upgrade readme_renderer
python -m readme_renderer README.rst
deactivate
```
It would be nice if we could catch these errors before release, which means that we should create a step in our CI tooling to check it before any bad changes get merged. | 0easy
|
Title: Add support for DRYAD data repository
Body: [Sorry if my issues sound like a broken record! I hope more contentproviders will open up BinderHub for more communities.]
### Proposed change
Support DRYAD (or [stash](https://github.com/CDL-Dryad/stash)) as a content provider (DOI + URL-based): https://datadryad.org/
Their guide to data sharing explicitly mentions code + data, see https://datadryad.org/docs/QuickstartGuideToDataSharing.pdf, so _they should be a good fit_.
_The biggest problem right now_ is that I did not yet find a suitable repository (chicken and egg problem?!), and DRYAD is [not for free](https://datadryad.org/stash/publishing_charges) unless you are affiliated with specific institutions.
**Anyone here who can upload a test binder to DRYAD?**
Relevant API documentation:
- https://datadryad.org/api/v2/docs/#/default/get_datasets__doi_
- https://datadryad.org/api/v2/docs/#/default/get_datasets__doi__download (zipped and direct file download possible, didn't see any limitation on ZIP archive size)
### Alternative options
None? :fist_raised:
### Who would use this feature?
DRYAD is widely known in some research communities, e.g. ecology, so any researcher publishing a dataset or workflow there would be familiar with it.
### How much effort will adding it take?
1 day, with the current examples of contentproviders as a boilerplate.
### Who can do this work?
A developer familiar with the content provider implementation.
| 0easy
|
Title: Docstring of align2d and estimate_shift2d should clarify units of shift
Body: Shifts are currently in pixels. Might be nice if they could also be in calibrated units if floats, but regardless should be clarified in the docstrings. | 0easy
|
Title: [Feature] facebook/contriever support requring
Body: ### Checklist
- [ ] 1. If the issue you raised is not a feature but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed.
- [ ] 2. Please use English, otherwise it will be closed.
### Motivation
I guess facebook/contriever somehow is a popular embedding model so if SGLang supports it, will be very cool.
But I guess it may be time-consuming to support a model that is not llama-structure, but it is indeed popular in RAG setting.
### Related resources
_No response_ | 0easy
|
Title: Add tests for mixed dimensions
Body: | 0easy
|
Title: UnboundLocalError: local variable 'notebook' referenced before assignment
Body: There might be very mighty looking error message if there are some tasks with old notebooks.
```python
[2022-07-28 11:17:14,676: ERROR/MainProcess] Task apps.tasks.tasks.task_execute[d9904836-2f1a-4f38-8d6d-69f1999dc13d] raised unexpected: UnboundLocalError("local variable 'notebook' referenced before assignment")
Traceback (most recent call last):
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 51, in task_execute
task = Task.objects.get(pk=job_params["db_id"])
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/django/db/models/manager.py", line 85, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/django/db/models/query.py", line 435, in get
raise self.model.DoesNotExist(
apps.tasks.models.Task.DoesNotExist: Task matching query does not exist.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 349, in task_execute
task.result = str(e)
UnboundLocalError: local variable 'task' referenced before assignment
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/celery/app/trace.py", line 450, in trace_task
R = retval = fun(*args, **kwargs)
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/celery/app/trace.py", line 731, in __protected_call__
return self.run(*args, **kwargs)
File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 356, in task_execute
print(notebook.schedule)
UnboundLocalError: local variable 'notebook' referenced before assignment
``` | 0easy
|
Title: Workflow
Body: Could someone comment, what is the workflow here.
If after the initial code generation, if user needs to modify or add new features to existing prompt text, next run of the command gpt-engineer <folder_name> will do what?
1. Completely remove existing code?
2. Append code with new changes?
3. Or something else?
docker-compose.yml is generated but it is looking for a local.yml file? So I renamed docker-compose.yml -> local.yml and docker-compose up seems to work, but why?
| 0easy
|
Title: [CI] ci job `Test Build For Component katib-ui` is flaky
Body: ### What happened?
ERROR: failed to solve: process "/bin/sh -c npm config set fetch-retry-mintimeout 20000 && npm config set fetch-retry-maxtimeout 120000 && npm config get registry && npm config set registry https://registry.npmjs.org/ && npm config get https-proxy && npm config rm https-proxy && npm ci" did not complete successfully: exit code: 1
Error: buildx failed with: ERROR: failed to solve: process "/bin/sh -c npm config set fetch-retry-mintimeout 20000 && npm config set fetch-retry-maxtimeout 120000 && npm config get registry && npm config set registry https://registry.npmjs.org/ && npm config get https-proxy && npm config rm https-proxy && npm ci" did not complete successfully: exit code: 1
ref: https://github.com/kubeflow/katib/actions/runs/12830311211/job/36110037074?pr=2496
### What did you expect to happen?
the ci to pass this step
### Environment
Kubernetes version:
```bash
$ kubectl version
```
Katib controller version:
```bash
$ kubectl get pods -n kubeflow -l katib.kubeflow.org/component=controller -o jsonpath="{.items[*].spec.containers[*].image}"
```
Katib Python SDK version:
```bash
$ pip show kubeflow-katib
```
### Impacted by this bug?
Give it a 👍 We prioritize the issues with most 👍 | 0easy
|
Title: Improve test coverage
Body: <!-- Please use the appropriate issue title format:
BUG REPORT
Bug: {Short description of bug}
SUGGESTION
Suggestion: {Short description of suggestion}
OTHER
{Question|Discussion|Whatever}: {Short description} -->
## Description
Currently, the test coverage is around 88% which is okay but we can improve it. This issue focuses on adding tests covering missing branches.
## People to notify
@manrajgrover
| 0easy
|
Title: Add blueprint model
Body: It would be nice to use it's flexibility. | 0easy
|
Title: ENH: implement `__array__` for index
Body: ### Is your feature request related to a problem? Please describe
Implement `__array__` for index.
The impl for dataframe could be helpful: #282 .
| 0easy
|
Title: How to use the KDJ indicator?
Body: Hello, I'm new to pandas/FreqTrade and want to use pandas-ta because this library provide the KDJ indicator.
But I'm really confused how to use the KDJ in my strategy - how I can measure if the KDJ is positive or negative?
Thx in advance. | 0easy
|
Title: Remove full stop at end of viewer/layer action names
Body: Related: #7231(https://github.com/napari/napari/pull/7231#discussion_r1741406517)
Some viewer/layer action names end with a full stop and some do not, e.g.,
https://github.com/napari/napari/blob/a9e7375e339a2240f6cc6affd05eceefa2e843e9/napari/components/_viewer_key_bindings.py#L200
I propose that we remove the full stops which would make the names consistent and it would improve readability of error messages, e.g. the one in #7231
If no objections will do this end of the week. | 0easy
|
Title: Managing multiple pipelines that import modules with the same name
Body: In a layout like this:
```sh
some-pipeline/
pipeline.yaml
tasks.py
another-pipeline/
pipeline.yaml
tasks.py
```
If a user loads in Jupyter `some-pipeline/pipeline.yaml` for cell injection and there are tasks imported from tasks.py, if they then try to open a file in `another-pipeline`, `another-pipeline/tasks.py` will not be imported and the one from the other pipeline will be used.
Perhaps remove the cache before processing specs? Although this will make processing slower.
| 0easy
|
Title: TextProfiler only do case sensitivity at report
Body: Instead of case sensitivity being checked at profile time, we can leave sensitivity on and at report time we can then output collapse it if the options require case-insensitive. | 0easy
|
Title: Bug: AsyncAPI 2.6.0 schema ignores messages schema overriding
Body: **Describe the bug**
1) In some way you describe the message, for example:
```python
@dataclass
class Input:
value: int
```
2) Describe the handler with this message:
```python
@broker.subscriber("in1")
async def handle_msg(a: Input) -> None: ...
```
3) Override your message:
```python
@dataclass
class Input:
not_in_schema: str
```
4) Describe the handler with overrided message:
```python
@broker.subscriber("in2")
async def please_good_schema(a: Input) -> None: ...
```
5) Create asyncapi schema and look at `#/components/schemas`
```json
"schemas": {
"Input": {
"properties": {
"value": {
"title": "Value",
"type": "integer"
}
},
"required": [
"value"
],
"title": "Input",
"type": "object"
}
}
```
Where is my `Input.not_in_schema: str`?
For example, it will be a problem if I have the same named dataclasses in different packages
**How to reproduce**
```python
from dataclasses import dataclass
from faststream import FastStream
from faststream.rabbit import RabbitBroker
broker = RabbitBroker()
app = FastStream(broker)
@dataclass
class Input:
value: int
@broker.subscriber("in1")
async def handle_msg(a: Input) -> None: ...
@dataclass
class Input:
not_in_schema: str
@broker.subscriber("in2")
async def please_good_schema(a: Input) -> None: ...
```
And/Or steps to reproduce the behavior:
1. faststream docs get main:app
**Expected behavior**
I expect two payload in schema:
- Input with `value: int`
- Input2IdkAnyNameHere with `not_in_schema: str`
**Observed behavior**
Message "overriding" just ignored, in the schema all handlers use only the first definition of the message
**Environment**
Running FastStream 0.5.15 with CPython 3.12.3 on Linux
**Additional context**
Same problem in the case of handlers + args names duplication:
```python
@broker.subscriber("in1")
async def handle_msg(a: int) -> None: ...
@broker.subscriber("in2")
async def handle_msg(a: str) -> None: ...
```
```json
"schemas": {
"HandleMsg:Message:Payload": {
"title": "HandleMsg:Message:Payload",
"type": "integer"
}
}
```
| 0easy
|
Title: [Documentation] Add examples how to apply transforms to videos
Body: We can apply the same transform to videos using `additional targets`.
But no one knows about this methods => need example notebook | 0easy
|
Title: Disable multiple dates/date types in dcc.DatePicker*
Body: I'll open this here but I am aware that you're using [AirBnB's react-dates](https://github.com/airbnb/react-dates/issues) so I'll start here and then open an issue there if it's not something that can be implemented on the dash side.
We'd like to disable "categories" of days from being selected. Our smallest request would be disabling weekdays or weekends, but generally supplying a list of available dates or unavailable dates to the DatePicker would be :+1: | 0easy
|
Title: Update Docs / Examples to Use yellowbrick.datasets.load_**
Body: Building off of:
https://github.com/DistrictDataLabs/yellowbrick/pull/417
We should update the documentation to use this much easier interface for using our example datasets.
Docs to Update:
- [x] Readme.md
- [x] Quick Start at http://www.scikit-yb.org/en/latest/quickstart.html
- [x] Model Selection Tutorial at http://www.scikit-yb.org/en/latest/tutorial.html
- [ ] Datasets in the API at http://www.scikit-yb.org/en/latest/api/datasets.html
| 0easy
|
Title: webhook secret_token is not verified
Body: ### Checklist
- [X] I am sure the error is coming from aiogram code
- [X] I have searched in the issue tracker for similar bug reports, including closed ones
### Operating system
Ubuntu 22.04.1 LTS
### Python version
3.10.6
### aiogram version
2.22.2
### Expected behavior
When secret_token is set through calling `await db.bot.set_webhook(..., secret_token=...)` there should be a check that it matches against `X-Telegram-Bot-Api-Secret-Token` header in every webhook request.
### Current behavior
Setting the secret_token gives false sense of security as it is not checked against `X-Telegram-Bot-Api-Secret-Token` header in every webhook request.
### Steps to reproduce
Grep all the codebase for `X-Telegram-Bot-Api-Secret-Token`.
### Code example
_No response_
### Logs
_No response_
### Additional information
_No response_ | 0easy
|
Title: url parameters are not available in `dcc.Location`
Body: we're only sending the `window.location.pathname` (https://github.com/plotly/dash-core-components/blob/master/src/components/Location.react.js#L36) but we could send an object of `window.location.search` (https://stackoverflow.com/questions/5448545/how-to-retrieve-get-parameters-from-javascript)
as reported in https://community.plot.ly/t/get-url-with-get-parameters/5878/1 | 0easy
|
Title: Add tests for ensemble save and load
Body: Add tests for ensemble save and load. It can be done:
- by using some existing learner
- or by writing simple learner framework mockup | 0easy
|
Title: Document Structures in GraphQL Types.
Body: The structures PositiveInt, QuerySetList, TagList, PaginationType, BasePaginatedType and type factory PaginatedQuerySet should be documented in docs/general-usage/graphql-types.rst and how they are used by the decorators should be documented in docs/general-usage/decorators.rst.
This would be a good starter issue as it will help new devs become familiar with the wagtail-grapple architecture and code base.
| 0easy
|
Title: Draw Diagrams
Body: ## Description
Draw Diagrams and make they available in the [New Comers Guide](https://github.com/scanapi/scanapi/wiki/Newcomers)
- [ ] Class
- [ ] Sequence | 0easy
|
Title: Defect Resolution Duration metric API
Body: The canonical definition is here: https://chaoss.community/?p=4727 | 0easy
|
Title: Issue Age metric API
Body: The canonical definition is here: https://chaoss.community/?p=3629 | 0easy
|
Title: @root_validator will not updated to model_validator
Body: ## Description
I have a @root_validator without having any parameter in my project like this:
```
from pydantic import BaseModel, root_validator
class Model(BaseModel):
name: str
@root_validator
def validate_root(cls, values):
return values
```
In this way, `bump-pydantic` can't detect it and will not update it to a new version and without any TODO or anything.
Also, if I define my root_validator like `@root_validator()`, it will be detected and converted to `@model_validator()` without defining `mode` parameter.
The only successful convert will happen if I have `@root_validator(pre=True)`. I've not checked other scenarios.
If it is an accepted bug, I will be happy to work on it and make a PR
## My Environment
Python 3.11.4
bump-pydantic==0.7.0 | 0easy
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.