text
stringlengths
20
57.3k
labels
class label
4 classes
Title: Where is function browserContext.overridePermissions(origin, permissions)? Body: I need to get the read and write permissions of the **clipboard**, but in the headless mode, I can only use this function to increase the permissions. The test is valid in puppeteer, but there is no such function in pyppeteer. I looked at the source code and the principle of pyppeteer looks like It's CDP, I don't know if this function can be implemented? https://pptr.dev/#?product=Puppeteer&version=master&show=api-browsercontextoverridepermissionsorigin-permissions If it can be achieved, may you add **browserContext.overridePermissions(origin, permissions)**? Thanks a lot!
0easy
Title: bug: SQS list_queues doesn't support nextToken Body: ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior using boto3 client: response = client.list_queues(MaxResults=100) response doesn't hold any nextToken field. - I have 2000 sqs queues ### Expected Behavior As mentioned in this: [doc](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html) Responce Syntax: { "[NextToken](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html#SQS-ListQueues-response-NextToken)": "string", "[QueueUrls](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/APIReference/API_ListQueues.html#SQS-ListQueues-response-QueueUrls)": [ "string" ] } ### How are you starting LocalStack? With a docker-compose file ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) docker-compose up --build (image: localstack/localstack) #### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands) AWS SDK: client.list_queues(MaxResults=100) ### Environment ```markdown - OS: - LocalStack: LocalStack version: 3.8.2.dev69 LocalStack Docker image sha: LocalStack build date: 2024-10-31 LocalStack build git hash: fb7cde9cc ``` ### Anything else? _No response_
0easy
Title: Marketplace - fix the navigation bar, margins are too big Body: ### Describe your issue. The nav bar should stretch almost the entire screen. Can we adjust so that the bar only has a 16px margin between the sides of the browser and the nav bar itself? The max that the nav bar should stretch is 1600px. Other than that, have the nav bar dynamically adjust based on the page width <img width="1725" alt="Screenshot 2024-12-16 at 21 17 08" src="https://github.com/user-attachments/assets/3710a582-ba1b-4bae-8342-7ef1c652def0" />
0easy
Title: Explore needed changes to support violinplots Body: This could intuitively straightforward, as they are very similar to boxplots... But, is it ?
0easy
Title: Legend colors are not updated when colors are set with format strings or color kwarg Body: The legend colors are set by the palette and not updated when the colors are specified manually. To replicate the bug: ``` # import import hypertools as hyp import numpy as np from scipy.stats import multivariate_normal # simulate clusters cluster1 = np.random.multivariate_normal(np.zeros(3), np.eye(3), size=100) cluster2 = np.random.multivariate_normal(np.zeros(3)+3, np.eye(3), size=100) # plot hyp.plot([cluster1, cluster2], ['k', 'r'], legend=['cluster1', 'cluster2']) ``` outputs this: ![download](https://user-images.githubusercontent.com/6596567/32500822-71e80086-c3a4-11e7-9ebb-2512fd9120a2.png)
0easy
Title: c_tf_idf_.indptr is None when attempting to save merged model Body: After using `BERTopic.merge_models`, I am unable to save the resulting model. Here is some repro code: ```python import bertopic from bertopic import BERTopic from sklearn.datasets import fetch_20newsgroups from sklearn.decomposition import PCA print(f'Using bertopic version {bertopic.__version__}') embedding_model = 'all-mpnet-base-v2' def new_topic_model(): return BERTopic( embedding_model=embedding_model, # Using PCA because UMAP segfaults on my machine umap_model=PCA(n_components=10, random_state=40), ) docs = fetch_20newsgroups(subset='all', remove=('headers', 'footers', 'quotes'))['data'] docs_a = docs[:400] docs_b = docs[400:500] topic_model_a = new_topic_model() topic_model_a.fit(docs_a) # No problem saving topic_model_a topic_model_a.save( './topic_model_a', serialization='safetensors', save_embedding_model=embedding_model, save_ctfidf=True, ) topic_model_b = new_topic_model() topic_model_b.fit(docs_b) # No problem saving topic_model_b topic_model_b.save( './topic_model_b', serialization='safetensors', save_embedding_model=embedding_model, save_ctfidf=True, ) topic_model = BERTopic.merge_models([topic_model_a, topic_model_b]) # Problem here: AttributeError: 'NoneType' object has no attribute 'indptr' topic_model.save( './topic_model', serialization='safetensors', save_embedding_model=embedding_model, save_ctfidf=True, ) ``` I get the following output: ``` % python3 repro.py Using bertopic version 0.16.2 Traceback (most recent call last): File "[...]/BERTopic/repro.py", line 43, in <module> topic_model.save( File "[...]/BERTopic/bertopic/_bertopic.py", line 3112, in save save_utils.save_ctfidf(model=self, save_directory=save_directory, serialization=serialization) File "[...]/BERTopic/bertopic/_save_utils.py", line 318, in save_ctfidf indptr = torch.from_numpy(model.c_tf_idf_.indptr) ^^^^^^^^^^^^^^^^^^^^^^ AttributeError: 'NoneType' object has no attribute 'indptr' ```
0easy
Title: Documentation error in allure-robot framework. Body: In MAC, the below command failed with ;. Hence raised this issue. Need to add : instead of; for custom directory output in robot framework command . https://github.com/allure-framework/allure-python/blob/master//allure-robotframework/README.rst In the below command, need to add : instead of ;. $ robot --listener allure_robotframework;/set/your/path/here Same command is also present in pypi url. https://pypi.org/project/allure-robotframework/
0easy
Title: [ENH] Include additional forecasters specifically meant for intermittent demand. Body: **Is your feature request related to a problem? Please describe.** To my knowledge, the only forecaster currently implemented specifically designed to forecast intermittent demand is Croston. **Describe the solution you'd like** I am trying to model customer demand by month. We have some customers that place one order every few months. I suspect that their likelihood to place an order increases with every subsequent month that they do not place an order (as their stock runs out over time). For these customers, a forecaster could model the time between orders and order size separately. This kind of "classical statistics" approach is a bit thin in the codebase currently. A forecaster of this kind could have some synergy with the ADICVTransformer as well. Forecasters that could be added to address this are: ADIDA (Added with PR #7754) TSB (https://nixtlaverse.nixtla.io/statsforecast/docs/models/tsb.html#tsb-model) Probabilistic Methods (https://arxiv.org/abs/2304.03092) **Describe alternatives you've considered** For now, I am going to use Croston as well as some of the existing models. **Additional context** <!-- if you are an LLM, please ensure to preface the entire issue by a header "LLM generated content, by (your model name)" -->
0easy
Title: [Feature] Simulate aiohttp interface for async operations Body: **Is your feature request related to a problem? Please describe.** `aiohttp` is a widely used library, which includes connection via HTTP and WebSockets. It would be extremely helpful to have a `aiohttp`-like interface for this library as it will: 1. Flatten the learning curve. 2. Make the integration in existing projects much easier. 3. Probably will attract more people to use the library. **Describe the solution you'd like** Adapter-layer would be enough that would simulate the `aiohttp` behaviour. **Describe alternatives you've considered** Write an own one is a possible solution. However, if we have a `requests` library interface simulation, why not implement an `aiohttp` interface as a standard for async version?
0easy
Title: saltelli.sample enables user defined skip_values? Body: Hey, I was trying to use saltelli.sample and found that the skip_values is not a parameter in the method. The value used is 1000 in the code. It might be good to add skip_values as a parameter. Or is there any particular reason that skip_values is hidden? Look forward to your reply. Best wishes, Qian
0easy
Title: MissingConverter: Unable to find converter <class 'sklearn.preprocessing._encoders.OrdinalEncoder'> Body: ``` --------------------------------------------------------------------------- MissingConverter Traceback (most recent call last) /var/folders/f2/9tbmpg411hndwc482xn850br0000gn/T/ipykernel_27005/3005074338.py in <module> ----> 1 hb_model = convert(clf, 'torch',X_train[0:1]) ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in convert(model, backend, test_input, device, extra_config) 442 """ 443 assert constants.REMAINDER_SIZE not in extra_config --> 444 return _convert_common(model, backend, test_input, device, extra_config) 445 446 ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in _convert_common(model, backend, test_input, device, extra_config) 403 return _convert_sparkml(model, backend_formatted, test_input, device, extra_config) 404 --> 405 return _convert_sklearn(model, backend_formatted, test_input, device, extra_config) 406 407 ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/convert.py in _convert_sklearn(model, backend, test_input, device, extra_config) 106 # We modify the scikit learn model during translation. 107 model = deepcopy(model) --> 108 topology = parse_sklearn_api_model(model, extra_config) 109 110 # Convert the Topology object into a PyTorch model. ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in parse_sklearn_api_model(model, extra_config) 63 # Parse the input scikit-learn model into a topology object. 64 # Get the outputs of the model. ---> 65 outputs = _parse_sklearn_api(topology, model, inputs) 66 67 # Declare output variables. ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs) 228 tmodel = type(model) 229 if tmodel in sklearn_api_parsers_map: --> 230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs) 231 else: 232 outputs = _parse_sklearn_single_model(topology, model, inputs) ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_pipeline(topology, model, inputs) 274 """ 275 for step in model.steps: --> 276 inputs = _parse_sklearn_api(topology, step[1], inputs) 277 return inputs 278 ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs) 228 tmodel = type(model) 229 if tmodel in sklearn_api_parsers_map: --> 230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs) 231 else: 232 outputs = _parse_sklearn_single_model(topology, model, inputs) ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_column_transformer(topology, model, inputs) 451 ) 452 else: --> 453 var_out = _parse_sklearn_api(topology, model_obj, transform_inputs)[0] 454 if model.transformer_weights is not None and name in model.transformer_weights: 455 # Create a Multiply node ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_api(topology, model, inputs) 230 outputs = sklearn_api_parsers_map[tmodel](topology, model, inputs) 231 else: --> 232 outputs = _parse_sklearn_single_model(topology, model, inputs) 233 234 return outputs ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/_parse.py in _parse_sklearn_single_model(topology, model, inputs) 250 raise RuntimeError("Parameter model must be an object not a " "string '{0}'.".format(model)) 251 --> 252 alias = get_sklearn_api_operator_name(type(model)) 253 this_operator = topology.declare_logical_operator(alias, model) 254 this_operator.inputs = inputs ~/opt/anaconda3/lib/python3.9/site-packages/hummingbird/ml/supported.py in get_sklearn_api_operator_name(model_type) 463 """ 464 if model_type not in sklearn_api_operator_name_map: --> 465 raise MissingConverter("Unable to find converter for model type {}.".format(model_type)) 466 return sklearn_api_operator_name_map[model_type] 467 MissingConverter: Unable to find a converter for model type <class 'sklearn.preprocessing._encoders.OrdinalEncoder'>. It usually means the pipeline being converted contains a transformer or a predictor with no corresponding converter implemented. Please fill an issue at https://github.com/microsoft/hummingbird. ``` **Which Scikit-learn pipeline operators do hummingbird support?**
0easy
Title: ENH: remove unreachable constructors wrappers from `xorbits.pandas` Body: ### Is your feature request related to a problem? Please describe We have several unreachable constructor wrappers that could be removed. https://github.com/UranusSeven/xorbits/blob/3281ef9b20222df7125093aa0d0503c8611862c0/python/xorbits/pandas/mars_adapters/core.py#L50-L71 ### Describe the solution you'd like Remove the unreachable wrappers.
0easy
Title: Add exponential backoff retry for API errors Body: When the upstream API returns an error that's out of our control, use exponential backoff to attempt to retry the original request and return it to the original return location
0easy
Title: [Feature Request] Make DecentralizedAverager resistant to KeyboardInterrupt Body: __Update:__ The issue is fixed for the DHT but remains unsolved for DecentralizedAverager. Currently, if you're running hivemind.DHT in jupyter (or an interactive console), you will sometimes need to hit "interrupt" when prototyping. Unfortunately, KeyboardInterrupt will also be caught by the DHT process and will shutdown that DHT instance. To reproduce: ```python import time import hivemind dht = DHT(start=True) while True: time.sleep(1) dht.store('key', 'value', expiration_time=hivemind.get_dht_time() + 999) print(end='.', flush=True) # [interrupt here via kernel -> interrupt] # and then check: assert dht.is_alive() ``` Notes: - Whatever fix we implement must still be able to shutdown DHT if it gets deallocated or when calling .shutdown() - This problem also applies to DecentralizedAverager (fix or create an issue)
0easy
Title: Virtualized table error when navigating past the lower edge of the viewport Body: Issue first described in https://github.com/plotly/dash-table/issues/447. The bug seems fixed in `dash==1.0.2` but that can only have happened by random luck. There's no actual fix to be done here, just adding a test to make sure navigating with the keyboard past the edge of the viewport (top or bottom) does not trigger a console error / behaves correctly.
0easy
Title: Add links to cookbook examples to the docs Body: We have a new cookbook section: https://github.com/ploomber/projects/tree/master/cookbook where we add minimal examples for certain use cases, we should add links to each one on the relevant sections in the docs
0easy
Title: ViewSet filter date picker missing margin Body: <!-- Found a bug? Please fill out the sections below. 👍 --> ### Issue Summary filtering a snippet by date shows the date field missing margins on one side. ![image](https://github.com/user-attachments/assets/89338c52-3e20-4b98-9ac0-2bc9c0008d2e) <!-- A summary of the issue. --> ### Steps to Reproduce I am using the model BreadPage from the bakery demo as a ModelViewSet to demo this issue since the other viewsets do not make use of any date fields. 1. In bakery demo bakerydemo > breads > wagtail_hooks.py Add the following to the bottom ```python from bakerydemo.breads.models import BreadPage class ExampleModelViewSet(ModelViewSet): model = BreadPage add_to_admin_menu = True panels = [ FieldPanel("title"), ] list_filter = ["last_published_at"] register_snippet(ExampleModelViewSet) ``` 2. in the admin sidebar click Bread Pages 3. Click filter options > Last published at. It will show below ![image](https://github.com/user-attachments/assets/01f4e069-8df0-4f4b-8715-528c5d1c2ad1) Any other relevant information. For example, why do you consider this a bug and what did you expect to happen instead? - I have confirmed that this issue can be reproduced as described on a fresh Wagtail project: yes ### Technical details - Wagtail version: 6.3 (dev) ### Working on this <!-- Do you have thoughts on skills needed? Are you keen to work on this yourself once the issue has been accepted? Please let us know here. --> Anyone can contribute to this. View our [contributing guidelines](https://docs.wagtail.org/en/latest/contributing/index.html), add a comment to the issue once you’re ready to start.
0easy
Title: Add Japanese translation of deployment guides #190 Body: Once #190 is closed, we can start working on Japanese translation of the new pages. ### The page URLs * https://slack.dev/bolt-python/ ## Requirements Please read the [Contributing guidelines](https://github.com/slackapi/bolt-python/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
0easy
Title: Make plot_duplicates output clearer Body: ## WHAT Make `plot_duplicates` output clearer. ## WHY The graph plotted by `plot_duplicates` output can be cluttered, especially when there are several duplicates found for an image. The titles of each image run into each other, which makes it hard to read the plot. Additionally, the titles can be hard to read, possibly due to a static dpi. An example can be seen below: ![imgdedup_plot_issue](https://user-images.githubusercontent.com/9531254/101158741-86070500-362c-11eb-87ed-eb5337cc00ac.png) ## HOW In the current implementation, a fixed number of columns in the layout is used (= 4), the top row is reserved for plotting only the original image and the number of rows are dynamic. There are several ways to remodel this: - Make the number columns also dynamic so that image titles do not run into each other. - Process the text output so that it is legible. - Come up with a completely new layout.
0easy
Title: Problems loading models Body: OK. I have trained models and I have them sitting in this directory. C:\Users\User\GBAW23 When I try to run this code. automl = AutoML(results_path=r"C:\Users\User\GBAW23") I get this error What am I doing wrong? 2023-05-19 20:24:33,807 supervised.exceptions ERROR Cannot load AutoML directory. Expecting value: line 1 column 1 (char 0) --------------------------------------------------------------------------- JSONDecodeError Traceback (most recent call last) File ~\anaconda3\lib\site-packages\supervised\base_automl.py:211, in BaseAutoML.load(self, path) 210 else: --> 211 m = ModelFramework.load(path, model_subpath, lazy_load) 212 self._models += [m] File ~\anaconda3\lib\site-packages\supervised\model_framework.py:570, in ModelFramework.load(results_path, model_subpath, lazy_load) 568 logger.info(f"Loading model framework from {model_path}") --> 570 json_desc = json.load(open(os.path.join(model_path, "framework.json"))) 571 mf = ModelFramework(json_desc["params"]) File ~\anaconda3\lib\json\__init__.py:293, in load(fp, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 276 """Deserialize ``fp`` (a ``.read()``-supporting file-like object containing 277 a JSON document) to a Python object. 278 (...) 291 kwarg; otherwise ``JSONDecoder`` is used. 292 """ --> 293 return loads(fp.read(), 294 cls=cls, object_hook=object_hook, 295 parse_float=parse_float, parse_int=parse_int, 296 parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) File ~\anaconda3\lib\json\__init__.py:346, in loads(s, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 343 if (cls is None and object_hook is None and 344 parse_int is None and parse_float is None and 345 parse_constant is None and object_pairs_hook is None and not kw): --> 346 return _default_decoder.decode(s) 347 if cls is None: File ~\anaconda3\lib\json\decoder.py:337, in JSONDecoder.decode(self, s, _w) 333 """Return the Python representation of ``s`` (a ``str`` instance 334 containing a JSON document). 335 336 """ --> 337 obj, end = self.raw_decode(s, idx=_w(s, 0).end()) 338 end = _w(s, end).end() File ~\anaconda3\lib\json\decoder.py:355, in JSONDecoder.raw_decode(self, s, idx) 354 except StopIteration as err: --> 355 raise JSONDecodeError("Expecting value", s, err.value) from None 356 return obj, end JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: AutoMLException Traceback (most recent call last) Cell In[13], line 2 1 automl = AutoML(results_path=r"C:\Users\User\GBAW23") ----> 2 automl.predict(X) File ~\anaconda3\lib\site-packages\supervised\automl.py:387, in AutoML.predict(self, X) 370 def predict(self, X: Union[List, numpy.ndarray, pandas.DataFrame]) -> numpy.ndarray: 371 """ 372 Computes predictions from AutoML best model. 373 (...) 385 AutoMLException: Model has not yet been fitted. 386 """ --> 387 return self._predict(X) File ~\anaconda3\lib\site-packages\supervised\base_automl.py:1369, in BaseAutoML._predict(self, X) 1367 def _predict(self, X): -> 1369 predictions = self._base_predict(X) 1370 # Return predictions 1371 # If classification task the result is in column 'label' 1372 # If regression task the result is in column 'prediction' 1373 return ( 1374 predictions["label"].to_numpy() 1375 if self._ml_task != REGRESSION 1376 else predictions["prediction"].to_numpy() 1377 ) File ~\anaconda3\lib\site-packages\supervised\base_automl.py:1301, in BaseAutoML._base_predict(self, X, model) 1299 if model is None: 1300 if self._best_model is None: -> 1301 self.load(self.results_path) 1302 model = self._best_model 1304 if model is None: File ~\anaconda3\lib\site-packages\supervised\base_automl.py:234, in BaseAutoML.load(self, path) 231 self.n_classes = self._data_info["n_classes"] 233 except Exception as e: --> 234 raise AutoMLException(f"Cannot load AutoML directory. {str(e)}") AutoMLException: Cannot load AutoML directory. Expecting value: line 1 column 1 (char 0)
0easy
Title: VWMA support for MA method. Body: Simple question, Is there a reason why not all MA's are supported in the ma method ? As an example, I would like to use the VWMA with ADX. Maybe I missed something. ? Regards,
0easy
Title: Enable PDF on Read the Docs Body: Provided one has the correct dependencies on, e.g., Ubuntu Linux 22.04, our docs can be built to a Latex PDF, but this is no longer enabled on Read the Docs (long ago, it was automatically included). Check what is needed to enable the PDF output again.
0easy
Title: Lazy decomposition raises error on mask check Body: [Noticed by @erh3cq](https://gitter.im/hyperspy/hyperspy?at=5fe024fd63fe0344960c993a) https://github.com/hyperspy/hyperspy/blob/7bdee7a854a52187a6f78c045184d0d9a783ee51/hyperspy/_signals/lazy.py#L883 The above line should read `if navigation_mask.any() or signal_mask.any()`. There might be more places with a similar error as well. Not sure if `navigation_mask is None` would raise the same error or not. Note that @erh3cq might have a more general error with lazy decomposition, worth taking a look at (see link above).
0easy
Title: Add internal API reference for datalab adapter modules in docs Body: <!-- Please be as detailed as possible in your issue. INCLUDE the exact code you are trying to run, if your question is related to code. --> Include (internal) API reference for the `clenalab.datalab.internal.adapter`  modules ([link](https://github.com/cleanlab/cleanlab/tree/master/cleanlab/datalab/internal/adapter)) in the docs. Just update this section and add the appropriate files. https://github.com/cleanlab/cleanlab/blob/325bbe2a77633c87b47f2b275128543f6a5cc980/docs/source/cleanlab/datalab/internal/index.rst#L15-L25
0easy
Title: Add `multiclass` fine-tuning example Body: It would be great to have more examples for fine-tuning in the library! Similar to current examples for [binary-segmentaion](https://github.com/qubvel-org/segmentation_models.pytorch/blob/main/examples/binary_segmentation_intro.ipynb) and [multi-label](https://github.com/qubvel-org/segmentation_models.pytorch/blob/main/examples/cars%20segmentation%20(camvid).ipynb) segmentation would be great to have multi-class segmentation example (the whole image is annotated). The same camvid dataset can be used. The example should showcase how to - fine-tune a model with pytorch-ligtning (or any other fine-tuning framework, even a plain pytorch) - use loss function with ignore index for unlabeled class - compute metrics per-class - visualize results In case anyone wants to work on this you are welcome! Just notify in this issue, and then ping me to review a PR when it's ready.
0easy
Title: [FEATURE] Add a parsable (JSON or XML) endpoint for results Body: <!-- DO NOT REQUEST UI/THEME/GUI/APPEARANCE IMPROVEMENTS HERE THESE SHOULD GO IN ISSUE #60 REQUESTING A NEW FEATURE SHOULD BE STRICTLY RELATED TO NEW FUNCTIONALITY --> **Describe the feature you'd like to see added** Current the `/search` endpoint returns a HTML file to display the results. This works well for UI users, which might be the first use case of Whoogle. But it's hard for programmatic use cases, as it requires parsing HTML. I propose to add a new `/search.{json|xml}` endpoint to return the same results in a structured format: JSON or XML ### Idea 1: JSON A `/search.json?q=%s` endpoint which would search results as JSON. I don't think there's a standard JSON format to use, but we could take inspiration for https://zenserp.com/ or https://serpapi.com/. ### Idea 2: OpenSearch XML format This project already uses opensearch, so we could leverage the same standard's response XML: https://github.com/dewitt/opensearch/blob/master/opensearch-1-1-draft-6.md#opensearch-response-elements **Additional context** I have a personal preference for JSON, as it seems to be today's web standard for APIs. Adding parsable results would allow integration with various projects, via API (e.g. Zapier, Integromat), RSS or simple scripts. Some use cases include: - keyword monitoring - market research - build a voice assistant - and the imagination of the community
0easy
Title: `allow_x00=False` not affecting Headers and Cookies [BUG] Body: ### Checklist - [x] I checked the [FAQ section](https://schemathesis.readthedocs.io/en/stable/faq.html#frequently-asked-questions) of the documentation - [x] I looked for similar issues in the [issue tracker](https://github.com/schemathesis/schemathesis/issues) - [x] I am using the latest version of Schemathesis ### Describe the bug When running Schemathesis tests, I get errors from Postgres about null bytes. Possibly related to #2072. ### To Reproduce 1. Install FastAPI and Schemathesis. 2. Run this code: ```python import warnings from typing import Annotated import schemathesis from fastapi import Cookie, FastAPI, Header from hypothesis.errors import NonInteractiveExampleWarning from schemathesis.generation import GenerationConfig warnings.filterwarnings("ignore", category=NonInteractiveExampleWarning) app = FastAPI() @app.post("/test") def test(x_test: Annotated[str, Header()], test_cookie: Annotated[str, Cookie()]): return x_test schemathesis.experimental.OPEN_API_3_1.enable() schema = schemathesis.from_asgi( "/openapi.json", app, generation_config=GenerationConfig(allow_x00=False), ) operation = schema["/test"]["POST"] strategy = operation.as_strategy() while True: case = strategy.example() if "\x00" in case.headers["x-test"]: print("Found null byte in header") break while True: case = strategy.example() if "\x00" in case.cookies["test_cookie"]: print("Found null byte in cookie") break ``` 3. After a few seconds, it should print that it found null bytes. I also checked query parameters and form data, but those seemed to respect the configuration option. Please include a minimal API schema causing this issue: The API schema from the code above: ```json {'openapi': '3.1.0', 'info': {'title': 'FastAPI', 'version': '0.1.0'}, 'paths': {'/test': {'post': {'summary': 'Test', 'operationId': 'test_test_post', 'parameters': [{'name': 'x-test', 'in': 'header', 'required': True, 'schema': {'type': 'string', 'title': 'X-Test'}}, {'name': 'test_cookie', 'in': 'cookie', 'required': True, 'schema': {'type': 'string', 'title': 'Test Cookie'}}], 'responses': {'200': {'description': 'Successful Response', 'content': {'application/json': {'schema': {}}}}, '422': {'description': 'Validation Error', 'content': {'application/json': {'schema': {'$ref': '#/components/schemas/HTTPValidationError'}}}}}}}}, 'components': {'schemas': {'HTTPValidationError': {'properties': {'detail': {'items': {'$ref': '#/components/schemas/ValidationError'}, 'type': 'array', 'title': 'Detail'}}, 'type': 'object', 'title': 'HTTPValidationError'}, 'ValidationError': {'properties': {'loc': {'items': {'anyOf': [{'type': 'string'}, {'type': 'integer'}]}, 'type': 'array', 'title': 'Location'}, 'msg': {'type': 'string', 'title': 'Message'}, 'type': {'type': 'string', 'title': 'Error Type'}}, 'type': 'object', 'required': ['loc', 'msg', 'type'], 'title': 'ValidationError'}}}} ``` ### Expected behavior Using the `allow_x00` configuration option should prevent null bytes in all strings. ### Environment ``` - OS: Linux - Python version: 3.12 - Schemathesis version: 3.28.1 - Spec version: Open API 3.1.0 ```
0easy
Title: [Tech debt] Improve interface for RandomSnow Body: Right now in the transform we have separate parameters for `snow_point_lower` and `snow_point_higher` Better would be to have one parameter `snow_point_range = [snow_point_lower, snow_point_higher]` => We can update transform to use new signature, keep old as working, but mark as deprecated. ---- PR could be similar to https://github.com/albumentations-team/albumentations/pull/1704
0easy
Title: Pydantic "exclude" option is not working anymore Body: **Describe the bug** After upgrading from version 1.22.6 to version 1.23.0, pydantic property exclude stopped working. **Expected behavior** I have a user model which have email and password and I don't want to save the password to mongo. I have the following model: ``` class UserCreate(User): """All fields needed to create a user""" password: Optional[str] = pydantic.Field( description="A user's password in plain text before it has been hashed", exclude=True, ) ``` I expected the password not to be saved into the database but after upgrading versions it is saving it.
0easy
Title: [DOCS] grouped metrics lack documentation Body: Feels like a good thing to add.
0easy
Title: Update notification email header image Body: ### Issue Summary When receiving an email notification from Wagtail, the image header appears to use the old logo and green/teal background. <img width="619" alt="Screenshot 2024-09-27 at 11 49 07 AM" src="https://github.com/user-attachments/assets/e907f1cb-e43d-401d-b7fe-6591fd6cd51d"> ### Steps to Reproduce 1. On a Wagtail site, submit a page for moderation in a workflow. 2. A moderator should receive a notification email from Wagtail 3. If applicable, click 'load images' in your email client. - I have confirmed that this issue can be reproduced as described on a fresh Wagtail project: yes ### Technical details - Python version: 3.12.4 - Django version: 5.0.9 - Wagtail version: 6.1.3 - Browser version: N/A ### Working on this Relevant image file is located at: https://github.com/wagtail/wagtail/blob/main/wagtail/admin/static_src/wagtailadmin/images/email-header.jpg I'd expect an image with the new Wagtail logo (see https://github.com/wagtail/wagtail/pull/11756) and a dark purple background, like Wagtail uses these days. The base notifications template: https://github.com/wagtail/wagtail/blob/abcb2da37259c16d1efd16d7233794f00bb29680/wagtail/admin/templates/wagtailadmin/notifications/base.html
0easy
Title: Marketplace - The headers of each section all have different text styles. Change it to Poppins font so they're consistent Body: ### Describe your issue. Marketplace - The headers of each section "Featured agents", "Top agents", "Featured creators" and "Become a creator" all have different text styles. Change it to Poppins font. Change it to the "large-Poppins" style.. as per the typography sheet. large-poppins: font: poppins size: 18px line-height: 28px Typography sheet is linked here: [https://www.figma.com/design/aw299myQfhiXPa4nWkXXOT/agpt-template?node-id=7-47&t=axoLiZIIUXifeRWU-1](url) <img width="1417" alt="Screenshot 2024-12-13 at 16 53 08" src="https://github.com/user-attachments/assets/2a1dcccb-1b78-4e49-9c6d-7522da5df7c6" /> <img width="1470" alt="Screenshot 2024-12-13 at 16 53 14" src="https://github.com/user-attachments/assets/aebc442b-46e4-42f3-be51-491a8a1d3df0" /> <img width="1404" alt="Screenshot 2024-12-13 at 16 53 21" src="https://github.com/user-attachments/assets/20b3c236-19c1-4b22-ae84-7c7157da4198" /> <img width="1369" alt="Screenshot 2024-12-13 at 16 53 42" src="https://github.com/user-attachments/assets/698ce7f9-21c1-44ec-96a0-cb2330a9b505" />
0easy
Title: explicit configuration file missing in readthedocs.yml Body: ### What is your issue? According to https://about.readthedocs.com/blog/2024/12/deprecate-config-files-without-sphinx-or-mkdocs-config/ `configuration: doc/conf.py` needs to be added to .readthedocs.yaml: https://github.com/pydata/xarray/blob/76120781f2b2205cc37bd97d0b4ed13d78521d75/.readthedocs.yaml#L17-L19
0easy
Title: 训练的问题 Body: 1.为什么8块和16块的GPU在训练的参数不同? 2.为什么针对不同测试数据集,训练的参数不同? 谢谢回复。
0easy
Title: Tab-completion fails when looking for some special strings in aliases ("less", "more" and "dir") Body: ## Steps to Reproduce type in xonsh: ```python "less" in ali ``` 2. press tab. Expected behavior: the commandline should read: ```python "less" in aliases ``` What happens: No autocompletion is done. Note: This happens for "more" in ali<tab> and "dir" in ali<tab> also but works as expected for almost every other string I have tested with. Example: "dirr" in ali<tab> --> "dirr" in aliases EDIT: Seems that all of the keys in the aliases dict cause the same behavior. <details> ``` <+------------------+---------------------------+ | xonsh | 0.14.4 | | Python | 3.12.1 | | PLY | 3.11 | | have readline | False | | prompt toolkit | 3.0.43 | | shell type | prompt_toolkit | | history backend | json | | pygments | 2.17.2 | | on posix | False | | on linux | False | | on darwin | False | | on windows | True | | on cygwin | False | | on msys2 | False | | is superuser | False | | default encoding | utf-8 | | xonsh encoding | utf-8 | | encoding errors | surrogateescape | | xontrib 1 | coreutils | | RC file 1 | C:\Users\<redacted>/.xonshrc | +------------------+---------------------------+ ``` <details> ## For community ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
0easy
Title: `List Should Contain Sub List` does not detect failure if only difference is empty string Body: I expect all of the following to raise AssertionError, as the second list is not a subset of the first: ```python from robot.libraries.Collections import Collections second = [""] Collections().list_should_contain_sub_list(["first"], second) second = ["first", ""] Collections().list_should_contain_sub_list(["first"], second) second = ["", "first"] Collections().list_should_contain_sub_list(["first"], second) second = ["", ""] Collections().list_should_contain_sub_list(["first"], second) ``` However, only the last one raises the error.
0easy
Title: Add data loader for HF oasst1 Body: Currently, data needs to be manually downloaded and a path specified via config to train on OpenAssistant datasets (both internal and https://huggingface.co/datasets/OpenAssistant/oasst1/blob/main/2023-04-12_oasst_ready.trees.jsonl.gz) so that the tree can be parsed by [`model_training.custom_datasets.oasst_dataset.load_oasst_export`](https://github.com/LAION-AI/Open-Assistant/blob/63937be7462ddc3a46399e9a813b36c2585a6fb3/model/model_training/custom_datasets/oasst_dataset.py#L21) We should write/refactor the data loader to work directly with the dataset returned by HF `datasets.load_dataset("OpenAssistant/oasst1")`
0easy
Title: Make moderations threshold parameters parameterized Body: Make moderations threshold parameters parameterized Make it such that users can enter specific values for the 6 openai moderations categories, and support for pre-defined values (like profiles) that can be loaded from file similar to `openers` in conversations
0easy
Title: 20bn website is not available, where can I download data? Body: Hi, I need to download the something-to-something and jester datasets. But the 20bn website "[https://20bn.com](https://20bn.com)" are not available for weeks, the error message is "503 Service Temporarily Unavailable". I have already downloaded the video data of something-to-something v2, and I need the label dataset. For the Jester, I need both video and label data. Can someone share me the download link? Please let me know if this post is not appropriate here, I will close it immediately. Thanks.
0easy
Title: Adopt Collectfasta Body: [Collectfast](https://github.com/antonagestam/collectfast) is archived, long live [Collectfasta](https://github.com/jasongi/collectfasta). The django newletter contained a link to the succssor of collectfast: https://jasongi.com/2024/03/04/speed-up-djangos-collectstatic-command-with-collectfasta/ Should be easy to adopt I think.
0easy
Title: Add documentation about the demo example and clean up `examples` Body: This is a good issue for someone interested in documentation. It doesn't require any particular knowledge of Vizro but does need an interest in poking through our [examples folder in within `vizro-core` for this repo](https://github.com/mckinsey/vizro/tree/main/vizro-core/examples) to see what is there and write about the layout and what the example shows. There's a[ basic style guide for docs creators here](https://vizro.readthedocs.io/en/stable/pages/development/documentation-style-guide/), but we are happy to take any contribution as we can run Vale over it to highlight any tweaks needed to follow the guidlines. The `examples` folder in the repo holds the code for [the demo](https://vizro.mckinsey.com). It isn't very well highlighted in the docs and we need to do this: 1. Add a section on the home page and a link to a page of text about the demo (mentioned in #378 which will restyle that whole page) 2. Add the page itself. That page should document the basics of the demo and how to work with it. 3. Add a section in the `examples` folder where any contributions of example code for #377 will be added.
0easy
Title: output to buffer Body: i need to save to io.buffer something like: ```python buf = io.BytesIO() mpf.plot( candles, type='candle', mav=(3, 6, 9), volume=True, show_nontrading=True, figratio=(6, 6), title="Graph", style='yahoo', savefig=buf ) buf.seek(0) ```
0easy
Title: 📝 Followup with documentation - External Features Body: I still support some old features here as an external service, you can help with extending the level of the documentation, such as update for the documentation ( more user-friendly 😞) ## External Features: - Instrument/profiler - https://authx-v0.yezz.me/configuration/middleware/profiling/ - Redis - https://authx-v0.yezz.me/configuration/cache/HTTPCache/ - Metrics - https://authx-v0.yezz.me/configuration/middleware/metrics/ - OAuth2 support - https://authx-v0.yezz.me/configuration/middleware/ (Bunch of Extra changes, that need my review)
0easy
Title: Require `--suite parent.suite` to match the full suite name Body: When you use `--test parent_name.test_name`, there must be a test named `test_name` in a suite named `parent_name` and that suite must also be the root suite. On the other hand, `--suite parent_name.suite_name` works so that `parent_name` can also be a child suite. This is pretty inconsistent and needs to be changed. The reason to requite `--test parent_name.test_name` to match the full suite name is that it makes it possible to explicitly pinpoint what test to run. If the name could match also partially, it could match also tests it shouldn't. For example Pabot relies on this functionality and it cannot be changed. If a partial match is needed, it's possible to use wildcards like `--test *.parent_name.test_name`. I don't remember there being any reason why `--suite parent_name.suite_name` should also support partial matches. To make `--test` and `--suite` behavior consistent, it ought to thus be ok to change how `--suite` behaves. The change is backwards incompatible but we can do it in Robot Framework 7.0. I doubt this functionality is used so often that it should be deprecated first.
0easy
Title: Tutorial to connect Supabase to Streamlit Body: **Is your feature request related to a problem? Please describe.** Tutorial instructions on how to use Streamlit with Supabase as a data source **Describe the solution you'd like** A PR or guide against https://github.com/streamlit/docs/blob/main/content/kb/tutorials/databases/index.md describing how to link supabase with streamlit. Should be similar to the Postgres example **Describe alternatives you've considered** N.A. **Additional context** I have filed an [issue on the streamlit docs repo](https://github.com/streamlit/docs/issues/325).
0easy
Title: posargs with `:` crashes virtualenv Body: ## Issue `posargs` with `:` are not properly passed, `virtualenv` crashes. ## Environment Provide at least: - OS: Ubuntu 20.04 - `pip list` of the host Python where `tox` is installed: ```console Package Version --------------------- --------- anyio 3.6.2 artifacts-keyring 0.3.2 attrs 22.2.0 bleach 5.0.1 cachetools 5.2.1 certifi 2022.12.7 cffi 1.15.1 chardet 5.1.0 charset-normalizer 2.1.1 click 8.1.3 colorama 0.4.6 commonmark 0.9.1 cryptography 39.0.0 distlib 0.3.6 docutils 0.19 exceptiongroup 1.1.0 filelock 3.9.0 h11 0.14.0 html5lib 1.1 httpcore 0.16.3 httpx 0.23.3 idna 3.4 importlib-metadata 6.0.0 importlib-resources 5.10.2 iniconfig 2.0.0 jaraco.classes 3.2.3 jeepney 0.8.0 keyring 23.13.1 more-itertools 9.0.0 packaging 23.0 pip 22.3.1 pkginfo 1.9.6 platformdirs 2.6.2 pluggy 1.0.0 pycparser 2.21 pydantic 1.10.4 Pygments 2.14.0 pyproject_api 1.4.0 pytest 7.2.0 pytest-httpx 0.21.2 readme-renderer 37.3 requests 2.28.1 requests-toolbelt 0.10.1 rfc3986 1.5.0 rich 13.0.1 SecretStorage 3.3.3 setuptools 57.5.0 simpleindex 0.5.0 simpleindex_artifacts 0.3 six 1.16.0 sniffio 1.3.0 starlette 0.23.1 toml 0.10.2 tomli 2.0.1 tox 4.2.8 twine 4.0.2 typing_extensions 4.4.0 urllib3 1.26.14 uvicorn 0.20.0 virtualenv 20.17.1 webencodings 0.5.1 wheel 0.38.4 zipp 3.11.0 ``` ## Output of running tox Provide the output of `tox -rvv`: ```console ROOT: 89 D setup logging to DEBUG on pid 26084 [tox/report.py:221] .pkg: 156 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56] .pkg: 156 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63] .pkg: 156 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65] .pkg: 158 D filesystem is case-sensitive [virtualenv/info.py:24] .pkg: 180 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56] .pkg: 180 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63] .pkg: 181 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65] .pkg: 184 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56] .pkg: 185 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63] .pkg: 185 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65] .pkg: 189 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56] .pkg: 190 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63] .pkg: 190 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65] .pkg: 204 I find interpreter for spec PythonSpec(path=/usr/local/bin/python) [virtualenv/discovery/builtin.py:56] .pkg: 204 I proposed PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63] .pkg: 205 D accepted PythonInfo(spec=CPython3.8.16.final.0-64, exe=/usr/local/bin/python, platform=linux, version='3.8.16 (default, Dec 8 2022, 03:18:29) \n[GCC 10.2.1 20210110]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65] usage: virtualenv [--version] [--with-traceback] [-v | -q] [--read-only-app-data] [--app-data APP_DATA] [--reset-app-data] [--upgrade-embed-wheels] [--discovery {builtin}] [-p py] [--try-first-with py_exe] [--creator {builtin,cpython3-posix,venv}] [--seeder {app-data,pip}] [--no-seed] [--activators comma_sep_list] [--clear] [--no-vcs-ignore] [--system-site-packages] [--symlinks | --copies] [--no-download | --download] [--extra-search-dir d [d ...]] [--pip version] [--setuptools version] [--wheel version] [--no-pip] [--no-setuptools] [--no-wheel] [--no-periodic-update] [--symlink-app-data] [--prompt prompt] [-h] dest virtualenv: error: argument dest: destination '--junitxml=unit_tests_results.xml --cov=autoavatar --cov-report=xml:coverage.xml' must not contain the path separator (:) as this would break the activation scripts ``` ## Minimal example If possible, provide a minimal reproducer for the issue: **UPDATE** After doing the investigation, it turned out that one environment influences the other. Here is a minimum `tox.ini` to reproduce the issue: `tox.int`: ```ini [tox] minversion = 4.3.1 [testenv:hello] commands = python ./helloworld.py {posargs:World} [testenv:dev] envdir = {posargs:venv} usedevelop = True commands = python ./helloworld.py World ``` `tox -e hello -- x:y` ``` ... usage: virtualenv [--version] [--with-traceback] [-v | -q] [--read-only-app-data] [--app-data APP_DATA] [--reset-app-data] [--upgrade-embed-wheels] [--discovery {builtin}] [-p py] [--try-first-with py_exe] [--creator {builtin,cpython3-posix,venv}] [--seeder {app-data,pip}] [--no-seed] [--activators comma_sep_list] [--clear] [--no-vcs-ignore] [--system-site-packages] [--symlinks | --copies] [--no-download | --download] [--extra-search-dir d [d ...]] [--pip version] [--setuptools version] [--wheel version] [--no-pip] [--no-setuptools] [--no-wheel] [--no-periodic-update] [--symlink-app-data] [--prompt prompt] [-h] dest virtualenv: error: argument dest: destination 'x:y' must not contain the path separator (:) as this would break the activation scripts ```
0easy
Title: Add caching to load_dag Body: `load_dag` is an expensive operation and many times it's called with the same argument, so we should cache the result here's the implementation: https://github.com/ploomber/ploomber/blob/2e0d764c8f914ba21480b275c545ea50ff800513/src/ploomber/jupyter/manager.py#L107 so every time the function is called, we should save the `starting_dir` value, and the returned `path`: https://github.com/ploomber/ploomber/blob/2e0d764c8f914ba21480b275c545ea50ff800513/src/ploomber/jupyter/manager.py#L120 next time the function is called, check if `starting_dir` is the same and the file in `path` hasn't changed. if so, skip the function's body add tests here: https://github.com/ploomber/ploomber/blob/master/tests/test_jupyter.py#L901
0easy
Title: tox quickstart: error: the following arguments are required: root Body: ## Issue I ran `tox quickstart` as described in https://tox.wiki/en/latest/user_guide.html. and get the error `tox quickstart: error: the following arguments are required: root`. It seems that either the default parameter that should select the current working environment seems to be broken or the documentation should be updated. ## Environment Provide at least: - OS: Windows 11 <details open> <summary>Output of <code>pip list</code> of the host Python, where <code>tox</code> is installed</summary> ```console Python version: 3.11 Packages: No idea, I installed with pipx ``` </details> Thanks a lot for developing this super useful library!
0easy
Title: [Feature request] Add apply_to_images to InvertImg Body:
0easy
Title: Scattering2D docstring contains an outdated example Body: See https://github.com/kymatio/kymatio/blob/master/kymatio/scattering2d/scattering2d.py#L37 ```python # 1) Define a Scattering object as: s = Scattering2D(J, M, N) # where (M, N) are the image sizes and 2**J the scale of the scattering ``` The constructor should read as `Scattering2D(J, shape=(M, N))`
0easy
Title: problem with azure, api key and/or config.yaml Body: ### Is your feature request related to a problem? Please describe. First, congratulations for this app, its terrific! :) My problem: my credits in openai finished, and I need to use my credits in azure. OpenInterpreter are not working with azure for me. I'm in Windows10. Before I run with the openai configuration without problem. I followed the PR [#786](https://github.com/KillianLucas/open-interpreter/pull/786) without success. My config.yaml (in c:\Users\Miguel\AppData\Local\Open Interpreter\Open Interpreter\ or :E:\Users\MFG\Desktop\Programming\Eos_2\open-interpreter\interpreter\terminal_interface\config.yaml) is: local: false temperature: 0 auto_run: true context_windows: 31000 max_tokens: 3000 openai_api_key: d1...43 api.base: https://oaiwestus.openai.azure.com/ api.type: azure model: azure/gpt-4-turbo azure.api_version: 2023-07-01-preview When I run "interpreter" in the terminal, appears: " OpenAI API key not found ", and a prompt to fill with the message: "OpenAI API key: " When I paste my key, it seems ok, and appear the prompt (>), but when I input a simple "hi", it returns a error message finished in : "There might be an issue with your API key(s). To reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc): Mac/Linux: 'export OPENAI_API_KEY=your-key-here', Windows: 'setx OPENAI_API_KEY your-key-here' then restart terminal." When I run "interpreter --model=azure/gpt-4-turbo" or "interpreter --m azure/gpt-4-turbo", it appears the prompt (>), I input "hi", and it returns: "We were unable to determine the context window of thismodel. Defaulting to 3000. If your model can handle more, run interpreter --context_window {token limit} --max_tokens {max tokens per response}. Continuing... # Interpreter Info Vision: False Model: azure/gpt-4-turbo Function calling: None Context window: None Max tokens: None Auto run: True API base: None Offline: False . . . File "C:\Users\Miguel\miniconda3\Lib\site-packages\litellm\utils.py", line 6567, in exception_type raise APIError( litellm.exceptions.APIError: AzureException - argument of type 'NoneType' is not iterable" Any idea to solve? Thank you in advance! ### Describe the solution you'd like run interpreter with azure openai ### Describe alternatives you've considered interpreter --reset_config install again ### Additional context _No response_
0easy
Title: [UX] `sky show-gpus` with an expired authentication of k8s shows long error message Body: <!-- Describe the bug report / feature request here --> `sky show-gpus` ``` sky show-gpus ERROR:root:exec: process returned 1. F0109 21:40:59.931088 61835 cred.go:145] print credential failed with error: Failed to retrieve access token:: failure while executing gcloud, with args [config config-helper --format=json]: exit status 1 (err: ERROR: (gcloud.config.config-helper) There was a problem refreshing your current auth tokens: Reauthentication failed. cannot prompt during non-interactive execution. Please run: $ gcloud auth login to obtain new credentials. If you have already logged in with a different account, run: $ gcloud config set account ACCOUNT to select an already authenticated account to use. ) COMMON_GPU AVAILABLE_QUANTITIES A10 1, 2, 4 A10G 1, 4, 8 A100 1, 2, 4, 8, 16 A100-80GB 1, 2, 4, 8 H100 1, 2, 4, 8, 12 L4 1, 2, 4, 8 L40S 1, 4, 8 P100 1, 2, 4 T4 1, 2, 4, 8 V100 1, 2, 4, 8 V100-32GB 1, 2, 4, 8 GOOGLE_TPU AVAILABLE_QUANTITIES tpu-v2-8 1 tpu-v3-8 1 tpu-v4-8 1 tpu-v4-16 1 tpu-v4-32 1 tpu-v5litepod-1 1 tpu-v5litepod-4 1 tpu-v5litepod-8 1 tpu-v5p-8 1 tpu-v5p-16 1 tpu-v5p-32 1 tpu-v6e-1 1 tpu-v6e-4 1 tpu-v6e-8 1 Hint: use -a/--all to see all accelerators (including non-common ones) and pricing. Note: No GPUs found in Kubernetes cluster. If your cluster contains GPUs, make sure nvidia.com/gpu resource is available on the nodes and the node labels for identifying GPUs (e.g., skypilot.co/accelerator) are setup correctly. To further debug, run: sky check ``` <!-- If relevant, fill in versioning info to help us troubleshoot --> _Version & Commit info:_ * `sky -v`: PLEASE_FILL_IN * `sky -c`: PLEASE_FILL_IN
0easy
Title: Add `id` and `finish_reason` to `ModelResponse` Body: These fields would be used to populate `gen_ai.response.id` and `gen_ai.response.finish_reasons` in https://opentelemetry.io/docs/specs/semconv/gen-ai/gen-ai-spans/#genai-attributes
0easy
Title: Bug (?) from #2041 Body: This new pattern in ``PlotCurveItem.setPen``, ``PlotCurveItem.setBrush``, and others has caused a number of breaks for me: https://github.com/pyqtgraph/pyqtgraph/blob/4e350d03ac2007c47cbfcca9d3b77a6a2a638e5c/pyqtgraph/graphicsItems/PlotCurveItem.py#L373-L380 I believe it would be safer as ```python if args and args[0] is None: self.opts['pen'] = None ``` As for now, I've had to change all my calls to give the color as a plain argument rather than keyword, which is fine, but inconsistent with the [docs for mkPen](https://pyqtgraph.readthedocs.io/en/latest/style.html#line-fill-and-color).
0easy
Title: Rename the API class to App Body: API can be confusing in that it is unclear that the class represents a WSGI "app", not the Falcon "api". Therefore, we have decided to rename the class and module. `falcon.API` should continue to be available as a deprecated alias. The docs will need to be updated accordingly.
0easy
Title: Add Python 3.13 to CI checks Body:
0easy
Title: Add DaysInMonth primitive Body: - This primitive determines the number of days in a month of a certain datetime
0easy
Title: Move from `setup.cfg` to `pyproject.toml` Body: References: - https://setuptools.pypa.io/en/latest/userguide/pyproject_config.html - https://github.com/abravalheri/ini2toml
0easy
Title: Multiple conversations in parallel without closing old ones Body: **Is your feature request related to a problem? Please describe.** Can we have multiple conversations/threads that are not closed and can be continued any time? **Describe the solution you'd like** Do not force user to use "End conversation" button on old conversations every time you need to start a new one. Also do not rename conversations to "Closed-GPT" - so it would be easier to navigate old convos. Maybe name them by first opener line? Also allow to continue any conversation any time. **Additional context** The original ChatGPT has all its conversations "open" for various topics. You don't need to click "end conversation" every time you start a new one. It's easy to continue any previous conversation there is.
0easy
Title: Whether the hyperparameter search algorithm will refer to the value of additionalMetricNames Body: If not, is the purpose of the additionalMetricNames parameter just for visualization? It will not affect the experimental results? Please help me answer this question, thank you very much !!!🙏
0easy
Title: [DOCS] Clarify difference in function between data/external and data/raw Body: Based on the `<-` descriptions in the 'Directory structure' section of the documentation, there doesn't seem to be clear-cut criterion for choosing between `data/external` and `data/raw` in those cases where the original data dump *originates* from a third-party source, i.e., fulfills the conditions for inclusion in either directory. What sort of criteria do you apply in such ambiguous cases? On a related note, where does `data/external` fit in your 'mental model' of the preprocessing pipeline? (Pick one) - A ``` raw ---> interim ---> processed ---+ | +---> [ analysis ] | external ----+ ``` - B ``` raw -------+ | +--> interim ---> processed ---> [ analysis ] | external --+ ```
0easy
Title: Make response short-circuiting docs clearer and more visual Body: Based on a question from our community (by @amarynets), where one is wondering whether setting `resp.complete = True` in `process_response` will stop any remaining `process_response` methods. The docs state clearly > However, any `process_response` middleware methods will still be called. The context is, however, setting `resp.complete` from `process_request` & `process_resource`, not `process_response`. I suggest rewording or improving docs to make it clear that any `process_response` middleware methods are unaffected by the `resp.complete` flag, be it set from any middleware method or inside the routed responder. Also, it would be nice to include visual diagrams like we have for middleware order, something along the lines of > mob1.process_request > &nbsp;&nbsp;mob2.process_request :arrow_right: `resp.complete = True` > &nbsp;&nbsp;&nbsp;&nbsp;~mob3.process_request~ > &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;~mob1.process_resource~ > &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;~mob2.process_resource~ > &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;~mob3.process_resource~ > &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;~`route to resource responder method`~ > &nbsp;&nbsp;&nbsp;&nbsp;mob3.process_response > &nbsp;&nbsp;mob2.process_response > mob1.process_response ```
0easy
Title: Add triangular git workflow helpers to documentation / contribution Body: On the [contributing page](http://www.scikit-yb.org/en/latest/contributing.html) on the website and in [contributing.md](https://github.com/DistrictDataLabs/yellowbrick/blob/develop/CONTRIBUTING.md) there are beginner-friendly instructions on how to make a fork of a repository. It would be useful to also include information about how to configure more than one remote repository (origin and upstream) so that contributors new to a forking workflow know how to get updates to the repo that occur after they forked the repo. ### Proposal/Issue We can use / adapt the writeup here that talks about this. http://housinginsights.org/resources/onboarding/triangular-git.html ### Background I don't know if anyone has had problems with this, but in Housing Insights I often had people confused why their local code was different when they had done `git pull`, but when it only pulls from their fork they won't get the updates that occur on the DDL upstream repo.
0easy
Title: Default value for custom input object type Body: I have a custom input object type ``` class Filters(graphene.InputObjectType): name = graphene.String() type = graphene.List(graphene.String) ``` and the query definition ``` class Query(graphene.ObjectType): search = graphene.Field( SearchResult, query=graphene.String(required=True), filters=graphene.Argument(Filters, default_value=None) ) ``` I want to be able to provide an empty default value for the `filters` argument. With the above definition, the argument is not set as optional and I'm getting this error message: ``` { "errors": [ { "message": "resolve_search() missing 1 required positional argument: 'filters'", "locations": [ { "line": 15, "column": 3 } ], "path": [ "search" ] } ], "data": { "search": null } } ``` What is the correct way to provide an empty default value for the argument of custom object type? I'm at these package versions: `graphene==2.1.3` `graphql-core==2.1` `graphql-server-core==1.1.1`
0easy
Title: More user information should be included in user mode scoreboard export Body: User mode scoreboard CSV exports are missing the user IDs as well as user emails. https://github.com/CTFd/CTFd/blob/2c32791c241cd9ee6bb3183ed2221154e70a927f/CTFd/utils/csv/__init__.py#L117-L126
0easy
Title: Clarify which methods allow keyword arguments or require a dict Body: In the README it [says][1]: > puppeteer uses an object for passing options to functions/methods. pyppeteer methods/functions accept both dictionary (python equivalent to JavaScript's objects) and keyword arguments for options. > > ### Dictionary style options (similar to puppeteer): > > ``` > browser = await launch({'headless': True}) > ``` > > ### Keyword argument style options (more pythonic, isn't it?): > > ``` > browser = await launch(headless=True) > ``` This is great and I 100% prefer the keyword argument-style options. But not all methods seem to support them, for example [`page.setViewport`][2]: ``` await page.setViewport(width=800, height=600, deviceScaleFactor=2) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-50-4d077b87a1ec> in <module> ----> 1 await page.setViewport(width=800, height=600, deviceScaleFactor=2) TypeError: setViewport() got an unexpected keyword argument 'width' ``` I think it would be best if all the methods in the pyppeteer API supported keyword arguments but, barring that, documentation on which ones do would be second best. Thanks for the great library, it's exactly what I was looking for! [1]: https://github.com/pyppeteer/pyppeteer#keyword-arguments-for-options [2]: https://pyppeteer.github.io/pyppeteer/reference.html#pyppeteer.page.Page.setViewport
0easy
Title: typed.List __repr__() Body: In Numba 0.60, I notice that the `__repr()__` for `typed.List` shows an ellipsis (as if there is more data) at the end of the list, even if the full array has been printed. Example: ``` from numba.typed import List l = List([1,2,3,4]) print(repr(l)) ``` Output: ``` ListType[int64]([1, 2, 3, 4, ...]) ``` This initially confused me as I thought there was more data in the list than what I put in.
0easy
Title: Utilize {% querystring %} template tag Body: ### Describe the problem The querystring handling pagination currently has quite complex logic in `weblate.trans.forms.SearchForm`. Most likely it can be fully eliminated by using the `{% querystring %}` template tag introduced in Django 5.1 (required by Weblate since https://github.com/WeblateOrg/weblate/pull/13636) see https://docs.djangoproject.com/en/5.1/releases/5.1/#querystring-template-tag. ### Describe the solution you would like I believe using `{% querystring %}` in the template can replace `items`, `urlencode` and `reset_offset` methods in the form. ### Describe alternatives you have considered _No response_ ### Screenshots _No response_ ### Additional context https://github.com/WeblateOrg/weblate/blob/b4c476f9ef7e38fd854169d270953abbee4c5d98/weblate/trans/forms.py#L809-L852
0easy
Title: Allow WHILE limit to be modified in listener V3 Body: Currently it is impossible to change WHILE limit settings using the listener interface- listener object is created before control is passed to attached listeners, thus modifying while body parameters related to the limit is pointless. It might be useful to allow listeners to change this- for example, an user might want to redefine default limit behaviour for WHILE loops in case it was not explicitly defined. For example, instead of having default 10k iterations as a default, no-configuration WHILE loop would abort after 10 seconds using this feature.
0easy
Title: [ENH] Coalesce function that doesn't delete columns Body: # Brief Description It seems the coalesce function must always replace all input columns with one output column. While this makes sense from the meaning of "coalesce" I can think of cases when this is not the desired outcome. For example, I have a column D that is a good default value for a few other columns and I might want to coalesce([A,D],A), coalesce([B, D], B) etc. This would break currently as D would disappear after the first coalesce In this case, I could use A.fillna(D, inplace=True) but if I have multiple default columns (e.g. I want to coalesce([A,B,D]) and coalesce([B,C,D]) then it becomes messy. Maybe there is a better way I am not aware of. I'm happy to try a PR for this but wanted to confirm that this is a reasonable thing to do. # Example API I would propose adding a delete_columns option which is defaulted to be True so existing behaviour is preserved. Or keep_columns which defaults to False. ```python def coalesce( df: pd.DataFrame, column_names: Union[str, Iterable[str], Any], new_column_name: str = None, delete_columns: bool = True ) -> pd.DataFrame: ... if delete_columns: df = df.drop(columns=column_names) ``` It would still replace an existing column if no column name is specified or the new column already exists.
0easy
Title: Setup.py uses the imp module that is deprecated Body: while trying to run `python setup.py test` I get the warning `DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses` Since python 2 is no longer supported I think we can migrate to `importlib`
0easy
Title: Refactoring: environment variables Body: Refactoring: ### First step - [x] In completion search by substring instead of search by lprefix. Solved in #5388. - [x] Fix setting variable using `xonsh -DVAR=VAL`. Solved in #5396. - [x] Group env variables by xonsh components. First step is in #5501 ### Second step Xonsh has [many environment variables](https://xon.sh/envvars.html) and part of them are marked as `XONSH_` and the rest not. I suggest to add `XONSH_` prefix to all environment variables that are related to xonsh itself i.e. rename `RAISE_SUBPROC_ERROR` to `XONSH_RAISE_SUBPROC_ERROR` etc. Pros: Clear understanding of `env`. The way to remove xonsh env variables by running `env | grep -v XONSH_ | grep -v XONTRIB_`. - [ ] We need to implemet backwards compatibility: the state of `RAISE_SUBPROC_ERROR` should update `XONSH_SUBPROC_RAISE_ERROR` values. After this the steps look like: 1. Add "Deprecated" to `RAISE_SUBPROC_ERROR` doc text. 2. Create `XONSH_SUBPROC_RAISE_ERROR`. 3. Enable sync values between these two i.e. if I set `RAISE_SUBPROC_ERROR=1` I expect to have `XONSH_SUBPROC_RAISE_ERROR=1` as well and vice versa. ## For community ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
0easy
Title: BUG: runtime error caused by `read_xxx` with relative path Body: ### Describe the bug In the interactive environment, the work dir of a local cluster is determined during initialization. If the current dir of IPython changed, loading a dataframe with a relative path may result in a FileNotFound error. ### To Reproduce ```python In [1]: import xorbits In [2]: xorbits.init() 2023-03-14 21:20:14,070 xorbits._mars.deploy.oscar.local 22259 WARNING Web service started at http://0.0.0.0:45351 In [3]: cd /path/to/my/dataset In [4]: import xorbits.pandas as pd In [5]: trips = pd.read_parquet("data.parquet") # will cause a FileNotFoundError since Xorbits is initalized under another directory In [6]: trips ``` ### Expected behavior Xorbits should better take the relative path and make it an absolute path.
0easy
Title: Add a CI lint task to check st2client's README.md Body: We need to make sure that the st2client `README.rst` file is acceptable to PyPI, since any syntax errors in it will cause the `push_st2client` task of the `st2cd.st2_finalize_release` workflow to fail. We can check the syntax using the same renderer that PyPI itself uses: ```bash # Use the same README renderer that PyPI uses to catch syntax issues in the # README.rst file # st2client uses README.rst # https://pypi.org/help/#description-content-type # https://pypi.org/project/readme-renderer # https://packaging.python.org/tutorials/packaging-projects/#description echo "Checking README.rst syntax" virtualenv venv-st2client-readme-checker . venv-st2client-readme-checker/bin/activate pip install --upgrade readme_renderer python -m readme_renderer README.rst deactivate ``` It would be nice if we could catch these errors before release, which means that we should create a step in our CI tooling to check it before any bad changes get merged.
0easy
Title: Add support for DRYAD data repository Body: [Sorry if my issues sound like a broken record! I hope more contentproviders will open up BinderHub for more communities.] ### Proposed change Support DRYAD (or [stash](https://github.com/CDL-Dryad/stash)) as a content provider (DOI + URL-based): https://datadryad.org/ Their guide to data sharing explicitly mentions code + data, see https://datadryad.org/docs/QuickstartGuideToDataSharing.pdf, so _they should be a good fit_. _The biggest problem right now_ is that I did not yet find a suitable repository (chicken and egg problem?!), and DRYAD is [not for free](https://datadryad.org/stash/publishing_charges) unless you are affiliated with specific institutions. **Anyone here who can upload a test binder to DRYAD?** Relevant API documentation: - https://datadryad.org/api/v2/docs/#/default/get_datasets__doi_ - https://datadryad.org/api/v2/docs/#/default/get_datasets__doi__download (zipped and direct file download possible, didn't see any limitation on ZIP archive size) ### Alternative options None? :fist_raised: ### Who would use this feature? DRYAD is widely known in some research communities, e.g. ecology, so any researcher publishing a dataset or workflow there would be familiar with it. ### How much effort will adding it take? 1 day, with the current examples of contentproviders as a boilerplate. ### Who can do this work? A developer familiar with the content provider implementation.
0easy
Title: Docstring of align2d and estimate_shift2d should clarify units of shift Body: Shifts are currently in pixels. Might be nice if they could also be in calibrated units if floats, but regardless should be clarified in the docstrings.
0easy
Title: [Feature] facebook/contriever support requring Body: ### Checklist - [ ] 1. If the issue you raised is not a feature but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed. - [ ] 2. Please use English, otherwise it will be closed. ### Motivation I guess facebook/contriever somehow is a popular embedding model so if SGLang supports it, will be very cool. But I guess it may be time-consuming to support a model that is not llama-structure, but it is indeed popular in RAG setting. ### Related resources _No response_
0easy
Title: Add tests for mixed dimensions Body:
0easy
Title: UnboundLocalError: local variable 'notebook' referenced before assignment Body: There might be very mighty looking error message if there are some tasks with old notebooks. ```python [2022-07-28 11:17:14,676: ERROR/MainProcess] Task apps.tasks.tasks.task_execute[d9904836-2f1a-4f38-8d6d-69f1999dc13d] raised unexpected: UnboundLocalError("local variable 'notebook' referenced before assignment") Traceback (most recent call last): File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 51, in task_execute task = Task.objects.get(pk=job_params["db_id"]) File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/django/db/models/query.py", line 435, in get raise self.model.DoesNotExist( apps.tasks.models.Task.DoesNotExist: Task matching query does not exist. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 349, in task_execute task.result = str(e) UnboundLocalError: local variable 'task' referenced before assignment During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/celery/app/trace.py", line 450, in trace_task R = retval = fun(*args, **kwargs) File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/celery/app/trace.py", line 731, in __protected_call__ return self.run(*args, **kwargs) File "/home/piotr/sandbox/bloxs_test/blox_test_env/lib/python3.8/site-packages/mercury/apps/tasks/tasks.py", line 356, in task_execute print(notebook.schedule) UnboundLocalError: local variable 'notebook' referenced before assignment ```
0easy
Title: Workflow Body: Could someone comment, what is the workflow here. If after the initial code generation, if user needs to modify or add new features to existing prompt text, next run of the command gpt-engineer <folder_name> will do what? 1. Completely remove existing code? 2. Append code with new changes? 3. Or something else? docker-compose.yml is generated but it is looking for a local.yml file? So I renamed docker-compose.yml -> local.yml and docker-compose up seems to work, but why?
0easy
Title: [CI] ci job `Test Build For Component katib-ui` is flaky Body: ### What happened? ERROR: failed to solve: process "/bin/sh -c npm config set fetch-retry-mintimeout 20000 && npm config set fetch-retry-maxtimeout 120000 && npm config get registry && npm config set registry https://registry.npmjs.org/ && npm config get https-proxy && npm config rm https-proxy && npm ci" did not complete successfully: exit code: 1 Error: buildx failed with: ERROR: failed to solve: process "/bin/sh -c npm config set fetch-retry-mintimeout 20000 && npm config set fetch-retry-maxtimeout 120000 && npm config get registry && npm config set registry https://registry.npmjs.org/ && npm config get https-proxy && npm config rm https-proxy && npm ci" did not complete successfully: exit code: 1 ref: https://github.com/kubeflow/katib/actions/runs/12830311211/job/36110037074?pr=2496 ### What did you expect to happen? the ci to pass this step ### Environment Kubernetes version: ```bash $ kubectl version ``` Katib controller version: ```bash $ kubectl get pods -n kubeflow -l katib.kubeflow.org/component=controller -o jsonpath="{.items[*].spec.containers[*].image}" ``` Katib Python SDK version: ```bash $ pip show kubeflow-katib ``` ### Impacted by this bug? Give it a 👍 We prioritize the issues with most 👍
0easy
Title: Improve test coverage Body: <!-- Please use the appropriate issue title format: BUG REPORT Bug: {Short description of bug} SUGGESTION Suggestion: {Short description of suggestion} OTHER {Question|Discussion|Whatever}: {Short description} --> ## Description Currently, the test coverage is around 88% which is okay but we can improve it. This issue focuses on adding tests covering missing branches. ## People to notify @manrajgrover
0easy
Title: Add blueprint model Body: It would be nice to use it's flexibility.
0easy
Title: ENH: implement `__array__` for index Body: ### Is your feature request related to a problem? Please describe Implement `__array__` for index. The impl for dataframe could be helpful: #282 .
0easy
Title: How to use the KDJ indicator? Body: Hello, I'm new to pandas/FreqTrade and want to use pandas-ta because this library provide the KDJ indicator. But I'm really confused how to use the KDJ in my strategy - how I can measure if the KDJ is positive or negative? Thx in advance.
0easy
Title: Remove full stop at end of viewer/layer action names Body: Related: #7231(https://github.com/napari/napari/pull/7231#discussion_r1741406517) Some viewer/layer action names end with a full stop and some do not, e.g., https://github.com/napari/napari/blob/a9e7375e339a2240f6cc6affd05eceefa2e843e9/napari/components/_viewer_key_bindings.py#L200 I propose that we remove the full stops which would make the names consistent and it would improve readability of error messages, e.g. the one in #7231 If no objections will do this end of the week.
0easy
Title: Managing multiple pipelines that import modules with the same name Body: In a layout like this: ```sh some-pipeline/ pipeline.yaml tasks.py another-pipeline/ pipeline.yaml tasks.py ``` If a user loads in Jupyter `some-pipeline/pipeline.yaml` for cell injection and there are tasks imported from tasks.py, if they then try to open a file in `another-pipeline`, `another-pipeline/tasks.py` will not be imported and the one from the other pipeline will be used. Perhaps remove the cache before processing specs? Although this will make processing slower.
0easy
Title: TextProfiler only do case sensitivity at report Body: Instead of case sensitivity being checked at profile time, we can leave sensitivity on and at report time we can then output collapse it if the options require case-insensitive.
0easy
Title: Bug: AsyncAPI 2.6.0 schema ignores messages schema overriding Body: **Describe the bug** 1) In some way you describe the message, for example: ```python @dataclass class Input: value: int ``` 2) Describe the handler with this message: ```python @broker.subscriber("in1") async def handle_msg(a: Input) -> None: ... ``` 3) Override your message: ```python @dataclass class Input: not_in_schema: str ``` 4) Describe the handler with overrided message: ```python @broker.subscriber("in2") async def please_good_schema(a: Input) -> None: ... ``` 5) Create asyncapi schema and look at `#/components/schemas` ```json "schemas": { "Input": { "properties": { "value": { "title": "Value", "type": "integer" } }, "required": [ "value" ], "title": "Input", "type": "object" } } ``` Where is my `Input.not_in_schema: str`? For example, it will be a problem if I have the same named dataclasses in different packages **How to reproduce** ```python from dataclasses import dataclass from faststream import FastStream from faststream.rabbit import RabbitBroker broker = RabbitBroker() app = FastStream(broker) @dataclass class Input: value: int @broker.subscriber("in1") async def handle_msg(a: Input) -> None: ... @dataclass class Input: not_in_schema: str @broker.subscriber("in2") async def please_good_schema(a: Input) -> None: ... ``` And/Or steps to reproduce the behavior: 1. faststream docs get main:app **Expected behavior** I expect two payload in schema: - Input with `value: int` - Input2IdkAnyNameHere with `not_in_schema: str` **Observed behavior** Message "overriding" just ignored, in the schema all handlers use only the first definition of the message **Environment** Running FastStream 0.5.15 with CPython 3.12.3 on Linux **Additional context** Same problem in the case of handlers + args names duplication: ```python @broker.subscriber("in1") async def handle_msg(a: int) -> None: ... @broker.subscriber("in2") async def handle_msg(a: str) -> None: ... ``` ```json "schemas": { "HandleMsg:Message:Payload": { "title": "HandleMsg:Message:Payload", "type": "integer" } } ```
0easy
Title: [Documentation] Add examples how to apply transforms to videos Body: We can apply the same transform to videos using `additional targets`. But no one knows about this methods => need example notebook
0easy
Title: Disable multiple dates/date types in dcc.DatePicker* Body: I'll open this here but I am aware that you're using [AirBnB's react-dates](https://github.com/airbnb/react-dates/issues) so I'll start here and then open an issue there if it's not something that can be implemented on the dash side. We'd like to disable "categories" of days from being selected. Our smallest request would be disabling weekdays or weekends, but generally supplying a list of available dates or unavailable dates to the DatePicker would be :+1:
0easy
Title: Update Docs / Examples to Use yellowbrick.datasets.load_** Body: Building off of: https://github.com/DistrictDataLabs/yellowbrick/pull/417 We should update the documentation to use this much easier interface for using our example datasets. Docs to Update: - [x] Readme.md - [x] Quick Start at http://www.scikit-yb.org/en/latest/quickstart.html - [x] Model Selection Tutorial at http://www.scikit-yb.org/en/latest/tutorial.html - [ ] Datasets in the API at http://www.scikit-yb.org/en/latest/api/datasets.html
0easy
Title: webhook secret_token is not verified Body: ### Checklist - [X] I am sure the error is coming from aiogram code - [X] I have searched in the issue tracker for similar bug reports, including closed ones ### Operating system Ubuntu 22.04.1 LTS ### Python version 3.10.6 ### aiogram version 2.22.2 ### Expected behavior When secret_token is set through calling `await db.bot.set_webhook(..., secret_token=...)` there should be a check that it matches against `X-Telegram-Bot-Api-Secret-Token` header in every webhook request. ### Current behavior Setting the secret_token gives false sense of security as it is not checked against `X-Telegram-Bot-Api-Secret-Token` header in every webhook request. ### Steps to reproduce Grep all the codebase for `X-Telegram-Bot-Api-Secret-Token`. ### Code example _No response_ ### Logs _No response_ ### Additional information _No response_
0easy
Title: url parameters are not available in `dcc.Location` Body: we're only sending the `window.location.pathname` (https://github.com/plotly/dash-core-components/blob/master/src/components/Location.react.js#L36) but we could send an object of `window.location.search` (https://stackoverflow.com/questions/5448545/how-to-retrieve-get-parameters-from-javascript) as reported in https://community.plot.ly/t/get-url-with-get-parameters/5878/1
0easy
Title: Add tests for ensemble save and load Body: Add tests for ensemble save and load. It can be done: - by using some existing learner - or by writing simple learner framework mockup
0easy
Title: Document Structures in GraphQL Types. Body: The structures PositiveInt, QuerySetList, TagList, PaginationType, BasePaginatedType and type factory PaginatedQuerySet should be documented in docs/general-usage/graphql-types.rst and how they are used by the decorators should be documented in docs/general-usage/decorators.rst. This would be a good starter issue as it will help new devs become familiar with the wagtail-grapple architecture and code base.
0easy
Title: Draw Diagrams Body: ## Description Draw Diagrams and make they available in the [New Comers Guide](https://github.com/scanapi/scanapi/wiki/Newcomers) - [ ] Class - [ ] Sequence
0easy
Title: Defect Resolution Duration metric API Body: The canonical definition is here: https://chaoss.community/?p=4727
0easy
Title: Issue Age metric API Body: The canonical definition is here: https://chaoss.community/?p=3629
0easy
Title: @root_validator will not updated to model_validator Body: ## Description I have a @root_validator without having any parameter in my project like this: ``` from pydantic import BaseModel, root_validator class Model(BaseModel): name: str @root_validator def validate_root(cls, values): return values ``` In this way, `bump-pydantic` can't detect it and will not update it to a new version and without any TODO or anything. Also, if I define my root_validator like `@root_validator()`, it will be detected and converted to `@model_validator()` without defining `mode` parameter. The only successful convert will happen if I have `@root_validator(pre=True)`. I've not checked other scenarios. If it is an accepted bug, I will be happy to work on it and make a PR ## My Environment Python 3.11.4 bump-pydantic==0.7.0
0easy