text
stringlengths
20
57.3k
labels
class label
4 classes
Title: Migrate from "yield from" to await (TypeError: object Engine can't be used in 'await' expression) Body: Hi, i replaced in my code "yield from" to "await", and received Traceback: "TypeError: object Engine can't be used in 'await' expression" ``` python async def db_psql_middleware(app, handler): async def middleware(request): db = app.get('db_psql') if not db: app['db_psql'] = db = await create_engine(app['psql_dsn'], minsize=1, maxsize=5) request.app['db_psql'] = db return (await handler(request)) return middleware async def psql_select(request): with (await request.app['db_psql']) as conn: result = await conn.execute(models.select()) ``` Traceback ``` python [2015-09-17 14:50:29 +0300] [26045] [ERROR] Error handling request Traceback (most recent call last): File "/Users/vvv/src/backend-tools/python/asyncio/venv35/lib/python3.5/site-packages/aiohttp/server.py", line 272, in start yield from self.handle_request(message, payload) File "/Users/vvv/src/backend-tools/python/asyncio/venv35/lib/python3.5/site-packages/aiohttp/web.py", line 85, in handle_request resp = yield from handler(request) File "/Users/vvv/src/backend-tools/python/asyncio/app.py", line 39, in middleware return (await handler(request)) File "/Users/vvv/src/backend-tools/python/asyncio/app.py", line 46, in psql_select with (await request.app['db_psql']) as conn: TypeError: object Engine can't be used in 'await' expression ```
1medium
Title: Proposal: Support for Failure-topic forwarding Body: Failure handling is an area where there is limited consensus within the Kafka community. One option for Faust would be adding support for failure forwarding in the same pattern as sinks. The API might look like: ```python # in faust.models.record # new Faust Record type specifically for error handling class FailureEventRecord(Record): errormsg: str exception: AgentException failed_record: Record # in app.py topic = app.topic('my.event', value_type=MyRecord) failure_topic = app.topic('my.event.failed', value_type=faust.FailureEventRecord) @app.agent(topic, sinks=[sink1, sink2], failures=[failed_record_operation]) async def my_exception_agent(records): with record in records: raising_operation(record) ``` ## Checklist - [x] I have included information about relevant versions - [x] I have verified that the issue persists when using the `master` branch of Faust. ## Steps to reproduce Documentations says: > "Crashing the instance to require human intervention is certainly a choice, but far from ideal considering how common mistakes in code or unexpected exceptions are. It may be better to log the error and have ops replay and reprocess the stream on notification." ## Expected behavior Any events which fail stream processing in an agent will be wrapped in a `FailureEventRecord` and delivered to the topic or topics specified in the `failures` parameter to `app.agent()`. # Versions * Python 3.6 * Faust 1.7.0 * MacOS 10.14.5 * Kafka ? * RocksDB version (if applicable)
2hard
Title: faster-rcnn performance Body: hi, wuyuxin Thank you for sharing this wonderful project! Here I have some performance problems of faster-rcnn. In coco dataset, experiments results show GN has better performance. R50-FPN | 38.9;35.4 R50-FPN-GN | 40.4;36.3 However, in my dataset(ocr icdar2017), GN performance is poor than FreezeBN. Scratch performance is also poor than pre-trainend which is contrary to your conclusion. Here are the results on my dataset. pre-train, resnet101, FreezeBN, freeze_AT=2, f-score:0.7742 pre-train, resnet101, GN, freeze_AT=2, f-score:0.7605 no pre-train(From Scratch), resnet101, GN, freeze_AT=0, f-score:0.7106 Hope for your suggestion.
1medium
Title: Docs breadcrumbs usually start with 'Intro' Body: e.g. 'Intro > Legal'. This doesn't make sense. ![Screenshot 2024-05-30 at 14 48 28](https://github.com/pydantic/logfire/assets/3627481/ce155de0-36bc-4956-bead-728661a674ad)
0easy
Title: linux agent disconnect after seconds Body: **Server Info :** - OS: Ubuntu 20.04 - Browser: Chrome - RMM Version (as shown in top left of web UI): 0.15.3 **Installation Method:** - [x] Standard **Agent Info (please complete the following information):** - Agent version (as shown in the 'Summary' tab of the agent from web UI): 2.4.2 - Agent OS: Ubuntu 20.04 **Describe the bug** when i install linux agent app on my linux vps, it connect to RMM panel and show in this panel but after some seconds the vps will disconnect and i cant do anything with RMM on that VPS how can i get some logs for debug the problem? or what shuld i do for fix the issue?
1medium
Title: Prediction widget shows no probabilities Body: **Describe the bug** The Prediction widget shows no probabilities for the predicted classes. **To Reproduce** **Expected behavior** See probabilities for predicted classes in prediction Widget. **Orange version:** 3.32 **Screenshots** ![grafik](https://user-images.githubusercontent.com/111502928/187162897-d7e4ea55-dc51-4b82-a6af-c548427f7887.png) **Operating system:** Windows 11 **Additional context**
1medium
Title: Wagtail choosers ignore Django's ForeignKey.limit_choices_to Body: ### Issue Summary If you set up a callable for a `ForeignKey.limit_choices_to`, It will get called by Django during form construction, but Wagtail's chooser system will never call that function, and thus not limit the choices presented in the chooser. This presumably also affects the other forms of `limit_choices_to` (see [here](https://docs.djangoproject.com/en/5.1/ref/models/fields/#django.db.models.ForeignKey.limit_choices_to)), but a callable is the easiest one to use as a robust example. ### Steps to Reproduce 1. Build a basic bakerydemo, then apply the changes from here: https://github.com/wagtail/bakerydemo/compare/main...coredumperror:bakerydemo:limit_choices_to 2. Open a BreadPage editor (which will cause `"limit_choices_to_bread"` to be printed to the console). 3. Choose a BreadType (which will _not_ cause that print statement to fire). 4. Every bread type is displayed in the chooser, instead of just ones with "bread" in their title. ### Technical details Python version: 3.12 Django version: 5.1 Wagtail version: 6.4.0 ### Working on this Shouldn't this Django feature be supported by Wagtail's chooser system? It's quite surprising that my project's existing code, which worked fine when I wrote it years ago, has now stopped limiting the available options because Wagtail's new(ish) chooser system apparently just ignores it.
1medium
Title: add benchmark vs mypyc ? Body:
1medium
Title: Question about code of `vit_for_small_dataset.py` Body: Hi, Thanks for the outstanding efforts to bring ViT to the framework of PyTorch, which is meaningful for me to learn about it! However, when I review the code in `vit_for_small_dataset.py`, in the `SPT`, I find code in the line `86` to be weird. ```python patch_dim = patch_size * patch_size * 5 * channels ``` I fail to understand the meaning of `5` here. Is it a typo? Based on the context here, it seems that `patch_size * patch_size * channels` is more appropriate. Thanks for your attention!
1medium
Title: Bug in functional model Body: I think there seems a bug in functional model. Case-1: When inputs=outputs for the model construction but training different output shape: Training success ``` import keras from keras import layers import numpy as np input_1 = layers.Input(shape=(3,)) input_2 = layers.Input(shape=(5,)) model_1 = keras.models.Model([input_1, input_2], [input_1, input_2]) print(model_1.summary()) model_1.compile(optimizer='adam',metrics=['accuracy','accuracy'],loss=['mse']) #Notice I am passing different output size for training but still training happens model_1.fit([np.random.normal(size=(10,3)),np.random.normal(size=(10,5))], [np.random.normal(size=(10,1)),np.random.normal(size=(10,2))]) print('Training completed') ``` Case 2: Same as Case-1 but different behavior with different mismatched output shapes(than case-1) for training: Error during loss calculation. But I expect Error during graph execution itself. ``` #With diffrent output shapes than model constructed its raising error while calculating the loss. #Instead it should have raised shape mismatch error during graph execution. model_1.fit([np.random.normal(size=(10,3)),np.random.normal(size=(10,5))], [np.random.normal(size=(10,2)),np.random.normal(size=(10,4))]) ``` Case 3: With Unconnected inputs and outputs ``` input_1 = layers.Input(shape=(3,)) input_2 = layers.Input(shape=(5,)) input_3 = layers.Input(shape=(1,)) input_4 = layers.Input(shape=(2,)) model_2 = keras.models.Model([input_1, input_2], [input_3, input_4]) model_2.compile(optimizer='adam',metrics=['accuracy','accuracy'],loss=['mse']) #Passing correct input and ouputs fails because these are not connected. model_2.fit([np.random.normal(size=(10,3)),np.random.normal(size=(10,5))], [np.random.normal(size=(10,1)),np.random.normal(size=(10,2))]) ``` Got error below which is correct but it is not useful for end users. Instead it should have raised error during graph construction. ``` 177 output_tensors = [] 178 for x in self.outputs: --> 179 output_tensors.append(tensor_dict[id(x)]) 180 181 return tree.pack_sequence_as(self._outputs_struct, output_tensors) KeyError: "Exception encountered when calling Functional.call().\n\n\x1b[1m139941182292272\x1b[0m\n\nArguments received by Functional.call():\n • inputs=('tf.Tensor(shape=(None, 3), dtype=float32)', 'tf.Tensor(shape=(None, 5), dtype=float32)')\n • training=True\n • mask=('None', 'None')" ``` I tried to fix an issue similar to case-3 by raising Error during graph build itself in PR #20705 where I noticed this issue related to case1(From failed Test case). Please refer the [gist](https://colab.research.google.com/gist/Surya2k1/88040ebc171b2154627fe54a3560d4b1/functional_model_bug.ipynb).
2hard
Title: ERNIE英文预训练模型 Body: ERNIE有公开的英文预训练模型吗
3misc
Title: Adding file.peek() Body: Hello! I noticed that in your README.md there is no mention of your aiofiles supporting [peek](https://docs.python.org/3/library/io.html#io.BufferedReader.peek). However in your code there are some tests for it. When attempting to use peek in my code, it did not seem to work as desired. Could you please add this functionality? Many Thanks
2hard
Title: TGI on NVIDIA GH200 (Arm64) Body: ### Feature request Would it be possible to build/publish an arm64 container image for the text-generation-inference? I would like to be able to run it on a NVIDIA GH200 which is an arm64-based system. Thanks, Jonathan ### Motivation Current images won't run on arm64. ### Your contribution I've tried to build the image myself, but I haven't been able to get it to build successfully.
2hard
Title: Dataset on Hub re-downloads every time? Body: ### Describe the bug Hi, I have a dataset on the hub [here](https://huggingface.co/datasets/manestay/borderlines). It has 1k+ downloads, which I sure is mostly just me and my colleagues working with it. It should have far fewer, since I'm using the same machine with a properly set up HF_HOME variable. However, whenever I run the below function `load_borderlines_hf`, it downloads the entire dataset from the hub and then does the other logic: https://github.com/manestay/borderlines/blob/4e161f444661e2ebfe643f3fe149d9258d63a57d/run_gpt/lib.py#L80 Let me know what I'm doing wrong here, or if it's a bug with the `datasets` library itself. On the hub I have my data stored in CSVs, but several columns are lists, so that's why I have the code to map splitting on `;`. I looked into dataset loading scripts, but it seemed difficult to set up. I have verified that other `datasets` and `models` on my system are using the cache properly (e.g. I have a 13B parameter model and large datasets, but those are cached and don't redownload). __EDIT: __ as pointed out in the discussion below, it may be the `map()` calls that aren't being cached properly. Supposing the `load_dataset()` retrieve from the cache, then it should be the case that the `map()` calls also retrieve from the cached output. But the `map()` commands re-execute sometimes. ### Steps to reproduce the bug 1. Copy and paste the function from [here](https://github.com/manestay/borderlines/blob/4e161f444661e2ebfe643f3fe149d9258d63a57d/run_gpt/lib.py#L80) (lines 80-100) 2. Run it in Python `load_borderlines_hf(None)` 3. It completes successfully, downloading from HF hub, then doing the mapping logic etc. 4. If you run it again after some time, it will re-download, ignoring the cache ### Expected behavior Re-running the code, which calls `datasets.load_dataset('manestay/borderlines', 'territories')`, should use the cached version ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-5.14.21-150500.55.7-default-x86_64-with-glibc2.31 - Python version: 3.10.13 - `huggingface_hub` version: 0.20.3 - PyArrow version: 15.0.0 - Pandas version: 1.5.3 - `fsspec` version: 2023.10.0
1medium
Title: OSError: When resolving region for bucket 'ursa-labs-taxi-data': AWS Error [code 99]: curlCode: 35, SSL connect error Body: ### Describe the bug When trying to read data from the example bucket (and any other bucket) the error "OSError: When resolving region for bucket 'ursa-labs-taxi-data': AWS Error [code 99]: curlCode: 35, SSL connect error" is displayed although aws environment variables are set and configured as well as the proxy ### How to Reproduce ``` import awswranger as wr df = wr.s3.read_parquet(path="s3://ursa-labs-taxi-data/2017/") ``` ### Expected behavior Read data from s3 into the dataframe ### Your project _No response_ ### Screenshots _No response_ ### OS Red Hat Enterprise Linux release 8.8 (Ootpa) ### Python version 3.8.16 ### AWS SDK for pandas version 3.4.0 ### Additional context Through the aws cli services are reachable
1medium
Title: pmtiles tooltip + layers control is broken with folium > 0.14.0 Body: <!-- Please search existing issues to avoid creating duplicates. --> ### Environment Information - leafmap version: 0.29.5 - Python version: 3.9 - Operating System: Windows/Linux ### Description When PMTilesLayer tooltip is enabled, it leads to console error and prevents layers control from loading. Layers control is able to load when tooltip=False This can be seen on the live site as well: https://leafmap.org/notebooks/82_pmtiles/ ### What I Did Followed the example in 82_pmtiles.ipynb and found out that it works properly on one notebook env but not the other. Realized that layers control is missing when I tried to add other layers to the map. Error in browser console suggests that it may be related to folium/jinja templates: ``` // below line is not present when using folium 0.14.0 macro_element_51d29068fa44734e98b3aa01b7f1db45.addTo(pm_tiles_vector_ff278fae2ca890eb72792f1dc892823e); ``` So the current workaround is to downgrade to folium==0.14.0 folium-pmtiles is affected as well (using this example: https://github.com/jtmiclat/folium-pmtiles/blob/master/example/pmtiles_vector_maplibre.ipynb) Confirmed to be related to this change in folium: https://github.com/python-visualization/folium/pull/1690#issuecomment-1377180410
1medium
Title: Replace ApolloTracing with simple tracing extension Body: Instead of `ApolloTracing` (which is deprecated) we could implement a simple extension that adds `extensions: {traces: []}` object where each trace would be a separate `{path: time: }`. I don't know how useful this would be to folk, but maybe it would have some utility as simple example extension?
1medium
Title: bug:main、dev code run main.py has error:lzma.LZMAError: Corrupt input data Body: I use the latest code, in main,dev;i find use api in api/main.py, the web result is error, no voice and server has:lzma.LZMAError: Corrupt input data 1.i try check model,code and pip package,it's run,but generate has no voice and is noise [audio_files (19).zip](https://github.com/user-attachments/files/17505105/audio_files.19.zip) 2. i set use_vllm, it's work, but generate has noise and other voice,like this: [audio_files (20).zip](https://github.com/user-attachments/files/17505111/audio_files.20.zip)
1medium
Title: CUDA Body: ### Your question CUDA error: no kernel image is available for execution on the device CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1 Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. ### Logs ```powershell ``` ### Other _No response_
2hard
Title: get Body: ## 🐛 Bug Description <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: 1. 1. 1. ## Expected Behavior <!-- A clear and concise description of what you expected to happen. --> ## Screenshot <!-- A screenshot of the error message or anything shouldn't appear--> ## Environment **Note**: User could run `cd scripts && python collect_info.py all` under project directory to get system information and paste them here directly. - Qlib version: - Python version: - OS (`Windows`, `Linux`, `MacOS`): - Commit number (optional, please provide it if you are using the dev version): ## Additional Notes <!-- Add any other information about the problem here. -->
1medium
Title: Questions about using the known exogenous variables to conduct forecasted values Body: Hi developers, I have a little confusion about **using the known exogenous variables** to conduct forecasted values. Firstly, I use the **multi-series function to "predict"** all the exogenous variables and the target output Y simultaneously. After that, I would like to use the 'predicted values' of the exogenous variables to conduct the direct/recursive forecasting for the target output Y, and I refer the related documents such as the example of [weather exogenous](https://www.cienciadedatos.net/documentos/py39-forecasting-time-series-with-skforecast-xgboost-lightgbm-catboost.html) variables. However, I have the confusion about dealing with the known future values in an appropriate place because they are the future known values and the format are "not consistent" with the known training data. How can I "combine" them together during using the skforecast framework?
1medium
Title: a way to make stream-bytes produce multiple http chunks Body: Currently (unless I'm missing something) the chunk_size parameter on stream-bytes controls the block size used to write the data to the socket. But the entire data is still sent as one HTTP chunk i.e. there is only one total size followed by the data. Would it be reasonable to have chunk_size control the size of the HTTP chunks? Or if not, maybe another parameter to control that? I don't know Python so I'm not sure how feasible that is. Thanks for a great tool.
1medium
Title: Keras should support bitwise ops Body: Numpy has bitwise ops https://numpy.org/doc/stable/reference/routines.bitwise.html TensorFlow has bitwise ops https://www.tensorflow.org/api_docs/python/tf/bitwise Jax has bitwise ops https://jax.readthedocs.io/en/latest/_autosummary/jax.numpy.bitwise_and.html PyTorch has bitwise ops https://pytorch.org/docs/stable/generated/torch.bitwise_and.html So it seems natural for Keras to support bitwise ops as well.
1medium
Title: 登陆报错:Remote end closed connection without response Body: TypeError: getresponse() got an unexpected keyword argument 'buffering' http.client.RemoteDisconnected: Remote end closed connection without response requests.packages.urllib3.exceptions.ProtocolError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response',)) requests.exceptions.ConnectionError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response',))
1medium
Title: [Bug/Feature]: long-short.py - Qty should be calculated relative to price of stock Body: ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior When opening new positions, the long-short.py opens all the positioning with approximately the same qty in terms of shares (ie: with 100k capital, each position is 160 shares). This does not make sense, the position should be calculated based on the price of the stock. For example, the code tries to open a long position for AMZN for 160 shares (2.880$/each) and this goes way above the available liquidity. Likewise, it opens a position of 160 shares for GM (44.85$/each). I've tried to fix the code but I'm not sure about all the references, so it'd be cool if someone from Alpaca could fix this. ### Expected Behavior _No response_ ### Steps To Reproduce _No response_ ### Anything else? _No response_
1medium
Title: by_sound method failing Body: **Describe the bug** by_sound method returns an error **The buggy code** ``` TikTokApi.get_instance(proxy=sys.argv[3]).by_sound(sys.argv[1], int(sys.argv[2])) ``` **Expected behavior** by_sound method returns an error **Error Trace (if any)** ``` raise EmptyResponseError(\nTikTokApi.exceptions.EmptyResponseError: Empty response from Tiktok to https://m.tiktok.com/api/music/item_list/?aid=1988&app_name=tiktok_web&device_platform=web&referer=&root_referer=&user_agent=Mozilla%252F5.0%2B%28iPhone%253B%2BCPU%2BiPhone%2BOS%2B12_2%2Blike%2BMac%2BOS%2BX%29%2BAppleWebKit%252F605.1.15%2B%28KHTML%2C%2Blike%2BGecko%29%2BVersion%252F13.0%2BMobile%252F15E148%2BSafari%252F604.1&cookie_enabled=true&screen_width=1858&screen_height=1430&browser_language=&browser_platform=&browser_name=&browser_version=&browser_online=true&ac=4g&timezone_name=&appId=1233&appType=m&isAndroid=False&isMobile=False&isIOS=False&OS=windows&secUid=&musicID=6763054442704145158&count=35&cursor=0&shareUid=&language=en&verifyFp=verify_khr3jabg_V7ucdslq_Vrw9_4KPb_AJ1b_Ks706M8zIJTq&did=4792150057109208114&_signature=_02B4Z6wo00f01X4-6mwAAIBDLD1T3YFqwoV-P-7AAD.80b\n ``` **Desktop (please complete the following information):** - OS: MacOS - TikTokApi Version 3.9.5
1medium
Title: [Overkiz] - Unsupported value CyclicSwingingGateOpener Body: ### The problem I have in my Tahoma an association with Netatmo station and legrand drivia switch (https://www.legrand.fr/pro/catalogue/pack-de-demarrage-drivia-with-netatmo-pour-installation-connectee-1-module-control-1-contacteur-connecte). In the system log, I see an error related my Netatmo on Overkiz. ### What version of Home Assistant Core has the issue? core-2025.3.3 ### What was the last working version of Home Assistant Core? _No response_ ### What type of installation are you running? Home Assistant OS ### Integration causing the issue Overkiz ### Link to integration documentation on our website _No response_ ### Diagnostics information Enregistreur: pyoverkiz.enums.ui Source: components/overkiz/__init__.py:87 S'est produit pour la première fois: 14 mars 2025 à 19:53:28 (3 occurrences) Dernier enregistrement: 14 mars 2025 à 19:53:28 Unsupported value CyclicSwingingGateOpener has been returned for <enum 'UIWidget'> Unsupported value NetatmoGateway has been returned for <enum 'UIWidget'> ### Example YAML snippet ```yaml ``` ### Anything in the logs that might be useful for us? ```txt ``` ### Additional information _No response_
1medium
Title: Error solving the challenge. Body: ### Have you checked our README? - [X] I have checked the README ### Have you followed our Troubleshooting? - [X] I have followed your Troubleshooting ### Is there already an issue for your problem? - [X] I have checked older issues, open and closed ### Have you checked the discussions? - [X] I have read the Discussions ### Environment ```markdown - FlareSolverr version:3.3.21 - Last working FlareSolverr version:none - Operating system:win10 - Are you using Docker: [yes/no]no - FlareSolverr User-Agent (see log traces or / endpoint)Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36 - Are you using a VPN: [yes/no]no - Are you using a Proxy: [yes/no]no - Are you using Captcha Solver: [yes/no]no - If using captcha solver, which one:no - URL to test this issue:https://unitedshop.ws/ ``` ### Description check pics for more info ### Logged Error Messages ```text 500 internal server error (Failed to recieve the response from the HTTP-server 'localhost'.)in silverbullet ``` ### Screenshots ![Screenshot_2](https://github.com/user-attachments/assets/0152a2e6-1062-47c6-9d3b-4f11ef4e823a) ![Screenshot_3](https://github.com/user-attachments/assets/6eea86d5-b49f-478e-bc4d-891642770178)
1medium
Title: Encode the video output in HEVC instead h264 ? Body: Hello there , how i can make RVM encode the video output in HEVC instead h264 ? And is it possible to get the sound from the original vid too ? Thank you in advance !
1medium
Title: make nlp work for individual dictionaries Body: right now it only works with dataframes. we'll probably want to move this whole block into fit: ``` col_names = self.text_columns[key].get_feature_names() # Make weird characters play nice, or just ignore them :) for idx, word in enumerate(col_names): try: col_names[idx] = str(word) except: col_names[idx] = 'non_ascii_word_' + str(idx) col_names = ['nlp_' + key + '_' + str(word) for word in col_names] ```
1medium
Title: nltk.lm.api entropy formula source? Body: Hi! I've been experimenting with training and testing a standard trigram language model on my own dataset. Upon investigating the `entropy` method of the LM class, I was a bit confused. The [docs](https://www.nltk.org/api/nltk.lm.api.html#nltk.lm.api.LanguageModel.entropy) only mention a very brief description: "Calculate cross-entropy of model for given evaluation text." It seems to me that this entropy measure just **averages the ngram negative log probability** (so trigram in my case) and **makes it positive** by multiplying by -1: ```python def _mean(items): """Return average (aka mean) for sequence of items.""" return sum(items) / len(items) def entropy(self, text_ngrams): """Calculate cross-entropy of model for given evaluation text. :param Iterable(tuple(str)) text_ngrams: A sequence of ngram tuples. :rtype: float """ return -1 * _mean( [self.logscore(ngram[-1], ngram[:-1]) for ngram in text_ngrams] ) ``` **So my general question is: Where does this formula for entropy stem from? Is there any paper referencing the method?** I'm just a bit stuck with the different versions of entropy that exist and I don't know which one is used here (and therefore I don't know how to interpret it correctly). ### Other formulas of entropy/perplexity As a formula I know the [Shannon entropy](https://en.wikipedia.org/wiki/Entropy_(information_theory)) which in Python would be: ```python def shannon_entropy(self, text_ngrams): return -1 * sum( [self.score(ngram[-1], ngram[:-1]) * self.logscore(ngram[-1], ngram[:-1]) for ngram in text_ngrams] ) ``` And there's also the perplexity formula of [Jurafsky](https://www.youtube.com/watch?v=NCyCkgMLRiY), which returns a different score than `lm.perplexity` (which is 2**entropy): ```python from numpy import prod def jurafsky_perplexity(self, text_ngrams): problist = [self.score(ngram[-1], ngram[:-1]) for ngram in text_ngrams] return pow(prod(problist), -1/len(problist) ``` Thanks! PS: My apologies if this issue lacks some information - I don't often open issues on Github 😅 Let me know if I need to update something! UPDATE: my Python implementation of the Jurafsky perplexity was totally wrong as I had quickly written it. I've updated it to reflect the actual scores from the web lecture.
1medium
Title: Database encryption Body: Add encryption at rest to the database so the server's owner can't access user's data. Account password should be the master key to decrypt the database.
2hard
Title: Docker compose stuck on waiting for my sql Body: I cloned the repo and did docker-compose build. It ran successfully. I then did docker-compose up. This is stuck on Waiting for mysql+pymysql... ``` docker-compose up Starting ctfd_db_1 ... done Starting ctfd_cache_1 ... done Starting ctfd_ctfd_1 ... done Starting ctfd_nginx_1 ... done Attaching to ctfd_cache_1, ctfd_db_1, ctfd_ctfd_1, ctfd_nginx_1 db_1 | 2021-07-24 17:57:00+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 1:10.4.12+maria~bionic started. cache_1 | 1:C 24 Jul 17:56:59.849 # oO0OoO0OoO0Oo Redis is starting oO0OoO0OoO0Oo cache_1 | 1:C 24 Jul 17:56:59.849 # Redis version=4.0.14, bits=64, commit=00000000, modified=0, pid=1, just started cache_1 | 1:C 24 Jul 17:56:59.849 # Warning: no config file specified, using the default config. In order to specify a config file use redis-server /path/to/redis.conf db_1 | 2021-07-24 17:57:01+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql' cache_1 | 1:M 24 Jul 17:56:59.903 * Running mode=standalone, port=6379. cache_1 | 1:M 24 Jul 17:56:59.903 # Server initialized cache_1 | 1:M 24 Jul 17:56:59.903 # WARNING overcommit_memory is set to 0! Background save may fail under low memory condition. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. cache_1 | 1:M 24 Jul 17:56:59.903 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled. db_1 | 2021-07-24 17:57:01+00:00 [Note] [Entrypoint]: Entrypoint script for MySQL Server 1:10.4.12+maria~bionic started. cache_1 | 1:M 24 Jul 17:56:59.903 * DB loaded from disk: 0.000 seconds cache_1 | 1:M 24 Jul 17:56:59.903 * Ready to accept connections db_1 | 2021-07-24 17:57:02 0 [Note] mysqld (mysqld 10.4.12-MariaDB-1:10.4.12+maria~bionic) starting as process 1 ... db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Using Linux native AIO db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Mutexes and rw_locks use GCC atomic builtins db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Uses event mutexes db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Compressed tables use zlib 1.2.11 db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Number of pools: 1 db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Using SSE2 crc32 instructions db_1 | 2021-07-24 17:57:02 0 [Note] mysqld: O_TMPFILE is not supported on /tmp (disabling future attempts) db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Initializing buffer pool, total size = 256M, instances = 1, chunk size = 128M db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: Completed initialization of buffer pool db_1 | 2021-07-24 17:57:02 0 [Note] InnoDB: If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: 128 out of 128 rollback segments are active. db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: Creating shared tablespace for temporary tables db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: File './ibtmp1' size is now 12 MB. db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: Waiting for purge to start db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: 10.4.12 started; log sequence number 61164; transaction id 21 db_1 | 2021-07-24 17:57:03 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool db_1 | 2021-07-24 17:57:04 0 [Note] Server socket created on IP: '::'. db_1 | 2021-07-24 17:57:04 0 [Warning] 'user' entry 'root@c6e0dcba624f' ignored in --skip-name-resolve mode. db_1 | 2021-07-24 17:57:04 0 [Warning] 'user' entry '@c6e0dcba624f' ignored in --skip-name-resolve mode. db_1 | 2021-07-24 17:57:04 0 [Warning] 'proxies_priv' entry '@% root@c6e0dcba624f' ignored in --skip-name-resolve mode. db_1 | 2021-07-24 17:57:04 0 [Note] InnoDB: Buffer pool(s) load completed at 210724 17:57:04 db_1 | 2021-07-24 17:57:04 0 [Note] mysqld: ready for connections. db_1 | Version: '10.4.12-MariaDB-1:10.4.12+maria~bionic' socket: '/var/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution ctfd_1 | Waiting for mysql+pymysql://ctfd:ctfd@db to be ready ```
1medium
Title: [Feature request] Implementation for FreeVC Body: <!-- Welcome to the 🐸TTS project! We are excited to see your interest and appreciate your support! ---> **🚀 Feature Description** Original implementation: https://github.com/OlaWod/FreeVC We can use it in combination with any 🐸TTS model and make it speak with any voice aka faking voice cloning. <!--A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> **Additional context** I'll initially implement only inference with the pre-trained models and later we can implement the training. It should be easy as it a quite the same as the VITS. <!-- Add any other context or screenshots about the feature request here. -->
1medium
Title: Why we don't use certifi root certificates by default? Body: Hey there, Today my colleauge step into the problem with ceritificate that time to time appears with other people who use your client. You can read about this problem more here: https://stackoverflow.com/questions/59808346/python-3-slack-client-ssl-sslcertverificationerror As I see the problem that people have some problem with root certificates on their machine. Python always had a problem with root certificatates and to solve this problem people started to use the project certifi. For example you can find usage of this project in the kinda popular library `requests` https://github.com/psf/requests/blob/main/src/requests/utils.py#L63 https://github.com/psf/requests/blob/main/src/requests/adapters.py#L294 Do you have any reason to don't use library certifi as requests library uses? I think it would simplify of using your library everywhere.
1medium
Title: Sliding window (mix of rolling and coarsen) Body: ### Is your feature request related to a problem? I often have functions to apply to sliding windows, such as a fft computation, but both coarsen and rolling do not fit. For example, generating a spectrogram with overlap between windows is complicated using xarray. Of course, one could use [scipy STFT](https://docs.scipy.org/doc/scipy/reference/generated/scipy.signal.ShortTimeFFT.html) if we consider the example of the spectrogram, but one still needs to manage the coordinates manually. ### Describe the solution you'd like A new window function that combines rolling and coarsen. The new window would have the current dim parameter of rolling and coarsen split into two: - window_size: the size of the window (equivalent to the current value of the dim parameter) - hop: how much do we shift in between each window (see the hop parameter of [scipy STFT](https://docs.scipy.org/doc/scipy/reference/generated/scipy.signal.ShortTimeFFT.html) for example). - perhaps we could add a window function to apply, but this is not necessary as this can be done without. This unifies rolling and coarsen as rolling is simply hop=1 and coarsen is hop=window_size. As for the implementation of this datastructure, I suspect we could use [as_strided](https://numpy.org/devdocs/reference/generated/numpy.lib.stride_tricks.as_strided.html) for very efficient implementation of construct. ### Describe alternatives you've considered - Using rolling + sel to only pick the windows with the correct hop, but this seems extremely inefficient - Handling the coordinates manually... ### Additional context _No response_
1medium
Title: Can't connect to mongo docker Body: okey, spend a day, but still can't make this work > .env ``` # Mongo DB MONGO_INITDB_ROOT_USERNAME=admin-user MONGO_INITDB_ROOT_PASSWORD=admin-password MONGO_INITDB_DATABASE=container ``` > docker-compose.yml ``` mongo-db: image: mongo:4.2.3 env_file: - .env ports: - 27017:27107 volumes: - ./bin/mongo-init.js:/docker-entrypoint-initdb.d/mongo-init.js:ro # restart: always add this production api: build: context: ./backend dockerfile: Dockerfile command: uvicorn app.main:app --host 0.0.0.0 --port 8006 --reload volumes: - ./backend:/app env_file: - .env depends_on: - mongo-db ports: - "8006:8006" ``` > mongo-init.js ``` db.auth('admin-user', 'admin-password') db = db.getSiblingDB('container') db.createUser({ user: 'test-user', pwd: 'test-password', roles: [ { role: 'root', db: 'admin', }, ], }); ``` using this as example - https://frankie567.github.io/fastapi-users/configuration/full_example/ changed few lines: `DATABASE_URL = "mongodb://test-user:test-password@mongo-db/container"` ``` client = motor.motor_asyncio.AsyncIOMotorClient(DATABASE_URL) db = client["container"] collection = db["users"] user_db = MongoDBUserDatabase(UserDB, collection) ``` so think Fatapi is linked to Mongo container ![image](https://user-images.githubusercontent.com/17315578/80109217-e78acf80-8585-11ea-9008-2dcfb4de3667.png) but when i try to register user using Fastapi docks section, i get "Internal server error" what I left unfinished?
1medium
Title: PROVIDE_AUTOMATIC_OPTIONS causes KeyError if not set Body: https://github.com/pallets/flask/blob/bc098406af9537aacc436cb2ea777fbc9ff4c5aa/src/flask/sansio/app.py#L641C12-L641C86 Simply changing this to : `self.config.get("PROVIDE_AUTOMATIC_OPTIONS", False)` should resolve the problem. This change now released is causing upstream trouble in other packages such as Quart: https://github.com/pallets/quart/issues/371
0easy
Title: `remove_columns` method used with a streaming enable dataset mode produces a LibsndfileError on multichannel audio Body: ### Describe the bug When loading a HF dataset in streaming mode and removing some columns, it is impossible to load a sample if the audio contains more than one channel. I have the impression that the time axis and channels are swapped or concatenated. ### Steps to reproduce the bug Minimal error code: ```python from datasets import load_dataset dataset_name = "zinc75/Vibravox_dummy" config_name = "BWE_Larynx_microphone" # if we use "ASR_Larynx_microphone" subset which is a monochannel audio, no error is thrown. dataset = load_dataset( path=dataset_name, name=config_name, split="train", streaming=True ) dataset = dataset.remove_columns(["sensor_id"]) # dataset = dataset.map(lambda x:x, remove_columns=["sensor_id"]) # The commented version does not produce an error, but loses the dataset features. sample = next(iter(dataset)) ``` Error: ``` Traceback (most recent call last): File "/home/julien/Bureau/github/vibravox/tmp.py", line 15, in <module> sample = next(iter(dataset)) ^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/iterable_dataset.py", line 1392, in __iter__ example = _apply_feature_types_on_example( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/iterable_dataset.py", line 1080, in _apply_feature_types_on_example encoded_example = features.encode_example(example) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/features/features.py", line 1889, in encode_example return encode_nested_example(self, example) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/features/features.py", line 1244, in encode_nested_example {k: encode_nested_example(schema[k], obj.get(k), level=level + 1) for k in schema} File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/features/features.py", line 1244, in <dictcomp> {k: encode_nested_example(schema[k], obj.get(k), level=level + 1) for k in schema} ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/features/features.py", line 1300, in encode_nested_example return schema.encode_example(obj) if obj is not None else None ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/datasets/features/audio.py", line 98, in encode_example sf.write(buffer, value["array"], value["sampling_rate"], format="wav") File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/soundfile.py", line 343, in write with SoundFile(file, 'w', samplerate, channels, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/soundfile.py", line 658, in __init__ self._file = self._open(file, mode_int, closefd) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/julien/.pyenv/versions/vibravox/lib/python3.11/site-packages/soundfile.py", line 1216, in _open raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name)) soundfile.LibsndfileError: Error opening <_io.BytesIO object at 0x7fd795d24680>: Format not recognised. Process finished with exit code 1 ``` ### Expected behavior I would expect this code to run without error. ### Environment info - `datasets` version: 2.18.0 - Platform: Linux-6.5.0-21-generic-x86_64-with-glibc2.35 - Python version: 3.11.0 - `huggingface_hub` version: 0.21.3 - PyArrow version: 15.0.0 - Pandas version: 2.2.1 - `fsspec` version: 2023.10.0
1medium
Title: Distributed Training with 4 machines on Translation Task Body: Hello , I want to do distributed training using four machines ,each one has 8 1080ti GPUs on En-Zh translation task, and the t2t-version is 1.6.5. I have seen the other similar issues , and the distributed_training.md ,but I still have some confusion. If the four machines named M1, M2 ,M3 ,M4 , the first step is to make the TF-Config. What I want is to make 4 machines to train together , thus I am not sure how to define the master and ps. And the second question is if four machines can do synchronous training ? Hope someone can give me some advice, thanks
2hard
Title: Can not install Flask-Restx in ARM Body: ## Description: I am trying to create a docker image in a RaspberryPi which runs `aarch64`, but when the Docker image starts installing the Flask-Restx package, it fails. !!**This was working correctly 4 months ago because that was the day of my last build.** ## **Environment** RaspberryPi 4 ```commandline $ uname -a Linux RPi 5.10.10-1-ARCH #1 SMP Sat Jan 23 16:26:04 MST 2021 aarch64 GNU/Linux $ docker --version Docker version 20.10.2, build 2291f610ae ``` ## **requirements.txt** ```text flask-restx==0.4.0 ``` ## **Dockerfile** ```dockerfile # Build FROM python:3.9 as build COPY requirements.txt / WORKDIR /pip-packages RUN pip download -r /requirements.txt --no-input # Prod FROM python:3.9 COPY --from=build /pip-packages/ /pip-packages/ RUN pip install --no-index --find-links=/pip-packages/ /pip-packages/* RUN pip list ``` ## **Expected Behavior** Docker image can be built correctly. ## **Actual Behavior** Docker image fails when building. ## **Error Messages/Stack Trace** ### **Message Error** ```text ERROR: Command errored out with exit status 1: command: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel cwd: None Complete output (4 lines): Looking in links: /pip-packages/ Processing /pip-packages/setuptools-57.2.0-py3-none-any.whl ERROR: Could not find a version that satisfies the requirement wheel ERROR: No matching distribution found for wheel ---------------------------------------- WARNING: Discarding file:///pip-packages/pyrsistent-0.18.0.tar.gz. Command errored out with exit status 1: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel Check the logs for full command output. ERROR: Command errored out with exit status 1: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel Check the logs for full command output. The command '/bin/sh -c pip install --no-index --find-links=/pip-packages/ /pip-packages/*' returned a non-zero code: 1 ``` ### Docker build output ```commandLine $ docker build --tag docker-arm-poc-3:latest --no-cache . ``` ```text Sending build context to Docker daemon 5.12kB Step 1/8 : FROM python:3.9 as build ---> 8ff7a2865c18 Step 2/8 : COPY requirements.txt / ---> 1f6972b1b83d Step 3/8 : WORKDIR /pip-packages ---> Running in 4d614d4db71b Removing intermediate container 4d614d4db71b ---> 45d2797d6000 Step 4/8 : RUN pip download -r /requirements.txt --no-input ---> Running in 53c81d589f32 Collecting flask-restx==0.4.0 Downloading flask_restx-0.4.0-py2.py3-none-any.whl (5.3 MB) Collecting six>=1.3.0 Downloading six-1.16.0-py2.py3-none-any.whl (11 kB) Collecting Flask<2.0.0,>=0.8 Downloading Flask-1.1.4-py2.py3-none-any.whl (94 kB) Collecting werkzeug<2.0.0 Downloading Werkzeug-1.0.1-py2.py3-none-any.whl (298 kB) Collecting jsonschema Downloading jsonschema-3.2.0-py2.py3-none-any.whl (56 kB) Collecting pytz Downloading pytz-2021.1-py2.py3-none-any.whl (510 kB) Collecting aniso8601>=0.82 Downloading aniso8601-9.0.1-py2.py3-none-any.whl (52 kB) Collecting itsdangerous<2.0,>=0.24 Downloading itsdangerous-1.1.0-py2.py3-none-any.whl (16 kB) Collecting Jinja2<3.0,>=2.10.1 Downloading Jinja2-2.11.3-py2.py3-none-any.whl (125 kB) Collecting click<8.0,>=5.1 Downloading click-7.1.2-py2.py3-none-any.whl (82 kB) Collecting MarkupSafe>=0.23 Downloading MarkupSafe-2.0.1-cp39-cp39-manylinux2014_aarch64.whl (31 kB) Collecting setuptools Downloading setuptools-57.2.0-py3-none-any.whl (818 kB) Collecting pyrsistent>=0.14.0 Downloading pyrsistent-0.18.0.tar.gz (104 kB) Installing build dependencies: started Installing build dependencies: finished with status 'done' Getting requirements to build wheel: started Getting requirements to build wheel: finished with status 'done' Preparing wheel metadata: started Preparing wheel metadata: finished with status 'done' Collecting attrs>=17.4.0 Downloading attrs-21.2.0-py2.py3-none-any.whl (53 kB) Saved ./flask_restx-0.4.0-py2.py3-none-any.whl Saved ./aniso8601-9.0.1-py2.py3-none-any.whl Saved ./Flask-1.1.4-py2.py3-none-any.whl Saved ./click-7.1.2-py2.py3-none-any.whl Saved ./itsdangerous-1.1.0-py2.py3-none-any.whl Saved ./Jinja2-2.11.3-py2.py3-none-any.whl Saved ./MarkupSafe-2.0.1-cp39-cp39-manylinux2014_aarch64.whl Saved ./six-1.16.0-py2.py3-none-any.whl Saved ./Werkzeug-1.0.1-py2.py3-none-any.whl Saved ./jsonschema-3.2.0-py2.py3-none-any.whl Saved ./attrs-21.2.0-py2.py3-none-any.whl Saved ./pyrsistent-0.18.0.tar.gz Saved ./pytz-2021.1-py2.py3-none-any.whl Saved ./setuptools-57.2.0-py3-none-any.whl Successfully downloaded flask-restx aniso8601 Flask click itsdangerous Jinja2 MarkupSafe six werkzeug jsonschema attrs pyrsistent pytz setuptools WARNING: You are using pip version 21.0.1; however, version 21.1.3 is available. You should consider upgrading via the '/usr/local/bin/python -m pip install --upgrade pip' command. Removing intermediate container 53c81d589f32 ---> 1467b30d43ef Step 5/8 : FROM python:3.9 ---> 8ff7a2865c18 Step 6/8 : COPY --from=build /pip-packages/ /pip-packages/ ---> 89c8bfc9a430 Step 7/8 : RUN pip install --no-index --find-links=/pip-packages/ /pip-packages/* ---> Running in a10bdd1551cb Looking in links: /pip-packages/ Processing /pip-packages/Flask-1.1.4-py2.py3-none-any.whl Processing /pip-packages/Jinja2-2.11.3-py2.py3-none-any.whl Processing /pip-packages/MarkupSafe-2.0.1-cp39-cp39-manylinux2014_aarch64.whl Processing /pip-packages/Werkzeug-1.0.1-py2.py3-none-any.whl Processing /pip-packages/aniso8601-9.0.1-py2.py3-none-any.whl Processing /pip-packages/attrs-21.2.0-py2.py3-none-any.whl Processing /pip-packages/click-7.1.2-py2.py3-none-any.whl Processing /pip-packages/flask_restx-0.4.0-py2.py3-none-any.whl Processing /pip-packages/itsdangerous-1.1.0-py2.py3-none-any.whl Processing /pip-packages/jsonschema-3.2.0-py2.py3-none-any.whl Processing /pip-packages/pyrsistent-0.18.0.tar.gz Installing build dependencies: started Installing build dependencies: finished with status 'error' ERROR: Command errored out with exit status 1: command: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel cwd: None Complete output (4 lines): Looking in links: /pip-packages/ Processing /pip-packages/setuptools-57.2.0-py3-none-any.whl ERROR: Could not find a version that satisfies the requirement wheel ERROR: No matching distribution found for wheel ---------------------------------------- WARNING: Discarding file:///pip-packages/pyrsistent-0.18.0.tar.gz. Command errored out with exit status 1: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel Check the logs for full command output. ERROR: Command errored out with exit status 1: /usr/local/bin/python /usr/local/lib/python3.9/site-packages/pip install --ignore-installed --no-user --prefix /tmp/pip-build-env-qf_d47c3/overlay --no-warn-script-location --no-binary :none: --only-binary :none: --no-index --find-links /pip-packages/ -- 'setuptools>=42' wheel Check the logs for full command output. The command '/bin/sh -c pip install --no-index --find-links=/pip-packages/ /pip-packages/*' returned a non-zero code: 1 ```
1medium
Title: How to uninstall it? Body: I'm trying to test this package but not good for my project. And this package needs comercial license, so I want to remove it all. How can I uninstall it?
0easy
Title: [🐛 BUG] Lambda expression not working with new line Body: ### What went wrong? 🤔 Lambda expressions are not working for no reason. The second selector is not working. ![image](https://github.com/user-attachments/assets/4f83a3f7-a6f8-433b-8a18-e9fdd69c03ff) ### Expected Behavior It should work. ### Steps to Reproduce Issue ```python from taipy.gui import Gui import taipy.gui.builder as tgb value: int = 10 value_selector = 10 def get_list(value): return [str(i) for i in range(value)] with tgb.Page() as page: # Works tgb.selector(value="{value_selector}", lov=lambda value: get_list(value)) # don't work tgb.selector(value="{value_selector}", lov=lambda value: get_list(value)) Gui(page=page).run(title="Frontend Demo") ``` ### Browsers Chrome ### OS Windows ### Version of Taipy develop - 7/15/24 ### Acceptance Criteria - [ ] Ensure new code is unit tested, and check code coverage is at least 90%. - [ ] Create related issue in taipy-doc for documentation and Release Notes. ### Code of Conduct - [X] I have checked the [existing issues](https://github.com/Avaiga/taipy/issues?q=is%3Aissue+). - [ ] I am willing to work on this issue (optional)
1medium
Title: [Question] Face verification does not meet expectations Body: I am very thankful for this open source project, it solved my immediate problem. I am trying to accomplish the task of face comparison with yolov8n-face and Facenet512. But for some samples, the results are not too good. I tried to replace the face detection with a paid api and this is my test result: |Face detection type|x|y|w|h|left eye|right eye|Facenet512 cosine distance | |---|---|---|---|---|---|---|---| |yolov8n-face.pt img1|43|167|242|334|(99, 288)|(216, 290)| 0.2299605098248957| |yolov8n-face.pt img2|170|165|405|541|(280, 377)|(457, 372)| | |yolov8n-face.onnx img1|43|166|241|333|(99, 289)|(216, 291)|0.22794579406637527 | |yolov8n-face.onnx img2|168|160|401|544|(278, 380)|(456, 376)| | |payment api img1|44|229|245|271|(100, 284)|(220, 283)| 0.3275035163254295| |payment api img2|169|307|385|414|(279, 376)|(453, 362)| | On the left is the yolov8 detection area and on the right is the paid api detection area: ![test](https://github.com/serengil/deepface/assets/43714368/7e8fc61b-eeed-4c30-83af-6adf007b8445) Now I'm confused. Based on the recommended threshold (0.3), the face region obtained from the paid api seems to be better suited to accomplish the face matching task. So should I find a way to shrink the face region or should I find a more suitable threshold based on my actual usage scenario? Can you provide optimization suggestions?
1medium
Title: Multiple authentication methods enabled Body: Hi, is it possible to have multiple authentication methods enabled at the same time? I noticed that if I set `AUTH_TYPE = AUTH_LDAP` DB authentication does not work anymore.
1medium
Title: Provide some basic Django Auth Factories (User, Group, Permission...) Body: #### The problem Any project who uses `django.contrib.auth` and/or `django.contrib.contenttypes` now has to create the factories for User, Group, Permission, ContentType for using them in his tests, right? Why we do not provide related Factories for them as they are very basic models... #### Proposed solution Something similar to: ``` import factory import factory.fuzzy from django.contrib.auth import get_user_model from django.contrib.auth.models import Group, Permission from django.contrib.contenttypes.models import ContentType from factory.django import DjangoModelFactory class ContentTypeFactory(factory.django.DjangoModelFactory): class Meta: model = ContentType django_get_or_create = ('app_label', 'model') app_label = factory.Faker("word") model = factory.Faker("word") class PermissionFactory(DjangoModelFactory): class Meta: model = Permission django_get_or_create = ("name",) name = factory.Faker("name") content_type = factory.SubFactory(ContentTypeFactory) class GroupFactory(DjangoModelFactory): class Meta: model = Group django_get_or_create = ("name",) name = factory.Faker("name") class UserFactory(DjangoModelFactory): username = factory.Faker("user_name") email = factory.Faker("email") password = factory.Faker("password") is_active = True is_staff = False is_superuser = False class Meta: model = get_user_model() django_get_or_create = ("username",) @classmethod def _create(cls, model_class, *args, **kwargs): """Override the default ``_create`` with our custom call.""" manager = cls._get_manager(model_class) # Some support to generate auth users.... if kwargs.get("is_superuser"): return manager.create_superuser(*args, **kwargs) return manager.create_user(*args, **kwargs) @factory.post_generation def groups(self, create, extracted, **kwargs): if not create: # Simple build, do nothing. return if extracted: # A list of groups were passed in, use them for group in extracted: self.groups.add(Group.objects.get_or_create(name=group)[0]) @factory.post_generation def user_permissions(self, create, extracted, **kwargs): if not create: # Simple build, do nothing. return if extracted: # A list of tuple (ModelClass, "my_permission") for model, permission in extracted: content_type = ContentType.objects.get_for_model(model) self.user_permissions.add( Permission.objects.get_or_create( codename=permission, content_type=content_type )[0] ) ``` #### Extra notes Please, notice this is only a draft... I have to refine them :) If could be interesting I can set up a PR draft :)
1medium
Title: Escape special characters Body: matplotlib2tikz does not correctly escape special or accented characters in plots. Specifically, including characters such as the percent sign (%) in labels or captions causes a `Extra }, or forgotten \endgroup. \end{tikzpicture}` error; on the other hand, including accented characters such as á, é, and the like causes an inpuntenc error `Package inputenc Error: Unicode char ícu (U+E5)(inputenc) not set up for use with LaTeX. \end{axis}` **(Not-so Minimal) Working Example:** **Python code:** ```Python def vehicles_vs_time(): cols = ['nvhcs', 't', 'demand', 'runID'] df = pandas.read_csv('NVeh_vs_T.csv', encoding='ISO-8859-1', sep=';') df.columns = cols df['nvhcs'] = pandas.to_numeric(df['nvhcs']) df['t'] = pandas.to_numeric(df['t']) df['runID'] = pandas.to_numeric(df['runID']) df['demand'] = df['demand'].map(lambda x: float(x.strip('%')) / 100) df100 = df.loc[df['demand'] == 1.00] df75 = df.loc[df['demand'] == 0.75] df50 = df.loc[df['demand'] == 0.50] df25 = df.loc[df['demand'] == 0.25] mean_df = pandas.DataFrame(columns=['demand', 'mean_vhcs', 'mean_time']) mean_df.loc[0] = [1.00, df100['nvhcs'].mean(), df100['t'].mean()] mean_df.loc[1] = [0.75, df75['nvhcs'].mean(), df75['t'].mean()] mean_df.loc[2] = [0.50, df50['nvhcs'].mean(), df50['t'].mean()] mean_df.loc[3] = [0.25, df25['nvhcs'].mean(), df25['t'].mean()] # from this point onward, plot fig, ax = pyplot.subplots() ax.set_facecolor('white') ax.grid(color='#a1a1a1', linestyle='-', alpha=0.1) pyplot.xlim([df['nvhcs'].min() - 50, df['nvhcs'].max() + 50]) pyplot.ylim(0, df['t'].max() + 120) yticks_mins = numpy.arange(0, df['t'].max() + 120, 120) yticks_10secs = numpy.arange(0, df['t'].max() + 120, 60) xticks = numpy.arange(200, 1500, 100) xticks_minor = numpy.arange(150, 1500, 10) ax.set_yticks(yticks_mins) ax.set_yticks(yticks_10secs, minor=True) ax.set_xticks(xticks) ax.set_xticks(xticks_minor, minor=True) # trendline z = numpy.polyfit(df['nvhcs'], df['t'], 2) p = numpy.poly1d(z) nx = range(0, int(df['nvhcs'].max()) + 200) ax.plot(nx, p(nx), '-.', alpha=0.3, label='Ajuste polinomial', color='#F06449') # scatter ax.plot(df100['nvhcs'], df100['t'], 'o', color='#17BEBB', label='Factor de demanda 100%') ax.plot(df75['nvhcs'], df75['t'], 'o', color='#EF2D56', label='Factor de demanda 75%') ax.plot(df50['nvhcs'], df50['t'], 'o', color='#8CD867', label='Factor de demanda 50%') ax.plot(df25['nvhcs'], df25['t'], 'o', color='#2F243A', label='Factor de demanda 25%') ax.legend(loc='upper left') pyplot.ylabel('Tiempo (MM:SS)') formatter = matplotlib.ticker.FuncFormatter(to_min_secs) ax.yaxis.set_major_formatter(formatter) pyplot.xlabel('Cantidad promedio vehículos en simulación') # pyplot.title('Scatterplot: Cantidad promedio de vehículos vs duración en tiempo real de simulación') # pyplot.savefig('n_vhcs_vs_time.pgf') # pyplot.show() tikz_save('n_vhcs_vs_time.tex', figureheight='\\figureheight', figurewidth='\\figurewidth') ``` **Generated output (note the labels and legends):** ```Latex % This file was created by matplotlib2tikz v0.6.10. \begin{tikzpicture} \definecolor{color0}{rgb}{0.941176470588235,0.392156862745098,0.286274509803922} \definecolor{color1}{rgb}{0.0901960784313725,0.745098039215686,0.733333333333333} \definecolor{color2}{rgb}{0.937254901960784,0.176470588235294,0.337254901960784} \definecolor{color3}{rgb}{0.549019607843137,0.847058823529412,0.403921568627451} \definecolor{color4}{rgb}{0.184313725490196,0.141176470588235,0.227450980392157} \begin{axis}[ xlabel={Cantidad promedio vehículos en simulación}, ylabel={Tiempo (MM:SS)}, xmin=150, xmax=1490, ymin=0, ymax=1713, width=\figurewidth, height=\figureheight, tick align=outside, tick pos=left, xmajorgrids, x grid style={lightgray!84.183006535947712!black}, ymajorgrids, y grid style={lightgray!84.183006535947712!black}, axis line style={white}, legend style={at={(0.03,0.97)}, anchor=north west, draw=white!80.0!black, fill=white!89.803921568627459!black}, legend cell align={left}, legend entries={{Ajuste polinomial},{Factor de demanda 100%},{Factor de demanda 75%},{Factor de demanda 50%},{Factor de demanda 25%}} ] \addplot [semithick, color0, opacity=0.3, dash pattern=on 1pt off 3pt on 3pt off 3pt] table {% 0 6.3172112993997 1 6.57640934437459 2 6.83677322746044 3 7.09830294865727 % ... % ... lots of points % ... 1611 1935.80824515293 1612 1937.94560839468 1613 1940.08413747453 1614 1942.2238323925 }; \addplot [semithick, color1, mark=*, mark size=3, mark options={solid}, only marks] table {% 1398.4 1593 1351.4 1439 1354.67 1388 1415.13 1466 }; \addplot [semithick, color2, mark=*, mark size=3, mark options={solid}, only marks] table {% 870.2 660 872.27 664 842.13 703 890.4 707 }; \addplot [semithick, color3, mark=*, mark size=3, mark options={solid}, only marks] table {% 521.13 281 504.4 268 512.2 272 520.6 282 }; \addplot [semithick, color4, mark=*, mark size=3, mark options={solid}, only marks] table {% 249.67 118 254.8 122 240.87 113 240.93 100 }; \end{axis} \end{tikzpicture} ``` **Expected (and working!) output:** ```Latex % This file was created by matplotlib2tikz v0.6.10. \begin{tikzpicture} \definecolor{color0}{rgb}{0.941176470588235,0.392156862745098,0.286274509803922} \definecolor{color1}{rgb}{0.0901960784313725,0.745098039215686,0.733333333333333} \definecolor{color2}{rgb}{0.937254901960784,0.176470588235294,0.337254901960784} \definecolor{color3}{rgb}{0.549019607843137,0.847058823529412,0.403921568627451} \definecolor{color4}{rgb}{0.184313725490196,0.141176470588235,0.227450980392157} \begin{axis}[ xlabel={Cantidad promedio veh\'iculos en simulaci\'on}, ylabel={Tiempo (MM:SS)}, xmin=150, xmax=1490, ymin=0, ymax=1713, width=\figurewidth, height=\figureheight, tick align=outside, tick pos=left, xmajorgrids, x grid style={lightgray!84.183006535947712!black}, ymajorgrids, y grid style={lightgray!84.183006535947712!black}, axis line style={white}, legend style={at={(0.03,0.97)}, anchor=north west, draw=white!80.0!black, fill=white!89.803921568627459!black}, legend cell align={left}, legend entries={{Ajuste polinomial},{Factor de demanda 100\%},{Factor de demanda 75\%},{Factor de demanda 50\%},{Factor de demanda 25\%}} ] \addplot [semithick, color0, opacity=0.3, dash pattern=on 1pt off 3pt on 3pt off 3pt] table {% 0 6.3172112993997 1 6.57640934437459 2 6.83677322746044 3 7.09830294865727 % ... % ... lots of points % ... 1611 1935.80824515293 1612 1937.94560839468 1613 1940.08413747453 1614 1942.2238323925 }; \addplot [semithick, color1, mark=*, mark size=3, mark options={solid}, only marks] table {% 1398.4 1593 1351.4 1439 1354.67 1388 1415.13 1466 }; \addplot [semithick, color2, mark=*, mark size=3, mark options={solid}, only marks] table {% 870.2 660 872.27 664 842.13 703 890.4 707 }; \addplot [semithick, color3, mark=*, mark size=3, mark options={solid}, only marks] table {% 521.13 281 504.4 268 512.2 272 520.6 282 }; \addplot [semithick, color4, mark=*, mark size=3, mark options={solid}, only marks] table {% 249.67 118 254.8 122 240.87 113 240.93 100 }; \end{axis} \end{tikzpicture} ```
1medium
Title: Add attachment to order_placed email Body: I need to add an attachment to the `order_placed` email. I see the function I would ideally hook into. The `Checkout.mixins.OrderPlacementMixin` there is a function called `send_order_placed_email`. So would I fork the whole Checkout app? And add my own `mixins.py` and override the `OrderPlacementMixin`?
1medium
Title: Add social preview image for documentation Body: Reference: https://squidfunk.github.io/mkdocs-material/setup/setting-up-social-cards/#meta-tags
0easy
Title: BUG: Exporting df as a ESRI shapefile error: Object of type 'float' has no len() Body: - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the latest version of geopandas. - [x] (optional) I have confirmed this bug exists on the main branch of geopandas. --- #### Code Sample, a copy-pastable example ```python # Your code here df.to_file(filename="test.shp", driver="ESRI Shapefile") ``` #### Problem description The problem is occurred when the column name is an `float` type. Below is the Traceback, ```python Traceback (most recent call last): File "/opt/sdss/django/sdss/exposure/views.py", line 353, in downloadExposureResult df.to_file(filename=shp_path, driver='ESRI Shapefile') File "/home/sdss/.virtualenvs/sdss/lib/python3.8/site-packages/geopandas/geodataframe.py", line 1114, in to_file _to_file(self, filename, driver, schema, index, **kwargs) File "/home/sdss/.virtualenvs/sdss/lib/python3.8/site-packages/geopandas/io/file.py", line 376, in _to_file if driver == "ESRI Shapefile" and any([len(c) > 10 for c in df.columns.tolist()]): File "/home/sdss/.virtualenvs/sdss/lib/python3.8/site-packages/geopandas/io/file.py", line 376, in <listcomp> if driver == "ESRI Shapefile" and any([len(c) > 10 for c in df.columns.tolist()]): TypeError: object of type 'float' has no len() ``` #### Expected Output It should write df into the shapefile. #### Output of ``geopandas.show_versions()`` 0.10.2 [paste the output of ``geopandas.show_versions()`` here leaving a blank line after the details tag] </details>
1medium
Title: WiringConfiguration(packages=[top_level_package]) affect subpackages only if they have __init__.py Body: Hello folks! I'm just started to learn dependency-injector. I was following this public template [https://github.com/teamhide/fastapi-boilerplate](https://github.com/teamhide/fastapi-boilerplate) and found that WiringConfiguration can be affect subpackages of designated top level package only if all of paths have __init__.py. I understand that from Python 3.6 we don't necessarily need to use __init__.py not if in some of situation but i believe this might not be one of the those cases. **app.container.py** ```python from dependency_injector.containers import DeclarativeContainer, WiringConfiguration from dependency_injector.providers import Factory, Singleton from app.auth.application.service.jwt import JwtService from app.user.adapter.output.persistence.repository_adapter import UserRepositoryAdapter from app.user.adapter.output.persistence.sqlalchemy.user import UserSQLAlchemyRepo from app.user.application.service.user import UserService class Container(DeclarativeContainer): wiring_config = WiringConfiguration(packages=["app"]) user_repo = Singleton(UserSQLAlchemyRepo) user_repo_adapter = Factory(UserRepositoryAdapter, user_repo=user_repo) user_service = Factory(UserService, repository=user_repo_adapter) jwt_service = Factory(JwtService) ```
1medium
Title: ATRTS not included in df.ta.study("volatility") Body: Using the latest version of develop, `Average True Range Trailing Stop: atrts`, isn't included in `df.ta.study("volatility")`. To be honest, I'm not entirely sure why. Things appears to be set up right. ```python Python 3.8.13 (default, Mar 18 2022, 02:13:38) [GCC 11.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import pandas_ta as ta >>> print(ta.version) 0.3.59b0 >>> import talib >>> import pytz >>> from datetime import datetime >>> from config import OUTPUT_DIR >>> import numpy as np >>> timezone = pytz.utc >>> date_range = ["20200801", "20200901"] >>> start_date = timezone.localize(datetime.strptime(date_range[0], "%Y%m%d")) >>> end_date = timezone.localize(datetime.strptime(date_range[1], "%Y%m%d")) >>> data = pd.read_feather(f"{OUTPUT_DIR}/BTC.feather") >>> df = data[data["Datetime"].between(left=start_date, right=end_date, inclusive="both")] >>> df.set_index(pd.DatetimeIndex(df["Datetime"]), inplace=True) >>> df.ta.study("volatility") >>> df.columns Index(['Datetime', 'Open', 'High', 'Low', 'Close', 'Volume', 'ABER_ZG_5_15', 'ABER_SG_5_15', 'ABER_XG_5_15', 'ABER_ATR_5_15', 'ACCBL_20', 'ACCBM_20', 'ACCBU_20', 'ATRr_14', 'BBL_5_2.0', 'BBM_5_2.0', 'BBU_5_2.0', 'BBB_5_2.0', 'BBP_5_2.0', 'DCL_20_20', 'DCM_20_20', 'DCU_20_20', 'HWM_1', 'HWU_1', 'HWL_1', 'KCLe_20_2', 'KCBe_20_2', 'KCUe_20_2', 'MASSI_9_25', 'NATR_14', 'PDIST', 'RVI_14', 'THERMO_20_2_0.5', 'THERMOma_20_2_0.5', 'THERMOl_20_2_0.5', 'THERMOs_20_2_0.5', 'TRUERANGE_1', 'UI_14'], dtype='object')```
1medium
Title: The dataset for lecture 5 practical is not available Body: I was watching practical of lecture 5 and I like to code along the video, but the dataset used in the video is not available anywhere. Its not present in your kaggle nor in the github repo. Would humbly request you to upload the same
0easy
Title: Stop automatically sorting `gr.Barplot` labels alphabetically Body: It seems that the `gr.Barplot` component ignores the actual order of the labels in your dataframe and instead sorts the x values alphabetically / numerically? It should respect the order of values in your dataframe: <img width="1237" alt="Image" src="https://github.com/user-attachments/assets/1b825a1c-61fa-4feb-9c14-1f5b5d414c92" /> See [playground link](https://www.gradio.app/playground?demo=Hello_World&code=aW1wb3J0IGdyYWRpbyBhcyBncgppbXBvcnQgcGFuZGFzIGFzIHBkCmltcG9ydCBudW1weSBhcyBucAoKIyBHZW5lcmF0ZSByYW5kb20gZGF0YSBmb3IgdGhlIGJhciBwbG90CmRhdGEgPSB7CiAgICAnQ2F0ZWdvcnknOiBbJ0YnLCAnRScsICdDJywgJ0InLCAnQUFBQUEnXSwKICAgICdWYWx1ZSc6IG5wLnJhbmRvbS5yYW5kaW50KDEsIDEwMCwgNSkKfQpkZiA9IHBkLkRhdGFGcmFtZShkYXRhKQoKIyBEZWZpbmUgdGhlIGZ1bmN0aW9uIHRvIGNyZWF0ZSB0aGUgYmFyIHBsb3QKZGVmIGNyZWF0ZV9iYXJfcGxvdCgpOgogICAgcmV0dXJuIGRmCgojIENyZWF0ZSB0aGUgR3JhZGlvIGludGVyZmFjZQp3aXRoIGdyLkJsb2NrcygpIGFzIGRlbW86CiAgICAjIENyZWF0ZSBhIEJhclBsb3QgY29tcG9uZW50CiAgICBiYXJfcGxvdCA9IGdyLkJhclBsb3QoCiAgICAgICAgdmFsdWU9Y3JlYXRlX2Jhcl9wbG90LAogICAgICAgIHg9IkNhdGVnb3J5IiwKICAgICAgICB5PSJWYWx1ZSIsCiAgICApCgojIExhdW5jaCB0aGUgaW50ZXJmYWNlCmRlbW8ubGF1bmNoKHNob3dfZXJyb3I9VHJ1ZSk%3D&reqs=cGFuZGFzCm51bXB5)
1medium
Title: [tabular] Speed-up learning-curve unit tests Body: Currently learning curve unit tests (introduced in #4411) take a long time (~25+ minutes). They comprise of 380 unit tests that each fit a model and evaluate it, which is time consuming. We should either speed up these unit tests or disable the majority of them and only run the disabled ones prior to release.
1medium
Title: Any support return dict from fetch? Body: In many synchronous IO libraries, cursor.fetch_all() is allowed to return a sequence of dictionaries. Do databases consider similar implementations? Just like: ```python rows: List[Dict[str, Any]] = await database.fetch_all(query=query, include_field_name=True) ```
1medium
Title: Document `column_default_sort` requires strings Body: ### Checklist - [X] The bug is reproducible against the latest release or `master`. - [X] There are no similar issues or pull requests to fix it yet. ### Describe the bug The examples in the [documentation](https://aminalaee.dev/sqladmin/configurations/#list-page) show `column_default_sort` as accepting class attributes. However, since https://github.com/aminalaee/sqladmin/commit/8f9d07d0705b241490a1f55fc69115ef598b8c50 `split()` is being called on them, and that fails unless they are strings ### Steps to reproduce the bug Visit an admin page with default sort defined via class attributes: ```python class UserAdmin(ModelView, model=User): ... column_default_sort = [(User.email, True), (User.name, False)] ``` ### Expected behavior The default sort is applied as defined ### Actual behavior The server crashes because it can't `split()` class attributes ### Debugging material _No response_ ### Environment Ubuntu 22.04, Python 3.11 ### Additional context `column_default_sort = [("email", True), ("name", False)]` works as expected
1medium
Title: sh: /usr/bin/dotnet: 权限不够 Body: 想要执行 dotnet core 5 ,提示 sh: /usr/bin/dotnet: 权限不够。 目录下执行也类似:sh: ./dotnet: 权限不够
1medium
Title: Feature: Filtering Rule for HTTP requests payload (like on emails) Body: Hi, I'd like to see it possible to filter the HTTP request for a specific value, as it's currently possible with emails. I've not spotted this in the current solution. I'm using a solution where I can create a webhook, but not control other than "on job run". So I'd like to use healthcheck to filter the content, on basis of the payload (like if it contains "success" or similar). I would be great to see a JSON parsed, however, using plain-text should possibly also work :) Thank you for making this. I actually enjoy how simple it is it work with. ## Example data Here's one Request Body for the application I'm trying to Filter (JSON formatted for easier reading). The first one are the "job done" post, where I'd like to look to match the `params.vms.id` and `result`. Solution one would be to look after `VM_ID_I_NEED_TO_MATCH_GOES_HERE` in the text, **and** `"result":true` in the text, too (so more than one check), or have the JSON parsed, and define the `result` nested_key, including a `mandatory` information to be part of the payload, before healthchecks will acknowledge the requests as related to the current `check`. ### Example for "Job Done", I need to acknowledge as a success Requirements: - `VM_ID_I_NEED_TO_MATCH_GOES_HERE` needs to be part of the payload - Result should be `true` For both requirements, it should be possible to just cheat the payload as a string, without having to use a JSON parser or similar. But two different requirements need to be allowed as a user-input (or a regex sting, maybe?... I mean.. Keep is simpel stupid, if that's the easiest thing to do). ```json { "callId":"CALL_ID_GOES_HERE", "method":"backupNg.runJob", "params":{ "id":"SECRET_ID_GOES_HERE", "schedule":"SCHEDULER_ID_GOES_HERE", "settings":{ "SETTINGS_ID_GOES_HERE":{ "snapshotRetention":5 } }, "vms":{ "id":"VM_ID_I_NEED_TO_MATCH_GOES_HERE" } }, "timestamp":1709391402615, "userId":"USER_ID_GOES_HERE", "userName":"admin", "result":true, "duration":15735, "type":"post" } ``` Here's another example. Note the difference in the `vms` "array", which looks a bit odd, but, that's how it's reported. Therefore, the plain-text validator are OK with me, I guess. Again, more real-world like data. I've just replaced letters and stuff like that. ```json { "callId":"222h2akasaa", "method":"backupNg.runJob", "params":{ "id":"fccd0fbb-5ba8-47fd-9e44-401a1f10c474", "schedule":"2a029a93-0f6d-4a10-891f-0a891041eeec", "settings":{ "6d564d15-3d24-40e4-aa81-5c7df522def9":{ "exportRetention":1 } }, "vms":{ "id":{ "__or":[ "bcdff841-c452-4eee-af61-5b8fddc3318b", "083ac012-06ea-44f8-96e4-93b4731328e4", "2c1ed51f-67bc-4306-b8f6-c080ee9221d9", "8577151d-c758-49ce-a177-ad29d3e5fec2", "f3c60996-418a-41f3-a246-aad3731e5c84", "25d04cd3-292e-430d-85db-c67f1ecc45b0" ] } } }, "timestamp":1709396008102, "userId":"9a9758c5-233d-4a69-84e6-7b62ba522895", "userName":"admin", "result":true, "duration":4645273, "type":"post" } ``` ### Another example Here's another example, which I'd need to maybe acknowledge,or maybe ignore, on basis of the "result" value, and the params.id value (`a22aa300-2310-11dd-a222-2201aa1101b1`) value. ```json { "callId":"CALL_ID_GOES_HERE", "userId":"USER_ID_GOES_HERE", "userName":"admin", "userIp":"::USER_IP_GOES_HERE", "method":"backupNg.runJob", "params":{ "id":"ID_GOES_HERE", "schedule":"SCHDEULE_ID_GOES_HERE" }, "timestamp":1709391402650, "duration":15776, "result":true, "type":"post" } ``` With more "real world" data: ```json { "callId":"1ab2abcdaa2a", "userId":"USER_ID_GOES_HERE", "userName":"admin", "userIp":"::REMOVED", "method":"backupNg.runJob", "params":{ "id":"a22aa300-2310-11dd-a222-2201aa1101b1", "schedule":"a3cfd101-faaa-12c2-bbd1-12b22c22eaa1" }, "timestamp":1709396008114, "duration":4645290, "result":true, "type":"post" }
1medium
Title: New release? Body: Hi, could you please make a new release on pypi? I'd like my deploy systems to grab the new fixes and it would probably be a nice idea to also have the updated author on there.
0easy
Title: docs: code snippet needed: DocumentArray construct Body: In [this section](https://docarray.jina.ai/fundamentals/documentarray/construct/#construct-from-local-files) there's no snippet for `read_mode`. Note: Please leave snippet as comment, don't fix directly. I'm working on docs right now and don't want merge conflicts.
0easy
Title: [Bug]: Executor performance degradation Body: ### Your current environment <details> <summary>The output of `python collect_env.py`</summary> ```text Versions of relevant libraries: [pip3] numpy==1.26.4 [pip3] nvidia-cublas-cu12==12.1.3.1 [pip3] nvidia-cuda-cupti-cu12==12.1.105 [pip3] nvidia-cuda-nvrtc-cu12==12.1.105 [pip3] nvidia-cuda-runtime-cu12==12.1.105 [pip3] nvidia-cudnn-cu12==9.1.0.70 [pip3] nvidia-cufft-cu12==11.0.2.54 [pip3] nvidia-curand-cu12==10.3.2.106 [pip3] nvidia-cusolver-cu12==11.4.5.107 [pip3] nvidia-cusparse-cu12==12.1.0.106 [pip3] nvidia-ml-py==12.560.30 [pip3] nvidia-nccl-cu12==2.20.5 [pip3] nvidia-nvjitlink-cu12==12.6.85 [pip3] nvidia-nvtx-cu12==12.1.105 [pip3] pyzmq==26.2.0 [pip3] torch==2.4.0 [pip3] torch_tensor_module==0.0.0 [pip3] torchvision==0.19.0 [pip3] transformers==4.47.0 [pip3] triton==3.0.0 [conda] numpy 1.26.4 pypi_0 pypi [conda] nvidia-cublas-cu12 12.1.3.1 pypi_0 pypi [conda] nvidia-cuda-cupti-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cuda-nvrtc-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cuda-runtime-cu12 12.1.105 pypi_0 pypi [conda] nvidia-cudnn-cu12 9.1.0.70 pypi_0 pypi [conda] nvidia-cufft-cu12 11.0.2.54 pypi_0 pypi [conda] nvidia-curand-cu12 10.3.2.106 pypi_0 pypi [conda] nvidia-cusolver-cu12 11.4.5.107 pypi_0 pypi [conda] nvidia-cusparse-cu12 12.1.0.106 pypi_0 pypi [conda] nvidia-ml-py 12.560.30 pypi_0 pypi [conda] nvidia-nccl-cu12 2.20.5 pypi_0 pypi [conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi [conda] nvidia-nvtx-cu12 12.1.105 pypi_0 pypi [conda] pyzmq 26.2.0 pypi_0 pypi [conda] torch 2.4.0 pypi_0 pypi [conda] torch-tensor-module 0.0.0 pypi_0 pypi [conda] torchvision 0.19.0 pypi_0 pypi [conda] transformers 4.47.0 pypi_0 pypi [conda] triton 3.0.0 pypi_0 pypi ROCM Version: Could not collect Neuron SDK Version: N/A vLLM Version: 0.6.3.post1 ``` </details> ### 🐛 Describe the bug I am currently using `vllm serve` to create an online server and using this server as a Scheduler to send and execute `ExecuteModelRequest` to one `LLM` instances created with `self.llm = LLM(model=self.model, gpu_memory_utilization=self.gpu_memory_utilization, enforce_eager=self.enforce_eager, tensor_parallel_size=self.tp_size)`. The inference is then performed by calling `self.llm.llm_engine.model_executor.execute_model(req)`. However, when I test using `vllm/benchmarks/benchmark_serving.py`, I notice a significant drop in inference performance, with both TTFT and TPOT increasing substantially, while directly using `vllm serve` for inference does not exhibit this issue. Is it possible that `vllm` has implemented some caching or optimizations above the Executor layer in the `LLMEngine`? ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
1medium
Title: AutoKeras with tensorflow.dataset gives TypeError: 'PrefetchDataset' object does not support indexing Body: ### Bug Description --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-18-e8d1a7f824a0> in <module> 1 # Feed the image classifier with training data. ----> 2 clf.fit(train_ds[0], train_ds[1]) TypeError: 'PrefetchDataset' object does not support indexing ### Reproducing Steps Following the steps for creating a tf.dataset from: https://www.tensorflow.org/tutorials/load_data/images I encountered an issue with AutoModel when calling: clf.fit(train_ds[0], train_ds[1]) and clf.fit(next(iter(train_ds))[0], next(iter(train_ds))[1]) gives : TypeError: Expect the data to ImageInput to be numpy.ndarray or tf.data.Dataset, but got <class 'tensorflow.python.framework.ops.EagerTensor'>. ### Expected Behavior I expect to be able to use AutoKeras with imagesets that are too large to fit in memory.
1medium
Title: TYPO Body: Astounding and Great Template for SQL-based Project with FastAPI. There's a small typo with `USERNAME` here: https://github.com/Aeternalis-Ingenium/FastAPI-Backend-Template/blob/571981c3d6af5feccb90de25a81e82442c22b40f/backend/src/config/settings/base.py#L35C11-L35C11 Consider making this for NoSQL-based aswell as it seems really becoming "Convention" to startup on any project!
0easy
Title: Missing data in HTML report table after upgrading to pytest-html 4.1.1 Body: Hi Team, I'm currently encountering a problem with pytest-html after upgrading from version 3.2.0 to 4.1.1. My HTML reports are now empty when running tagged tests, although untagged tests generate reports correctly. Environment: Python version: 3.10 pytest version: 7.4.4 pytest-html version: 4.1.1 (previously 3.2.0) I'm suspecting since pytest_html_results_table_header hook is not invoking in pytest-html 4.1.1 when running tests with tags. With pytest_html_results_table_header hook I’m trying to insert a snapshot column for failed test , since the table is not properly customized I'm getting an empty report. This hook was invoked in the previous version and is still invoking in 4.1.1 when running tests without tags. The lack of this hook prevents proper table customization, resulting in an empty table. Observed Behavior: The HTML report's results table is empty when running the tests with tags. ![empty_report](https://github.com/user-attachments/assets/472c9b9c-9775-440f-badc-09b0dd81bae3) Expected Behavior: The results table should display all test results (name, outcome, duration), including the newer column 'snapshot' when running the tests with tags. If you have a moment, I'd appreciate your help.
1medium
Title: Save nparray as list Body: ### Describe the bug When I use the `map` function to convert images into features, datasets saves nparray as a list. Some people use the `set_format` function to convert the column back, but doesn't this lose precision? ### Steps to reproduce the bug the map function ```python def convert_image_to_features(inst, processor, image_dir): image_file = inst["image_url"] file = image_file.split("/")[-1] image_path = os.path.join(image_dir, file) image = Image.open(image_path) image = image.convert("RGBA") inst["pixel_values"] = processor(images=image, return_tensors="np")["pixel_values"] return inst ``` main function ```python map_fun = partial( convert_image_to_features, processor=processor, image_dir=image_dir ) ds = ds.map(map_fun, batched=False, num_proc=20) print(type(ds[0]["pixel_values"]) ``` ### Expected behavior (type < list>) ### Environment info - `datasets` version: 2.16.1 - Platform: Linux-4.19.91-009.ali4000.alios7.x86_64-x86_64-with-glibc2.35 - Python version: 3.11.5 - `huggingface_hub` version: 0.23.4 - PyArrow version: 14.0.2 - Pandas version: 2.1.4 - `fsspec` version: 2023.10.0
1medium
Title: Documentation Hidden by Menu Body: ### Is there an existing issue for this? - [X] I have searched the existing issues ### Describe the bug When scrolling and zooming on sanic.dev via safari on OSX (is it MacOS now?) half the documentation gets obscured and it is not possible to correct it without a refresh. <img width="979" alt="Screenshot 2024-03-13 at 17 07 46" src="https://github.com/sanic-org/sanic/assets/159023003/fdd4c1aa-2e17-418d-9658-f10e0645326a"> ### Code snippet _No response_ ### Expected Behavior The documentation is not obscured. ### How do you run Sanic? As a module ### Operating System MacOS ### Sanic Version current ### Additional context _No response_
1medium
Title: Custom fields serialisation & deserialisation Body: Hi, I am trying to create a custom field to implement client side encryption. I am able to perform serialisation during save, however I dont see a way to deserialise the data after a find (decrypt it). Is this possible by some means? This is how I created the custom field `class EncryptedFieldType(str): @classmethod def __get_validators__(cls): yield cls.validate @classmethod def validate(cls, v): if isinstance(v, bytes): # Handle data coming from MongoDB print("In isinstance(v, bytes) ...") return "hello" if not isinstance(v, str): raise TypeError("string required") if not v.isascii(): raise ValueError("Only ascii characters are allowed") return v @classmethod def __bson__(cls, v) -> str: print("In __bson__") return "*******" class CollectionWithEncField(Model): encrypted_field: EncryptedFieldType`
1medium
Title: Doc error: on Docs » Visualizers and API » Feature Analysis Visualizers » PCA Projection "principle component analysis" should be "principal component analysis" Body: In http://www.scikit-yb.org/en/latest/api/features/pca.html the phrase "principle component analysis" should be "principal component analysis"
0easy
Title: Tokenization in 8.2.2 Body: I think in tokenize function, if it's tokenizing words, it should add space_character to tokens too. Otherwise, in predict function, it will assume '<unk>' for spaces and the predictions doesn't have spaces between them (which can be solved by manipulating the predict function to this line: `return ''.join([vocab.idx_to_token[i] + ' ' for i in outputs])`) I think tokenized should change like this: `[line.split() for line in lines] + [[' ']]` If I'm right, I can make a pr for both tokenize and predict funcs. (although for predict I might have to change inputs of function as well to recognize if it's a char level or word lever rnn)
1medium
Title: API Calls not working Body: I am not getting the expected response from Alpha Vantage APIs. The following are the details, could someone help? Infrastructure - GCP Windows server GCP Region: asia-south1-c Error - { "Note": "Thank you for using Alpha Vantage! Our standard API call frequency is 5 calls per minute and 500 calls per day. Please visit https://www.alphavantage.co/premium/ if you would like to target a higher API call frequency." } Key - Free trial keys Pls note our url /python code works well to fetch the details however through GCP infrastructure is not working. Code: def __init__(self, api_key, part_url, query, function, exchange, symbol, interval, format): self.api_key = api_key self.part_url = part_url self.query = query self.function = function self.exchange = exchange self.symbol = symbol self.interval = interval self.format = format def get_data(self): #data_url = 'https://www.alphavantage.co/query?function=TIME_SERIES_INTRADAY&symbol=NSE:TATAMOTORS&interval=1min&apikey=STPO0LS61SH9DZ0J&datatype=csv' data_url = self.part_url + self.query + self.function + '&symbol' + '=' + self.exchange + ':' + self.symbol + '&' + 'interval=' + self.interval + '&apikey=' + self.api_key + '&datatype=' + self.format print (data_url) try: data=requests.get(data_url) except Exception as e: print ("Failed with error {0}".format(e)) raise SystemExit
1medium
Title: Detected short number Body: Hello. I have image: ![cell](https://user-images.githubusercontent.com/20105785/137730636-f16acbe0-31a4-4fe5-9231-761038b326b7.png) but easyOCR do not work. I try use `detect` with diferrent parameters, but method return empty. then i tried to make my simple-detector: ``` def recognition_number(self, img: np.ndarray) -> str: image = self._filtration(img) thresh = cv2.adaptiveThreshold(image, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY_INV, 3, 1) cnts, _ = cv2.findContours(thresh, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE) ox = [i[0][0] for i in cnts[0]] oy = [i[0][1] for i in cnts[0]] data = self.reader.recognize(image, horizontal_list=[[min(ox), max(ox), min(oy), max(oy)]], free_list=[], allowlist="1234567890.,") return "".join([el[1] for el in data]) ``` this solved the problem and now the number is recognized. is it possible to do something with a easyOCR `detect`?
1medium
Title: [RFC]typed: Turn typed.Dynaconf into a class (currently it is just a constructor) Body: Current implementation of `typed.Dynaconf` is just a pass-through constructor, this class intercepts passed arguments and then created an instance of `LazySettings` with `cast`ed type annotation for auto complete. ## Proposal - Turn that class into a real class with a `__init__` method, that will perform all the processes currently performed on `__new__` but instead of returning a `LazySettings` will instantiate `self` and assign `self._dynaconf_instance = LazySettings(...)` - Stop inheriting from `BaseSettings` - Implement attribute and method `stubs` that will feed the type check and auto completion - Implement lookup for missing keys on the `self._dynaconf_instance` The advantages will be: - Expose only selected attributes for type annotation - cache the values on the instance itself, after the first access, assign the value to the instance - LazySettings will be just the `settings storage backend` that will eventually be replaced
2hard
Title: [ChunkStore] Decompression error on concurrent read & write Body: #### Arctic Version ``` 1.70.0 ``` #### Arctic Store ``` ChunkStore ``` #### Platform and version CentOS 7.2 #### Description of problem and/or code sample that reproduces the issue I'm getting an error when I have two threads simultaneously reading and writing from one symbol. The reading thread will periodically throw a LZ4BlockError: ```LZ4BlockError: Decompression failed: corrupt input or insufficient space in destination buffer. Error code: 24049``` Here is the code to reproduce it: ``` def get_library(): return Arctic('some_host:1234').get_library(LIB_NAME) def write_loop(): library = get_library() while True: data = create_data(100) # Creates 100 rows of random data library.append('SYMBOL', data) def read_loop(): library = get_library() while True: df = library.read('SYMBOL') proc = mp.Process(target=write_loop) proc.start() try: read_loop() finally: proc.terminate() ``` From a quick check, it seems that the data being passed to decompress(_str) in _compression.py is not valid lz4 - could the the block metadata and data be out of sync?
2hard
Title: FastAPI integration error Body: ### Description I apologize if this belongs in the FastAPI issues instead of logfire. I'm not really sure who is the culprit here. I'm attaching a [sample project](https://github.com/user-attachments/files/16090524/example.zip) to demonstrate an error when `logfire[fastapi]` is added to a FastAPI project. > **Note:** forcing a downgrade to pydantic v1 fixes the issue _(or removing logfire all together)_ ### Error Reproduction A simple API without any input works: ```shell curl "http://localhost:8000/hello/" ``` Calling an API with a pydantic model as input fails: ```shell curl -X POST "http://localhost:8000/test/" -H "Content-Type: application/json" -d '{"name": "test"}' ```` Error log: ``` INFO: 127.0.0.1:58776 - "POST /test/ HTTP/1.1" 500 Internal Server Error ERROR: Exception in ASGI application Traceback (most recent call last): File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/uvicorn/protocols/http/httptools_impl.py", line 399, in run_asgi result = await app( # type: ignore[func-returns-value] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__ return await self.app(scope, receive, send) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/fastapi/applications.py", line 1054, in __call__ await super().__call__(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/applications.py", line 123, in __call__ await self.middleware_stack(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/middleware/errors.py", line 186, in __call__ raise exc File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/middleware/errors.py", line 164, in __call__ await self.app(scope, receive, _send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/opentelemetry/instrumentation/asgi/__init__.py", line 631, in __call__ await self.app(scope, otel_receive, otel_send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/middleware/exceptions.py", line 65, in __call__ await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/routing.py", line 756, in __call__ await self.middleware_stack(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/routing.py", line 776, in app await route.handle(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/routing.py", line 297, in handle await self.app(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/routing.py", line 77, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app raise exc File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app await app(scope, receive, sender) File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/starlette/routing.py", line 72, in app response = await func(request) ^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/fastapi/routing.py", line 269, in app solved_result = await solve_dependencies( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/logfire/_internal/integrations/fastapi.py", line 111, in patched_solve_dependencies return await instrumentation.solve_dependencies(request, original) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/logfire/_internal/integrations/fastapi.py", line 173, in solve_dependencies result = await original ^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/fastapi/dependencies/utils.py", line 628, in solve_dependencies ) = await request_body_to_args( # body_params checked above ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/fastapi/dependencies/utils.py", line 758, in request_body_to_args v_, errors_ = field.validate(value, values, loc=loc) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/fastapi/_compat.py", line 127, in validate self._type_adapter.validate_python(value, from_attributes=True), ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/pydantic/type_adapter.py", line 142, in wrapped return func(self, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/pydantic/type_adapter.py", line 373, in validate_python return self.validator.validate_python(object, strict=strict, from_attributes=from_attributes, context=context) ^^^^^^^^^^^^^^ File "/Users/mcantrell/.pyenv/versions/3.11.9/lib/python3.11/functools.py", line 1001, in __get__ val = self.func(instance) ^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/pydantic/type_adapter.py", line 142, in wrapped return func(self, *args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/acme/fastapi-pydantic-error/.venv/lib/python3.11/site-packages/pydantic/type_adapter.py", line 318, in validator assert isinstance(self._validator, SchemaValidator) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ AssertionError ``` To fix the problem, either remove logfire completely or downgrade pydantic to v1. ### Python, Logfire & OS Versions, related packages (not required) ```TOML requests="2.32.3" pydantic="2.8.0" fastapi="0.111.0" protobuf="4.25.3" rich="13.7.1" executing="2.0.1" opentelemetry-api="1.25.0" opentelemetry-exporter-otlp-proto-common="1.25.0" opentelemetry-exporter-otlp-proto-http="1.25.0" opentelemetry-instrumentation="0.46b0" opentelemetry-instrumentation-asgi="0.46b0" opentelemetry-instrumentation-fastapi="0.46b0" opentelemetry-proto="1.25.0" opentelemetry-sdk="1.25.0" opentelemetry-semantic-conventions="0.46b0" opentelemetry-util-http="0.46b0" ```
1medium
Title: geemap.ee_to_geopandas allow more than 5,000 features Body: It seems that it is not possible to convert from EE to geopandas using the geemap.ee_to_geopandas function if there are more than 5,000 features. If so it fails with: Exception: Collection query aborted after accumulating over 5000 elements. Is it possible to allow more than 5,000 features?
1medium
Title: Suppress Matplotlib 2.0 warning Body: cf. #45
1medium
Title: Set video to main video in edit screen throws: 'File' object has no attribute 'content_type' Body: ## Steps to Reproduce <!-- Please include as many steps to reproduce so that we can replicate the problem. --> 1. Upload a video without setting it as main video 2. Edit the uploaded video 3. set it as main video 4. get this error message whole error message in html: [AttributeError at _de_exercise_video_677_1_edit.html.txt](https://github.com/wger-project/wger/files/8497979/AttributeError.at._de_exercise_video_677_1_edit.html.txt) PS: Aren't this too much sensible informations for a productivity error page? EDIT: I don't choose an video for replacing the old one.
1medium
Title: Latest requirements.txt seems to be over-specifying dependencies Body: This is the latest version requirements.txt file: ``` black>=19.3b0 darglint hypothesis>=4.4.0 interrogate ipykernel isort>=4.3.18 jupyter_client lxml natsort nbsphinx>=0.4.2 pandas-flavor pandas-vet pre-commit pyspark pytest-azurepipelines pytest-cov pytest>=3.4.2 scikit-learn seaborn setuptools>=38.5.2 sphinxcontrib-fulltoc==1.2.0 unyt xarray ``` Those are inject in the setup.py as mandatory dependencies but a lot in there looks like docs, dev, or optional. Is pyjanitor really require all that at run time?
1medium
Title: Request/Guidance - HowTo make using pywinauto easy Body: **This is not an issue - but a request** I'm a novice in python, perhaps I was still able to use pywinauto and do a good set of automation in the Unified Service Desk of Dynamics 365. Now that I struggle a lot to progress quickly to cover more scenarios, I was looking for the below: - A good tutorial with examples on how to use pywinauto libraries including how to use the uia_control wrappers - A utility similar to inpect.exe but generates the pywinauto code to the action that we do on the uia control - The same utility to give the hierarchical path of the uia control. because using the print_control_identifiers output to navigate the controls is way too difficult in complex application like USD - How to switch between wpf controls and html components @vasily-v-ryabov @airelil @enjoysmath
3misc
Title: Where are natures of pokemons? Body: not able to find nature a pokemon in the pokemon endpoint
3misc
Title: 【百度深度学习开源平台飞桨合作】 Body: 您好,我是百度飞桨的产品经理施依欣,看到transbigdata提供了较丰富的交通时空大数据分析能力,飞桨作为国内首个开源的深度学习平台,希望能不断深耕各行各业,因此想要和您进一步沟通,看看是否能够有深入结合/合作的机会,如果您方便的话可以添加我的微信(同电话):18108656919 期待您的回复~ 施依欣 | 百度飞桨产品经理
3misc
Title: Texttiling and paragraphs Body: Hello. I have a block of text that I want to segment. However I found out that the texttiling implementation requires that the input text is split into paragraphs. My questions is why does this happen? Doesn't the algorithm create pseudosentences and blocks on its own? After all, Hearst didn't rely on sentences and paragraphs to segment text.
1medium
Title: Add support to Julia notebooks Body: Please add option to convert Julia notebooks into standalone web apps.
1medium
Title: broken with httpx 0.10 Body: **Describe the bug** authlib uses old classes of httpx. New version of httpx 0.10 removed `AsyncRequest` `AsyncResponse` and also there's no middleware. **Environment:** - OS: Linux - Python Version: 3.7 - Authlib Version: 0.13
1medium
Title: [BUG] ERROR: Failed building wheel for duckdb Body: **Describe the bug** When attempting to install via `pip install pygwalker` I receive the following errors: `ERROR: Failed building wheel for duckdb` `ERROR: Could not build wheels for duckdb, which is required to install pyproject.toml-based projects` **To Reproduce** Steps to reproduce the behavior: 1. Run ` pip install pygwalker ` 2. See error **Expected behavior** The library should install without issue. **Screenshots** <img width="1513" alt="Screenshot 2024-01-19 at 8 06 33 PM" src="https://github.com/Kanaries/pygwalker/assets/7577457/65938774-af79-4325-a1b6-1aa4c1ee888f"> **Versions** - pygwalker version: 0.4.2 - python version: 3.12.1
1medium
Title: [🐛 BUG] Problem of intercompability between environment Docker/Un*x and Windows for path Body: ### What went wrong? 🤔 First, create a storage folder, ".taipy", built when running Taipy with Windows. Then, try to run your code in a Linux-based Docker with this ".taipy". This will create an issue for File Based Datanodes. ``` Traceback (most recent call last): File "/app/main.py", line 20, in <module> print(scenario.data.read()) ^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/data_node.py", line 411, in read return self.read_or_raise() ^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/data_node.py", line 402, in read_or_raise return self._read() ^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/pickle.py", line 99, in _read return self._read_from_path() ^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/pickle.py", line 105, in _read_from_path with open(path, "rb") as pf: ^^^^^^^^^^^^^^^^ FileNotFoundError: [Errno 2] No such file or directory: 'user_data\\pickles\\DATANODE_data_dec56298-e438-4fcd-a1d9-313ad7bfa33e.p' ``` ### Expected Behavior This should work on all supported versions of Taipy. ### How to reproduce Dockerfile ``` # Copyright 2021-2024 Avaiga Private Limited # # Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on # an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the # specific language governing permissions and limitations under the License. FROM python:3.11 WORKDIR /app # Install application dependencies. COPY src/requirements.txt . RUN pip install -r requirements.txt # Copy the application source code. COPY src . CMD ["taipy", "run", "--no-debug", "--no-reloader", "main.py", "--host", "0.0.0.0", "-P", "5000", "--experiment", "1", "--force"] ``` src/requirements.txt ``` taipy==4.0.1 ``` src/main.py ```python import taipy as tp from taipy import Config, Scope data_cfg = Config.configure_data_node( id="data", storage_type="pickle", scope=Scope.GLOBAL, default_data=1 ) scenario_cfg = Config.configure_scenario( id="scenario", additional_data_node_configs=[data_cfg], ) if __name__ == "__main__": tp.Orchestrator().run() scenarios = tp.get_scenarios() if len(scenarios) == 0: scenario = tp.create_scenario(scenario_cfg) else: scenario = scenarios[0] print(scenario.data.read()) ``` Go into `src` and run the main.py locally (Windows) with : ``` taipy run main.py --experiment 1 --force ``` This will create the `.taipy` and `user_data` folders inside of src. Then, go in the root folder and build and run the docker: ```bash docker build -t issue2261 . docker run -it -p 5000:5000issue2261 ``` Here is the error raised: ``` (dev) PS C:\Users\jacta\OneDrive\Bureau\taipy\Code\basic-demo> docker run -it -p 5000:5000 issue [2024-11-21 09:02:51.005][Taipy][INFO] Updating configuration with command-line arguments... [2024-11-21 09:02:51.005][Taipy][INFO] Managing application's version... [2024-11-21 09:02:51.014][Taipy][INFO] Checking application's version... [2024-11-21 09:02:51.015][Taipy][WARNING] tasks field of ScenarioConfig `scenario` is empty. [2024-11-21 09:02:51.015][Taipy][INFO] Blocking configuration update... [2024-11-21 09:02:51.015][Taipy][INFO] Starting job dispatcher... [2024-11-21 09:02:51.016][Taipy][INFO] Orchestrator service has been started. Traceback (most recent call last): File "/app/main.py", line 20, in <module> print(scenario.data.read()) ^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/data_node.py", line 411, in read return self.read_or_raise() ^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/data_node.py", line 402, in read_or_raise return self._read() ^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/pickle.py", line 99, in _read return self._read_from_path() ^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/taipy/core/data/pickle.py", line 105, in _read_from_path with open(path, "rb") as pf: ^^^^^^^^^^^^^^^^ FileNotFoundError: [Errno 2] No such file or directory: 'user_data\\pickles\\DATANODE_data_dec56298-e438-4fcd-a1d9-313ad7bfa33e.p' ``` ### Version of Taipy 4.0.1 ### Acceptance Criteria - [x] A unit test reproducing the bug is added. - [x] Any new code is covered by a unit tested. - [x] Check code coverage is at least 90%. - [ ] The bug reporter validated the fix. - [ ] Related issue(s) in taipy-doc are created for documentation and Release Notes are updated. ### Code of Conduct - [X] I have checked the [existing issues](https://github.com/Avaiga/taipy/issues?q=is%3Aissue+). - [ ] I am willing to work on this issue (optional)
1medium
Title: From future import annotations breaks lazy types Body: ## Describe the Bug `from __future__ import annotations` breaks type resolver which results in `TypeError: Model fields cannot be resolved. Unexpected type 'typing.Any'` I did not manage to track down what exactly causes this issue, but whenever I add `from __future__ import annotations` to the top of the file with strawberry types - strawberry type resolver breaks <!-- A clear and concise description of what the bug is. --> ## System Information - Operating system: - Strawberry version (if applicable): 0.235.2 ## Additional Context <!-- Add any other relevant information about the problem here. -->
2hard
Title: Add intensity_median to regionprops Body: ### Description: We have `intensity_min`, `intensity_mean` and `intensity_max` in `regionprops` already: https://github.com/scikit-image/scikit-image/blob/5305ae2ae1fa5d8c01a45e77bb0c44cbabfe1102/skimage/measure/_regionprops.py#L573-L585 What about adding `intensity_median` as well, backed by `np.median` in a similar way as the other props? I can open a pull request if there are no general objections against this idea.
1medium
Title: [Bug][V1]: allenai/OLMo-2-0325-32B-Instruct - unexpected keyword argument 'inputs_embeds' Body: ### Your current environment <details> INFO 03-17 17:57:21 [__init__.py:256] Automatically detected platform cuda. Collecting environment information... PyTorch version: 2.5.1+cu124 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.4 LTS (x86_64) GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 Clang version: 14.0.0-1ubuntu1.1 CMake version: version 3.30.4 Libc version: glibc-2.35 Python version: 3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0] (64-bit runtime) Python platform: Linux-6.5.0-35-generic-x86_64-with-glibc2.35 Is CUDA available: True CUDA runtime version: 12.5.82 CUDA_MODULE_LOADING set to: LAZY GPU models and configuration: GPU 0: NVIDIA H100 80GB HBM3 GPU 1: NVIDIA H100 80GB HBM3 GPU 2: NVIDIA H100 80GB HBM3 GPU 3: NVIDIA H100 80GB HBM3 GPU 4: NVIDIA H100 80GB HBM3 GPU 5: NVIDIA H100 80GB HBM3 GPU 6: NVIDIA H100 80GB HBM3 GPU 7: NVIDIA H100 80GB HBM3 Nvidia driver version: 555.42.02 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.2.1 /usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.2.1 HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 46 bits physical, 57 bits virtual Byte Order: Little Endian CPU(s): 128 On-line CPU(s) list: 0-127 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Platinum 8462Y+ CPU family: 6 Model: 143 Thread(s) per core: 2 Core(s) per socket: 32 Socket(s): 2 Stepping: 8 CPU max MHz: 4100.0000 CPU min MHz: 800.0000 BogoMIPS: 5600.00 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf tsc_known_freq pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb cat_l3 cat_l2 cdp_l3 invpcid_single intel_ppin cdp_l2 ssbd mba ibrs ibpb stibp ibrs_enhanced tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm rdt_a avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb intel_pt avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local split_lock_detect avx_vnni avx512_bf16 wbnoinvd dtherm ida arat pln pts hfi vnmi avx512vbmi umip pku ospke waitpkg avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg tme avx512_vpopcntdq la57 rdpid bus_lock_detect cldemote movdiri movdir64b enqcmd fsrm md_clear serialize tsxldtrk pconfig arch_lbr ibt amx_bf16 avx512_fp16 amx_tile amx_int8 flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 3 MiB (64 instances) L1i cache: 2 MiB (64 instances) L2 cache: 128 MiB (64 instances) L3 cache: 120 MiB (2 instances) NUMA node(s): 2 NUMA node0 CPU(s): 0-31,64-95 NUMA node1 CPU(s): 32-63,96-127 Vulnerability Gather data sampling: Not affected Vulnerability Itlb multihit: Not affected Vulnerability L1tf: Not affected Vulnerability Mds: Not affected Vulnerability Meltdown: Not affected Vulnerability Mmio stale data: Not affected Vulnerability Retbleed: Not affected Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; Enhanced / Automatic IBRS; IBPB conditional; RSB filling; PBRSB-eIBRS SW sequence; BHI BHI_DIS_S Vulnerability Srbds: Not affected Vulnerability Tsx async abort: Not affected Versions of relevant libraries: [pip3] mypy==1.11.1 [pip3] mypy-extensions==1.0.0 [pip3] numpy==1.26.4 [pip3] nvidia-cublas-cu12==12.4.5.8 [pip3] nvidia-cuda-cupti-cu12==12.4.127 [pip3] nvidia-cuda-nvrtc-cu12==12.4.127 [pip3] nvidia-cuda-runtime-cu12==12.4.127 [pip3] nvidia-cudnn-cu12==9.1.0.70 [pip3] nvidia-cufft-cu12==11.2.1.3 [pip3] nvidia-curand-cu12==10.3.5.147 [pip3] nvidia-cusolver-cu12==11.6.1.9 [pip3] nvidia-cusparse-cu12==12.3.1.170 [pip3] nvidia-ml-py==12.560.30 [pip3] nvidia-nccl-cu12==2.21.5 [pip3] nvidia-nvjitlink-cu12==12.4.127 [pip3] nvidia-nvtx-cu12==12.4.127 [pip3] onnx==1.14.1 [pip3] onnxruntime==1.18.1 [pip3] pyzmq==26.2.1 [pip3] sentence-transformers==3.2.1 [pip3] torch==2.5.1 [pip3] torchaudio==2.5.1 [pip3] torchvision==0.20.1 [pip3] transformers==4.48.3 [pip3] transformers-stream-generator==0.0.5 [pip3] triton==3.1.0 [pip3] tritonclient==2.51.0 [pip3] vector-quantize-pytorch==1.21.2 [conda] Could not collect ROCM Version: Could not collect Neuron SDK Version: N/A vLLM Version: 0.7.4.dev448+g977a16772.d20250314 vLLM Build Flags: CUDA Archs: Not Set; ROCm: Disabled; Neuron: Disabled GPU Topology: GPU0 GPU1 GPU2 GPU3 GPU4 GPU5 GPU6 GPU7 NIC0 NIC1 NIC2 NIC3 NIC4 NIC5 NIC6 NIC7 CPU AffinityNUMA Affinity GPU NUMA ID GPU0 X NV18 NV18 NV18 NV18 NV18 NV18 NV18 PIX NODE NODE NODE SYS SYS SYS SYS 0-31,64-95 0 N/A GPU1 NV18 X NV18 NV18 NV18 NV18 NV18 NV18 NODE PIX NODE NODE SYS SYS SYS SYS 0-31,64-95 0 N/A GPU2 NV18 NV18 X NV18 NV18 NV18 NV18 NV18 NODE NODE PIX NODE SYS SYS SYS SYS 0-31,64-95 0 N/A GPU3 NV18 NV18 NV18 X NV18 NV18 NV18 NV18 NODE NODE NODE PIX SYS SYS SYS SYS 0-31,64-95 0 N/A GPU4 NV18 NV18 NV18 NV18 X NV18 NV18 NV18 SYS SYS SYS SYS PIX NODE NODE NODE 32-63,96-127 1 N/A GPU5 NV18 NV18 NV18 NV18 NV18 X NV18 NV18 SYS SYS SYS SYS NODE PIX NODE NODE 32-63,96-127 1 N/A GPU6 NV18 NV18 NV18 NV18 NV18 NV18 X NV18 SYS SYS SYS SYS NODE NODE PIX NODE 32-63,96-127 1 N/A GPU7 NV18 NV18 NV18 NV18 NV18 NV18 NV18 X SYS SYS SYS SYS NODE NODE NODE PIX 32-63,96-127 1 N/A NIC0 PIX NODE NODE NODE SYS SYS SYS SYS X NODE NODE NODE SYS SYS SYS SYS NIC1 NODE PIX NODE NODE SYS SYS SYS SYS NODE X NODE NODE SYS SYS SYS SYS NIC2 NODE NODE PIX NODE SYS SYS SYS SYS NODE NODE X NODE SYS SYS SYS SYS NIC3 NODE NODE NODE PIX SYS SYS SYS SYS NODE NODE NODE X SYS SYS SYS SYS NIC4 SYS SYS SYS SYS PIX NODE NODE NODE SYS SYS SYS SYS X NODE NODE NODE NIC5 SYS SYS SYS SYS NODE PIX NODE NODE SYS SYS SYS SYS NODE X NODE NODE NIC6 SYS SYS SYS SYS NODE NODE PIX NODE SYS SYS SYS SYS NODE NODE X NODE NIC7 SYS SYS SYS SYS NODE NODE NODE PIX SYS SYS SYS SYS NODE NODE NODE X Legend: X = Self SYS = Connection traversing PCIe as well as the SMP interconnect between NUMA nodes (e.g., QPI/UPI) NODE = Connection traversing PCIe as well as the interconnect between PCIe Host Bridges within a NUMA node PHB = Connection traversing PCIe as well as a PCIe Host Bridge (typically the CPU) PXB = Connection traversing multiple PCIe bridges (without traversing the PCIe Host Bridge) PIX = Connection traversing at most a single PCIe bridge NV# = Connection traversing a bonded set of # NVLinks NIC Legend: NIC0: mlx5_0 NIC1: mlx5_1 NIC2: mlx5_2 NIC3: mlx5_3 NIC4: mlx5_4 NIC5: mlx5_5 NIC6: mlx5_6 NIC7: mlx5_7 CUDA_HOME=/usr/local/cuda CUDA_HOME=/usr/local/cuda CUDA_TOOLKIT_ROOT_DIR=/usr/local/cuda LD_LIBRARY_PATH=:/usr/local/cuda/lib64 NCCL_CUMEM_ENABLE=0 TORCHINDUCTOR_COMPILE_THREADS=1 CUDA_MODULE_LOADING=LAZY </details> ### 🐛 Describe the bug I am hitting the following ``` ERROR 03-17 17:56:39 [core.py:340] TypeError: Olmo2ForCausalLM.forward() got an unexpected keyword argument 'inputs_embeds' ``` When running `vllm serve allenai/OLMo-2-0325-32B-Instruct --tensor-parallel-size 2` in V1. V0 is fine: ``` ERROR 03-17 17:56:39 [core.py:340] EngineCore hit an exception: Traceback (most recent call last): ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/engine/core.py", line 332, in run_engine_core ERROR 03-17 17:56:39 [core.py:340] engine_core = EngineCoreProc(*args, **kwargs) ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/engine/core.py", line 287, in __init__ ERROR 03-17 17:56:39 [core.py:340] super().__init__(vllm_config, executor_class, log_stats) ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/engine/core.py", line 62, in __init__ ERROR 03-17 17:56:39 [core.py:340] num_gpu_blocks, num_cpu_blocks = self._initialize_kv_caches( ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/engine/core.py", line 121, in _initialize_kv_caches ERROR 03-17 17:56:39 [core.py:340] available_gpu_memory = self.model_executor.determine_available_memory() ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/executor/abstract.py", line 66, in determine_available_memory ERROR 03-17 17:56:39 [core.py:340] output = self.collective_rpc("determine_available_memory") ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/executor/multiproc_executor.py", line 133, in collective_rpc ERROR 03-17 17:56:39 [core.py:340] raise e ERROR 03-17 17:56:39 [core.py:340] File "/home/tms/vllm/vllm/v1/executor/multiproc_executor.py", line 122, in collective_rpc ERROR 03-17 17:56:39 [core.py:340] raise result ERROR 03-17 17:56:39 [core.py:340] TypeError: Olmo2ForCausalLM.forward() got an unexpected keyword argument 'inputs_embeds' ERROR 03-17 17:56:39 [core.py:340] CRITICAL 03-17 17:56:39 [core_client.py:269] Got fatal signal from worker processes, shutting down. See stack trace above for root cause issue. ``` ### Before submitting a new issue... - [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions.
1medium
Title: QString' object has no attribute 'rfind' Body: Ubuntu 16.04 python 2.7 qt4 when I load the image, the terminal show that "QString' object has no attribute 'rfind'.So ,I can't create the rectbox.
1medium
Title: Use of matplotlib's colormap viridis Body: When I use matplotlibs new colormap _viridis_ and add a colorbar in a figure, matplotlib2tikz returns an AssertionError, because _viridis_ is not an linear-segemented-colormap. ``` python import numpy as np import matplotlib.pyplot as plt import matplotlib.cm as cm import matplotlib2tikz x, y = np.meshgrid(np.linspace(0, 1), np.linspace(0, 1)) z = x**2 - y**2 plt.pcolormesh(x, y, z, cmap=cm.viridis) plt.colorbar() matplotlib2tikz.save('test.tex') ``` Is it possible to use pgfplots built-in colormap viridis for the colorbar specification with `colormap/viridis`, `point meta min=x` and `point meta max=y` in the axis options or is there a problem with different colormap-calculations/-interpolations?
1medium
Title: Validation runs during overfit, even when turned off Body: ### Bug description I am attempting to overfit a model for demonstration. I am using the CLI and I am using trainer.overfit_batches=.125 and trainer.limit_val_batches=0. If trainer.limit_val_batches=0 is run without the overfit_batches, the desired effect of turning off the validation dataloader and epoch is achieved. ### What version are you seeing the problem on? v2.1 ### How to reproduce the bug ```python python cli.py fit -c config.yaml --trainer.limit_val_batches=0 --trainer.overfit_batches=.125 ``` ``` ### Error messages and logs ``` No error message, undesired behavior ``` ### Environment <details> <summary>Current environment</summary> ``` #- Lightning Component (e.g. Trainer, LightningModule, LightningApp, LightningWork, LightningFlow): #- PyTorch Lightning Version (e.g., 1.5.0): 2.1 #- Lightning App Version (e.g., 0.5.2): N/A #- PyTorch Version (e.g., 2.0):2.1 #- Python version (e.g., 3.9): 3.10 #- OS (e.g., Linux): Linux #- CUDA/cuDNN version: 12.1 #- GPU models and configuration: A100 #- How you installed Lightning(`conda`, `pip`, source): PIP #- Running environment of LightningApp (e.g. local, cloud): SLURM ``` </details> ### More info _No response_ cc @borda @justusschock @awaelchli
1medium
Title: Pre prompt like GPT Body: Just like OPENAI library has parameters like prompt where we put instruction to it. For example: Act like a mychatbot who's name is Toranto. so it will act like Toranto. Can't we do the same with bardapi?
1medium
Title: 获取用户的粉丝数等数据 Body: 期望提供一个获取用户的粉丝数等数据的接口,谢谢大佬 <img width="499" alt="image" src="https://user-images.githubusercontent.com/16174533/211499289-88218d27-b738-449f-957c-39ea9436c617.png">
1medium
Title: [Proposal] Allowing environments with different Box action/observation space limits in a vector environment Body: ### Proposal https://github.com/Farama-Foundation/Gymnasium/blob/8333df8666811d1d0f87f1ca71803cc58bcf09c6/gymnasium/vector/sync_vector_env.py#L251-L266 https://github.com/Farama-Foundation/Gymnasium/blob/8333df8666811d1d0f87f1ca71803cc58bcf09c6/gymnasium/vector/async_vector_env.py#L576-L596 Currently, these methods checks if all the observation and action spaces in a vector environment are identical, and raises an error if they are not. I'm assuming this is the case because we want to ensure that we can stack the observations and actions into one numpy array. I'm proposing a change to allow differences in the observation and action spaces as long as the shapes are consistent (e.g. the values in the `low` and `high` portions of a Box space). The change can be implemented with an optional parameter to enable/disable it when creating the vector environments to preserve current default behaviours for now. ### Motivation I want to vectorize environments with different action space boundaries but the current implementation of vector environments does not allow for that. ### Pitch _No response_ ### Alternatives _No response_ ### Additional context _No response_ ### Checklist - [ ] I have checked that there is no similar [issue](https://github.com/Farama-Foundation/Gymnasium/issues) in the repo
1medium
Title: seaborn objects scale with two visualisations with same kwargs? Body: Hello, I ran into a problem with scale, when I'm trying to display two visualizations with color mapped by column. I'm trying to create bar plot with labels on bars. Position of labels and color of labels depends on column of dataframe. Also, I would like to color bars by column. here is my question on stack overflow: https://stackoverflow.com/questions/75161245/how-to-use-seaborn-objects-scale-with-two-visualisations-with-same-kwargs Is there a way to do this? Thank you for answer
1medium
Title: Is it possible to load my own quantized model from local Body: Here is the code I tried for the coreference resolution ``` from allennlp.predictors.predictor import Predictor model_url = 'https://storage.googleapis.com/pandora-intelligence/models/crosslingual-coreference/minilm/model.tar.gz' predictor = Predictor.from_path(model_url) text = "Eva and Martha didn't want their friend Jenny \ to feel lonely so they invited her to the party." prediction = predictor.predict(document=text) print(prediction['clusters']) print(predictor.coref_resolved(text)) ``` And it worked well I got the output with solved coreference. like below ``` Eva and Martha didn't want Eva and Martha's friend Jenny to feel lonely so Eva and Martha invited their friend Jenny to the party. ``` Now I have quantized the model used here (https://storage.googleapis.com/pandora-intelligence/models/crosslingual-coreference/minilm/model.tar.gz) and the new quantized model is stored in a specific path in my local machine. Shall I use that customized(quantized) model from my local path in ```model_url``` value and use this prediction command like below? ``` model_url = <Path to the quantized model in my local machine> predictor = Predictor.from_path(model_url) ```
1medium
Title: Day 10 Exercises level 3 Body: I'm trying to solve exercise number 3, but I can't iterate through the file: `countries_data.py` I got this error message: `non iterable value is used in an iterating context` any ideas that could help me to solve this issue? and also How can I iterate a list of dictionaries like the one in `countries_data.py`. Thanks in advance!
1medium
Title: can I run the code in windows and without GPU Body: can I run the code in windows and without GPU?
0easy