text
stringlengths 20
57.3k
| labels
class label 4
classes |
---|---|
Title: payjs支付宝用不了
Body: payjs支付宝用不了,获取不到二维码
| 0easy
|
Title: [BUG] Adding new page ignores setting CMS_DEFAULT_IN_NAVIGATION
Body: ## Description
When someone adds a new page, the setting parameter `CMS_DEFAULT_IN_NAVIGATION` is ignored.
## Steps to reproduce
* Add or change `CMS_DEFAULT_IN_NAVIGATION = False` in your project's `settings.py`.
* Add a new page.
## Expected behaviour
The checkbox "Menu" for the new page in the page tree should be unchecked.
## Actual behaviour
The checkbox "Menu" for the new page in the page tree is checked.
## Additional information (CMS/Python/Django versions)
Django-CMS version 4.1
## Do you want to help fix this issue?
I have a [patch](https://github.com/jrief/django-cms/commit/5365d291838136ac0c00d3c0a1e1e92551e1bb83).
* [x] Yes, I want to help fix this issue and I will join #workgroup-pr-review on [Slack](https://www.django-cms.org/slack) to confirm with the community that a PR is welcome.
* [ ] No, I only want to report the issue.
| 0easy
|
Title: Make `byteswap` an nplike function
Body: ### Version of Awkward Array
main
### Description and code to reproduce
`to_buffers` expects the array object to have a `byteswap` method, which we aren't supported across all ndarray implementations. We should add this to our nplikes instead. | 0easy
|
Title: variable assingment order inconsistency when using file| in set_env
Body: When not using `file|` in `set_env`, the variables are assigned in order (with the last value replacing any previous one, as in `FOO` being overwriten with `QUX` in my `example1`).
However, this behaviour is not consistent when using `file|`, when assingment of variables from `file` always take precedence (see `example2`, where `FOO` gets the value `BAR` from `.env`, and is not overwriten by subsequent assignment of `QUX`).
Expected:
- consistent beavhiour, so that the last assingment of variable takes precedence, regardles if the previous one was set directly or using `file|`
- or at least document the priority of variables when using `file|`.
_tox.ini:_
```
[tox]
env_list = example1,example2
[testenv]
allowlist_externals = bash
commands =
bash -c 'env | egrep "FOO"'
[testenv:example1]
set_env =
FOO=BAZ
FOO=QUX
[testenv:example2]
set_env =
FOO=BAZ
file|.env
FOO=QUX
```
_.env:_
```
FOO=BAR
```
_Output:_
```
$ tox -v
4.18.0
$ tox -e example1
FOO=QUX
$ tox -e example2
FOO=BAR
``` | 0easy
|
Title: add VARIMA, ... probabilistic support to readme
Body: | 0easy
|
Title: support loading documents with `from_files` signature
Body: **Is your feature request related to a problem? Please describe.**
In docarray v1, we support:
```python
da = DocumentArray.from_files('imgs/*.jpg')
```
While it is missing in the current v2. Could be very useful for us to loading the documents from path.
| 0easy
|
Title: [Feature request] Add apply_to_images to PlankianJitter
Body: | 0easy
|
Title: Add a --version switch to the command line interface
Body: ## Classification:
Enhancement
## Summary
Add a `-V` and `--version` command line switch that prints the current version.
Use the `argparse` version action to implement it.
## Version
AutoKey version: <=0.95.3
Used GUI (Gtk, Qt, or both): both
| 0easy
|
Title: tox 3.27.1 to 4.0.9, environment var substitution failing
Body: ## Issue
All of our build pipelines are failing due to what appears to be a bug with environment variable substitution in our tox.ini. The variable substitution works as expected with tox==3.27.1, but we start having issues with tox==4.0.9.
## Environment
Provide at least:
- OS: Windows
- `pip list` of the host Python where `tox` is installed:
```console
(.venv) C:\tmp\tox_4.0.9_bug>pip list
Package Version
------------- -------
cachetools 5.2.0
chardet 5.1.0
colorama 0.4.6
distlib 0.3.6
filelock 3.8.2
packaging 22.0
pip 22.3.1
platformdirs 2.6.0
pluggy 1.0.0
pyproject_api 1.2.1
setuptools 65.6.3
tomli 2.0.1
tox 4.0.9
virtualenv 20.17.1
wheel 0.38.4
```
## Output of running tox
Provide the output of `tox -rvv`:
```console
(.venv) C:\tmp\tox_4.0.9_bug>tox -rvv
default: 2108 I find interpreter for spec PythonSpec(path=C:\tmp\tox_4.0.9_bug\.venv\Scripts\python.exe) [virtualenv\discovery\builtin.py:56]
default: 2134 D got python info of C:\Users\nws2293\AppData\Local\Programs\Python\Python39\python.exe from C:\Users\nws2293\AppData\Local\pypa\virtualenv\py_info\1\72aa99cb99a91779d7fa96aaf09f6d3b1207abe0138f99ab0af57c8a95fa9102.json [virtualenv\app_data\via_disk_folder.py:129]
default: 2173 I proposed PythonInfo(spec=CPython3.9.10.final.0-64, system=C:\Users\nws2293\AppData\Local\Programs\Python\Python39\python.exe, exe=C:\tmp\tox_4.0.9_bug\.venv\Scripts\python.exe, platform=win32, version='3.9.10 (tags/v3.9.10:f2f3f53, Jan 17 2022, 15:14:21) [MSC v.1929 64 bit (AMD64)]', encoding_fs_io=utf-8-cp1252) [virtualenv\discovery\builtin.py:63]
default: 2174 D accepted PythonInfo(spec=CPython3.9.10.final.0-64, system=C:\Users\nws2293\AppData\Local\Programs\Python\Python39\python.exe, exe=C:\tmp\tox_4.0.9_bug\.venv\Scripts\python.exe, platform=win32, version='3.9.10 (tags/v3.9.10:f2f3f53, Jan 17 2022, 15:14:21) [MSC v.1929 64 bit (AMD64)]', encoding_fs_io=utf-8-cp1252) [virtualenv\discovery\builtin.py:65]
default: 2203 D symlink on filesystem does not work [virtualenv\info.py:43]
default: 2207 D filesystem is not case-sensitive [virtualenv\info.py:24]
default: 2420 I create virtual environment via CPython3Windows(dest=C:\tmp\tox_4.0.9_bug\.tox\default, clear=False, no_vcs_ignore=False, global=False) [virtualenv\run\session.py:48]
default: 2422 D create folder C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages [virtualenv\util\path\_sync.py:9]
default: 2427 D create folder C:\tmp\tox_4.0.9_bug\.tox\default\Scripts [virtualenv\util\path\_sync.py:9]
default: 2429 D write C:\tmp\tox_4.0.9_bug\.tox\default\pyvenv.cfg [virtualenv\create\pyenv_cfg.py:30]
default: 2430 D home = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 2431 D implementation = CPython [virtualenv\create\pyenv_cfg.py:34]
default: 2431 D version_info = 3.9.10.final.0 [virtualenv\create\pyenv_cfg.py:34]
default: 2432 D virtualenv = 20.17.1 [virtualenv\create\pyenv_cfg.py:34]
default: 2433 D include-system-site-packages = false [virtualenv\create\pyenv_cfg.py:34]
default: 2434 D base-prefix = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 2434 D base-exec-prefix = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 2434 D base-executable = C:\Users\nws2293\AppData\Local\Programs\Python\Python39\python.exe [virtualenv\create\pyenv_cfg.py:34]
default: 2439 D copy C:\Users\nws2293\AppData\Local\Programs\Python\Python39\Lib\venv\scripts\nt\python.exe to C:\tmp\tox_4.0.9_bug\.tox\default\Scripts\python.exe [virtualenv\util\path\_sync.py:36]
default: 2448 D copy C:\Users\nws2293\AppData\Local\Programs\Python\Python39\Lib\venv\scripts\nt\pythonw.exe to C:\tmp\tox_4.0.9_bug\.tox\default\Scripts\pythonw.exe [virtualenv\util\path\_sync.py:36]
default: 2466 D create virtualenv import hook file C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\_virtualenv.pth [virtualenv\create\via_global_ref\api.py:89]
default: 2473 D create C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\_virtualenv.py [virtualenv\create\via_global_ref\api.py:92]
default: 2480 D ============================== target debug ============================== [virtualenv\run\session.py:50]
default: 2481 D debug via 'C:\tmp\tox_4.0.9_bug\.tox\default\Scripts\python.exe' 'C:\tmp\tox_4.0.9_bug\.venv\lib\site-packages\virtualenv\create\debug.py' [virtualenv\create\creator.py:197]
default: 2480 D {
"sys": {
"executable": "C:\\tmp\\tox_4.0.9_bug\\.tox\\default\\Scripts\\python.exe",
"_base_executable": "C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39\\python.exe",
"prefix": "C:\\tmp\\tox_4.0.9_bug\\.tox\\default",
"base_prefix": "C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39",
"real_prefix": null,
"exec_prefix": "C:\\tmp\\tox_4.0.9_bug\\.tox\\default",
"base_exec_prefix": "C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39",
"path": [
"C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39\\python39.zip",
"C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39\\DLLs",
"C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39\\lib",
"C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39",
"C:\\tmp\\tox_4.0.9_bug\\.tox\\default",
"C:\\tmp\\tox_4.0.9_bug\\.tox\\default\\lib\\site-packages"
],
"meta_path": [
"<class '_virtualenv._Finder'>",
"<class '_frozen_importlib.BuiltinImporter'>",
"<class '_frozen_importlib.FrozenImporter'>",
"<class '_frozen_importlib_external.PathFinder'>"
],
"fs_encoding": "utf-8",
"io_encoding": "cp1252"
},
"version": "3.9.10 (tags/v3.9.10:f2f3f53, Jan 17 2022, 15:14:21) [MSC v.1929 64 bit (AMD64)]",
"makefile_filename": "C:\\Users\\nws2293\\AppData\\Local\\Programs\\Python\\Python39\\Lib\\config\\Makefile",
"os": "<module 'os' from 'C:\\\\Users\\\\nws2293\\\\AppData\\\\Local\\\\Programs\\\\Python\\\\Python39\\\\lib\\\\os.py'>",
"site": "<module 'site' from 'C:\\\\Users\\\\nws2293\\\\AppData\\\\Local\\\\Programs\\\\Python\\\\Python39\\\\lib\\\\site.py'>",
"datetime": "<module 'datetime' from 'C:\\\\Users\\\\nws2293\\\\AppData\\\\Local\\\\Programs\\\\Python\\\\Python39\\\\lib\\\\datetime.py'>",
"math": "<module 'math' (built-in)>",
"json": "<module 'json' from 'C:\\\\Users\\\\nws2293\\\\AppData\\\\Local\\\\Programs\\\\Python\\\\Python39\\\\lib\\\\json\\\\__init__.py'>"
} [virtualenv\run\session.py:51]
default: 3034 I add seed packages via FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=C:\Users\nws2293\AppData\Local\pypa\virtualenv) [virtualenv\run\session.py:55]
default: 3051 D got embed update of distribution setuptools from C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\embed\3\setuptools.json [virtualenv\app_data\via_disk_folder.py:129]
default: 3052 D got embed update of distribution pip from C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\embed\3\pip.json [virtualenv\app_data\via_disk_folder.py:129]
default: 3053 D got embed update of distribution wheel from C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\embed\3\wheel.json [virtualenv\app_data\via_disk_folder.py:129]
default: 3060 D install setuptools from wheel C:\tmp\tox_4.0.9_bug\.venv\lib\site-packages\virtualenv\seed\wheels\embed\setuptools-65.6.3-py3-none-any.whl via CopyPipInstall [virtualenv\seed\embed\via_app_data\via_app_data.py:47]
default: 3062 D install pip from wheel C:\tmp\tox_4.0.9_bug\.venv\lib\site-packages\virtualenv\seed\wheels\embed\pip-22.3.1-py3-none-any.whl via CopyPipInstall [virtualenv\seed\embed\via_app_data\via_app_data.py:47]
default: 3063 D install wheel from wheel C:\tmp\tox_4.0.9_bug\.venv\lib\site-packages\virtualenv\seed\wheels\embed\wheel-0.38.4-py3-none-any.whl via CopyPipInstall [virtualenv\seed\embed\via_app_data\via_app_data.py:47]
default: 3077 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\pip-22.3.1-py3-none-any\pip to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\pip [virtualenv\util\path\_sync.py:36]
default: 3078 D copy C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\distutils-precedence.pth to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\distutils-precedence.pth [virtualenv\util\path\_sync.py:36]
default: 3080 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\wheel-0.38.4-py3-none-any\wheel to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\wheel [virtualenv\util\path\_sync.py:36]
default: 3094 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\pkg_resources to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\pkg_resources [virtualenv\util\path\_sync.py:36]
default: 3399 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\wheel-0.38.4-py3-none-any\wheel-0.38.4.dist-info to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\wheel-0.38.4.dist-info [virtualenv\util\path\_sync.py:36]
default: 3515 D copy C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\wheel-0.38.4-py3-none-any\wheel-0.38.4.virtualenv to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\wheel-0.38.4.virtualenv [virtualenv\util\path\_sync.py:36]
default: 3583 D generated console scripts wheel3.9.exe wheel-3.9.exe wheel3.exe wheel.exe [virtualenv\seed\embed\via_app_data\pip_install\base.py:41]
default: 3911 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\setuptools to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\setuptools [virtualenv\util\path\_sync.py:36]
default: 6578 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\setuptools-65.6.3.dist-info to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\setuptools-65.6.3.dist-info [virtualenv\util\path\_sync.py:36]
default: 6686 D copy C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\setuptools-65.6.3.virtualenv to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\setuptools-65.6.3.virtualenv [virtualenv\util\path\_sync.py:36]
default: 6693 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\setuptools-65.6.3-py3-none-any\_distutils_hack to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\_distutils_hack [virtualenv\util\path\_sync.py:36]
default: 6737 D generated console scripts [virtualenv\seed\embed\via_app_data\pip_install\base.py:41]
default: 10295 D copy directory C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\pip-22.3.1-py3-none-any\pip-22.3.1.dist-info to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\pip-22.3.1.dist-info [virtualenv\util\path\_sync.py:36]
default: 10382 D copy C:\Users\nws2293\AppData\Local\pypa\virtualenv\wheel\3.9\image\1\CopyPipInstall\pip-22.3.1-py3-none-any\pip-22.3.1.virtualenv to C:\tmp\tox_4.0.9_bug\.tox\default\Lib\site-packages\pip-22.3.1.virtualenv [virtualenv\util\path\_sync.py:36]
default: 10413 D generated console scripts pip.exe pip3.9.exe pip-3.9.exe pip3.exe [virtualenv\seed\embed\via_app_data\pip_install\base.py:41]
default: 10413 I add activators for Bash, Batch, Fish, Nushell, PowerShell, Python [virtualenv\run\session.py:61]
default: 10513 D write C:\tmp\tox_4.0.9_bug\.tox\default\pyvenv.cfg [virtualenv\create\pyenv_cfg.py:30]
default: 10513 D home = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 10514 D implementation = CPython [virtualenv\create\pyenv_cfg.py:34]
default: 10514 D version_info = 3.9.10.final.0 [virtualenv\create\pyenv_cfg.py:34]
default: 10514 D virtualenv = 20.17.1 [virtualenv\create\pyenv_cfg.py:34]
default: 10515 D include-system-site-packages = false [virtualenv\create\pyenv_cfg.py:34]
default: 10515 D base-prefix = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 10516 D base-exec-prefix = C:\Users\nws2293\AppData\Local\Programs\Python\Python39 [virtualenv\create\pyenv_cfg.py:34]
default: 10516 D base-executable = C:\Users\nws2293\AppData\Local\Programs\Python\Python39\python.exe [virtualenv\create\pyenv_cfg.py:34]
default: 10534 W commands[0]> pytest [tox\tox_env\api.py:417]
default: 10713 E failed with pytest is not allowed, use allowlist_externals to allow it [tox\session\cmd\run\single.py:54]
default: FAIL code 1 (8.64 seconds)
evaluation failed :( (9.83 seconds)
```
## Minimal example
This is a stripped down tox.ini file, "C:\tmp\tox_4.0.9_bug\tox.ini".
```ini
[tox]
minversion = 3.24
envlist = default
isolated_build = True
[testenv]
description = Invoke pytest to run automated tests
passenv =
HOME
SETUPTOOLS_*
setenv =
TOXINIDIR = {toxinidir}
[testenv:{docs}]
description =
docs: Pretend to invoke sphinx-build to build the docs
skip_install = True
passenv =
SETUPTOOLS_*
setenv =
DOCSDIR = {toxinidir}{/}docs
BUILDDIR = {toxinidir}{/}docs{/}_build
docs: BUILD = html
commands =
; sphinx-build --color -b {env:BUILD} -d "{env:BUILDDIR}{/}doctrees" "{env:DOCSDIR}" "{env:BUILDDIR}{/}{env:BUILD}" {posargs}
python -c 'print("{env:BUILD} -- {env:BUILDDIR}{/}{env:BUILD}")'
```
If I run "tox -e docs" with tox 4.0.9, this is the output. Notice that the substitution of `{env:BUILD}` works in the first part of the line but not in the second part following the `{/}`:
```console
(.venv) C:\tmp\tox_4.0.9_bug>tox -e docs
docs: commands[0]> python -c "print(\"html -- C:\tmp\tox_4.0.9_bug\docs\_build{env:BUILD}\")"
html -- C: mp ox_4.0.9_bug\docs\_build{env:BUILD}
docs: OK (5.73=setup[5.16]+cmd[0.58] seconds)
congratulations :) (6.73 seconds)
```
If I run "tox -e docs" with tox 3.27.1, this is the output:
```console
C:\tmp\tox_4.0.9_bug>tox -e docs
docs run-test-pre: PYTHONHASHSEED='297'
docs run-test: commands[0] | python -c 'print("html -- C:\tmp\tox_4.0.9_bug\docs\_build\html")'
html -- C: mp ox_4.0.9_bug\docs\_build\html
docs: commands succeeded
congratulations :)
```
| 0easy
|
Title: Add Docker support
Body: It will help people run the app without LaTeX issues | 0easy
|
Title: Alternate image publish location (vs docker.io)
Body: ### What docker image(s) is this feature applicable to?
base-notebook
### What changes are you proposing?
Publish JupyterHub to an additional - or alternate - location
See https://web.docker.com/rs/790-SSB-375/images/privatereposfaq.pdf
- Docker has recently announced that 'free' organization repositories are going away in April. This will affect some open source projects that publish to dockerhub. I do not know if it affects Jupyter lab, or if the team has a paid account
- The docker faq does state that open source projects can apply to get a free accouunt
- Some projects publish to multiple repos. One option is 'quay.io'. I also know nginx use amazon. Even if docker.io is good for now, it may be worth considering publishing to an alternate repo
Ref: https://github.com/odpi/egeria/issues/7530
### How does this affect the user?
User would be unable to access a prebuilt docker stacks image
### Anything else?
_No response_ | 0easy
|
Title: 3.2.1 sphinx warnings `reference target not found`
Body: Firs of all currently it is npt possible to use spraight `sphinx-build` because missing path to module in copy.py
```console
+ /usr/bin/sphinx-build -n -T -b man docs build/sphinx/man
Running Sphinx v4.5.0
Traceback (most recent call last):
File "/usr/lib/python3.8/site-packages/sphinx/config.py", line 332, in eval_config_file
exec(code, namespace)
File "/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/conf.py", line 12, in <module>
import factory
ModuleNotFoundError: No module named 'factory'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/lib/python3.8/site-packages/sphinx/cmd/build.py", line 272, in build_main
app = Sphinx(args.sourcedir, args.confdir, args.outputdir,
File "/usr/lib/python3.8/site-packages/sphinx/application.py", line 202, in __init__
self.config = Config.read(self.confdir, confoverrides or {}, self.tags)
File "/usr/lib/python3.8/site-packages/sphinx/config.py", line 165, in read
namespace = eval_config_file(filename, tags)
File "/usr/lib/python3.8/site-packages/sphinx/config.py", line 345, in eval_config_file
raise ConfigError(msg % traceback.format_exc()) from exc
sphinx.errors.ConfigError: There is a programmable error in your configuration file:
Traceback (most recent call last):
File "/usr/lib/python3.8/site-packages/sphinx/config.py", line 332, in eval_config_file
exec(code, namespace)
File "/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/conf.py", line 12, in <module>
import factory
ModuleNotFoundError: No module named 'factory'
Configuration error:
There is a programmable error in your configuration file:
Traceback (most recent call last):
File "/usr/lib/python3.8/site-packages/sphinx/config.py", line 332, in eval_config_file
exec(code, namespace)
File "/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/conf.py", line 12, in <module>
import factory
ModuleNotFoundError: No module named 'factory'
```
Thi can be fixed by patch like below
```patch
--- a/docs/conf.py~ 2021-10-26 14:02:25.000000000 +0000
+++ b/docs/conf.py 2022-05-14 17:15:28.667904146 +0000
@@ -9,13 +9,12 @@
import os
import sys
-import factory
-
# If extensions (or modules to document with autodoc) are in another directory,
# add these directories to sys.path here. If the directory is relative to the
# documentation root, use os.path.abspath to make it absolute, like shown here.
-sys.path.insert(0, os.path.dirname(os.path.abspath('.')))
+sys.path.insert(0, os.path.abspath('..'))
+import factory
# -- Project information -----------------------------------------------------
```
Tan .. on building my packages I'm using `sphinx-build` command with `-n` switch which shows warmings about missing references. These are not critical issues.
Here is the output with warnings:
```console
+ /usr/bin/sphinx-build -n -T -b man docs build/sphinx/man
Running Sphinx v4.5.0
making output directory... done
loading intersphinx inventory from https://docs.python.org/objects.inv...
loading intersphinx inventory from https://docs.djangoproject.com/en/dev/_objects/...
loading intersphinx inventory from https://docs.sqlalchemy.org/en/latest/objects.inv...
intersphinx inventory has moved: https://docs.python.org/objects.inv -> https://docs.python.org/3/objects.inv
intersphinx inventory has moved: https://docs.sqlalchemy.org/en/latest/objects.inv -> https://docs.sqlalchemy.org/en/14/objects.inv
building [mo]: targets for 0 po files that are out of date
building [man]: all manpages
updating environment: [new config] 11 added, 0 changed, 0 removed
reading sources... [100%] reference
looking for now-outdated files... none found
pickling environment... done
checking consistency... done
writing... python-factory-boy.3 { introduction reference orms recipes fuzzy examples internals changelog credits ideas } /home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/introduction.rst:107: WARNING: py:meth reference target not found: factory.@sequence
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/introduction.rst:145: WARNING: py:meth reference target not found: factory.@lazy_attribute
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/introduction.rst:179: WARNING: py:meth reference target not found: factory.@lazy_attribute
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:206: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:216: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:226: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:231: WARNING: py:obj reference target not found: strategy
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:235: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:244: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:599: WARNING: py:obj reference target not found: strategy
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:668: WARNING: py:class reference target not found: file
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:835: WARNING: py:meth reference target not found: LazyAttribute.lazy_attribute
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:868: WARNING: py:class reference target not found: builder.Resolver
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:1127: WARNING: py:attr reference target not found: factory.SubFactory.factory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:1334: WARNING: py:attr reference target not found: builder.Resolver.factory_parent
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:1788: WARNING: py:attr reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2008: WARNING: py:mod reference target not found: factory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2016: WARNING: py:class reference target not found: factory.base.Factory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2024: WARNING: py:obj reference target not found: klass
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2024: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2034: WARNING: py:class reference target not found: class
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2038: WARNING: py:obj reference target not found: klass
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2038: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2048: WARNING: py:class reference target not found: class
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2052: WARNING: py:obj reference target not found: klass
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2052: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2062: WARNING: py:class reference target not found: class
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2066: WARNING: py:obj reference target not found: klass
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2066: WARNING: py:obj reference target not found: strategy
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2066: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2077: WARNING: py:class reference target not found: class
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2081: WARNING: py:obj reference target not found: klass
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2081: WARNING: py:obj reference target not found: size
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2091: WARNING: py:class reference target not found: class
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2100: WARNING: py:mod reference target not found: factory.random
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/reference.rst:2116: WARNING: py:meth reference target not found: random.seed
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:129: WARNING: py:class reference target not found: file
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:129: WARNING: py:class reference target not found: func
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:166: WARNING: py:class reference target not found: file
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:166: WARNING: py:class reference target not found: func
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:270: WARNING: py:attr reference target not found: associated class <factory.FactoryOptions.model
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:270: WARNING: py:class reference target not found: mongoengine.EmbeddedDocument
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:270: WARNING: py:meth reference target not found: MongoEngineFactory.create
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:322: WARNING: py:meth reference target not found: sqlalchemy.orm.session.Session.add
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:327: WARNING: py:class reference target not found: factory.base.FactoryOptions
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:342: WARNING: py:meth reference target not found: sqlalchemy.orm.session.Session.flush
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:343: WARNING: py:meth reference target not found: sqlalchemy.orm.session.Session.commit
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:351: WARNING: py:meth reference target not found: sqlalchemy.orm.query.Query.one_or_none
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/orms.rst:351: WARNING: py:meth reference target not found: sqlalchemy.orm.session.Session.add
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/recipes.rst:90: WARNING: py:class reference target not found: UserFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/recipes.rst:102: WARNING: py:meth reference target not found: factory.Factory._generate
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/recipes.rst:213: WARNING: py:class reference target not found: RelatedFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/recipes.rst:213: WARNING: py:class reference target not found: RelatedFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/fuzzy.rst:364: WARNING: py:obj reference target not found: factory.random.randgen
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/fuzzy.rst:364: WARNING: py:meth reference target not found: get_random_state
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/examples.rst:80: WARNING: py:class reference target not found: Account
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/examples.rst:80: WARNING: py:class reference target not found: Profile
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/internals.rst:56: WARNING: py:meth reference target not found: Factory.__call__
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/internals.rst:60: WARNING: py:meth reference target not found: Factory._generate
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/internals.rst:62: WARNING: py:meth reference target not found: Factory._generate_batch`
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/internals.rst:64: WARNING: py:meth reference target not found: Factory._generate
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/internals.rst:84: WARNING: py:meth reference target not found: FactoryOptions.instantiate
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:17: WARNING: py:meth reference target not found: factory.alchemy.SQLAlchemyModelFactory._save
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:17: WARNING: py:meth reference target not found: factory.alchemy.SQLAlchemyModelFactory._get_or_create
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:103: WARNING: py:attr reference target not found: factory.Factory.model
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:105: WARNING: py:meth reference target not found: get_random_state
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:110: WARNING: py:class reference target not found: factory.declarations.SubFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:128: WARNING: py:attr reference target not found: factory.fuzzy.FuzzyChoice.getter
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:130: WARNING: py:meth reference target not found: factory.faker.Faker.generate
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:132: WARNING: py:class reference target not found: factory.faker.Faker
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:154: WARNING: py:class reference target not found: factory.declarations.ParameteredAttribute
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:169: WARNING: py:class reference target not found: factory.PostGenerationDeclaration
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:291: WARNING: py:func reference target not found: django.apps.apps.get_model
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:292: WARNING: py:meth reference target not found: mogo.model.Model.new
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:401: WARNING: py:attr reference target not found: factory.Factory.FACTORY_FOR
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:402: WARNING: py:attr reference target not found: factory.Factory.ABSTRACT_FACTORY
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:403: WARNING: py:attr reference target not found: factory.Factory.FACTORY_STRATEGY
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:404: WARNING: py:attr reference target not found: factory.Factory.FACTORY_ARG_PARAMETERS
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:405: WARNING: py:attr reference target not found: factory.Factory.FACTORY_HIDDEN_ARGS
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:409: WARNING: py:attr reference target not found: factory.django.DjangoModelFactory.FACTORY_DJANGO_GET_OR_CREATE
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:413: WARNING: py:attr reference target not found: factory.alchemy.SQLAlchemyModelFactory.FACTORY_SESSION
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:435: WARNING: py:class reference target not found: mongoengine.EmbeddedDocument
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:454: WARNING: py:mod reference target not found: factory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:475: WARNING: py:class reference target not found: factory.Factory.ABSTRACT_FACTORY
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:475: WARNING: py:class reference target not found: factory.Factory.FACTORY_FOR
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:498: WARNING: py:attr reference target not found: factory.builder.Resolver.factory_parent
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:498: WARNING: py:class reference target not found: factory.builder.Resolver
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:501: WARNING: py:mod reference target not found: factory.django
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:501: WARNING: py:mod reference target not found: factory.mogo
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:530: WARNING: py:attr reference target not found: factory.django.DjangoModelFactory.FACTORY_DJANGO_GET_OR_CREATE
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:541: WARNING: py:attr reference target not found: factory.django.DjangoModelFactory.FACTORY_DJANGO_GET_OR_CREATE
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:554: WARNING: py:attr reference target not found: factory.Sequence.type
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:555: WARNING: py:attr reference target not found: factory.Factory.FACTORY_HIDDEN_ARGS
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:557: WARNING: py:attr reference target not found: factory.django.DjangoModelFactory.FACTORY_DJANGO_GET_OR_CREATE
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:566: WARNING: py:class reference target not found: factory.InfiniteIterator
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:566: WARNING: py:func reference target not found: factory.infinite_iterator
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:567: WARNING: py:class reference target not found: factory.CircularSubFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:571: WARNING: py:meth reference target not found: factory.Factory.set_building_function
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:571: WARNING: py:meth reference target not found: factory.Factory.set_creation_function
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:594: WARNING: py:attr reference target not found: factory.Factory.FACTORY_ARG_PARAMETERS
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:596: WARNING: py:attr reference target not found: factory.Factory.FACTORY_FOR
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:617: WARNING: py:class reference target not found: factory.InfiniteIterator
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:618: WARNING: py:class reference target not found: factory.CircularSubFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:622: WARNING: py:meth reference target not found: factory.Factory.set_creation_function
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:622: WARNING: py:meth reference target not found: factory.Factory.set_building_function
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:624: WARNING: py:attr reference target not found: factory.Factory.FACTORY_FOR
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:661: WARNING: py:class reference target not found: factory.CircularSubFactory
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:671: WARNING: py:class reference target not found: factory.PostGenerationDeclaration
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:705: WARNING: py:class reference target not found: factory.InfiniteIterator
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:728: WARNING: py:class reference target not found: factory.ContainerAttribute
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:738: WARNING: py:attr reference target not found: factory.Factory.FACTORY_FOR
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:751: WARNING: py:data reference target not found: factory.MOGO_BUILD
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/changelog.rst:753: WARNING: py:attr reference target not found: factory.Factory.ABSTRACT_FACTORY
/home/tkloczko/rpmbuild/BUILD/factory_boy-3.2.1/docs/ideas.rst:7: WARNING: py:class reference target not found: Factory
done
build succeeded, 114 warnings.
```
You can peak on fixes that kind of issues in other projects
https://github.com/latchset/jwcrypto/pull/289
https://github.com/click-contrib/sphinx-click/commit/abc31069
| 0easy
|
Title: TTM Squeeze value results are wrong
Body: I am using the TTM Squeeze indicator via tradingview. The lazybear version of the squeeze indicator is exactly the same as the values in tradingview, but I would like to point out that these values are incorrect for "squeeze_pro" in Pandas_ta.

As in the image, the lazybear version values of Squeeze are actually the same and correct values as tradingview but squeeze_pro values are not. | 0easy
|
Title: Global disable_web_page_preview parameter
Body: **Is your feature request related to a problem? Please describe.**
Set `disable_web_page_preview` in `aiogram.Bot`:
`bot = aiogram.Bot(token=..., disable_web_page_preview=...)`
**Describe the solution you'd like**
Like logic with `parse_mode`, but with `disable_web_page_preview` (https://github.com/aiogram/aiogram/blob/358ecc78213183adeb70809d7f3e95f1524657fb/aiogram/bot/bot.py#L311-L312):
```
if self.disable_web_page_preview and entities is None:
payload.setdefault('disable_web_page_preview', self.disable_web_page_preview)
```
In ideal, need paste that piece of code to all methods with messages where supports `disable_web_page_preview` | 0easy
|
Title: Update supported Python versions to include 3.13, remove 3.8
Body: ### Problem
Per https://devguide.python.org/versions/, on 2024-10-07, Python 3.13 was officially released, and Python 3.8 is now end of life (EOL). The documentation mentions 3.8 but does not mention 3.12.
### Proposed Solution
Test this extension with Python 3.13. Add support for 3.13 and remove 3.8 as a supported Python version.
| 0easy
|
Title: sklearn-0.19: deprecation warnings of y_train_mean
Body: In sklearn-0.19, the `y_train_mean` of `GaussianProcessorRegressor` has been deprecated. See:
https://github.com/scikit-learn/scikit-learn/blob/ec0dbf0399c4187a1d17e66de17265c67588373f/sklearn/gaussian_process/gpr.py#L153
This results in tons of warnings, as scikit-optimize itself depends on this attribute, at https://github.com/scikit-optimize/scikit-optimize/blob/e08a067abfa759fa60f279eaf2df569bafc79147/skopt/learning/gaussian_process/gpr.py#L305
Reported in #461. | 0easy
|
Title: Correct description of Saltelli's sampling
Body: Sergei Kucherenko shared the following with me (sorry I am laking time right now to make a PR, though I thought that could be also left as a good first PR 😅):
> I have a remark with regards to SaLib, namely in the Section on Sobol’ Sensitivity Analysis it says: “Saltelli’s scheme extends the Sobol’ sequence in a way to reduce the error rates in the resulting sensitivity index calculations”. This is incorrect. Saltelli indeed proposed a formula to reduce the number of model runs from N ( 2D +1) to N(D+1) ( it was also published at the same time in [1]) but it was I who proposed a formula to reduce the error rates in the resulting sensitivity index calculations. It was firstly published in [1] but typically we give reference [2].
>
> [1] Mauntz W. Global sensitivity analysis of general nonlinear systems. Master’s thesis, Imperial College London, CPSE; 2002. [Supervisors: C. Pantelides and S. Kucherenko.].
>
> [2] Sobol’ I, Tarantola S, Gatelli D, Kucherenko S, Mauntz W. Estimating the approximation error when fixing unessential factors in global sensitivity analysis. Reliab Eng Syst Safety 2007;92:957–60
>
> [3] [S. Kucherenko](http://arxiv.org/find/math/1/au:+Kucherenko_S/0/1/0/all/0/1), [S. Song](http://arxiv.org/find/math/1/au:+Song_S/0/1/0/all/0/1), Different numerical estimators for main effect global sensitivity indices. Reliability Engineering and System Safety, 165 (2017) 222-238
>
> I has one trivial but important improvement: in [3] see Improved formula of Sobol’ and Myshetzskay | 0easy
|
Title: Add `Typing :: Typed` Trove classifier
Body: Add `Typing :: Typed` [Trove classifier](https://pypi.org/classifiers/) to our package's metadata.
Not sure what the exact requirements are, but I think once the current typing workload is completed, we should qualify. | 0easy
|
Title: Elders Thermometer Indicator
Body: **Which version are you running? The lastest version is on Github. Pip is for major releases.**
```python
import pandas_ta as ta
print(ta.version)
```
Version 0.2.01b
**Upgrade.**
```sh
$ pip install -U git+https://github.com/twopirllc/pandas-ta
```
**Is your feature request related to a problem? Please describe.**
No
**Describe the solution you'd like**
[Elders Thermometer Psuedocode](https://www.motivewave.com/studies/elders_thermometer.htm)
[Elders Thermometer by LazyBear](https://www.tradingview.com/script/HqvTuEMW-Elder-s-Market-Thermometer-LazyBear/)
[MQL5 Description](https://www.mql5.com/en/market/product/126#!tab=overview)
**Describe alternatives you've considered**
None at this time
**Additional context**
None
Thanks for using Pandas TA!
😎 | 0easy
|
Title: better document that -s (capture=no) doesn't work with xdist
Body: Hi, I want to print some info in stdout, so I use `-s` pytest option. It works fine unless I also use xdist with something like `-n 2`. In this case, logs got captured and I don't see anything.
Am I missing something? | 0easy
|
Title: Using pyppeteer behind a proxy
Body: Hi,
When we launch pyppeteer from behind a proxy for the first time, the downloader can't fetch the chromium zip.
```bash
MaxRetryError: HTTPSConnectionPool(host='storage.googleapis.com', port=443): Max retries exceeded with url: /chromium-browser-snapshots/Linux_x64/588429/chrome-linux.zip (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7ffa71bf6e90>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution'))
```
By using an `urllib3.ProxyManager` instead of the `urllib3.PoolManager` and setting the proxy manually in https://github.com/pyppeteer/pyppeteer/blob/bc4b62555da445b252e0c9f1d7eefbd6dd8e52a0/pyppeteer/chromium_downloader.py#L82 the issue can be solved | 0easy
|
Title: Add the JupyterLab trove classifiers
Body: <!--
Welcome! Thanks for thinking of a way to improve Voilà. If this solves a problem for you, then it probably solves that problem for lots of people! So the whole community will benefit from this request.
Before creating a new feature request please search the issues for relevant feature requests.
-->
### Problem
The new trove classifiers for JupyterLab prebuilt extensions have just been added to PyPI. See the following issues and PRs for more context:
- https://github.com/pypa/trove-classifiers/pull/72
- https://github.com/jupyterlab/extension-cookiecutter-ts/pull/174
This will help discover prebuilt JupyterLab extensions on PyPI. For example with this search request:
https://pypi.org/search/?c=Framework+%3A%3A+Jupyter+%3A%3A+JupyterLab+%3A%3A+Extensions+%3A%3A+Prebuilt
<!-- Provide a clear and concise description of what problem this feature will solve. For example:
* I'm always frustrated when [...] because [...]
* I would like it if [...] happened when I [...] because [...]
-->
### Proposed Solution
We should add them to the `voila` package too here:
https://github.com/voila-dashboards/voila/blob/56fe4cf4bfe114168819f065e22afcb760f51ac6/setup.py#L111-L120
```
Framework :: Jupyter :: JupyterLab
Framework :: Jupyter :: JupyterLab :: 3
Framework :: Jupyter :: JupyterLab :: Extensions
Framework :: Jupyter :: JupyterLab :: Extensions :: Prebuilt
```
### Additional context
| 0easy
|
Title: Support for bar plots with "positive" and "negative" colors
Body: These types of plots are very popular -- where bars are one color for positive data, and another color for negative data. Example:

Something similar is already implemented for `fill_between` and `fill_betweenx` (a.k.a. `area` and `areax`) via `fill_between_wrapper`in `proplot/wrappers.py`.
This could be implemented with a simple override of `bar_wrapper`. | 0easy
|
Title: receiving CORS error [BUG]
Body: 
I used your described method of using docker image here https://github.com/FujiwaraChoki/MoneyPrinter/issues/140. and I can see containers are running without any issue.

here's the backend log

Environment:
Docker version 24.0.4, build 3713ee1
Running on linux host:
Linux XXXXX 4.18.0-477.15.1.el8_8.x86_64 #1 SMP Fri Jun 2 08:27:19 EDT 2023 x86_64 x86_64 x86_64 GNU/Linux
This is a node on same network I'm accessing it from a laptop with 192.168.1.202 IP . the Linux has IP 192.168.1.21 . do you think it needs to be accessed from same host i.e., 192.168.1.21 GUI ?
Any kind of help is highly appreciated.
Thanks | 0easy
|
Title: Mark `iternal` as a typo of `internal`
Body: ### Summary
To address https://github.com/mlflow/mlflow/pull/14450#discussion_r1942071959, add `iternal` to the typo list:
```
diff --git a/pyproject.toml b/pyproject.toml
index aa6bb2aa3..82c83d53b 100644
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -294,6 +294,7 @@ fpr = "fpr" # false positive rate
gam = "gam" # generalized additive models
ser = "ser" # serialization
yhat = "yhat" # ŷ
+iternal = "internal"
```
### Notes
- Make sure to open a PR from a **non-master** branch.
- Sign off the commit using the `-s` flag when making a commit:
```sh
git commit -s -m "..."
# ^^ make sure to use this
```
- Include `#{issue_number}` (e.g. `#123`) in the PR description when opening a PR.
| 0easy
|
Title: Initially empty test fails even if pre-run modifier adds content to it
Body: It seems that adding a keyword to an empty test case fails when using prerunmodifier. Below example was working with RF 6.0.2, but not any more with 6.1.1. If test case has some keywords in it, then the modifier works fine.
.robot file
```
*** Test Cases ***
Pass with Log
# Modifier adds keyword to this test case
Log Hello from .robot
Pass with Modifier
# Modifier fails to add keyword
# Log Hello from .robot
```
k.py file
```python
from robot.api import SuiteVisitor
from robot.running.model import Keyword
class K(SuiteVisitor):
def start_suite(self, suite):
for test in suite.tests:
kw = Keyword('Log', args=['Hello from MODIFIER'])
test.body.append(kw)
def start_keyword(self, keyword: Keyword) -> bool | None:
print(f'KEYWORD: {keyword.parent.name}:{keyword.name}')
```
run `robot --pythonpath . --prerunmodifier k.K .` | 0easy
|
Title: Add more tutorials with use cases
Body: Currently, the tutorials only display the basic functionality of the `mljar`.It would be great if someone can add more tutorials exhibiting specific functionalities and use-cases. This also should display and explain the outputs generated by the package such as SHAP explainability, leaderboard, etc. What do you think? | 0easy
|
Title: cancel current PR build is another commit is pushed
Body: https://twitter.com/adamchainz/status/1485572084617121796?s=11 | 0easy
|
Title: Support long command line options with hyphens like `--pre-run-modifier`
Body: To make long option names like `--prerunmodifier` easier to read, Robot handles them case-insensitively so that also `--PreRunModifier` works. This is an old feature and works nicely, but it's inconsistent with the `--my-long-option` format other tools typically use nowadays. We could easily support also that format simply by removing hyphens from long option values at the same time as we convert them to lower case.
Removing hyphens would make `--prerunmodifier`, `--PreRunModifier` and `--pre-run-modifier` all equivalent. It would also allow stupid usages like `--p-r-eru-n-mod-i-fier`, but we support also `--PReruNmodIfier` so it shouldn't be a big deal. The leading two hyphens obviously shouldn't be removed.
Supporting multiple different formats is perhaps not a good idea in general, though, and it could be better to require using a single format instead. I personally prefer the `--my-long-option` format over the other alternatives because it is consistent with shorter options like `--option` and also what other tools typically use. I'm not sure is requiring that worth the deprecation effort, but making that format valid now certainly helps if we ever want to do that. | 0easy
|
Title: ANSIBLE_DEBUG with add_host in a loop tracebacks
Body: ### Summary
```yaml
- hosts: localhost
gather_facts: false
tasks:
- add_host:
name: foo
loop:
- 1
```
```
ANSIBLE_DEBUG=1 ansible-playbook playbook.yml
```
```pytb
Traceback (most recent call last):
File "/Users/sivel/venvs/ansibledev/bin/ansible-playbook", line 8, in <module>
sys.exit(main())
^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/cli/playbook.py", line 240, in main
PlaybookCLI.cli_executor(args)
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/cli/__init__.py", line 659, in cli_executor
exit_code = cli.run()
^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/cli/playbook.py", line 156, in run
results = pbex.run()
^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/executor/playbook_executor.py", line 190, in run
result = self._tqm.run(play=play)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/executor/task_queue_manager.py", line 348, in run
play_return = strategy.run(iterator, play_context)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/plugins/strategy/linear.py", line 231, in run
results.extend(self._wait_on_pending_results(iterator))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/plugins/strategy/__init__.py", line 826, in _wait_on_pending_results
results = self._process_pending_results(iterator)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/plugins/strategy/__init__.py", line 172, in inner
results = func(self, iterator, one_pass=one_pass, max_passes=max_passes)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/plugins/strategy/__init__.py", line 700, in _process_pending_results
all_task_vars = combine_vars(found_task_vars, item_vars)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/sivel/projects/ansibledev/ansible/lib/ansible/utils/vars.py", line 91, in combine_vars
result = a | b
~~^~~
TypeError: unsupported operand type(s) for |: 'VarsWithSources' and 'dict'
```
This is an extension of what was fixed in https://github.com/ansible/ansible/pull/79764
### Issue Type
Bug Report
### Component Name
lib/ansible/utils/vars.py
### Ansible Version
```console
$ ansible --version
```
### Configuration
```console
# if using a version older than ansible-core 2.12 you should omit the '-t all'
$ ansible-config dump --only-changed -t all
```
### OS / Environment
N/A
### Steps to Reproduce
See above
### Expected Results
No TB
### Actual Results
```console
TB
```
### Code of Conduct
- [X] I agree to follow the Ansible Code of Conduct | 0easy
|
Title: rename jinja2 templates to `.j2` suffix
Body: seems like that's the standard, and it will stop confusing tools. | 0easy
|
Title: [Feature] Allow Serving Requests During CUDA Graph Capture
Body: ### Checklist
- [x] 1. If the issue you raised is not a feature but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed.
- [x] 2. Please use English, otherwise it will be closed.
### Motivation
#### Background
When `--enable-cuda-graph`, service restart (after crashes/unexpected exits) currently requires about 10-minute CUDA Graph capture process before becoming operational (longer with torch.compile enabled). This creates significant service downtime despite even with external health check and automatic restart, unless hot-standbys are deployed.
I raised this question in community discussion group and received reply from @ baizhou: Theoretically feasible to let model runner choose whether to replay existing CUDA Graph.
#### Requested Feature
Add an option or fallback mechanism to allow serving requests without CUDA Graph replay during initialization, prioritizing availability over performance.
#### Discussion
I am open to discussing the details further and would appreciate any thoughts on implementation from the community. Thank you.
### Related resources
_No response_ | 0easy
|
Title: dcc.DatePickerRange and dcc.DatePickerSingle should allow type in
Body: It takes forever if a user need to move 10 years, especially since the popup calendar does not allow moving by years (only months) | 0easy
|
Title: Reformat demo application list
Body: The list of apps in the demo is getting long and untidy; it could do with reformatting.
In addition, the corresponding list of apps in the documentation should provide links to the online demo.
| 0easy
|
Title: ploomber examples into local directory
Body: When running !ploomber examples -n guides/spec-api-python -o . it outputs an error:
Local copy does not exist...
Cloning into '/root/.ploomber/projects'...
remote: Enumerating objects: 416, done.
remote: Counting objects: 100% (416/416), done.
remote: Compressing objects: 100% (335/335), done.
remote: Total 416 (delta 64), reused 283 (delta 34), pack-reused 0
Receiving objects: 100% (416/416), 1.31 MiB | 8.96 MiB/s, done.
Resolving deltas: 100% (64/64), done.
================= Copying example guides/spec-api-python to ./ =================
Error: '.' already exists in the current working directory, please rename it or move it to another location and try again.
| 0easy
|
Title: 几个bug与新功能的建议
Body: BUG1:邮箱昵称设置了压根没用,发件人姓名是发件邮箱@前的内容
BUG2:从编辑商品的编辑器页面直接返回时,鼠标滚轮滑动网页功能失效
建议1:添加验证码系统,不是腾讯云这类让同行恶意刷房子赔进去的验证码,尽量是recaptcha或者hcaptcha这类**免费**并且还不是**过于过分**的验证码。(点名epic和origin的数骰子
建议2:添加优惠劵系统,可以设定单张优惠劵优惠金额或优惠折扣,限制最低使用额,限制使用商品。
建议3:改进批发价系统:虽说能看懂,但是还是有点麻烦,可以参考值联云卡的 件数=批发价的设置
(举个例子 5=2.8 也就是满五件每件2.8元) | 0easy
|
Title: Cleanup plugins documentation
Body: - Move most content of `inference/worker/PLUGINS.md` to `docs/`.
- Move technical content to `inference/worker/README.md`.
- Cleanup plugins docs and add more explanation on using them in the web UI. | 0easy
|
Title: Dummy_minimize in benchmarks
Body: After the work of @iaroslav-ai , we should have a dummy_minimize function in benchmarks to see if we are performing better than random.
| 0easy
|
Title: When creating `sysadmin` user flow aborts, the user is still created
Body: ## CKAN version
CKAN version: 2.10.3
Python version: 3.9.6
OS: Mac
## Describe the bug
When creating a `sysadmin` user for an install of CKAN on my local machine and it fails, the user is still created in the user table.
### Steps to reproduce
Steps to reproduce the behavior:
1. With the virtual environment activated, run `ckan sysadmin add <admin>`
2. There is a prompt to enter an email, enter any email
3. There is a prompt to enter a password, enter in a password that does NOT meet the criteria (longer than 8 characters)
4. Confirm the password
5. There is a prompt with the password criteria which needs to be more than 8 characters and that the user flow has been aborted.
6. Now re-run `ckan sysadmin add <admin>`
7. Enter an email
8. Enter a valid password with at least 8 characters
9. Confirm the password
10. Then you should see a note that says:
```
None - {'name': ['That login name is not available.']}
Aborted!
```
11. Once you create the a sysadmin user, run the app then go to the admin panel and view users to confirm that the first user you tried to create was created even though the user creation flow was aborted.
### Expected behavior
If the user creation flow was aborted, a user should not be created.
### Additional details
This is a screenshot of the 2nd time attempting to create a user with the same name which illustrated that the first time it was created even though the user creation flow was aborted.

| 0easy
|
Title: DataChannel-only 'feature' without codec dependencies?
Body: Great stuff! I'd love to use this in [magic-wormhole](https://github.com/warner/magic-wormhole).. I only need the DataChannel, though. Any idea how hard it'd be to define a 'feature' for the codecs, so `pip install aiortc[video]` gets you the dependencies on the codecs, but `pip install aiortc[datachannel]` does not?
| 0easy
|
Title: Add information on how to view the docs
Body: ### Is your feature request related to a problem? Please describe.
There seem to be extensive documentation in the doc repo, but those are under some specific format (mdx) that doesn't seem to be compatible with md viewers. I also tried the specific mdx extension for vscode, but this one complains on pretty much every file that some of the descriptor is missing, so I assume the docs need to be built somehow.
### Describe the solution you'd like
I case of local use without Internet access, it would be useful to add pointers about how to build the documentation.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_ | 0easy
|
Title: Typing error in generated client when response attribute use reserved keywords
Body: **Describe the bug**
I am using this library to generate a Python client for Jira using [their OpenAPI spec](https://developer.atlassian.com/cloud/jira/platform/swagger-v3.v3.json).
I get a strange bug at import time:
```
File "D:\Repos\jira_generated\the-jira-cloud-platform-rest-api-client\the_jira_cloud_platform_rest_api_client\models\json_node.py", line 38, in JsonNode
int_value: Union[Unset, int] = UNSET
File "C:\Apps\Python38\lib\typing.py", line 261, in inner
return func(*args, **kwds)
File "C:\Apps\Python38\lib\typing.py", line 358, in __getitem__
parameters = tuple(_type_check(p, msg) for p in parameters)
File "C:\Apps\Python38\lib\typing.py", line 358, in <genexpr>
parameters = tuple(_type_check(p, msg) for p in parameters)
File "C:\Apps\Python38\lib\typing.py", line 149, in _type_check
raise TypeError(f"{msg} Got {arg!r:.100}.")
TypeError: Union[arg, ...]: each arg must be a type. Got <the_jira_cloud_platform_rest_api_client.types.Unset object at 0x0000018CF4EEA070>.
```
After investigating for a while, I came to the conclusion that it's because the `JsonNode` object has an attribute named "int". The attribute shadows the built-in name `int` and causes type hinting to fail:
```python
@attr.s(auto_attribs=True)
class JsonNode:
""" """
... # Other attributes
int: Union[Unset, bool] = UNSET
... # Other attributes
int_value: Union[Unset, int] = UNSET # Breaks here!
... # Other attributes
```
**To Reproduce**
Steps to reproduce the behavior:
1. Generate the client for Jira: openapi-python-client generate --url https://developer.atlassian.com/cloud/jira/platform/swagger-v3.v3.json
2. cd to the generated directory and open a Python shell
3. Import `get_changelogs`: `from the_jira_cloud_platform_rest_api_client.api.issues import get_change_logs` or anything else indirectly using `JsonNode`
**Expected behavior**
A different name may need to be used here like `int_` or similar.
**OpenAPI Spec File**
https://developer.atlassian.com/cloud/jira/platform/swagger-v3.v3.json
**Desktop (please complete the following information):**
- OS: Windows 10
- Python Version: 3.8.8
- openapi-python-client version: 0.8.0
**Additional context**
Thanks for the great tool!
| 0easy
|
Title: Bollinger Bands Bandwidth
Body: **Which version are you running? The lastest version is on Github. Pip is for major releases.**
```python
import pandas_ta as ta
print(ta.version)
```
Version: 0.2.xxb
**Upgrade.**
```sh
$ pip install -U git+https://github.com/twopirllc/pandas-ta
```
Yup.
**Is your feature request related to a problem? Please describe.**
Bollinger Bands Bandwidth
**Describe the solution you'd like**
[Bollinger Bands Bandwidth](https://www.sierrachart.com/index.php?page=doc/StudiesReference.php&ID=135&Name=Bollinger_Bands_-_Bandwidth)
**Describe alternatives you've considered**
None
**Additional context**
None | 0easy
|
Title: Disable some of the extensions by default
Body: We should think more carefully about which extensions are available by default. One option is to disable the Assignment List (which requires the exchange to be properly setup) and the Courses extension (which is really only used for JupyterHub) and only keep the Formgrader, Create Assignment, and Validate extensions. Then in the installation docs (both in the user and development docs) we can indicate that if people want to install these extensions they should follow the instructions for a server-based setup. | 0easy
|
Title: Add `float("inf")` support
Body: `float("inf")` is probably the widest used variant for specifying infinity, so would be nice to support. Workaround is to use `np.inf` instead, so not super important. | 0easy
|
Title: Add docker support
Body: ### Description
create a docker image so that people can just docker pull the igel container and therefore they can use igel without installing anything.
If someone wants to contribute and work on this, then feel free to create a docker folder in the repo and work on this feature there
| 0easy
|
Title: v0.13, django client: lazy-loading server metadata breaks logout flows
Body: **Describe the bug**
Version 0.13 intoduces lazy loading of server metadata on first auth request, but this breaks logout flows for some OIDC providers (in my case, Keycloak).
Consider this logout view (django):
```
def logout(request):
""" log user out of django and redirect to the Keycloak logout url, so that he gets logged out everywhere """
log_out(request)
return_to = urlencode({'redirect_uri': request.build_absolute_uri('/')})
logout_url = oauth.keycloak.server_metadata['end_session_endpoint']
logout_url += '?' + return_to
return HttpResponseRedirect(logout_url)
```
This logic counts on `server_metadata['end_session_endpoint']` being available (this is a special Keycloak feature).
However, when the django app restarts, `server metadata` is empty until someone performs a login flow. Until that, all logout flows will end up with an error.
**Error Stacks**
```
Internal Server Error: /logout
Traceback (most recent call last):
File "/Users/1111/Library/Caches/pypoetry/virtualenvs/demo-keycloak-client-py3.7/lib/python3.7/site-packages/django/core/handlers/exception.py", line 34, in inner
response = get_response(request)
File "/Users/1111/Library/Caches/pypoetry/virtualenvs/demo-keycloak-client-py3.7/lib/python3.7/site-packages/django/core/handlers/base.py", line 115, in _get_response
response = self.process_exception_by_middleware(e, request)
File "/Users/1111/Library/Caches/pypoetry/virtualenvs/demo-keycloak-client-py3.7/lib/python3.7/site-packages/django/core/handlers/base.py", line 113, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/Users/1111/_projects/fasttrack/demo_keycloak_client/auth0login/views.py", line 28, in logout
logout_url = oauth.keycloak.server_metadata['end_session_endpoint']
KeyError: 'end_session_endpoint'
```
**To Reproduce**
Instantiate the client like so:
```
from authlib.integrations.django_client import OAuth
oauth = OAuth()
oauth.register(
'keycloak',
server_metadata_url='https://sso.fstrk.io/auth/realms/fasttrack/.well-known/openid-configuration',
client_kwargs={'scope': 'openid profile email'}
)
assert oauth.keycloak.server_metadata['end_session_endpoint']
```
**Suggestions**
Currently I am force-loading the metadata on Django startup:
```
oauth.keycloak._load_server_metadata()
```
Since `end_session_endpoint` is Keycloak-specific and the library cannot be extented to provide a `deauthorize_redirect()` method, my suggestion is make `load_server_metadata()` public and document it.
Thanks!
| 0easy
|
Title: CI: script that runs all Colab notebooks
Body: This script can be Github Action that is only executed when a new release is tagged.
We want to test that all Colab notebooks run without execution errors (particularly because they may not have installed the right version of dependencies, eg. if a dependency got updated without us knowing).
Cf: https://stackoverflow.com/questions/61762045/running-google-colab-every-day-at-a-specific-time | 0easy
|
Title: New style streaming route handler improvements
Body: **Is your feature request related to a problem? Please describe.**
The new streaming API allows streaming responses in the route handler without having to use a callback:
```python
@app.route("/")
async def test(request):
response = await request.respond(content_type="text/csv")
await response.send("foo,")
await response.send("bar")
await response.send("", True)
return response
```
**Describe the solution you'd like**
A simpler method to closing the stream before returning in place of `await response.send("", True)`
```python
await response.eof()
```
Under the hood, `eof` should simply just be a convenience call to `send("", True)`. | 0easy
|
Title: User guide/FAQ: Send JSON & URL-encoded requests to the App
Body: Hey, I'm trying to fetch out the data with JSON I've spent way too much time on google and tried every single solution with no success.. when I send the request it raises an json.decoder.JSONDecodeError Exception.
maybe can you help?
```
class TextOcrRes:
def on_post(self, req, resp):
"""Handles Text POST requests"""
json_data = json.loads(req.bounded_stream.read().decode("utf-8")) # json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
print(json_data)
```
**What I'm sending:**
`
r = requests.post("http://x.xx.xx.xx/translate", data={'image_data':'message'})
`
I'm on python 3.6.8 | 0easy
|
Title: Marketplace - agent page - fix margins in the agent description area
Body:
### Describe your issue.
Please fix margins to reflect the following
<img width="611" alt="Screenshot 2024-12-17 at 19 08 03" src="https://github.com/user-attachments/assets/36d2fb0f-a954-4e29-bed6-9fd314dc0875" />
| 0easy
|
Title: Bug: RMQ parrent test message
Body: Test `publish` doesn't support case with `#` not at the end (like `#.text`)
The problem should be solved here: https://github.com/airtai/faststream/blob/main/faststream/rabbit/test.py#L245 | 0easy
|
Title: The start/end_listening sounds should be called with the pulse media.role=phone to enable ducking
Body: **Describe the bug**
When pulse ducking is enabled, the background audio is only ducked when TTS audio is played. But I think the start and end listening sounds should also be ducked
**To Reproduce**
Steps to reproduce the behavior:
1. Enable ducking
2. Say "Hey Mycroft.."
Result: The start_listening sound is played, but the existing background music/audio is not ducked.
**Expected behavior**
Play the start/end_listening audio files with the proper pulse audio environment to enable ducking (`media.role=phone`)
**Log files**
If possible, add log files from `/var/log/mycroft/` to help explain your problem.
You may also include screenshots, however screenshots of log files are often difficult to read and parse.
If you are running Mycroft, the [Support Skill](https://github.com/MycroftAI/skill-support) helps to automate gathering this information. Simply say "Create a support ticket" and the Skill will put together a support package and email it to you.
**Environment (please complete the following information):**
- Device type: RPI 4
- OS: Raspbian
- Mycroft-core version: ab242a2c821553e6ecde0fa656f3be0331aef738
| 0easy
|
Title: Improve test coverage from 91% to 100%
Body: Test coverage is always important, and my goal is to eventually be at 100%. I'll be sure to do this for PRs from now on, or at least try to, but there is some existing code that is not yet covered.
The good thing is that all files have existing tests, tests for `__main__.py` for example can be found in the [`main_test.py`](https://github.com/python-security/pyt/blob/master/tests/main_test.py) file.
By running `make test` you can see where we aren't yet at 100%:
```python
py36 runtests: commands[3] | coverage report --include=pyt/* --fail-under 91
Name Stmts Miss Cover Missing
---------------------------------------------------------------------------------
pyt/__init__.py 0 0 100%
pyt/__main__.py 68 13 81% 45, 55-57, 70-71, 83, 86, 105-108, 129
pyt/analysis/constraint_table.py 9 0 100%
pyt/analysis/definition_chains.py 16 0 100%
pyt/analysis/fixed_point.py 21 0 100%
pyt/analysis/lattice.py 31 0 100%
pyt/analysis/reaching_definitions_taint.py 27 0 100%
pyt/cfg/__init__.py 2 0 100%
pyt/cfg/alias_helper.py 33 0 100%
pyt/cfg/expr_visitor.py 212 20 91% 70, 99-100, 111-122, 131, 136, 141, 149, 154, 407, 560
pyt/cfg/expr_visitor_helper.py 9 0 100%
pyt/cfg/make_cfg.py 9 0 100%
pyt/cfg/stmt_visitor.py 456 46 90% 111, 168, 174, 217, 235, 247-248, 277-278, 297-298, 386-399, 658-661, 668-671, 678, 684, 702-704, 780, 816-823, 868, 893, 919-920, 951, 972, 988, 1010
pyt/cfg/stmt_visitor_helper.py 72 6 92% 83, 95-96, 133-136
pyt/core/ast_helper.py 53 4 92% 56, 101, 103, 105
pyt/core/module_definitions.py 47 4 91% 32, 108-110
pyt/core/node_types.py 70 4 94% 51, 142-145
pyt/core/project_handler.py 30 1 97% 23
pyt/formatters/__init__.py 0 0 100%
pyt/helper_visitors/__init__.py 4 0 100%
pyt/helper_visitors/label_visitor.py 206 40 81% 31-37, 93-98, 179-184, 186-188, 200, 203, 212, 215, 218, 221, 224, 227, 230, 246, 249, 252, 255, 261, 265, 271, 274, 278, 281
pyt/helper_visitors/right_hand_side_visitor.py 13 0 100%
pyt/helper_visitors/vars_visitor.py 104 4 96% 66-68, 102
pyt/usage.py 39 1 97% 124
pyt/vulnerabilities/__init__.py 3 0 100%
pyt/vulnerabilities/trigger_definitions_parser.py 30 0 100%
pyt/vulnerabilities/vulnerabilities.py 160 15 91% 233-235, 237, 241-243, 310, 312-321, 424, 427
pyt/vulnerabilities/vulnerability_helper.py 76 19 75% 84, 116-119, 132-134, 178-182, 204-209
pyt/web_frameworks/__init__.py 3 0 100%
pyt/web_frameworks/framework_adaptor.py 41 0 100%
pyt/web_frameworks/framework_helper.py 23 0 100%
---------------------------------------------------------------------------------
TOTAL 1867 177 91%
```
If you would like help with writing tests for any of these files let me know. | 0easy
|
Title: Collections and String: Add `ignore_case` as alias for `case_insensitive`
Body: We added case-insensitive comparison support to various Collections keywords so that keywords got an optional `ignore_case` argument (#4343). While doing that, we noticed that four Collections keywords already supported case-insensitivity and had `case_insensitive` argument for that purpose. This kind of inconsistency within a library isn't great, so something needs to be done to it.
An easy solution would be using `case_insensitive` also with the new keywords. That would make Collections consistent, but because we use `ignore_case` in BuiltIn (#2439), different libraries would be inconsistent. I believe we should aim for consistency with all standard libraries and also believe that that `ignore_case` is a better argument name than `case_insensitive`. It is also consistent with other similar arguments like `ignore_order` and `ignore_keys` in Collections.
The plan is to do the following:
- Add `ignore_case` argument to Collections keywords that accept `case_insensitive`. The old argument should be preserved for now, but it should be mentioned in the documentation that the new one is preferred. We shouldn't yet emit deprecation warnings, but we can start doing that in RF 8.0.
- For consistency reasons, add `ignore_whitespace` to these same keywords as an alias for `whitespace_insensitive`. It should be treated the same way a `case_insensitive`.
- For consistency reasons, add `ignore_case` as an alias also for the two String library keywords that accept `case_insensitive`.
Although we won't be emitting deprecation warnings, it's good to mention this under the deprecated features section in the release notes. | 0easy
|
Title: Clean up legacy dependecies from `eval/`
Body: We don't use codeblue anymore. It's just legacy code. Just need to verify and clean up any dangling references, and remove this dependency.
To remove: https://github.com/ShishirPatil/gorilla/tree/main/eval/eval-scripts/codebleu
Check for dependencies: https://github.com/ShishirPatil/gorilla/tree/main/eval/eval-scripts/* | 0easy
|
Title: Spaces are not normalized when matching keywords with embedded arguments
Body: When matching normal keywords, different kind of spaces are considered equal. For example, using a keyword so that there's the no-break space instead of the normal ASCII space in its name is fine. No-break spaces have been normalized for ages (#1264) and I believe we started to normalize all spaces when the new parser was taken into use in RF 3.2.
This kind of normalization isn't done, however, with keywords containing embedded arguments. In practice this means that using a keyword like `Given a good example` so that one of the spaces is a no-break space, or some other Unicode space, fails. A related problem is that the emitted error is `No keyword with name 'Given a good example' found.` and it's hard to understand why there's no match when the name looks just fine.
This is relatively easy to fix, but different spaces need to be taken into account both in the keyword name itself and in the possible BDD prefix. I actually noticed this issue when investigating ways to make matching BDD prefixes faster after PR #5096 changed keyword matching logic so that BDD style keywords are matched first.
Although we can normalize spaces pretty easily, we cannot make keywords with embedded arguments space-insensitive similarly as normal keywords are. I believe that's fine. | 0easy
|
Title: Replace `assert_almost_equal` with `assert_allclose`
Body: `assert_allclose` was added to TensorLy in #420 and `assert_almost_equal` (which is currently used in many tests in TensorLy) is discouraged by [NumPy](https://numpy.org/doc/stable/reference/generated/numpy.testing.assert_almost_equal.html). We should therefore update all tests that use `assert_almost_equal` to use `assert_allclose` instead.
The change should be relatively straightforward: Find a test with `assert_almost_equal`, and change to `assert_allclose`, possibly modifying the tolerance accordingly. | 0easy
|
Title: Trino split_statements is default to False
Body: ### Apache Airflow Provider(s)
trino
### Versions of Apache Airflow Providers
6.0.1
### Apache Airflow version
2.9.0
### Operating System
Red Hat Enterprise Linux 9.5
### Deployment
Virtualenv installation
### Deployment details
_No response_
### What happened
I'm following this [document](https://airflow.apache.org/docs/apache-airflow-providers-trino/stable/operators/trino.html#using-the-operator) to use SQLExecuteQueryOperator to issue multiple queries separated with semicolon to Trino.
I got a syntax error complaining `mismatched input ';'`. I have to add an additional `split_statements=True` parameter to get it work.
### What you think should happen instead
SQLExecuteQueryOperator defers ``split_statements`` to the default value in the ``run`` method of the configured hook. Such is not defined in TrinoHook, so it uses default False from DbApiHook and throws the error.
### How to reproduce
```
SQLExecuteQueryOperator(
task_id='t1',
sql='SELECT 1; SELECT 2;',
conn_id='loading_user_trino' # Adding split_statements=True will work
)
```
### Anything else
_No response_
### Are you willing to submit PR?
- [x] Yes I am willing to submit a PR!
### Code of Conduct
- [x] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
| 0easy
|
Title: Previous Outputs drop down: list of files is not sorted
Body: The already generated file list in the drop down is not sorted. This makes it difficult to find recently generated clips.
Suggestion: sort the files by generation date and have most recent first. | 0easy
|
Title: Drag and Drop problem (Pointer.down() / Pointer.up())
Body: Hey,
as already shared in Discord, Iam not able to create a Drag and Drop movement. Maybe I do something wrong, but the documentation didnt really help me there, for up() and down() it just asks for (**kwargs).
I made a small **example** of the problem and of what I tried:
```
import asyncio
from selenium_driverless import webdriver
async def test_drag_and_drop():
# Create driver
driver = await webdriver.Chrome()
try:
# Navigate to test page
await driver.get("https://kaliiiiiiiiii.github.io/Selenium-Driverless/index.html")
await asyncio.sleep(1) # Wait for page load
# Find source and target elements
source_element = await driver.find_element(
"xpath",
"//div[@class='wy-menu wy-menu-vertical']//a[@href='api/WebELement.html']"
)
target_element = await driver.find_element(
"xpath",
"//*[@id='rtd-search-form']"
)
if not source_element or not target_element:
print("Could not find required elements")
return
print("Calculating element positions...")
source_pos = await source_element.mid_location()
target_pos = await target_element.mid_location()
print(f"Source position: {source_pos}")
print(f"Target position: {target_pos}")
pointer = driver.current_pointer
# Variant 1: No arguments for down/up
print("\nTesting Variant 1: down()/up() without arguments")
try:
await pointer.move_to(source_pos[0], source_pos[1])
await driver.sleep(0.2)
await pointer.down() # No arguments
await driver.sleep(0.2)
await pointer.move_to(target_pos[0], target_pos[1])
await driver.sleep(0.2)
await pointer.up() # No arguments
print("Variant 1 completed")
except Exception as e:
print(f"Variant 1 error: {e}")
await asyncio.sleep(1)
# Variant 2: Complete tuple from mid_location
print("\nTesting Variant 2: down()/up() with complete tuple")
try:
await pointer.move_to(source_pos[0], source_pos[1])
await driver.sleep(0.2)
await pointer.down(source_pos) # Complete tuple
await driver.sleep(0.2)
await pointer.move_to(target_pos[0], target_pos[1])
await driver.sleep(0.2)
await pointer.up(target_pos) # Complete tuple
print("Variant 2 completed")
except Exception as e:
print(f"Variant 2 error: {e}")
await asyncio.sleep(1)
# Variant 3: Separate x,y coordinates from mid_location
print("\nTesting Variant 3: down()/up() with separate x,y coordinates")
try:
await pointer.move_to(source_pos[0], source_pos[1])
await driver.sleep(0.2)
await pointer.down(source_pos[0], source_pos[1]) # Separate coordinates
await driver.sleep(0.2)
await pointer.move_to(target_pos[0], target_pos[1])
await driver.sleep(0.2)
await pointer.up(target_pos[0], target_pos[1]) # Separate coordinates
print("Variant 3 completed")
except Exception as e:
print(f"Variant 3 error: {e}")
await asyncio.sleep(2) # Wait to see the results
finally:
await driver.quit()
if __name__ == "__main__":
asyncio.run(test_drag_and_drop())
```
**Result:**
> Calculating element positions...
> Source position: [133, 327]
> Target position: [210, 62]
>
> Testing Variant 1: down()/up() without arguments
> Variant 1 error: PointerEvent.__init__() missing 2 required positional arguments: 'x' and 'y'
>
> Testing Variant 2: down()/up() with complete tuple
> Variant 2 error: Pointer.down() takes 1 positional argument but 2 were given
>
> Testing Variant 3: down()/up() with separate x,y coordinates
> Variant 3 error: Pointer.down() takes 1 positional argument but 3 were given
>
> Process finished with exit code 0
I think it would be great to add a Pointer.drag_n_drop() with a start element (or coordinate) and end element (or coordinate). I hope you can fix the bug or tell me, when it was my fault ^^ | 0easy
|
Title: Misleading error messages when users enter invalid Groq model
Body: ### Describe the bug
This occurs when trying to use one of the new Groq models, such as `llama-3.1-405b-reasoning` and non-paying users do not have access. This might occur with other invalid models as well, but I haven't tested.
When the user runs `interpreter --model groq/llama-3.1-405b-reasoning --api_base https://api.groq.com/openai/v1 --api_key $GROQ_API_KEY`
the error is
```
You do not have access to openai/groq/llama-3.1-405b-reasoning. You will need to add a payment method and purchase credits for the OpenAI API billing page (different from ChatGPT) to use `GPT-4`.
.com/account/billing/overview
Would you like to try GPT-3.5-TURBO instead? (y/n)
```
### Reproduce
Run `interpreter --model groq/llama-3.1-405b-reasoning --api_base https://api.groq.com/openai/v1 --api_key $GROQ_API_KEY`
Enter any prompt
### Expected behavior
An error message that is related to the provider of the model attempted to be used
### Screenshots
_No response_
### Open Interpreter version
0.3.4
### Python version
3.11.7
### Operating System name and version
MacOS
### Additional context
_No response_ | 0easy
|
Title: Execution fails at the end when using `--output NONE` and console hyperlinks are enabled
Body: ```
robot --log NONE --output NONE asdf.robot
```
```
[ ERROR ] Unexpected error: AttributeError: 'NoneType' object has no attribute 'as_uri'
Traceback:
File "C:\Users\user\project\.venv\Lib\site-packages\robot\utils\application.py", line 81, in _execute
rc = self.main(arguments, **options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\user\project\.venv\Lib\site-packages\robot\run.py", line 463, in main
result = suite.run(settings)
^^^^^^^^^^^^^^^^^^^
File "C:\Users\user\project\.venv\Lib\site-packages\robot\running\model.py", line 776, in run
output.close(runner.result)
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\output.py", line 58, in close
LOGGER.output_file(self._settings['Output'])
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\logger.py", line 423, in output_file
logger.output_file(path)
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\loggerapi.py", line 172, in output_file
self.result_file('Output', path)
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\console\dotted.py", line 66, in result_file
self.stdout.result_file(kind, path)
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\console\highlighting.py", line 122, in result_file
path = self._highlighter.link(path)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\user\project\.venv\Lib\site-packages\robot\output\console\highlighting.py", line 162, in link
uri = path.as_uri()
^^^^^^^^^^^
AttributeError: 'NoneType' object has no attribute 'as_uri'
```
this only seems to happen when running in a TTY | 0easy
|
Title: Concerns about the calculation of return
Body: ## 🐛 Bug Description
I found the calculation of return in backtest confusing. The return is defined on each day's basis [here](https://github.com/microsoft/qlib/blob/2ed806c8469904e441cac494fad2af104a4dbd1d/qlib/backtest/account.py#L260). The return on each day are summed up directly to obtain the annualized return [here](https://github.com/microsoft/qlib/blob/2ed806c8469904e441cac494fad2af104a4dbd1d/qlib/contrib/evaluate.py#L58). Shouldn't the total return be a product of (1 + r_i) minus 1? What's the problem here?
## To Reproduce
Steps to reproduce the behavior:
1.
1.
1.
## Expected Behavior
<!-- A clear and concise description of what you expected to happen. -->
## Screenshot
<!-- A screenshot of the error message or anything shouldn't appear-->
## Environment
**Note**: User could run `cd scripts && python collect_info.py all` under project directory to get system information
and paste them here directly.
- Qlib version: 2ed806c8469904e441cac494fad2af104a4dbd1d
- Python version: 3.7
- OS (`Windows`, `Linux`, `MacOS`): Linux
- Commit number (optional, please provide it if you are using the dev version): 2ed806c8469904e441cac494fad2af104a4dbd1d
## Additional Notes
<!-- Add any other information about the problem here. -->
| 0easy
|
Title: [BUG] 150% make the flag huge........
Body: One twitter found it.
Try to fix it maybe by CSS?

@geekplux
https://twitter.com/Lynx_0C/status/1337688269140545537
Really huge........
| 0easy
|
Title: EPUB file translation broken
Body: It sounds like Epub file translation is broken. EPUB translation is done by [Argos Translate Files](https://github.com/LibreTranslate/argos-translate-files).
https://community.libretranslate.com/t/file-translation/366 | 0easy
|
Title: Writing specific fields
Body: I think it would be very useful to be able to specify the fields to write out. Something along the lines of
pyart.io.write(radar,fields = ['reflectivity','differential_reflectivity','velocity']")
| 0easy
|
Title: AttributeError: module 'os' has no attribute 'geteuid'
Body: ### Bug description
#### Expected behaviour
Expect to produce docker image.
#### Actual behaviour
```
Traceback (most recent call last):
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\traitlets\traitlets.py", line 528, in get
value = obj._trait_values[self.name]
KeyError: 'user_id'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "c:\appdata\local\continuum\anaconda3\lib\runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "c:\appdata\local\continuum\anaconda3\lib\runpy.py", line 85, in _run_code
exec(code, run_globals)
File "C:\AppData\Local\Continuum\anaconda3\Scripts\repo2docker.exe\__main__.py", line 9, in <module>
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\repo2docker\__main__.py", line 345, in main
r2d = make_r2d()
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\repo2docker\__main__.py", line 300, in make_r2d
if r2d.user_id == 0 and not r2d.dry_run:
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\traitlets\traitlets.py", line 556, in __get__
return self.get(obj, cls)
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\traitlets\traitlets.py", line 535, in get
value = self._validate(obj, dynamic_default())
File "c:\appdata\local\continuum\anaconda3\lib\site-packages\repo2docker\app.py", line 206, in _user_id_default
return os.geteuid()
AttributeError: module 'os' has no attribute 'geteuid'
```
### How to reproduce
1. On a Windows 10 PC, open the Anaconda Command Prompt
2. Run `repo2docker https://github.com/matthewolckers/fbr2020`
- OS: Windows 10
- Docker version: 19.03.8,
- repo2docker version 0.11.0
| 0easy
|
Title: [Feature]: [V1] Validate / Fix Load Formats on V1
Body: ### 🚀 The feature, motivation and pitch
We are not sure if `--load-format sharded_state` of `--load-format tensorizer` work with V1
This issue asks to look into it and fix any issues that occur, including for TP>1
### Alternatives
_No response_
### Additional context
_No response_
### Before submitting a new issue...
- [x] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the [documentation page](https://docs.vllm.ai/en/latest/), which can answer lots of frequently asked questions. | 0easy
|
Title: `kick_chat_member` should be removed
Body: ### Problem
`kick_chat_member` is the same as `ban_chat_member`, but having two methods that does the same confuses: Kicking is just removing a person from a chat, and a banning is removing a person from a conversation + blacklisting him.
https://core.telegram.org/bots/api#banchatmember
### Possible solution
Deleting or changing the `kick_chat_member` method
### Alternatives
_No response_
### Code example
_No response_
### Additional information
This method no longer exists in Telegram Bot API
https://core.telegram.org/bots/api-changelog#june-20-2022 | 0easy
|
Title: Disable the caplog fixture by default
Body: When profiling a service that generates logs pytest by default automatically collects all log records. This makes the memory profile not true to real life as all log lines are part of the peak memory usage, though in practice after every log that record would be removed. This feature can be disabled for pytest via `-p no:logging`. At the very least we should document this under gotcha's or even better provide a marker that does this? | 0easy
|
Title: Update Readthedocs Documentation
Body: - Update the documentation which can be found in [Readthedocs](https://scrapegraph-ai.readthedocs.io/en/latest/).
- [Sphinx](https://www.sphinx-doc.org/en/master/) is used for handling the documents generation.
- The files can be modified and added inside the folder docs/source of the repo | 0easy
|
Title: Inline completer displays ``` at the end of suggested code when using provider
Body: As flagged in this [Jupyterlab forum post](https://discourse.jupyter.org/t/inline-completer-using-jupyter-ai/24388); should be removed in post-processing. | 0easy
|
Title: Some patch actions raise `ckan.logic.NotAuthorized` even though `context['ignore_auth'] = True`
Body: ## CKAN version
2.9+
## Describe the bug
The patch action functions in [ckan/logic/action/patch.py](https://github.com/ckan/ckan/tree/master/ckan/logic/action/patch.py) create a separate `show_context: Context` object that is used with a show action to retrieve the resource that is being patched. For almost all of these patch functions, the `'ignore_auth'` value from the patch action's input `context: Context` argument is not propagated to the `show_context` object. As a result, patching some resource types with `'ignore_auth': True` in the patch action's `Context` unexpectedly fails with a `ckan.logic.NotAuthorized` error.
Only [`package_patch()`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L14) correctly propagates this value. The other four patch action functions are affected:
* [`resource_patch()`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L57)
* [`group_patch()`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L88)
* [`organization_patch()`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L122)
* [`user_patch()`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L157)
## Example
The following code snippet uses the Plugin Toolkit to access the [`user_patch()` function](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L157). This will fail if `user` is not authorized to perform `'user_show'`, because `'ignore_auth'` [is not propagated to `show_context`](https://github.com/ckan/ckan/blob/master/ckan/logic/action/patch.py#L171) in `user_patch()`.
```python
toolkit.get_action('user_patch')(
context={
'user': user,
'ignore_auth': True,
},
)
```
A problem like this showed up while I was modifying some code in the `ckanext-ldap` plugin. I believe the reason is that at the time this is being called, a currently not-logged-in user is being passed, and such a user cannot perform `'user_show'`. Regardless, I would have expected that with `'ignore_auth'` being passed into the patch function, the action would succeed, or at least would not return an authorization error.
## Suggested fix
### Easy
The easiest thing to do is just add `'ignore_auth': context.get('ignore_auth', False)` to each of the `show_context` definitions that are missing them.
### Robust
A more robust fix would be to introduce a helper function, `_create_show_context()` (defined below), that each function can call to create the `show_context` object. That way, future changes to the `show_context` will be propagated to all of the patch functions.
It is worth noting that I have absolutely no clue what the `'for_update'` key does. I couldn't find any documentation about it. It seems to be used in the database interaction code, but I'm not really familiar with working with databases. In any case: this key is not set consistently in the `show_context` objects across the various patch functions, so in the code below, it is an optional parameter that can be passed into the new function.
```python
def _create_show_context(context: Context, for_update: bool = False) -> Context:
'''Create a Context that can be used with a user_show action call.
This method is internal. It is meant to be used by the patch action
functions to generate a Context that can be used with a show action
corresponding to the type of the patch action. The show action is
used to retrieve the item that will be patched.
The ``show_context`` is derived from the original patch Context,
which is the ``context`` input argument. Certain values are propagated
from the input ``context`` to the returned ``show_context``.
:param context: Context from the original patch request
:type context: Context
:param for_update: if ``True``, then ``show_context['for_update'] = True``.
If ``False`` (the default), then ``'for_update'`` will not be
explicitly set in ``show_context``.
:type for_update: bool
:returns: A Context, ``show_context``, with the appropriate settings.
'''
show_context: Context = {
'model': context['model'],
'session': context['session'],
'user': context['user'],
'auth_user_obj': context['auth_user_obj'],
'ignore_auth': context.get('ignore_auth', False),
}
if for_update:
show_context['for_update'] = True
return show_context
``` | 0easy
|
Title: sentry raven lib is deprecated
Body: Sentry related documentation looks outdated.
raven package is outdated (3y ago) in favor of [sentry-sdk](https://docs.sentry.io/platforms/python/) lib | 0easy
|
Title: Add example for how to extend support for other serialization formats
Body: This can be done by [writing a custom converter factory](https://uplink.readthedocs.io/en/stable/user/serialization.html#writing-a-custom-converter). We should add an example to the [`examples`](https://github.com/prkumar/uplink/tree/master/examples) directory on how to do this. | 0easy
|
Title: Allow access to sub-factory fields from build method.
Body: It would be great to be able to customize the field of a sub-factory (or sub-sub-factory, etc.) right from the parent's build method. For example:
```python
class Bar(BaseModel):
x: int
class Foo(BaseModel):
a: int
bar: Bar
class FooFactory(ModelFactory):
__model__ = Foo
test_foo = FooFactory.build(bar__x=42)
print(f'{test_foo!r}') # prints Foo(a=9817, bar=Bar(x=42))
```
A real example of code that I have to write is something like this:
```python
def test_the_lambda_handler(item_of_interest_fixture):
test_event = SqsModelFactory.build((Records=[
SqsRecordsFactory.build(
body=WebhookEventFactory.build(
details=WebhookEventDetailsFactory.build(resource_id=item_of_interest_fixture.id)
).json()
)
]))
```
I'd love to be able to do something like this:
```python
def test_the_lambda_handler(item_of_interest_fixture):
test_event = SqsModelFactory.build(
Records = SqsRecordsFactory.batch(size=1, body__details__resource_id=item_of_interest_fixture.id)
)
```
Everything else can be either the default values or whatever I've customized in the various sub-factories and the test has the value I'm interested in. | 0easy
|
Title: StaticRouteAsync leaves open files
Body: When using static routes with a [`falcon.asgi.App`](https://falcon.readthedocs.io/en/stable/api/app.html#asgi-app), it seems that the `_AsyncFileReader` wrapper does not implement any `.close()` method, so files are left open.
On CPython, I wasn't able to demonstrate any practical impact of this bug as the file object in question is refcounted to 0 and garbage collected as soon as it goes out of scope. However, that isn't the case when running `uvicorn` on PyPy 3.7, as PyPy uses a different GC implementation.
Test case in point:
```
import io
import logging
import os.path
import unittest.mock
import falcon.asgi
logging.basicConfig(
format='%(asctime)s [%(levelname)s] %(message)s', level=logging.INFO)
class DebugIO(io.BytesIO):
@classmethod
def open(cls, *args, **kwargs):
return cls(b'Test data!\n')
def close(self):
logging.info(f'{self}.close()')
super().close()
app = falcon.asgi.App()
app.add_static_route('/files', '/tmp')
debug = unittest.mock.patch('io.open', DebugIO.open)
debug.start()
``` | 0easy
|
Title: IdentityGate Visualization
Body: **Description of the issue**
IdentityGate constructor accepts a single parameter which is the number of qubits it acts on.
When the parameter value is set to 0, the circuit works fine until the circuit is printed. Program crashes with an ValueError.
**How to reproduce the issue**
```python
import cirq as c
idgate = c.IdentityGate(0)
qubit = c.LineQubit(0)
circuit = c.Circuit()
print(idgate())
circuit.append(idgate())
print(circuit)
circuit += c.inverse(circuit)
circuit.append(c.measure_each(qubit))
print(c.Simulator().run(circuit, repetitions=100))
```
<details>
```bash
File "/Users/xxx/identity_gate_cirq.py", line 11, in <module>
print(circuit)
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/circuits/circuit.py", line 285, in __str__
return self.to_text_diagram()
^^^^^^^^^^^^^^^^^^^^^^
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/circuits/circuit.py", line 1197, in to_text_diagram
diagram = self.to_text_diagram_drawer(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/circuits/circuit.py", line 1276, in to_text_diagram_drawer
_draw_moment_in_diagram(
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/circuits/circuit.py", line 2612, in _draw_moment_in_diagram
_draw_moment_annotations(
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/circuits/circuit.py", line 2543, in _draw_moment_annotations
info = get_circuit_diagram_info(annotation, args)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/xxx/cirqenv/lib/python3.11/site-packages/cirq/protocols/circuit_diagram_info_protocol.py", line 345, in _op_info_with_fallback
raise ValueError(f'Wanted diagram info from {op!r} for {rows!r}) but got {info!r}')
ValueError: Wanted diagram info from cirq.IdentityGate(0)() for []) but got cirq.CircuitDiagramInfo(wire_symbols=(), exponent=1, connected=True, exponent_qubit_index=None, auto_exponent_parens=True)
```
</details>
**Cirq version**
You can get the cirq version by printing `cirq.__version__`. From the command line:
```
Cirq: 1.4.1
```
| 0easy
|
Title: CLI arguments inconsistent: --inputfile and --inputdir
Body: Suggested solution: change `--outputdir` command-line argument to `--output-dir`:
1. Change behavior in `cli.py`
2. Update the tests in `tests/cli_tests.py` | 0easy
|
Title: Add a chat preamble feature
Body: It would be cool to be able to add a preamble that gets inserted at the start of every prompt, even when the rest gets summarized. This can be used for example to define the character that the bot should play.
A possible way in which this could be set by the user is by issuing a command: !gp <preamble>
| 0easy
|
Title: Should chunked data keep current chunk shape by default when saving?
Body: If a hyperspy signal already contains chunked dask array data, it will currently not automatically save the array with those chunk shapes unless specified in the `chunks` argument to `s.save()`. Should it? I think it would be the expected behaviour to keep the chunk shape.
It would be good to clarify this in the documentation, since people typically save large files.
**Example code:**
```python
import hyperspy.api as hs
import dask.array as da
data = da.random.random((10000, 3000))
data = data.rechunk((100, 3000))
s = hs.signals.Signal1D(data)
s = s.as_lazy()
s.data # chunks: (100, 3000)
s.save('test.hspy')
s2 = hs.load('test.hspy', lazy=True)
s2.data # chunks: (40, 3000)
s.save('test2.hspy', chunks = (100, 3000))
s3 = hs.load('test2.hspy', lazy=True)
s3.data # chunks: (100, 3000)
``` | 0easy
|
Title: Mutation filters not working
Body: When specifying a mutation filter as per the docs, the filter does not appear in the generated schema:
```python
@strawberry.type
class Mutation:
updateFruits: List[Fruit] = mutations.update(FruitPartialInput, filters=FruitFilter)
deleteFruits: List[Fruit] = mutations.delete(filters=FruitFilter)
```
- Operating system:
- Strawberry version (if applicable): 0.195.2 | 0easy
|
Title: Document requirements for custom dissimilarity functions
Body: I am trying to implement hamming distance for categorical data, but I am getting an error
`C:\Users\Mukul.Sharma\AppData\Local\Continuum\Anaconda3\lib\site-packages\kmodes\kmodes.py in init_huang(X, n_clusters, dissim)
39 # so set centroid to closest point in X.
40 for ik in range(n_clusters):
---> 41 ndx = np.argsort(dissim(X, centroids[ik]))
42 # We want the centroid to be unique, if possible.
43 while np.all(X[ndx[0]] == centroids, axis=1).any() and ndx.shape[0] > 1:
`
my hamming distance is:
`def hamming_distance(s1, s2):
"""Return the Hamming distance between equal-length sequences"""
if len(s1) != len(s2):
raise ValueError("Undefined for sequences of unequal length")
return sum(el1 != el2 for el1, el2 in zip(s1, s2))
`
I am having issues with scipy hamming as well (scipy.spatial.distance.hamming)
Here the error says
`ValueError: Input vector should be 1-D`
Can you please help me ?
Also give me an idea for writing my custom distance metric, like telling me the internal working of this algo (K-prototypes? | 0easy
|
Title: Improve `TestClient` to automatically add `raw_path` to ASGI scope
Body: Thanks to @liborjelinek, the WSGI flavour of the framework does this out of the box as per #2159.
To make things consistent, we ought to add `raw_path` to ASGI scope. | 0easy
|
Title: Login Credentials Gone upon Navigation
Body: from https://talk.desec.io/t/login-session-expires-when-navigating-the-site/468:
> When logged in on the [desec.io](http://desec.io/) site, and i navigate to some of the other menus, and when i return to the DNS overview, i am no longer logged in. Thats very annoying when i have to login every time i navigate somewhere else. You state that i am logged out by inactivity, but this is a matter of seconds after logging in.
>
> Please also consider the ability to save login credentials in the browser. Almost all sites have that capability, and its up to the user if they think about security.
>
> I use momentarily the chrome browser on a mac computer.
I confirmed the issue with Chrome on Ubuntu. | 0easy
|
Title: 🐛Fix table in export
Body: ## Bug Report
When we export a document the tables are not rendered correctly.
| 0easy
|
Title: First class support for GraphQL frameworks
Body: ## Problem
Prisma + GraphQL is a very good [combination](https://www.prisma.io/graphql), we should at the very least provide examples for using Prisma Client Python with a GraphQL framework and at the most we should provide code generation support for basic CRUD queries.
List of potential GraphQL frameworks:
- [ ] [Graphene](https://docs.graphene-python.org)
- [ ] [Ariadne](https://github.com/mirumee/ariadne)
- [ ] [Strawberry](https://github.com/strawberry-graphql/strawberry)
## Suggested solution
* Add examples for using GraphQL frameworks with prisma
* Create extension modules for generating GraphQL framework schemas
The generated schemas and objects must be very easily extendable.
| 0easy
|
Title: Create own types instead of using `asgiref`
Body: We are using `asgiref` only for typing, but it has been super bothersome over the years.
PR welcome to create a new `_typing.py`/`_types.py` module on Uvicorn with the needed ASGI types. | 0easy
|
Title: User keyword teardown missing from running model JSON schema
Body: The running model got JSON serialization support in RF 6.1 (#3902) and the JSON format got a schema. The schema currently doesn't include teardown with user keywords.
The reason our tests didn't notice this issue is that the schema allows extra attributes. Thus the teardown included in actual spec files was accepted although it wasn't in the schema. My intention was to not allow extra attributes in the schema, but apparently configuring the model root element wasn't enough and instead the configuration needs to be done in a common base class. This needs to be fixed as well. | 0easy
|
Title: [BUG] X Frame Options Removed from Advanced Form
Body: Hello Django CMS team, we recently upgraded our project from 3.11 to 4.1. The migrations work perfectly, but we noticed that X Frame Options in Advanced Page settings removed, I see this commit: https://github.com/django-cms/django-cms/pull/6503
## Description
X Frame Options in Advanced Page settings removed from Advanced Settings Form. I see this field available in PageContent model, but now it's not editable everywhere
## Steps to reproduce
1. Go to Page Tree
2. Select one page, then click "Advanced Settings" from hamburger menu
3. No X Frame Options displayed, previously we have one.
## Expected behaviour
X Frame Options field available in Advance Settings Form.
## Actual behaviour
X Frame Options field not displayed in Advance Settings Form.
## Screenshots
Django 4.1

Django 3.11

## Additional information (CMS/Python/Django versions)
Python 3.11
Django 4.2
Django CMS 4.1 | 0easy
|
Title: Activity Dates and Times metric API
Body: The canonical definition is here: https://chaoss.community/?p=3444 | 0easy
|
Title: provide a way to init a .yaml file
Body:
### Description
An initialized yaml file should be created automatically for the user. The user should not waste time creating a yaml file and copy pasting commands from igel's docs.
A better way would be to provide some command (e.g `igel init` or `igel start`) that will create a yaml file in the current working directory, where default commands are written.
| 0easy
|
Title: Low Volume
Body: **Describe the feature you'd like**
An environment variable to set the output volume level. The volume seems low on the output files.
| 0easy
|
Title: Allow PATCH HTTP Method
Body: Hi, I see that HTTP requests PATCH method are not implemented, for any particular reason?
Thank you.
Congratulations on the project, it is a great idea. 👌 | 0easy
|
Title: Revamp things-advanced (WSGI/ASGI) examples
Body: We ought to showcase media handling in these instead of manually wrangling (de-)serialization. It is fine to show what is possible, but the emphasis should still be on media.
Revise App vs API usage: `Useful for debugging problems in your API;` :arrow_right: `Useful for debugging problems in your App;`.
Also check what's going on with the missing `Content-Length` in the ASGI variant, it looks like the proposed middleware doesn't handle it particularly well. | 0easy
|
Title: Fix a missing piece in documentation
Body: **Describe the bug**
A bit of documentation is not displayed due to a formatting error:
<img width="626" alt="Image" src="https://github.com/user-attachments/assets/db045709-69c0-4169-90bb-212df3b14493" />
Relevant piece of code https://github.com/tortoise/tortoise-orm/blob/develop/docs/index.rst?plain=1#L62
**To Reproduce**
`make docs`
| 0easy
|
Title: "Not implemented" exception in TypeRef.to_json()
Body: Input:
```
(Pdb) pp self.__dict__
{'is_list': True,
'is_list_item_nullable': True,
'is_nullable': False,
'kind': 'SCALAR',
'list': True,
'name': 'String',
'non_null': True,
'non_null_item': False}
```
Output:
```
(Pdb) where
...
> /Users/nikitastupin/Documents/repos/clairvoyance/clairvoyance/graphql.py(234)to_json()
-> raise Exception("Not implemented")
```
Expected output:
```
{
"name": null,
"kind": "NON_NULL",
"ofType": {
"name": null,
"kind": "LIST",
"ofType": {
"name": "String",
"kind": "SCALAR",
"ofType": null
}
}
}
``` | 0easy
|
Title: Foreign key input type is not in any way checked to actually represent the same model as in ForeignKey(model=...) parameter
Body: **Describe the bug**
If fields.ForeignKeyField model parameter is not the same as in fields.ForeignKeyRelation, there's no help pointing at this inconsistency. It will work sometimes, sometimes it will fail
It is deterministic but took me like 2 hours using various search engines and LLMs to actually spot that by sanity check, when I managed to minimise one passing test's preamble (creating initial objects in database) close enough to second, failing test.
It also should be said somewhere in the start of this issue that I tested this bug only on **SQLite**, using `aiosqlite`.
**To Reproduce**
Create empty folder. Create `testapp` folder in it.
Create file models.py:
```python
from tortoise import fields
from tortoise import Model
class TableA(Model):
id = fields.IntField(primary_key=True)
name = fields.CharField(max_length=255)
class TableB(Model):
id = fields.IntField(primary_key=True)
name = fields.CharField(max_length=255)
class TableC(Model):
id = fields.IntField(primary_key=True)
fkrelation: fields.ForeignKeyRelation[TableA] = fields.ForeignKeyField(
# Look at it as it is a typo and it should be "models.TableA" like in type annotation
# In my case it was an actual typo - I copied code and changed only related_name (and strangely changed to absolutely unrelated name (no pun intended), so I changed related_name again when I found that "model" parameter is incorrect).
model_name="models.TableB",
related_name="c_list"
)
name = fields.CharField(max_length=255)
```
Create file main.py:
<details>
<summary>Long file so I put it under spoiler</summary>
```python
import asyncio
from tortoise import Tortoise
from testapp.models import TableA, TableB, TableC
async def will_fail():
a = await TableA.create(name="tableA")
b = await TableB.create(id=5, name="tableB") # unrelated id
await TableC.create(name="tableC", fkrelation=a)
async def will_probably_pass():
a = await TableA.create(name="tableA")
b = await TableB.create(name="tableB")
assert a.id == b.id # if this is false, next line will fail
await TableC.create(name="tableC", fkrelation=a) # Here TableA as I intended provided in place of TableB as expected by library
async def will_definitely_pass():
a = await TableA.create(id=10, name="tableA")
b = await TableB.create(id=10, name="tableB")
await TableC.create(name="tableC", fkrelation=a) # Here TableA as I intended provided in place of TableB as expected by library
async def main():
config = {
"connections": {"default": "sqlite://:memory:"},
"apps": {
"models": {
"models": ["testapp.models"],
"default_connection": "default",
}
}
}
await Tortoise.init(config=config)
await Tortoise.generate_schemas()
try:
await will_fail()
except Exception as e:
print("1. Failed as excepted")
print("1.", e)
# run in isolation - close_connections will remove database from memory
await Tortoise.close_connections()
await Tortoise.init(config=config)
await Tortoise.generate_schemas()
try:
await will_probably_pass()
print("2. Passed as expected")
except Exception as e:
print("2. Failed as not expected but possible")
print("2.", e)
await Tortoise.close_connections()
await Tortoise.init(config=config)
await Tortoise.generate_schemas()
try:
await will_definitely_pass()
print("3. Passed as excepted")
except Exception as e:
print("3. Failed as not excepted")
print("3.", e)
# exit does not work here, and it is not obvious that I need to close connections as it is an in-memory test
await Tortoise.close_connections()
if __name__ == '__main__':
asyncio.run(main())
```
Also I naively tried to exit(0) at the end when I see that script hangs, but it does not work (idk why). It is another (documentation?) issue but is not as frustrating as the one I wrote this reproducer for.
</details>
Run `main.py`. It will output this:
```
1. Failed as excepted
1. FOREIGN KEY constraint failed
2. Passed as expected
3. Passed as excepted
```
Second test is probabilistic but more close to production code (as it does not have hardcoded id values). Third test is deterministic. Otherwise they are the same.
It may be confusing that this script prints "as expected" in all three cases, but let me clarify that it is expected *to reproduce bug*.
~~Bonus points to the one who will make this reproducer one-file, it complains that default connection cannot be None otherwise.~~
**Expected behavior**
I think that all three tests should fail with invalid type passed. Or probably this behavior is documented somewhere (didn't found, documentation is very confusing), okay, then Tortoise ORM should verify that type annotation is consistent with actual runtime type, it will at least catch errors in the class of "changed in one place and forgot in other".
**Additional context**
Add any other context about the problem here.
I ran `pipenv update` and I have no version pinning, so I am on latest version as verified by `pipenv requirements`:
```
...
tortoise-orm==0.22.1; python_version >= '3.8' and python_version < '4.0'
...
```
This bug is pretty mischief. As you see in third case, I create rows in both tables in the same id:
```
async def will_definitely_pass():
a = await TableA.create(id=10, name="tableA")
b = await TableB.create(id=10, name="tableB")
...
```
It is the requirement for this bug to hide: there's row with the same id in another table. Idk how about production deployments, but it did not interfere with my first test as I first created rows in tables without foreign keys, and when I started writing second test where I needed to modify only *some* of all tables first test touched, I started thinking that `sqlite://:memory:` is cleared between Tortoise.init and Tortoise.close_connections, but it is not true. And the reproducer above works with in-file databases too, just remember to delete database file between all case.
If you look to generated schema, you'll see foreignkey from TableC to TableB, so foreign key violation is obvious (once you take a look at generated schema of course).
Also, type annotations are not a requirement - the only inconsistency required is between `model` parameter in `fields.ForeignKeyField` and actual type passed to this field at runtime. | 0easy
|
Title: [Bug] Please help! I encountered an error when using the command to deploy the model, but there was no error message.
Body: ### Checklist
- [ ] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
- [ ] 3. Please note that if the bug-related issue you submitted lacks corresponding environment info and a minimal reproducible demo, it will be challenging for us to reproduce and resolve the issue, reducing the likelihood of receiving feedback.
- [ ] 4. If the issue you raised is not a bug but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed.
- [ ] 5. Please use English, otherwise it will be closed.
### Describe the bug
I got error
`/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
INFO 02-22 01:37:05 __init__.py:190] Automatically detected platform cuda.
[2025-02-22 01:37:14] server_args=ServerArgs(model_path='sophosympatheia/Midnight-Miqu-70B-v1.0', tokenizer_path='sophosympatheia/Midnight-Miqu-70B-v1.0', tokenizer_mode='auto', load_format='auto', trust_remote_code=False, dtype='auto', kv_cache_dtype='auto', quantization_param_path=None, quantization=None, context_length=None, device='cuda', served_model_name='sophosympatheia/Midnight-Miqu-70B-v1.0', chat_template=None, is_embedding=False, revision=None, skip_tokenizer_init=False, host='127.0.0.1', port=30000, mem_fraction_static=0.9, max_running_requests=None, max_total_tokens=None, chunked_prefill_size=8192, max_prefill_tokens=16384, schedule_policy='lpm', schedule_conservativeness=1.0, cpu_offload_gb=0, prefill_only_one_req=False, tp_size=8, stream_interval=1, stream_output=False, random_seed=288994143, constrained_json_whitespace_pattern=None, watchdog_timeout=300, download_dir=None, base_gpu_id=0, log_level='info', log_level_http=None, log_requests=False, show_time_cost=False, enable_metrics=False, decode_log_interval=40, api_key=None, file_storage_pth='sglang_storage', enable_cache_report=False, dp_size=1, load_balance_method='round_robin', ep_size=1, dist_init_addr=None, nnodes=1, node_rank=0, json_model_override_args='{}', lora_paths=None, max_loras_per_batch=8, lora_backend='triton', attention_backend='flashinfer', sampling_backend='flashinfer', grammar_backend='outlines', speculative_draft_model_path=None, speculative_algorithm=None, speculative_num_steps=5, speculative_num_draft_tokens=64, speculative_eagle_topk=8, enable_double_sparsity=False, ds_channel_config_path=None, ds_heavy_channel_num=32, ds_heavy_token_num=256, ds_heavy_channel_type='qk', ds_sparse_decode_threshold=4096, disable_radix_cache=False, disable_jump_forward=False, disable_cuda_graph=False, disable_cuda_graph_padding=False, enable_nccl_nvls=False, disable_outlines_disk_cache=False, disable_custom_all_reduce=False, disable_mla=False, disable_overlap_schedule=False, enable_mixed_chunk=False, enable_dp_attention=False, enable_ep_moe=False, enable_torch_compile=False, torch_compile_max_bs=32, cuda_graph_max_bs=160, cuda_graph_bs=None, torchao_config='', enable_nan_detection=False, enable_p2p_check=False, triton_attention_reduce_in_fp32=False, triton_attention_num_kv_splits=8, num_continuous_decode_steps=1, delete_ckpt_after_loading=False, enable_memory_saver=False, allow_auto_truncate=False, return_hidden_states=False, enable_custom_logit_processor=False, tool_call_parser=None, enable_hierarchical_cache=False, enable_flashinfer_mla=False)
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
/home/ubuntu/sglang/lib/python3.10/site-packages/transformers/models/auto/image_processing_auto.py:590: FutureWarning: The image_processor_class argument is deprecated and will be removed in v4.42. Please use `slow_image_processor_class`, or `fast_image_processor_class` instead
warnings.warn(
INFO 02-22 01:37:18 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:18 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
INFO 02-22 01:37:19 __init__.py:190] Automatically detected platform cuda.
[2025-02-22 01:37:30 TP2] Init torch distributed begin.
[2025-02-22 01:37:32 TP7] Init torch distributed begin.
[2025-02-22 01:37:32 TP3] Init torch distributed begin.
[2025-02-22 01:37:32 TP0] Init torch distributed begin.
[2025-02-22 01:37:32 TP4] Init torch distributed begin.
[2025-02-22 01:37:32 TP5] Init torch distributed begin.
[2025-02-22 01:37:32 TP6] Init torch distributed begin.
[2025-02-22 01:37:32 TP1] Init torch distributed begin.
[2025-02-22 01:37:33 TP5] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP1] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP0] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP2] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP3] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP4] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP7] sglang is using nccl==2.21.5
[2025-02-22 01:37:33 TP6] sglang is using nccl==2.21.5
Fatal Python error: Segmentation fault
Current thread 0x000072862798f1c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
[2025-02-22 01:37:39] Rank 0 scheduler is dead. Please check if there are relevant logs.
Fatal Python error: Segmentation fault
Current thread 0x00007c21b539c1c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
Fatal Python error: Segmentation fault
Current thread 0x000078dfe49a51c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
Fatal Python error: Segmentation fault
Current thread 0x0000763bcdcde1c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
Fatal Python error: Segmentation fault
Current thread 0x0000726fb547b1c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
[2025-02-22 01:37:39] Exit code: -11
Fatal Python error: Segmentation fault
Current thread 0x000071453395c1c0 (most recent call first):
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl_wrapper.py", line 351 in ncclCommInitRank
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/device_communicators/pynccl.py", line 108 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 233 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 890 in init_model_parallel_group
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/distributed/parallel_state.py", line 1055 in initialize_model_parallel
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 262 in init_torch_distributed
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/model_executor/model_runner.py", line 187 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker.py", line 68 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/tp_worker_overlap_thread.py", line 63 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 240 in __init__
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/managers/scheduler.py", line 1816 in run_scheduler_process
File "/usr/lib/python3.10/multiprocessing/process.py", line 108 in run
File "/usr/lib/python3.10/multiprocessing/process.py", line 314 in _bootstrap
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 129 in _main
File "/usr/lib/python3.10/multiprocessing/spawn.py", line 116 in spawn_main
File "<string>", line 1 in <module>
Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, uvloop.loop, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, psutil._psutil_linux, psutil._psutil_posix, setproctitle, zmq.backend.cython._zmq, yaml._yaml, markupsafe._speedups, PIL._imaging, PIL._imagingft, msgspec._core, msgpack._cmsgpack, google._upb._message, ray._raylet, sentencepiece._sentencepiece, regex._regex (total: 50)
Traceback (most recent call last):
File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
exec(code, run_globals)
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/launch_server.py", line 14, in <module>
launch_server(server_args)
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/entrypoints/http_server.py", line 491, in launch_server
tokenizer_manager, scheduler_info = _launch_subprocesses(server_args=server_args)
File "/home/ubuntu/sglang/lib/python3.10/site-packages/sglang/srt/entrypoints/engine.py", line 449, in _launch_subprocesses
data = scheduler_pipe_readers[i].recv()
File "/usr/lib/python3.10/multiprocessing/connection.py", line 250, in recv
buf = self._recv_bytes()
File "/usr/lib/python3.10/multiprocessing/connection.py", line 414, in _recv_bytes
buf = self._recv(4)
File "/usr/lib/python3.10/multiprocessing/connection.py", line 383, in _recv
raise EOFError
EOFError`
### Reproduction
I use the commend ` python -m sglang.launch_server --model-path sophosympatheia/Midnight-Miqu-70B-v1.0 --mem-fraction-static 0.9 --tp 8`
### Environment
INFO 02-22 01:40:04 __init__.py:190] Automatically detected platform cuda.
Python: 3.10.12 (main, Nov 6 2024, 20:22:13) [GCC 11.4.0]
CUDA available: True
GPU 0,1,2,3,4,5,6,7: NVIDIA H100 80GB HBM3
GPU 0,1,2,3,4,5,6,7 Compute Capability: 9.0
CUDA_HOME: /usr
NVCC: Cuda compilation tools, release 12.4, V12.4.131
CUDA Driver Version: 550.127.08
PyTorch: 2.5.1+cu124
sglang: 0.4.3.post2
sgl_kernel: 0.0.3.post6
flashinfer: 0.2.1.post2+cu124torch2.5
triton: 3.1.0
transformers: 4.48.3
torchao: 0.8.0
numpy: 1.26.4
aiohttp: 3.11.12
fastapi: 0.115.8
hf_transfer: 0.1.9
huggingface_hub: 0.29.1
interegular: 0.3.3
modelscope: 1.23.1
orjson: 3.10.15
packaging: 24.2
psutil: 7.0.0
pydantic: 2.10.6
multipart: 0.0.20
zmq: 26.2.1
uvicorn: 0.34.0
uvloop: 0.21.0
vllm: 0.7.2
openai: 1.63.2
tiktoken: 0.9.0
anthropic: 0.46.0
decord: 0.6.0
NVIDIA Topology:
GPU0 GPU1 GPU2 GPU3 GPU4 GPU5 GPU6 GPU7 NIC0 CPU Affinity NUMA Affinity GPU NUMA ID
GPU0 X NV18 NV18 NV18 NV18 NV18 NV18 NV18 SYS 0-103 0 N/A
GPU1 NV18 X NV18 NV18 NV18 NV18 NV18 NV18 SYS 0-103 0 N/A
GPU2 NV18 NV18 X NV18 NV18 NV18 NV18 NV18 SYS 0-103 0 N/A
GPU3 NV18 NV18 NV18 X NV18 NV18 NV18 NV18 SYS 0-103 0 N/A
GPU4 NV18 NV18 NV18 NV18 X NV18 NV18 NV18 SYS 104-207 1 N/A
GPU5 NV18 NV18 NV18 NV18 NV18 X NV18 NV18 SYS 104-207 1 N/A
GPU6 NV18 NV18 NV18 NV18 NV18 NV18 X NV18 SYS 104-207 1 N/A
GPU7 NV18 NV18 NV18 NV18 NV18 NV18 NV18 X SYS 104-207 1 N/A
NIC0 SYS SYS SYS SYS SYS SYS SYS SYS X
Legend:
X = Self
SYS = Connection traversing PCIe as well as the SMP interconnect between NUMA nodes (e.g., QPI/UPI)
NODE = Connection traversing PCIe as well as the interconnect between PCIe Host Bridges within a NUMA node
PHB = Connection traversing PCIe as well as a PCIe Host Bridge (typically the CPU)
PXB = Connection traversing multiple PCIe bridges (without traversing the PCIe Host Bridge)
PIX = Connection traversing at most a single PCIe bridge
NV# = Connection traversing a bonded set of # NVLinks
NIC Legend:
NIC0: mlx5_0
Hypervisor vendor: KVM
ulimit soft: 1024 | 0easy
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.