repo_name
stringlengths
9
75
topic
stringclasses
30 values
issue_number
int64
1
203k
title
stringlengths
1
976
body
stringlengths
0
254k
state
stringclasses
2 values
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
url
stringlengths
38
105
labels
listlengths
0
9
user_login
stringlengths
1
39
comments_count
int64
0
452
huggingface/datasets
pytorch
7,337
One or several metadata.jsonl were found, but not in the same directory or in a parent directory of
### Describe the bug ImageFolder with metadata.jsonl error. I downloaded liuhaotian/LLaVA-CC3M-Pretrain-595K locally from Hugging Face. According to the tutorial in https://huggingface.co/docs/datasets/image_dataset#image-captioning, only put images.zip and metadata.jsonl containing information in the same folder. However, after loading, an error was reported: One or several metadata.jsonl were found, but not in the same directory or in a parent directory of. The data in my jsonl file is as follows: > {"id": "GCC_train_002448550", "file_name": "GCC_train_002448550.jpg", "conversations": [{"from": "human", "value": "<image>\nProvide a brief description of the given image."}, {"from": "gpt", "value": "a view of a city , where the flyover was proposed to reduce the increasing traffic on thursday ."}]} ### Steps to reproduce the bug from datasets import load_dataset image = load_dataset("imagefolder",data_dir='data/opensource_data') ### Expected behavior success ### Environment info datasets==3.2.0
open
2024-12-17T12:58:43Z
2025-01-03T15:28:13Z
https://github.com/huggingface/datasets/issues/7337
[]
mst272
1
zappa/Zappa
django
740
[Migrated] Getting 'AttributeError' Exception Thrown
Originally from: https://github.com/Miserlou/Zappa/issues/1864 by [canada4663](https://github.com/canada4663) I am getting uncaught AttributeError thrown at the line below: https://github.com/Miserlou/Zappa/blob/3ccf7490a8d8b8fa74a61ee39bf44234f3567739/zappa/handler.py#L309 It seems that the code is designed to catch exceptions where the 'command' key doesn't exist, however, if the message doesn't included an object with the get method and is only a 'str' then this fails.
closed
2021-02-20T12:41:37Z
2022-07-16T06:23:36Z
https://github.com/zappa/Zappa/issues/740
[]
jneves
1
AUTOMATIC1111/stable-diffusion-webui
deep-learning
16,807
[Bug]: Updated installation instructions for installing Stable Diffusion using ROCm (Linux) (Documentation and webui.sh needs updating)
### Checklist - [ ] The issue exists after disabling all extensions - [x] The issue exists on a clean installation of webui - [ ] The issue is caused by an extension, but I believe it is caused by a bug in the webui - [x] The issue exists in the current version of the webui - [ ] The issue has not been reported before recently - [ ] The issue has been reported before but has not been fixed yet ### What happened? Run the same instructions as the documentation says for the first part. (Debian): `sudo apt install git python3.10-venv -y` (Fedora): `sudo dnf install python-3.10` `git clone https://github.com/AUTOMATIC1111/stable-diffusion-webui && cd stable-diffusion-webui` `python3.10 -m venv venv` Then update line 156 in webui.sh `pip install torch torchvision --index-url https://download.pytorch.org/whl/nightly/rocm5.7` --> `pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2` Run webui.sh with this command `HSA_OVERRIDE_GFX_VERSION=11.0.0 HIP_VISIBLE_DEVICES=0 ./webui.sh --precision full --no-half` VERSION=11.0.0 is specific to the 7900XTX, version number may change depending on GPU model so check ROCm documentation just in case. If you have a 7900XTX, follow instructions exactly. ### Steps to reproduce the problem 1. Follow the official documentation ### What should have happened? Documentation and ROCm in webui.sh needs updating to make webui.sh work error free. ### What browsers do you use to access the UI ? Brave ### Sysinfo Not needed as the program is running with all features when instructions above is followed. ### Console logs ```Shell ################################################################ Install script for stable-diffusion + Web UI Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer. ################################################################ ################################################################ Running on user user ################################################################ ################################################################ Repo already cloned, using it as install directory ################################################################ ################################################################ Create and activate python venv ################################################################ ################################################################ Launching launch.py... ################################################################ glibc version is 2.40 Check TCMalloc: libtcmalloc_minimal.so.4 libtcmalloc_minimal.so.4 is linked with libc.so,execute LD_PRELOAD=/lib64/libtcmalloc_minimal.so.4 Python 3.10.16 (main, Dec 4 2024, 00:00:00) [GCC 14.2.1 20240912 (Red Hat 14.2.1-3)] Version: v1.10.1 Commit hash: 82a973c04367123ae98bd9abdf80d9eda9b910e2 ControlNet init warning: Unable to install insightface automatically. Please try run `pip install insightface` manually. Launching Web UI with arguments: --precision full --no-half no module 'xformers'. Processing without... no module 'xformers'. Processing without... No module 'xformers'. Proceeding without it. ControlNet preprocessor location: /home/user/AI/stable-diffusion-webui/extensions/sd-webui-controlnet/annotator/downloads 2025-01-21 14:56:37,167 - ControlNet - INFO - ControlNet v1.1.455 Loading weights [a31be20e08] from /home/user/AI/stable-diffusion-webui/models/Stable-diffusion/v1-5-pruned-emaonly.safetensors 2025-01-21 14:56:37,546 - ControlNet - INFO - ControlNet UI callback registered. Running on local URL: http://127.0.0.1:7860 To create a public link, set `share=True` in `launch()`. Creating model from config: /home/user/AI/stable-diffusion-webui/configs/v1-inference.yaml Startup time: 12.6s (prepare environment: 7.1s, import torch: 1.7s, import gradio: 0.4s, setup paths: 1.6s, other imports: 0.3s, load scripts: 0.6s, create ui: 0.4s, gradio launch: 0.3s). ``` ### Additional information _No response_
open
2025-01-21T20:02:02Z
2025-03-01T10:33:19Z
https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/16807
[ "bug-report" ]
theman23290
3
man-group/arctic
pandas
197
Libraries use cases
Hi, I am working with historical minute data and was wondering which is the right library to store it. Most of the columns are numerical but I have some metadata on each transaction as well, some of which that I could encode as numerical just adding more columns (dummy variables), and some that are just strings adding information on the transaction. First thing I thought was to use TickStoreV3, but I'm struggling a bit with the metadata. Also, in a more general way, for which kind of data are the different libraries optimized for? Thanks!
closed
2016-08-12T15:28:45Z
2016-08-20T09:36:55Z
https://github.com/man-group/arctic/issues/197
[]
felipe-ducau7c
8
pallets/flask
python
4,910
Deprecation of before_first_request leads to issue with Gunicorn --preload while launching a Thread in the Setup Code
Flask 2.2.2 has deprecated `before_first_request`, advising us to instead run the setup code in the the factory function that creates the `app`. I've found a case where this solution does not work. The prerequisite is that the application is launched from Gunicorn with the `--preload` option, and that you want to have a global variable which is shared among all the workers/threads. Running generic setup code results in the memory of the global variable apparently not being shared, while running the code in `before_first_request` behaves according to expectations. ----- Here is the code to reproduce. It has a global variable called `global_list`. There is a thread that is supposed to be launched in the setup code that append values onto the global variable. A GET API will return the result of the global variable. When using `before_first_request`, it runs as expected. Calling GET will return the populated values in the list. When using independent setup code, the global variable is not global and is not shared. ``` from flask import Flask, Blueprint import threading import time global_list = [] bp = Blueprint('hello', __name__) @bp.get('/') def home(): return global_list def create_app(): app = Flask(__name__) app.register_blueprint(bp) run_setup_code() # this works fine but is deprecated in 2.2.2 # app.before_first_request(run_setup_code) return app def run_setup_code(): t = threading.Thread(target=work, daemon=True) t.start() def work(): for i in range(100): global_list.append(i) print('in _setup', global_list) time.sleep(2) ``` Run this code via: gunicorn --preload --workers=1 --threads=1 "app:create_app()" --bind=0.0.0.0:3000 You will see from the print statement that the global variable is being appended to. You can confirm this variable is not being shared correctly by calling the following, which will not return the values in the global list. http GET :3000/ ------- The desired behavior here is that the global_variable will be shared between the `work` thread and the gunicorn worker process. Calling `GET /` should return a list of all the values that have been appended in the work thread. The actual behavior is that calling `GET /` will not return the list of values. ----- This works fine with `before_first_request` while using gunicorn's `--preload` option. It also works fine if you remove the `--preload` option and just run the setup code in the factory function. ----- Environment: - Python version:3.9,3.10 - Flask version:2.2.2
closed
2022-12-15T17:13:53Z
2023-01-10T00:05:52Z
https://github.com/pallets/flask/issues/4910
[]
mkmoisen
1
pydantic/bump-pydantic
pydantic
104
Detect missing pydantic dependency (e.g. pydantic-settings or extra types)
Hey team, I wasn't sure if you would be comfortable with this, but what's your opinion on telling the user when they are missing a pydantic dependency, such as `pydantic-settings` is using `BaseSettings` in Pydantic V1? This could spit out a warning (whether that's stdout, stderr, log.txt, or somewhere else) or a `# TODO` comment if the `pydantic-settings` package isn't installed. This will help users more quickly identify when their application requires packages that are not currently installed. To detect if a package is installed without importing it (for safety reasons), `importlib.util` can be used: ``` from importlib.util import find_spec print(find_spec("pydantic-settings")) # '' print(find_spec("pydantic")) # 'ModuleSpec(name='pydantic', loader=<_frozen_importlib_external.SourceFileLoader object at 0x104e18240>, origin='/Users/kkirsche/.asdf/installs/python/3.11.4/lib/python3.11/site-packages/pydantic/__init__.py', submodule_search_locations=['/Users/kkirsche/.asdf/installs/python/3.11.4/lib/python3.11/site-packages/pydantic'])' ``` Documentation on `find_spec`: https://docs.python.org/3/library/importlib.html#importlib.util.find_spec
open
2023-07-18T15:01:16Z
2024-07-21T07:39:26Z
https://github.com/pydantic/bump-pydantic/issues/104
[ "good first issue" ]
kkirsche
3
DistrictDataLabs/yellowbrick
matplotlib
727
Not able to import UMAPVisualizer from yellowbrick.text
I downloaded the latest package of yellowbrick but it was file of umap_vis.py was missing from the package of yellowbrick.text
closed
2019-02-06T05:43:53Z
2019-02-11T03:19:43Z
https://github.com/DistrictDataLabs/yellowbrick/issues/727
[ "type: question" ]
soniaarora
2
flasgger/flasgger
rest-api
590
Latest version 0.9.7.1 break Marshmallow apispec dumping
Maybe it's just an incompatible versions, but i must use this set of packages ``` apispec==5.2.2 marshmallow==3.15 flasgger==0.9.5 ``` i can't upgrade apispec or marshmallow because there are other dependency. with flasgger 0.9.7.1 ``` TypeError: Object of type SchemaMeta is not JSON serializable ``` I suggest to: - fix version requirements with apispec/marshmallow - write versions changelog - better error output when generating apispec. It's struggle to find which API cause the problem. I use swag_from decorator Thanks
open
2023-08-02T09:20:09Z
2023-08-02T09:20:09Z
https://github.com/flasgger/flasgger/issues/590
[]
overbost
0
matplotlib/mplfinance
matplotlib
659
Bug Report: Typo
**Describe the bug** mplfinance/plotting.py", line 946, in plot fig.suptitle(title,**title_kwargs) ^^^^^^^^^^^^ AttributeError: 'NoneType' object has no attribute 'suptitle'
closed
2024-02-15T22:09:17Z
2024-02-16T12:48:30Z
https://github.com/matplotlib/mplfinance/issues/659
[ "bug" ]
aidenkwong
1
NVIDIA/pix2pixHD
computer-vision
129
No module named 'apex'
Do someone know how to solve this problem?
open
2019-05-31T13:34:37Z
2019-06-03T02:12:05Z
https://github.com/NVIDIA/pix2pixHD/issues/129
[]
ashergaga
1
explosion/spaCy
machine-learning
12,034
TypeError: dataclass_transform() got an unexpected keyword argument 'field_specifiers'
After latest pydantic-1.10.3 release, spacy no longer can be imported, raising a TypeError. The root cause is an incompatibility between pydantic-1.10.3 and your requirement `typing_extensions>=3.7.4,<4.2.0; python_version < "3.8"`. See issue in pydantic: - https://github.com/pydantic/pydantic/issues/4885 See fixing PR in pydantic (`typing-extensions>=4.2.0`), which will be incompatible with your requirement `typing_extensions>=3.7.4,<4.2.0; python_version < "3.8"`: - https://github.com/pydantic/pydantic/pull/4886 <!-- Include a code example or the steps that led to the problem. Please try to be as specific as possible. --> To reproduce the error: ```shell pip install spacy python -c "import spacy" ``` raises: ``` Traceback (most recent call last): File "<string>", line 1, in <module> File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/spacy/__init__.py", line 6, in <module> from .errors import setup_default_warnings File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/spacy/errors.py", line 2, in <module> from .compat import Literal File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/spacy/compat.py", line 3, in <module> from thinc.util import copy_array File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/thinc/__init__.py", line 5, in <module> from .config import registry File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/thinc/config.py", line 2, in <module> import confection File ".../venvs/venv-py37-spacy/lib/python3.7/site-packages/confection/__init__.py", line 10, in <module> from pydantic import BaseModel, create_model, ValidationError, Extra File "pydantic/__init__.py", line 2, in init pydantic.__init__ File "pydantic/dataclasses.py", line 46, in init pydantic.dataclasses # | None | Attribute is set to None. | File "pydantic/main.py", line 121, in init pydantic.main TypeError: dataclass_transform() got an unexpected keyword argument 'field_specifiers' ``` ## Your Environment <!-- Include details of your environment. You can also type `python -m spacy info --markdown` and copy-paste the result here.--> * Operating System: * Python Version Used: 3.7.15 * spaCy Version Used: 3.4.4 * Environment Information:
closed
2022-12-30T06:51:22Z
2023-03-05T00:02:40Z
https://github.com/explosion/spaCy/issues/12034
[ "install", "third-party" ]
albertvillanova
9
rougier/from-python-to-numpy
numpy
44
Typos in 7.3
Excellent work! I noticed two minor grammatical issues in section 7.3, "Scipy & co" (file `07-beyond-numpy.rst`): - "there is a trillion" -> "there are a trillion" - "it was not the goal" -> "that was not the goal"
closed
2017-01-12T20:01:19Z
2017-01-13T14:16:51Z
https://github.com/rougier/from-python-to-numpy/issues/44
[]
SeanDS
3
yzhao062/pyod
data-science
355
Question: Does this repo include smoothing?
Referencing https://github.com/cerlymarco/tsmoothie
closed
2021-11-03T06:25:45Z
2021-11-07T22:29:43Z
https://github.com/yzhao062/pyod/issues/355
[]
BradKML
1
tensorpack/tensorpack
tensorflow
961
Accuracy of Res-18 is lower than expected on ImageNet
1. I run the Res-18 by + run examples/ResNet/imagenet-resnet.py only with modification on reading data: + `def get_imagenet_dataflow( datadir, name, batch_size, augmentors, parallel=None): assert name in ['train', 'val', 'test'] assert datadir is not None assert isinstance(augmentors, list) isTrain = name == 'train' if parallel is None: parallel = min(40, multiprocessing.cpu_count() // 2) # assuming hyperthreading lmdb_data = os.path.join(datadir, 'ILSVRC-%s.lmdb'%name) ds = LMDBData(lmdb_data, shuffle=False) if isTrain: ds = LocallyShuffleData(ds, 50000) ds = PrefetchData(ds, 5000, 1) ds = LMDBDataPoint(ds) ds = MapDataComponent(ds, lambda x: cv2.imdecode(x, cv2.IMREAD_COLOR), 0) ds = AugmentImageComponent(ds, augmentors, copy=False) if parallel < 16: logger.warn("DataFlow may become the bottleneck when too few processes are used.") ds = BatchData(ds, batch_size, remainder=False) ds = PrefetchDataZMQ(ds, parallel) ds.reset_state() return ds` + All hyper-parameters are set as default. 2. The top-1 accuracy is 67.5% after 105 epochs, which is 3% lower than reported. 3. Your environment: + Python version: 3.6.7 + TF version: 1.10.0 + Tensorpack version: 0.8.9 + Hardware information: 4 * V100
closed
2018-11-01T06:56:35Z
2019-01-16T23:47:00Z
https://github.com/tensorpack/tensorpack/issues/961
[ "examples" ]
Sunasity
8
hankcs/HanLP
nlp
1,027
分词算法问题
我想把 “机器人” 分词成 “机器” “机器人” “机” “器” “人” 这五个词应该选择哪个分词算法
closed
2018-11-21T02:58:46Z
2018-12-06T08:18:07Z
https://github.com/hankcs/HanLP/issues/1027
[]
quicksandznzn
2
proplot-dev/proplot
matplotlib
28
Stable tagged release?
It would be great if you could make a tagged version release on pip that's stable. The API changes so frequently right now that I feel like I'm rewriting my notebook cells every week when I pull down updates. Then folks could just install a specific version from pip while you continue development. Releasing on PyPI takes < 5 minutes: https://github.com/bradyrx/climpred/blob/master/HOWTORELEASE.rst. You could tag a v1.0.0 and then follow semantic versioning (https://semver.org/) as you release more features.
closed
2019-09-02T16:01:56Z
2019-12-18T04:20:51Z
https://github.com/proplot-dev/proplot/issues/28
[ "high priority", "distribution" ]
bradyrx
12
roboflow/supervision
tensorflow
1,567
Bug in git-committers-plugin-2, v2.4.0
At the moment, an error is observed when running the `mkdocs build` action from develop. ``` File "/opt/hostedtoolcache/Python/3.10.15/x64/lib/python3.10/site-packages/mkdocs_git_committers_plugin_2/plugin.py", line 121, in get_contributors_to_file 'avatar': commit['author']['avatar_url'] if user['avatar_url'] is not None else '' UnboundLocalError: local variable 'user' referenced before assignment ``` This is due to: https://github.com/ojacques/mkdocs-git-committers-plugin-2/issues/72
closed
2024-10-03T21:29:35Z
2024-10-04T23:41:07Z
https://github.com/roboflow/supervision/issues/1567
[ "bug", "documentation", "github_actions" ]
LinasKo
4
Lightning-AI/pytorch-lightning
pytorch
19,992
Fabric example trainer fails with validation
### Bug description Using the example fabric trainer with validation would result in: > AttributeError: Your LightningModule code tried to access `self.trainer.model` but this attribute is not available when using Fabric with a LightningModule. Which would occur during validation when 'on_validation_model_eval' and 'on_validation_model_train' were called, as they do not work with fabric. ### What version are you seeing the problem on? master ### How to reproduce the bug _No response_ ### Error messages and logs ``` # AttributeError: Your LightningModule code tried to access `self.trainer.model` but this attribute is not available when using Fabric with a LightningModule. ``` ### Environment <details> <summary>Current environment</summary> ``` #- Lightning Component (e.g. Trainer, LightningModule, LightningApp, LightningWork, LightningFlow): #- PyTorch Lightning Version (e.g., 1.5.0): #- Lightning App Version (e.g., 0.5.2): #- PyTorch Version (e.g., 2.0): #- Python version (e.g., 3.9): #- OS (e.g., Linux): #- CUDA/cuDNN version: #- GPU models and configuration: #- How you installed Lightning(`conda`, `pip`, source): #- Running environment of LightningApp (e.g. local, cloud): ``` </details> ### More info _No response_
closed
2024-06-18T21:01:03Z
2024-06-21T14:43:31Z
https://github.com/Lightning-AI/pytorch-lightning/issues/19992
[ "bug", "needs triage", "ver: 2.2.x" ]
liambsmith
0
MagicStack/asyncpg
asyncio
1,066
Authentication issue with postgres14 : InternalClientError: unexpected error while performing authentication: Incorrect padding
<!-- Thank you for reporting an issue/feature request. If this is a feature request, please disregard this template. If this is a bug report, please answer to the questions below. It will be much easier for us to fix the issue if a test case that reproduces the problem is provided, with clear instructions on how to run it. Thank you! --> * **asyncpg version**: 0.27.1 * **PostgreSQL version**: 14.6 * **Do you use a PostgreSQL SaaS? If so, which? Can you reproduce the issue with a local PostgreSQL install?**: No. * **Python version**: python 3.8 * **Platform**: linux/x86-64 * **Do you use pgbouncer?**: No * **Did you install asyncpg with pip?**: Yes * **If you built asyncpg locally, which version of Cython did you use?**: No * **Can the issue be reproduced under both asyncio and [uvloop](https://github.com/magicstack/uvloop)?**: No, we are unable to reproduce the issue consistently <!-- Enter your issue details below this comment. --> We have run into an exception `InternalClientError: unexpected error while performing authentication: Incorrect padding` while trying to establish connection to Aurora PostgresSQL. This issue is not happening consistently, as more often than not, request to DB goes through successfully, but scouring through the logs, we've found few instances in past as well. We are using username/password authentication to connect to postgres instance. Below is the error trace ``` asyncpg.exceptions._base.InternalClientError: unexpected error while performing authentication: Incorrect padding return fut.result() File "/usr/local/lib/python3.8/site-packages/asyncpg/connection.py", line 2092, in connect return dialect.connect(*cargs, **cparams) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ dbapi_connection = rec.get_connection() File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 424, in checkout File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 3166, in connect return super(Engine, self).connect() File "/usr/local/lib/python3.8/site-packages/sqlalchemy/orm/session.py", line 747, in _connection_for_bind result = context.throw(*sys.exc_info()) return await greenlet_spawn( File "/usr/local/lib/python3.8/site-packages/sqlalchemy/ext/asyncio/session.py", line 158, in scalar File "/app/./service_handler/api/deps.py", line 80, in get_account_or_404 solved_result = await solve_dependencies( raise exc from None return await self.app(scope, receive, send) result = await app(self.scope, self.receive, self.send) File "/usr/local/lib/python3.8/site-packages/asyncpg/compat.py", line 56, in wait_for value = await result File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 115, in greenlet_spawn return current.driver.switch(awaitable) await_only(self.asyncpg.connect(*arg, **kw)), return self.dbapi.connect(*cargs, **cparams) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 599, in __connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ compat.raise_( compat.raise_( return fn() conn = bind.connect() return self._transaction._connection_for_bind( File "/usr/local/lib/python3.8/site-packages/sqlalchemy/ext/asyncio/session.py", line 139, in execute response = await func(request) File "/usr/local/lib/python3.8/site-packages/starlette/routing.py", line 52, in app await self.app(scope, receive, sender) await self.app(scope, receive, send) asyncpg.exceptions._base.InternalClientError: unexpected error while performing authentication: Incorrect padding return await _connect_addr( await_only(self.asyncpg.connect(*arg, **kw)), raise exception compat.raise_( pool.logger.debug("Error on connect(): %s", e) self.__connect() File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 574, in get_connection File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 421, in checkout File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 424, in checkout File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 301, in connect conn = self._connection_for_bind(bind, close_with_result=True) result = context.throw(*sys.exc_info()) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/ext/asyncio/session.py", line 139, in execute account: Optional[models.Account] = await db.scalar(statement=get_account_stmt) solved_result = await solve_dependencies( await route.handle(scope, receive, send) raise exc from None return fut.result() File "/usr/local/lib/python3.8/asyncio/tasks.py", line 494, in wait_for File "/usr/local/lib/python3.8/site-packages/asyncpg/connect_utils.py", line 881, in _connect return await connect_utils._connect( value = await result File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 115, in greenlet_spawn connection = pool._invoke_creator(self) return fn() return self._wrap_pool_connect(self.pool.connect, _connection) return await greenlet_spawn( File "/app/./service_handler/crud/account.py", line 83, in lookup_account_by_company_id File "/app/./service_handler/api/deps.py", line 80, in get_account_or_404 File "/usr/local/lib/python3.8/site-packages/starlette/routing.py", line 580, in __call__ File "/usr/local/lib/python3.8/site-packages/asyncpg/compat.py", line 56, in wait_for await compat.wait_for(connected, timeout=timeout) return current.driver.switch(awaitable) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/default.py", line 584, in connect return dialect.connect(*cargs, **cparams) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/create.py", line 578, in connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 599, in __connect raise exception compat.raise_( File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ return super(Engine, self).connect() File "/usr/local/lib/python3.8/site-packages/sqlalchemy/orm/session.py", line 1676, in execute File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 120, in greenlet_spawn File "/usr/local/lib/python3.8/site-packages/sqlalchemy/ext/asyncio/session.py", line 158, in scalar solved = await call(**sub_values) File "/usr/local/lib/python3.8/site-packages/fastapi/routing.py", line 204, in app await self.app(scope, receive, send) return await asyncio.wait_for(fut, timeout) File "/usr/local/lib/python3.8/site-packages/asyncpg/connect_utils.py", line 831, in __connect_addr return await __connect_addr(params, timeout, True, *args) File "/usr/local/lib/python3.8/site-packages/asyncpg/connection.py", line 2092, in connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/_concurrency_py3k.py", line 62, in await_only return self.dbapi.connect(*cargs, **cparams) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 605, in __connect dbapi_connection = rec.get_connection() rec._checkin_failed(err, _fairy_was_created=False) return _ConnectionFairy._checkout(self) else engine.raw_connection() return self._transaction._connection_for_bind( result = await self.execute( result = await account.lookup_account_by_company_id(db=session, account_id=account_id, company_id=company_id) File "/usr/local/lib/python3.8/site-packages/fastapi/dependencies/utils.py", line 548, in solve_dependencies response = await func(request) File "/usr/local/lib/python3.8/site-packages/starlette/routing.py", line 52, in app File "/usr/local/lib/python3.8/site-packages/starlette/routing.py", line 241, in handle File "/usr/local/lib/python3.8/site-packages/starlette/exceptions.py", line 71, in __call__ File "/usr/local/lib/python3.8/site-packages/asyncpg/connect_utils.py", line 773, in _connect_addr File "/usr/local/lib/python3.8/site-packages/sqlalchemy/dialects/postgresql/asyncpg.py", line 747, in connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ File "/usr/local/lib/python3.8/site-packages/sqlalchemy/util/langhelpers.py", line 70, in __exit__ fairy = _ConnectionRecord.checkout(pool) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/pool/base.py", line 761, in _checkout File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 3212, in _wrap_pool_connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 3245, in raw_connection File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 96, in __init__ return self._connection_cls(self, close_with_result=close_with_result) File "/usr/local/lib/python3.8/site-packages/sqlalchemy/engine/base.py", line 3166, in connect File "/usr/local/lib/python3.8/site-packages/sqlalchemy/future/engine.py", line 419, in connect conn = bind.connect() File "/usr/local/lib/python3.8/site-packages/sqlalchemy/orm/session.py", line 747, in _connection_for_bind File "/usr/local/lib/python3.8/site-packages/sqlalchemy/orm/session.py", line 1526, in _connection_for_bind await self.app(scope, receive, sender) await self.app(scope, receive, send) File "/usr/local/lib/python3.8/site-packages/starlette/middleware/errors.py", line 159, in __call__ Traceback (most recent call last): File "/usr/local/lib/python3.8/site-packages/starlette/exceptions.py", line 82, in __call__ File "/usr/local/lib/python3.8/site-packages/starlette/middleware/cors.py", line 78, in __call__ await self.app(scope, receive, _send) ```
open
2023-08-17T06:59:26Z
2023-10-12T15:27:01Z
https://github.com/MagicStack/asyncpg/issues/1066
[]
prakharg05
2
AntonOsika/gpt-engineer
python
231
Ability to resume after an error occurs
I didn't see a way to do this, apologies if it is described somewhere. I am attempting to build an application and I noticed any time it exits due to an error the process stops and then I'm stuck. So for example: ``` ... npm install To run the codebase: npm run dev Do you want to execute this code? npm install npm run dev If yes, press enter. Otherwise, type "no" Executing the code... npm WARN deprecated @types/googlemaps@3.43.3: Types for the Google Maps browser API have moved to @types/google.maps. Note: these types are not for the googlemaps npm package, which is a Node API. npm WARN deprecated @material-ui/styles@4.11.5: Material UI v4 doesn't receive active development since September 2021. See the guide https://mui.com/material-ui/migration/migration-v4/ to upgrade to v5. npm WARN deprecated @material-ui/core@4.12.4: Material UI v4 doesn't receive active development since September 2021. See the guide https://mui.com/material-ui/migration/migration-v4/ to upgrade to v5. added 39 packages, removed 347 packages, and audited 354 packages in 5s 44 packages are looking for funding run `npm fund` for details 5 critical severity vulnerabilities To address all issues (including breaking changes), run: npm audit fix --force Run `npm audit` for details. > scavenger@1.0.0 dev > next dev ready - started server on 0.0.0.0:3000, url: http://localhost:3000 info - Using webpack 5. Reason: Enabled by default https://nextjs.org/docs/messages/webpack5 It looks like you're trying to use TypeScript but do not have the required package(s) installed. Please install @types/react by running: yarn add --dev @types/react If you are not trying to use TypeScript, please remove the tsconfig.json file from your package root (and any TypeScript files in your pages directory). (venv) ➜ gpt-engineer git:(main) ✗ ``` Notice the typescript error above, and the process ends and I'm back at the command prompt. Is it possible to resume after adding the missing package with `yarn add --dev @types/react`? If not, it'd be a great feature.
closed
2023-06-19T21:26:15Z
2023-09-12T09:31:58Z
https://github.com/AntonOsika/gpt-engineer/issues/231
[ "good first issue", "triage" ]
danb235
7
recommenders-team/recommenders
data-science
2,089
[ASK] Is binary relevance the only option in RankingMetric class for pyspark evaluation?
While testing metrics for pyspark evaluation, I've noticed that the ranking metrics like NDCG seems to be using binary relevances only, while [python evaluation](https://github.com/recommenders-team/recommenders/blob/main/recommenders/evaluation/python_evaluation.py#L573) has a parameter to chose between binary, exponential or raw relevances. The snippet below shows that behavior (it will only consider which items are relevant, but not accessing their relevances): https://github.com/recommenders-team/recommenders/blob/c2ea583d27bb1a4d58a09a1621d5ce95672ef1dc/recommenders/evaluation/spark_evaluation.py#L292-L295 Is it possible to use exponencial or raw relevances in spark evaluation currently or am I wrong in this analysis?
open
2024-04-18T20:17:53Z
2024-04-18T20:22:14Z
https://github.com/recommenders-team/recommenders/issues/2089
[]
lgabs
0
home-assistant/core
python
140,700
Jellyfin "TypeError: 'NoneType' object is not iterable"
### The problem Getting this error since last update. on Jellyfin side I see regular successful auth in logs. > 2025-03-16 00:43:30.600 ERROR (SyncWorker_11) [JELLYFIN.jellyfin_apiclient_python.api] Invalid URL 'jellyfin.mydomain.net/Users/AuthenticateByName': No scheme supplied. Perhaps you meant https://jellyfin.mydomain.net/Users/AuthenticateByName? 2025-03-16 00:45:08.175 ERROR (SyncWorker_19) [Jellyfin.jellyfin_apiclient_python.http] 500 Server Error: Internal Server Error for url: https://jellyfin.mydomain.net//Sessions 2025-03-16 00:45:08.176 ERROR (SyncWorker_19) [Jellyfin.jellyfin_apiclient_python.http] --[ 500 response ] 500 Server Error: Internal Server Error for url: https://jellyfin.mydomain.net//Sessions 2025-03-16 00:45:08.177 ERROR (MainThread) [homeassistant.components.jellyfin] Unexpected error fetching jellyfin data Traceback (most recent call last): File "/usr/src/homeassistant/homeassistant/helpers/update_coordinator.py", line 380, in _async_refresh self.data = await self._async_update_data() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/src/homeassistant/homeassistant/components/jellyfin/coordinator.py", line 59, in _async_update_data for session in sessions ^^^^^^^^ TypeError: 'NoneType' object is not iterable ### What version of Home Assistant Core has the issue? core-2025.3.3 ### What was the last working version of Home Assistant Core? core-2025.3.2 ### What type of installation are you running? Home Assistant Container ### Integration causing the issue Jellyfin ### Link to integration documentation on our website https://www.home-assistant.io/integrations/jellyfin ### Diagnostics information [home-assistant_jellyfin_2025-03-15T23-48-41.223Z.log](https://github.com/user-attachments/files/19267274/home-assistant_jellyfin_2025-03-15T23-48-41.223Z.log) ### Example YAML snippet ```yaml ``` ### Anything in the logs that might be useful for us? ```txt ``` ### Additional information _No response_
open
2025-03-15T23:57:13Z
2025-03-15T23:58:11Z
https://github.com/home-assistant/core/issues/140700
[ "integration: jellyfin" ]
ytugarev
1
keras-team/autokeras
tensorflow
1,326
The dataset should at least contain 2 batches to be split
``` import pandas as pd import numpy as np import autokeras as ak from tensorflow.keras.datasets import cifar10 from tensorflow.python.keras.utils.data_utils import Sequence from tensorflow.keras.models import model_from_json import os def build_model(): input_layer =ak.Input() cnn_layer = ak.ConvBlock()(input_layer) cnn_layer2 =ak.ConvBlock()(cnn_layer) dense_layer =ak.DenseBlock()(cnn_layer2) dense_layer2 =ak.DenseBlock()(dense_layer) output_layer =ak.ClassificationHead(num_classes=10)(dense_layer2) automodel =ak.auto_model.AutoModel(input_layer,output_layer,max_trials=20,seed=123,project_name="automl") return automodel def build(): ((trainX,trainY),(testX,testY))=cifar10.load_data() automodel = build_model() automodel.fit(trainX,trainY,validation_split=0.2,epochs=40,batch_size=64)#error here if __name__ == '__main__': build() ``` i got this error even trying the example in the docs ``` automodel.fit(trainX,trainY,validation_split=0.2,epochs=40,batch_size=64) File "S:\Anaconda\envs\tensor37\lib\site-packages\autokeras\auto_model.py", line 276, in fit validation_split=validation_split, File "S:\Anaconda\envs\tensor37\lib\site-packages\autokeras\auto_model.py", line 409, in _prepare_data dataset, validation_split File "S:\Anaconda\envs\tensor37\lib\site-packages\autokeras\utils\data_utils.py", line 47, in split_dataset "The dataset should at least contain 2 batches to be split." ValueError: The dataset should at least contain 2 batches to be split. ``` autokeras 1.0.8 keras 2.3.1 tensorflow 2.1.0 numpy 1.19.1 pandas 1.1.1 keras-tuner 1.0.2rc1 python 3.7.7
closed
2020-09-01T03:14:07Z
2021-12-03T20:30:39Z
https://github.com/keras-team/autokeras/issues/1326
[ "bug report", "wontfix" ]
Cariaga
19
graphql-python/graphene-django
django
756
How to use throttling in django-graphene?
The answer says to use django-throttle-requests [Issue 676](https://github.com/graphql-python/graphene-django/issues/676) But how to use it, it gives the following error: This the error with django-throttle-requests `'ResolveInfo' object has no attribute 'META'` for the code: ``` from django.utils.decorators import method_decorator from throttle.decorators import throttle class CreatePageView(graphene.Mutation): pageview = graphene.Field(PageViewType) class Arguments: subscriptionkey = graphene.String(required=True) @method_decorator(throttle(zone='default')) def mutate(self, info, subscriptionkey): pv = PageView(subscriptionkey=subscriptionkey) pv.save() return CreatePageView(pageview=pv) ```
closed
2019-08-23T06:40:04Z
2019-12-26T21:38:58Z
https://github.com/graphql-python/graphene-django/issues/756
[ "wontfix" ]
amiyatulu
2
d2l-ai/d2l-en
deep-learning
2,494
French spelling to fix in chapter_attention-mechanisms-and-transformers/index.md
In file "chapter_attention-mechanisms-and-transformers/index.md" > ...translating the sentence “my feet hurt” to “j’ai mal au pieds” the french sentence should be > ...translating the sentence “my feet hurt” to “j’ai mal aux pieds” using "aux" which is the plural form of "au" as "pieds" is plural.
open
2023-05-17T17:00:12Z
2023-05-17T17:00:12Z
https://github.com/d2l-ai/d2l-en/issues/2494
[]
Serge-45
0
ray-project/ray
machine-learning
51,443
Release test aws_cluster_launcher_minimal failed
Release test **aws_cluster_launcher_minimal** failed. See https://buildkite.com/ray-project/release/builds/36144#0195a7e1-6c9b-4796-ae13-1d73d2b6e00d for more details. Managed by OSS Test Policy
closed
2025-03-18T06:20:19Z
2025-03-24T16:42:37Z
https://github.com/ray-project/ray/issues/51443
[ "bug", "P0", "triage", "release-test", "jailed-test", "ray-test-bot", "weekly-release-blocker", "stability", "clusters" ]
can-anyscale
1
dask/dask
numpy
11,383
New "auto" rechunking can break with Zarr
It looks like the updates in https://github.com/dask/dask/pull/11354 can result in non-uniform chunks when using "auto" rechunking. This is problematic for then writing to Zarr, which needs uniform chunks. For example, this rechunking snippet used to run with the latest `2024.8.2` release ```python import xarray as xr # Load dataset ds = xr.open_zarr( "gs://weatherbench2/datasets/era5/1959-2023_01_10-full_37-1h-0p25deg-chunk-1.zarr", ).drop_encoding() # Subset time_range = slice("2020-01-01", "2020-01-10") subset = ds.sea_surface_temperature.sel(time=time_range) # Rechunk result = subset.chunk({"time": -1, "longitude": "auto", "latitude": "auto"}) # Write result to cloud storage result.to_zarr("gs://coiled-scratch-space/jrbourbeau/test-rechunking-era5.zarr/", mode="w") ``` but now raises this error with `main`: ``` Traceback (most recent call last): File "/Users/james/projects/demos/rechunking/test.py", line 16, in <module> result.to_zarr("gs://coiled-scratch-space/jrbourbeau/test-rechunking-era5.zarr/", mode="w") File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/core/dataarray.py", line 4328, in to_zarr return to_zarr( # type: ignore[call-overload,misc] File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/api.py", line 1697, in to_zarr dump_to_store(dataset, zstore, writer, encoding=encoding) File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/api.py", line 1384, in dump_to_store store.store(variables, attrs, check_encoding, writer, unlimited_dims=unlimited_dims) File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/zarr.py", line 726, in store self.set_variables( File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/zarr.py", line 772, in set_variables encoding = extract_zarr_variable_encoding( File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/zarr.py", line 285, in extract_zarr_variable_encoding chunks = _determine_zarr_chunks( File "/Users/james/mambaforge/envs/examples/lib/python3.10/site-packages/xarray/backends/zarr.py", line 136, in _determine_zarr_chunks raise ValueError( ValueError: Zarr requires uniform chunk sizes except for final chunk. Variable named 'sea_surface_temperature' has incompatible dask chunks: ((240,), (241, 240, 240), (480, 480, 480)). Consider rechunking using `chunk()`. ``` cc @hendrikmakait @phofl
closed
2024-09-10T16:45:12Z
2024-09-12T15:08:23Z
https://github.com/dask/dask/issues/11383
[ "array", "bug" ]
jrbourbeau
0
sherlock-project/sherlock
python
1,798
Cloning into kali, Sherlock cloning is stack at 30% , anyone faced this before?
closed
2023-05-19T07:22:17Z
2023-08-29T12:41:59Z
https://github.com/sherlock-project/sherlock/issues/1798
[]
acrambot
0
huggingface/text-generation-inference
nlp
2,239
Can I somehow change attention type from 'FlashAttention' in the text-server-launcher?
closed
2024-07-16T18:37:45Z
2024-08-24T01:52:31Z
https://github.com/huggingface/text-generation-inference/issues/2239
[ "question", "Stale" ]
wasifmasood
2
flasgger/flasgger
rest-api
574
Please add a security policy
As per GitHub's best practices, please add a https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository to this repo.
open
2023-06-05T01:06:07Z
2023-06-05T01:06:07Z
https://github.com/flasgger/flasgger/issues/574
[]
reedy
0
geopandas/geopandas
pandas
3,434
ENH: Have Geoseries methods called on a GeoDataFrame return a GeoDataFrame?
#### Describe the solution you'd like When a `GeoSeries` function is called on a `GeoDataFrame`, a `GeoSeries` of the result of running the method on the geometry column is returned. I think it would be more practical and intuitif if a copy of the `GeoDataframe` with the `geometry` column replaced with the result of the function would be returned. This is rather a question on whether there is an explicit reason of why the behaviour is as it is or if it is rather historical or ? Because... the non-backwards-compatibility impact would most likely be large to existing code bases, and the gains might not be worth the trouble. #### Additional context ``` python import geopandas as gpd from shapely import Polygon # Input poly1 = Polygon([(1, 1), (8, 1), (9, 9), (1, 8)]) poly2 = Polygon([(2, 1), (8, 1), (8, 8), (2, 8)]) poly3 = Polygon([(1, 2), (9, 2), (7, 9), (2, 8)]) gdf = gpd.GeoDataFrame(geometry=[poly1, poly2, poly3]) repr_point_gdf = gdf.representative_point() print(type(repr_point_gdf)) repr_point_series = gdf.geometry.representative_point() print(type(repr_point_series)) ``` Output: ``` <class 'geopandas.geoseries.GeoSeries'> <class 'geopandas.geoseries.GeoSeries'> ```
open
2024-10-05T06:41:23Z
2024-10-06T12:05:42Z
https://github.com/geopandas/geopandas/issues/3434
[ "enhancement", "wontfix" ]
theroggy
11
strawberry-graphql/strawberry-django
graphql
82
Copied Docs, but getting: module 'strawberry_django.auth' has no attribute 'register'
All other uses of `strawberry.django.auth` are working except for `register()`. Sure enough, even when I debug, the `strawberry.django.auth` global only has three function variables: `current_user`, `login`, and `logout`. **types.py:** ``` from django.contrib.auth import get_user_model from strawberry.django import type, auto, input @type(get_user_model()) class User: username: auto email: auto @input(get_user_model()) class UserInput: username: auto password: auto ``` **schema.py:** @strawberry.type class Query: me: User = auth.current_user() @strawberry.type class Mutation: login: User = auth.login() register: User = auth.register(UserInput) logout = auth.logout() schema = strawberry.Schema(query=Query, mutation=Mutation) manage.py runserver: ``` ... "/usr/local/Cellar/python@3.9/3.9.9/Frameworks/Python.framework/Versions/3.9/lib/python3.9/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1030, in _gcd_import File "<frozen importlib._bootstrap>", line 1007, in _find_and_load File "<frozen importlib._bootstrap>", line 986, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 680, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 850, in exec_module File "<frozen importlib._bootstrap>", line 228, in _call_with_frames_removed File "/root_dir/app_name/urls.py", line 4, in <module> from app_name.schema import schema File "/root_dir/app_name/schema.py", line 18, in <module> class Mutation: File "/root_dir/app_name/schema.py", line 20, in Mutation register: User = auth.register(UserInput) AttributeError: module 'strawberry_django.auth' has no attribute 'register' ```
closed
2021-12-29T06:40:01Z
2022-02-19T19:22:56Z
https://github.com/strawberry-graphql/strawberry-django/issues/82
[]
evanheckert
2
gee-community/geemap
jupyter
1,226
There is shift in X and Y direction of 1 pixel while downloading data using geemap.download_ee_image()
<!-- Please search existing issues to avoid creating duplicates. --> ### Environment Information Please run the following code on your computer and share the output with us so that we can better debug your issue: ```python import geemap geemap.Report() ``` ### Description I am trying to download NASADEM data in EPSG:4326 coordinate system using geemap.download_ee_image(), but the downloaded data has pixel shift both in X and Y direction. The reason of error is due to the absence of crs transformation parameter. The geemap.ee_export_image() gives correct output, but has a limitation on downloadable data. I am looking for a solution to download large image as 1 tile. ### What I Did ``` #!/usr/bin/env python # coding: utf-8 # In[14]: import ee,geemap,os ee.Initialize() # In[15]: # NASADEM Digital Elevation 30m - version 001 elevdata=ee.Image("NASA/NASADEM_HGT/001").select('elevation') # In[16]: spatial_resolution_m=elevdata.projection().nominalScale().getInfo() print(spatial_resolution_m) # In[17]: Map = geemap.Map() Map # In[23]: # Draw any shape on the map using the Drawing tools before executing this code block AOI=Map.user_roi # In[21]: print(elevdata.projection().getInfo()) # In[29]: # geemap.ee_export_image( # elevdata, # r'C:\Users\rbapna\Downloads\nasadem_ee_export_image4.tif', # scale=spatial_resolution_m, # crs=elevdata.projection().getInfo()['crs'], # crs_transform=elevdata.projection().getInfo()['transform'], # region=AOI, # dimensions=None, # file_per_band=False, # format='ZIPPED_GEO_TIFF', # timeout=300, # proxies=None, # ) geemap.download_ee_image( elevdata, r'C:\Users\rbapna\Downloads\nasadem5.tif', region=AOI, crs=elevdata.projection().getInfo()['crs'], scale=spatial_resolution_m, resampling=None, dtype='int16', overwrite=True, num_threads=None ) ```
closed
2022-08-26T11:55:52Z
2022-08-30T16:02:18Z
https://github.com/gee-community/geemap/issues/1226
[ "bug" ]
ravishbapna
9
fastapi/sqlmodel
fastapi
203
The field name is python keywords, the field type is Mysql8.0 JSON
### First Check - [X] I added a very descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I searched the SQLModel documentation, with the integrated search. - [X] I already searched in Google "How to X in SQLModel" and didn't find any information. - [X] I already read and followed all the tutorial in the docs and didn't find an answer. - [X] I already checked if it is not related to SQLModel but to [Pydantic](https://github.com/samuelcolvin/pydantic). - [X] I already checked if it is not related to SQLModel but to [SQLAlchemy](https://github.com/sqlalchemy/sqlalchemy). ### Commit to Help - [X] I commit to help with one of those options 👆 ### Example Code ```python #!/usr/bin/env python3 # -*- coding: utf-8 -*- # author: TingHsi from typing import Optional, List, Any from sqlmodel import Field, Session, SQLModel, create_engine from time import time # the table is created in MySQL 8.0 # SQLModel.metadata.create_all(engine) class Logs_Push(SQLModel, table=True): id: Optional[int] = Field(default=None, primary_key=True) content: str createdAt: int = int(time()) _from: str remark: str result: str type: str userIds: List[str] engine = create_engine("mysql+cymysql://root:123@127.0.0.1:3308/db1?charset=utf8mb4") log_1 = Logs_Push(content="log_1", createdAt=time(),_from='from',remark='remark',result='result',type='im',userIds=['0','1']) log_2 = Logs_Push(content="log_2", createdAt=time(),_from='from',remark='remark',result='result',type='im',userIds=['2','3']) if __name__ == '__main__': with Session(engine) as session: session.add(log_1) session.add(log_2) session.commit() ``` ### Description * I have a field, the name is "from",rename to "_from" that's work in sqlalchemy, but sqlmodel not support; * I have a field "userIds", the type is "json" in mysql8.0, use code "userIds = Column(JSON)" is work in sqlalchemy, use "userIds: List[str]" in sqlmodel not work. ### Operating System macOS ### Operating System Details _No response_ ### SQLModel Version 0.0.5 ### Python Version Python 3.7.3 ### Additional Context I try patch "sqlmodel/main.py" line 378 ,function get_sqlachemy_type add code > if issubclass(field.type_, list) or issubclass(field.type_, dict): > return JSON And change my code "userIds: List[str]" to "userIds: Any" Run this code return error: > if issubclass(field.type_, str): > TypeError: issubclass() arg 1 must be a class my code not work, so i need your helps.
open
2021-12-27T08:12:41Z
2021-12-29T09:44:54Z
https://github.com/fastapi/sqlmodel/issues/203
[ "question" ]
TingHsi
2
wagtail/wagtail
django
12,459
Replace Twitter links with Mastodon links
<!-- Summarise the documentation change you’re suggesting in the Issue title. --> ### Details <!-- Provide a clear and concise description of what you want to happen. --> Wagtail has decided to move away from Twitter/X to Mastodon (specifically Fostodon): - https://x.com/WagtailCMS/status/1834237886285103507 - https://fosstodon.org/@wagtail <!-- If you're suggesting a very specific change to the documentation, feel free to directly submit a pull request. --> ### Working on this <!-- Do you have thoughts on skills needed? Are you keen to work on this yourself once the issue has been accepted? Please let us know here. --> For the files we need to change, refer to past PRs that changed `twitter.com` to `x.com`, e.g. - #12205 - #12234 There may be some others I have missed. Anyone can contribute to this. View our [contributing guidelines](https://docs.wagtail.org/en/latest/contributing/index.html), add a comment to the issue once you’re ready to start.
closed
2024-10-24T10:30:22Z
2024-10-24T11:10:00Z
https://github.com/wagtail/wagtail/issues/12459
[ "Documentation", "good first issue" ]
laymonage
0
graphdeco-inria/gaussian-splatting
computer-vision
695
Issue with Simple-KNN's distCUDA2 function
I found a strange problem when using the ```distCUDA2``` function. If the input tensor is on my second GPU ```cuda:1```, this function will cause the following error (please also see attached screenshot): >terminate called after throwing an instance of 'thrust::system::system_error' > what(): CUDA free failed: cudaErrorIllegalAddress: an illegal memory access was encountered > Aborted (core dumped) <img width="605" alt="image" src="https://github.com/graphdeco-inria/gaussian-splatting/assets/34731246/924183b4-18d4-4049-b06c-3673c67d3976"> However, it works fine when I use ```cuda:0```.
closed
2024-03-06T23:12:17Z
2024-03-06T23:25:11Z
https://github.com/graphdeco-inria/gaussian-splatting/issues/695
[]
PeizhiYan
1
biolab/orange3
data-visualization
7,009
Metavariables are not excluded in feature selection methods
A longstanding issue is that metavariables are not excluded from methods. For example, in "find informative projections" for scatter plots, they appear as suggestions. Also, in feature suggestions, the metas are included. If there are many, the automatic feature selection breaks down. This is a nuisance, as metas often contain the solution to a classification problem. "Find informative mosaics" has the same issue, as does the violin plot where ordering by relevance also includes metas. Tree prediction does ignore them, though. I am currently using version 338 on a Mac, and this error is present in the PC version as well. This issue has existed in every version of Orange that I can recall. Best larerooreal
open
2025-01-30T13:21:18Z
2025-02-19T10:01:28Z
https://github.com/biolab/orange3/issues/7009
[ "needs discussion", "bug report" ]
lareooreal
4
CorentinJ/Real-Time-Voice-Cloning
deep-learning
626
Is this complete app ?
Can i use this on windows 10 ?
closed
2021-01-14T10:40:37Z
2021-01-15T18:37:55Z
https://github.com/CorentinJ/Real-Time-Voice-Cloning/issues/626
[]
Ud4ba
2
tflearn/tflearn
tensorflow
1,078
ModuleNotFoundError: No module named 'tensorflow.contrib.framework'
when i import tflearn, it says `ModuleNotFoundError: No module named 'tensorflow.contrib.framework'`, does someone knows how to deal with this issues? thanks.
open
2018-07-23T11:39:06Z
2019-08-24T14:56:26Z
https://github.com/tflearn/tflearn/issues/1078
[]
ahbon123
2
plotly/dash
dash
2,249
[BUG] dash deployed with gunicorn keeps spamming _reload-hash requests
And this is how it looks like. ``` shaco_1 | 172.18.0.1 - - [24/Sep/2022:18:17:57 +0000] "GET /shaco/_reload-hash HTTP/1.1" 200 112 "http://localhost:8050/shaco/" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/105.0.0.0 Safari/537.36" shaco_1 | 172.18.0.1 - - [24/Sep/2022:18:18:00 +0000] "GET /shaco/_reload-hash HTTP/1.1" 200 112 "http://localhost:8050/shaco/" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/105.0.0.0 Safari/537.36" shaco_1 | 172.18.0.1 - - [24/Sep/2022:18:18:03 +0000] "GET /shaco/_reload-hash HTTP/1.1" 200 112 "http://localhost:8050/shaco/" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/105.0.0.0 Safari/537.36" shaco_1 | 172.18.0.1 - - [24/Sep/2022:18:18:06 +0000] "GET /shaco/_reload-hash HTTP/1.1" 200 112 "http://localhost:8050/shaco/" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/105.0.0.0 Safari/537.36" shaco_1 | 172.18.0.1 - - [24/Sep/2022:18:18:09 +0000] "GET /shaco/_reload-hash HTTP/1.1" 200 112 "http://localhost:8050/shaco/" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/105.0.0.0 Safari/537.36" ``` Why? Shaco is my base_url
closed
2022-09-24T18:25:09Z
2022-09-24T18:32:10Z
https://github.com/plotly/dash/issues/2249
[]
SnoozeFreddo
0
databricks/koalas
pandas
2,163
databricks.koalas.info() doesn't show memory usage
The documentation states that _databricks.koalas.info()_ shows memory usage too but it doesn't in practice. It doesn't show memory usage in the examples given in the documentation either.
closed
2021-05-20T14:25:29Z
2021-05-23T23:36:30Z
https://github.com/databricks/koalas/issues/2163
[ "docs" ]
AliWaheed
2
ultralytics/yolov5
machine-learning
12,715
imgsz calculation
### Search before asking - [X] I have searched the YOLOv5 [issues](https://github.com/ultralytics/yolov5/issues) and [discussions](https://github.com/ultralytics/yolov5/discussions) and found no similar questions. ### Question Hi, I would like to ask about the imgsz calculation. For example, I have an image(Dimension=Width 1143pixelsx Height 1499pixels). Based on my calculation, imgsz default as 640, so 1499 converted to 640 and 1143 is 488. Thus, the image shape should be 488x640. However, the input size torch.Size([1,3,640,512]) when I run TensorRT engine model. May I know why and how to calculate correctly? Thanks ### Additional _No response_
closed
2024-02-07T08:40:00Z
2024-03-19T00:20:32Z
https://github.com/ultralytics/yolov5/issues/12715
[ "question", "Stale" ]
KnightInsight
2
pytorch/pytorch
python
149,131
nn.GaussianNLLLoss and F.gaussian_nll_loss do not work with scalar `var`
### 🐛 Describe the bug The documentation for [nn.GaussianNLLLoss](https://pytorch.org/docs/stable/generated/torch.nn.GaussianNLLLoss.html) states that the `var` input can be a scalar value, but an error occurs if a float is used. Similarly, the documentation for the functional version [nn.functional.gaussian_nll_loss](https://pytorch.org/docs/stable/generated/torch.nn.functional.gaussian_nll_loss.html) says `var` can be a scalar, but throws an error if a float is used. # nn.GaussianNLLLoss ```python import torch import torch.nn as nn loss = nn.GaussianNLLLoss() input = torch.randn(5, 2, requires_grad=True) target = torch.randn(5, 2) var = 1.0 output = loss(input, target, var) ``` ``` Traceback (most recent call last): File "/Users/connorkrill/PycharmProjects/natural_hazards/burgers/scratch/torch_bug.py", line 8, in <module> output = loss(input, target, var) File "/opt/anaconda3/envs/natural_hazards/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1511, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/opt/anaconda3/envs/natural_hazards/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1520, in _call_impl return forward_call(*args, **kwargs) File "/opt/anaconda3/envs/natural_hazards/lib/python3.10/site-packages/torch/nn/modules/loss.py", line 377, in forward return F.gaussian_nll_loss(input, target, var, full=self.full, eps=self.eps, reduction=self.reduction) File "/opt/anaconda3/envs/natural_hazards/lib/python3.10/site-packages/torch/nn/functional.py", line 2858, in gaussian_nll_loss raise ValueError("var is of incorrect size") ValueError: var is of incorrect size ``` # nn.functional.gaussian_nll_loss ```python import torch import torch.nn.functional as F input = torch.randn(5, 2, requires_grad=True) target = torch.randn(5, 2) var = 1.0 output = F.gaussian_nll_loss(input, target, var) ``` ``` Traceback (most recent call last): File "/Users/connorkrill/PycharmProjects/natural_hazards/burgers/scratch/torch_bug.py", line 16, in <module> output = F.gaussian_nll_loss(input, target, var) File "/opt/anaconda3/envs/natural_hazards/lib/python3.10/site-packages/torch/nn/functional.py", line 2841, in gaussian_nll_loss if var.size() != input.size(): AttributeError: 'float' object has no attribute 'size' ``` ### Versions PyTorch version: 2.2.2 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: macOS 14.7.3 (x86_64) GCC version: Could not collect Clang version: 15.0.0 (clang-1500.3.9.4) CMake version: Could not collect Libc version: N/A Python version: 3.10.14 (main, May 6 2024, 14:47:20) [Clang 14.0.6 ] (64-bit runtime) Python platform: macOS-10.16-x86_64-i386-64bit Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Apple M2 Versions of relevant libraries: [pip3] hamiltorch==0.4.1 [pip3] numpy==1.26.4 [pip3] torch==2.2.2 [pip3] torchdiffeq==0.2.4 [pip3] torchinfo==1.8.0 [pip3] torchsummary==1.5.1 [pip3] torchvision==0.17.2 [conda] hamiltorch 0.4.1 pypi_0 pypi [conda] numpy 1.26.4 pypi_0 pypi [conda] torch 2.2.2 pypi_0 pypi [conda] torchdiffeq 0.2.4 pypi_0 pypi [conda] torchinfo 1.8.0 pypi_0 pypi [conda] torchsummary 1.5.1 pypi_0 pypi [conda] torchvision 0.17.2 pypi_0 pypi cc @albanD
closed
2025-03-13T16:42:09Z
2025-03-17T18:16:41Z
https://github.com/pytorch/pytorch/issues/149131
[ "module: loss", "triaged", "module: python frontend" ]
connor-krill
3
raphaelvallat/pingouin
pandas
185
Deprecate tail='one-sided' and rename to alternative
Since SciPy 1.6, several functions in [scipy.stats](https://docs.scipy.org/doc/scipy/reference/stats.html) accept `alternative="two-sided"` (or "greater" or "less"). To be consistent with these changes, I propose to: 1) In all relevant functions, rename the `tail` argument to `alternative` (similar to SciPy and R) 2) In all relevant functions, remove the option to pass `tail="one-sided"` (which currently would automatically select greater or less depending on the sign of the statistic). Instead, users must manually pass either "greater" or "less" to calculate a one-sided test. Therefore, `alternative` will only accept "two-sided", "greater" or "less", which is consistent with SciPy and R.
closed
2021-07-18T20:49:26Z
2021-08-13T18:02:31Z
https://github.com/raphaelvallat/pingouin/issues/185
[ "deprecation :skull:" ]
raphaelvallat
3
ultralytics/ultralytics
machine-learning
19,077
Don't know whether this is a bug.
### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report. ### Ultralytics YOLO Component _No response_ ### Bug I create my own custom dataset and have trained it with yolov8 successfully, but when I tried to predict the images with the best model that I got. An image in the test set was completely wrong, other images were predicted successfully, how does that happen? Here is the bug image in the test set. ![Image](https://github.com/user-attachments/assets/11a45100-6962-433d-ad4f-3cb1a7a884be) The result image: ![Image](https://github.com/user-attachments/assets/d3d04257-0c27-49f7-9017-38d7800f5e1c) ### Environment Ultralytics YOLOv8.2.90 🚀 Python-3.11.9 torch-2.4.0+cu121 CUDA:0 (NVIDIA A100-PCIE-40GB, 40339MiB) Setup complete ✅ (96 CPUs, 754.4 GB RAM, 401.0/438.9 GB disk) OS Linux-5.15.0-113-generic-x86_64-with-glibc2.35 Environment Linux Python 3.11.9 Install pip RAM 754.35 GB CPU Intel Xeon Gold 6248R 3.00GHz CUDA 12.1 ### Minimal Reproducible Example ``` from ultralytics import YOLO # Load a model model = YOLO("2_4/weights/best.pt") # Run batched inference on a list of images results = model(["./0007.jpg",]) # return a list of Results objects #results = model(["./0007.jpg",]) # return a list of Results objects # Process results list for result in results: boxes = result.boxes # Boxes object for bounding box outputs masks = result.masks # Masks object for segmentation masks outputs keypoints = result.keypoints # Keypoints object for pose outputs probs = result.probs # Probs object for classification outputs obb = result.obb # Oriented boxes object for OBB outputs result.show() # display to screen result.save(filename="result.jpg") # save to disk ``` ### Additional _No response_ ### Are you willing to submit a PR? - [ ] Yes I'd like to help by submitting a PR!
open
2025-02-05T01:40:17Z
2025-02-07T02:09:07Z
https://github.com/ultralytics/ultralytics/issues/19077
[ "bug", "detect" ]
Ilovecode93
3
kizniche/Mycodo
automation
626
Trigger failure
## Mycodo Issue Report: - Specific Mycodo Version: 7.2.1 #### Problem Description Please list: All triggers are producing the same error. Will not function. - what were you trying to do: Create a trigger ### Errors 2019-02-08 15:53:34,652 - mycodo.trigger_6745106f - INFO - Activated in 89.4 ms 2019-02-08 15:53:35,038 - mycodo.trigger_6745106f - ERROR - Run Error: trigger_function_actions() got an unexpected keyword argument 'last_measurement' Traceback (most recent call last): File "/var/mycodo-root/mycodo/controller_trigger.py", line 167, in run self.check_triggers() File "/var/mycodo-root/mycodo/controller_trigger.py", line 431, in check_triggers edge=gpio_state) TypeError: trigger_function_actions() got an unexpected keyword argument 'last_measurement' ### Steps to Reproduce the issue: How can this issue be reproduced? Create a trigger with a start point and wait for it to cycle/trigger. ### Additional Notes Sorry for another bug report! No worries or rush on my end.
closed
2019-02-08T23:59:40Z
2019-02-09T00:33:51Z
https://github.com/kizniche/Mycodo/issues/626
[]
ofernander
2
docarray/docarray
pydantic
1,025
V2: rich display of `Document` and `DocumentArray`
Document: - `doc.display_schema()` to display the Document's schema - `doc.display()` to display the Document instance - key, value (including type and shape) - maybe display shorter id: `2184b5 ...` instead of `2184b53f977f566d72f72a6e706edb00 ` - if eg. list of 200 elements is part of the Document, don't show all 200 elements, keep max size of such display manageable. DocumentArray: - schema and length - redundant fields for v2: - `Homogenous Documents` - `Common Attributes`, instead show the schema - `Multimodal dataclass`
closed
2023-01-17T10:35:46Z
2023-01-25T13:23:50Z
https://github.com/docarray/docarray/issues/1025
[]
anna-charlotte
0
huggingface/diffusers
pytorch
11,134
Implement caching on LTX and WAN video models
`CacheConfig` is used to enable **FasterCache** and **PyramidAttentionBroadcast** CacheConfig is present in Hunyuan, Mochi, Latte, Allegro, Cog transformer modules but its not present in WAN or LTX transformer modules: - `WanTransformer3DModel` - `LTXVideoTransformer3DModel` ask is to enable caching functionality for WAN and LTX models. cc @a-r-r-o-w
open
2025-03-21T17:48:50Z
2025-03-21T17:48:50Z
https://github.com/huggingface/diffusers/issues/11134
[]
vladmandic
0
huggingface/datasets
deep-learning
7,220
Custom features not compatible with special encoding/decoding logic
### Describe the bug It is possible to register custom features using datasets.features.features.register_feature (https://github.com/huggingface/datasets/pull/6727) However such features are not compatible with Features.encode_example/decode_example if they require special encoding / decoding logic because encode_nested_example / decode_nested_example checks whether the feature is in a fixed list of encodable types: https://github.com/huggingface/datasets/blob/16a121d7821a7691815a966270f577e2c503473f/src/datasets/features/features.py#L1349 This prevents the extensibility of features to complex cases ### Steps to reproduce the bug ```python class ListOfStrs: def encode_example(self, value): if isinstance(value, str): return [str] else: return value feats = Features(strlist=ListOfStrs()) assert feats.encode_example({"strlist": "a"})["strlist"] = feats["strlist"].encode_example("a")} ``` ### Expected behavior Registered feature types should be encoded based on some property of the feature (e.g. requires_encoding)? ### Environment info 3.0.2
open
2024-10-11T19:20:11Z
2024-11-08T15:10:58Z
https://github.com/huggingface/datasets/issues/7220
[]
alex-hh
2
autokey/autokey
automation
4
Error while installation
When I use command `pip install autokey-py3` It fails with message: ``` pip install autokey-py3 Downloading/unpacking autokey-py3 Downloading autokey-py3-0.93.2.tar.gz (130kB): 130kB downloaded Running setup.py (path:/tmp/pip_build_root/autokey-py3/setup.py) egg_info for package autokey-py3 Traceback (most recent call last): File "<string>", line 17, in <module> File "/tmp/pip_build_root/autokey-py3/setup.py", line 24, in <module> except FileNotFoundError: pass NameError: name 'FileNotFoundError' is not defined Complete output from command python setup.py egg_info: Traceback (most recent call last): File "<string>", line 17, in <module> File "/tmp/pip_build_root/autokey-py3/setup.py", line 24, in <module> except FileNotFoundError: pass NameError: name 'FileNotFoundError' is not defined ```
closed
2014-12-06T18:34:34Z
2014-12-06T21:31:45Z
https://github.com/autokey/autokey/issues/4
[]
dvdvdmt
1
RobertCraigie/prisma-client-py
asyncio
258
Improve partial model generation API for dynamic creation
## Problem <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> A user on the Prisma Python discord showcased their partial model generator which is highly dynamic and makes use of internal features, we should improve support for this so that no internal features have to be used. Their use case: > build a scaffolding tool for generating fully typed RESTAPI endpoints that are glassboxed Link to discussion: https://discord.com/channels/933860922039099444/933875073448804383/937759740061179935 ## Suggested solution <!-- A clear and concise description of what you want to happen. --> Proposed API is still a work in progress.
open
2022-01-31T17:42:40Z
2022-02-01T18:19:45Z
https://github.com/RobertCraigie/prisma-client-py/issues/258
[ "kind/improvement", "level/advanced", "priority/medium" ]
RobertCraigie
0
dgtlmoon/changedetection.io
web-scraping
1,887
[feature] Filter URLs that have errors
Be able to show listing only with URLs having errors. It would be useful to debug and fix URLs that are having some error. At present it is taking a lot of time scrolling and finding URLs that have errors. We have a URL list that has over 5000 URLs. Or alternatively, a button to recheck all URLs that have errors.
closed
2023-10-23T12:05:32Z
2023-10-23T15:55:53Z
https://github.com/dgtlmoon/changedetection.io/issues/1887
[ "enhancement" ]
jgupta
1
mwaskom/seaborn
matplotlib
3,104
stripplot edgecolor ignored
The `stripplot` function is ignoring my `edgecolor` choice. I'm using the Jupyter extension in VS Code, which defaults to the dark matplotlib style. Maybe that has something to do with it? An example is below, using seaborn 0.12.1 ``` import seaborn as sns tips = sns.load_dataset('tips') sns.stripplot( data=tips, x='day', y='total_bill', edgecolor='black' ) ``` ![example](https://user-images.githubusercontent.com/18745581/197267866-1d413be6-3ef8-484d-80a8-25b5f723c1cd.png)
closed
2022-10-21T18:48:45Z
2022-10-25T02:36:13Z
https://github.com/mwaskom/seaborn/issues/3104
[ "mod:distributions", "ux" ]
markmbaum
4
deepinsight/insightface
pytorch
2,656
insightface.app的get函数返回空列表
open
2024-09-29T10:31:46Z
2024-10-25T14:27:52Z
https://github.com/deepinsight/insightface/issues/2656
[]
Qiuhao-Wu
3
OpenInterpreter/open-interpreter
python
813
命令interpreter运行GTP4无法运行
### Describe the bug C:\Users\16121\Desktop\open-interpreter-env_vars>interpreter ▌ Model set to gpt-4-1106-preview Open Interpreter will require approval before running code. Use interpreter -y to bypass this. Press CTRL-C to exit. > interpreter -y Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new LiteLLM.Info: If you need to debug this error, use `litellm.set_verbose=True'. You do not have access to gpt-4-1106-preview. Would you like to try gpt-3.5-turbo instead? (y/n) ### Reproduce ![Uploading QQ截图20231204152243.png…]() ### Expected behavior ![QQ截图20231204152243](https://github.com/KillianLucas/open-interpreter/assets/151594014/765a6e8a-a8d3-4ec1-8447-07d39578b529) ### Screenshots > > ![QQ截图20231204152243](https://github.com/KillianLucas/open-interpreter/assets/151594014/08f26198-0527-4aa1-96fb-a3f6676380d3) ### Open Interpreter version 0.1.1 ### Python version 3.11 ### Operating System name and version win11 ### Additional context ![QQ截图20231204152243](https://github.com/KillianLucas/open-interpreter/assets/151594014/49ef1dbd-d4c1-49b4-9ff9-8390c9417534)
closed
2023-12-04T07:24:38Z
2023-12-04T15:30:49Z
https://github.com/OpenInterpreter/open-interpreter/issues/813
[ "Bug" ]
af6464
0
fastapi-users/fastapi-users
fastapi
1,229
Missing type information when dependencies are specified using AsyncIterator
## Describe the bug The type information for classes such as AuthenticationBackend causes mypy errors when using a dependency whose return type is AsyncIterator rather than AsyncGenerator. ## To Reproduce Consider a dependency: ```python async def get_redis_strategy( config: t.Annotated[Config, config_dependency], redis: t.Annotated[Redis, redis_dependency], # type: ignore[type-arg] ) -> t.AsyncGenerator[RedisStrategy[User, UUID], None]: yield RedisStrategy(redis, lifetime_seconds=config.SESSION_EXPIRY_SECONDS) ``` Which could also be written more succinctly as: ```python async def get_redis_strategy( config: t.Annotated[Config, config_dependency], redis: t.Annotated[Redis, redis_dependency], # type: ignore[type-arg] ) -> t.AsyncIterator[RedisStrategy[User, UUID]]: yield RedisStrategy(redis, lifetime_seconds=config.SESSION_EXPIRY_SECONDS) ``` In the latter case mypy reports a type issue, as DependencyCallable defined in fastapi_users/types.py does not allow for AsyncIterator (even though it's semantically the same in this case). ## Expected behavior Using a dependency that returns an AsyncIterator as the `get_strategy` argument to `AuthenticationBackend` should not cause a mypy error. ## Configuration - Python version : 3.11.3 - FastAPI version : 0.95.2 - FastAPI Users version : 11.0.0
closed
2023-06-16T17:54:09Z
2023-06-23T08:38:31Z
https://github.com/fastapi-users/fastapi-users/issues/1229
[ "enhancement", "good first issue" ]
jameswilliams1
2
marimo-team/marimo
data-visualization
3,302
Autocomplete does not find suggestions in polars namespaces
### Describe the bug When using polars the marimo autocomplete is unable to find any suggestions for names that are within namespaces. For example, here is what the autocomplete shows for the `dt` namespace. ![Image](https://github.com/user-attachments/assets/71169f44-6c31-4da1-8109-b2feb3c8d41a) Here is what VS Code shows: ![Image](https://github.com/user-attachments/assets/df01004c-1bb8-495d-94a0-131240dfad10) ### Environment { "marimo": "0.10.7", "OS": "Darwin", "OS Version": "24.2.0", "Processor": "arm", "Python Version": "3.13.1", "Binaries": { "Browser": "--", "Node": "v23.5.0" }, "Dependencies": { "click": "8.1.3", "docutils": "0.21.2", "itsdangerous": "2.2.0", "jedi": "0.19.2", "markdown": "3.7", "narwhals": "1.19.1", "packaging": "24.2", "psutil": "6.1.1", "pygments": "2.18.0", "pymdown-extensions": "10.13", "pyyaml": "6.0.2", "ruff": "0.6.9", "starlette": "0.42.0", "tomlkit": "0.13.2", "typing-extensions": "4.12.2", "uvicorn": "0.34.0", "websockets": "14.1" }, "Optional Dependencies": { "altair": "5.5.0", "duckdb": "1.1.3", "pandas": "2.2.3", "polars": "1.17.1", "pyarrow": "18.1.0" } } ### Code to reproduce ```python import polars as pl pl.col("col_name").dt ```
open
2024-12-27T21:03:15Z
2025-01-03T10:17:17Z
https://github.com/marimo-team/marimo/issues/3302
[ "bug", "upstream" ]
kjgoodrick
3
databricks/spark-sklearn
scikit-learn
59
AttributeError: 'function' object has no attribute '_input_kwargs'
I am using python 2.7, spark-2.2.0 with hadoop2.7 and sklearn 0.19. I get the following error: Traceback (most recent call last): File "test.py", line 19, in <module> km = KeyedEstimator(sklearnEstimator=LinearRegression(), yCol="y").fit(df) File "C:\spark-2.2.0-bin-hadoop2.7\python\pyspark\\_\_init\_\_.py", line 104, in wrapper return func(self, **kwargs) File "C:\Python27\lib\site-packages\spark_sklearn\keyed_models.py", line 323, in \_\_init\_\_ kwargs = KeyedEstimator._inferredParams(sklearnEstimator, self.\_\_init\_\_._input_kwargs) AttributeError: 'function' object has no attribute '_input_kwargs' when I try to run the code: km = KeyedEstimator(sklearnEstimator=LinearRegression(), yCol="y").fit(df) From the origian example code available in the welcome page. I also tried the Kmeans clustering, but it caused the same error. I downloaded the source code and checked line 323 in keyed_models.py, which was: kwargs = KeyedEstimator._inferredParams(sklearnEstimator, self._input_kwargs) Please correct me if I'm wrong but the two, do not seem to match
closed
2017-08-22T14:26:54Z
2018-12-09T21:56:41Z
https://github.com/databricks/spark-sklearn/issues/59
[]
sounakban
1
ScottfreeLLC/AlphaPy
pandas
28
AttributeError: type object 'DataFrame' has no attribute 'from_items'
Winbash python3.7 running mflow: Traceback (most recent call last): File "/home/freefall/.local/bin/mflow", line 8, in <module> sys.exit(main()) File "/home/d/.local/lib/python3.7/site-packages/alphapy/market_flow.py", line 430, in main model = market_pipeline(model, market_specs) File "/home/d/.local/lib/python3.7/site-packages/alphapy/market_flow.py", line 318, in market_pipeline tfs = run_system(model, system, group, intraday) File "/home/fredefall/.local/lib/python3.7/site-packages/alphapy/system.py", line 370, in run_system tf = DataFrame.from_items(gtlist, orient='index', columns=Trade.states) AttributeError: type object 'DataFrame' has no attribute 'from_items'
closed
2020-03-01T02:59:07Z
2020-03-03T01:57:19Z
https://github.com/ScottfreeLLC/AlphaPy/issues/28
[]
thegamecat
2
awesto/django-shop
django
697
Ordering is ignored for product.images.all()
Expected behavior: `product.images.all()` should be ordered based on the `order` column defined in the m2m-through model. Actual behavior: No ordering is applied by default. `ProductImage` model which is a m2m-through model between `Product` and `Image` has `order` column defined, but that column is not used when making ORM queries. ```python >>> from myshop.all_models import * >>> product1 = Product.objects.all()[0] >>> product1.images.all().ordered False >>> str(product1.images.all().query) 'SELECT "filer_file"."id", "filer_file"."polymorphic_ctype_id", "filer_file"."folder_id", "filer_file"."file", "filer_file"."_file_size", "filer_file"."sha1", "filer_file"."has_all_mandatory_data", "filer_file"."original_filename", "filer_file"."name", "filer_file"."description", "filer_file"."owner_id", "filer_file"."uploaded_at", "filer_file"."modified_at", "filer_file"."is_public", "filer_image"."_height", "filer_image"."_width", "filer_image"."default_alt_text", "filer_image"."default_caption", "filer_image"."subject_location", "filer_image"."file_ptr_id", "filer_image"."date_taken", "filer_image"."author", "filer_image"."must_always_publish_author_credit", "filer_image"."must_always_publish_copyright" FROM "filer_image" INNER JOIN "myshop_productimage" ON ("filer_image"."file_ptr_id" = "myshop_productimage"."image_id") INNER JOIN "filer_file" ON ("filer_image"."file_ptr_id" = "filer_file"."id") WHERE "myshop_productimage"."product_id" = 1' ``` `product1.images.all().ordered` returns `False`. It's not clear to me why the ordering is not applied. It is specified in the `Meta` of the `ProductImage` model.
open
2017-12-28T11:45:01Z
2021-01-04T13:52:43Z
https://github.com/awesto/django-shop/issues/697
[ "bug" ]
sniku
2
InstaPy/InstaPy
automation
5,885
not sure whats going on
## Expected Behavior thought the code would execute ## Current Behavior Traceback (most recent call last): File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\quickstart.py", line 13, in <module> session = InstaPy(username='Dakota_cardone', password='dakota') File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\instapy\instapy.py", line 322, in __init__ self.browser, err_msg = set_selenium_local_session( File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\instapy\browser.py", line 132, in set_selenium_local_session browser = webdriver.Firefox( File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\selenium\webdriver\firefox\webdriver.py", line 170, in __init__ RemoteWebDriver.__init__( File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 157, in __init__ self.start_session(capabilities, browser_profile) File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 252, in start_session response = self.execute(Command.NEW_SESSION, parameters) File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 321, in execute self.error_handler.check_response(response) File "C:\Users\Dakot\AppData\Local\Programs\Python\Python39\Lib\site-packages\selenium\webdriver\remote\errorhandler.py", line 242, in check_response raise exception_class(message, screen, stacktrace) selenium.common.exceptions.SessionNotCreatedException: Message: Expected browser binary location, but unable to find binary in default location, no 'moz:firefoxOptions.binary' capability provided, and no binary flag set on the command line ## Possible Solution (optional) ## InstaPy configuration
closed
2020-11-12T03:22:27Z
2020-11-12T04:30:13Z
https://github.com/InstaPy/InstaPy/issues/5885
[]
VoidNebula13
0
waditu/tushare
pandas
906
停复牌数据不准确
002506.SZ在20140429到20150812期间都是停牌的,停复牌数据没有体现出来 000950.SZ在20170426到20180828期间都是停牌的,停复牌数据没有体现出来 目前获取到的停复牌信息数据(suspend接口)有点奇怪,有些有resume_date,有些没有,让人觉得比较困惑,能否调整一下,有两种方案:1、所有停复牌区间都记suspend_date和resume_date,没有resume_date说明依旧处于停牌中;2、针对每个停牌日,都记录一条数据,不记录resume_date。十分感谢。
closed
2019-01-24T08:18:21Z
2019-01-24T14:48:29Z
https://github.com/waditu/tushare/issues/906
[]
wangyaochong
1
lexiforest/curl_cffi
web-scraping
181
Support `base_url` when initializing sessions
**Describe alternatives you've considered**] https://www.python-httpx.org/advanced/#other-client-only-configuration-options
closed
2023-12-18T05:18:26Z
2024-03-27T16:09:53Z
https://github.com/lexiforest/curl_cffi/issues/181
[ "enhancement", "good first issue" ]
T-256
2
tinyfish-io/agentql
web-scraping
108
Recommended deployment pipeline for self-hosting
At the moment it appears the REST API / cloud service is limited to scraping single pages (doesn't handle pagination, cookies etc.). Given this, what is the team's recommended deployment pipeline for self-hosting? FastAPI + AgentQL + Playwright Stealth seems like a good combination. Are there any recommended proxy services in addition? Super impressed by the Python SDK so far - just looking to get this into production as soon as.
open
2024-12-20T17:30:29Z
2025-01-27T09:51:34Z
https://github.com/tinyfish-io/agentql/issues/108
[]
Ches-ctrl
1
neuml/txtai
nlp
750
Translation: AttributeError: 'ModelInfo' object has no attribute 'modelId'
Translation stopped working for me and I'm not sure what changed. ```python from txtai.pipeline import Translation # Create and run pipeline translate = Translation() translate("This is a test translation into Spanish", "es") ``` ``` --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) Cell In[17], line 4 1 from txtai.pipeline import Translation 3 # Create and run pipeline ----> 4 translate = Translation() 5 translate("This is a test translation into Spanish", "es") File ~/Development/notebooks/.venv/lib/python3.10/site-packages/txtai/pipeline/text/translation.py:54, in Translation.__init__(self, path, quantize, gpu, batch, langdetect, findmodels) 52 # Language models 53 self.models = {} ---> 54 self.ids = self.modelids() File ~/Development/notebooks/.venv/lib/python3.10/site-packages/txtai/pipeline/text/translation.py:115, in Translation.modelids(self) 107 def modelids(self): 108 """ 109 Runs a query to get a list of available language models from the Hugging Face API. 110 111 Returns: 112 list of source-target language model ids 113 """ --> 115 ids = [x.modelId for x in HfApi().list_models(author="Helsinki-NLP")] if self.findmodels else [] 116 return set(ids) File ~/Development/notebooks/.venv/lib/python3.10/site-packages/txtai/pipeline/text/translation.py:115, in <listcomp>(.0) 107 def modelids(self): 108 """ 109 Runs a query to get a list of available language models from the Hugging Face API. 110 111 Returns: 112 list of source-target language model ids 113 """ --> 115 ids = [x.modelId for x in HfApi().list_models(author="Helsinki-NLP")] if self.findmodels else [] 116 return set(ids) AttributeError: 'ModelInfo' object has no attribute 'modelId' ``` ``` > pip freeze | grep -e huggingface -e txtai huggingface-hub==0.24.0 txtai==7.3.0 ``` It looks like the last version of huggingface-hub that used `modelId` was [v0.18 ](https://huggingface.co/docs/huggingface_hub/v0.18.0.rc0/en/package_reference/hf_api#huggingface_hub.hf_api.ModelInfo) It isn't available in [v0.19](https://huggingface.co/docs/huggingface_hub/v0.19.3/en/package_reference/hf_api#huggingface_hub.hf_api.ModelInfo) or [later](https://huggingface.co/docs/huggingface_hub/v0.24.0/en/package_reference/hf_api#huggingface_hub.hf_api.ModelInfo)
closed
2024-07-22T17:03:29Z
2024-07-23T08:59:12Z
https://github.com/neuml/txtai/issues/750
[ "bug" ]
adieyal
5
piskvorky/gensim
machine-learning
2,735
Word2vec: loss tally maxes at 134217728.0 due to float32 limited-precision
<!-- **IMPORTANT**: - Use the [Gensim mailing list](https://groups.google.com/forum/#!forum/gensim) to ask general or usage questions. Github issues are only for bug reports. - Check [Recipes&FAQ](https://github.com/RaRe-Technologies/gensim/wiki/Recipes-&-FAQ) first for common answers. Github bug reports that do not include relevant information and context will be closed without an answer. Thanks! --> #### Cumulative loss of word2vec maxes out at 134217728.0 I'm training a word2vec model with 2,793,404 sentences / 33,499,912 words, vocabulary size 162,253 (words with at least 5 occurrences). Expected behaviour: with `compute_loss=True`, gensim's word2vec should compute the loss in the expected way. Actual behaviour: the cumulative loss seems to be maxing out at `134217728.0`: Building vocab... Vocab done. Training model for 120 epochs, with 16 workers... Loss after epoch 1: 16162246.0 / cumulative loss: 16162246.0 Loss after epoch 2: 11594642.0 / cumulative loss: 27756888.0 [ - snip - ] Loss after epoch 110: 570688.0 / cumulative loss: 133002056.0 Loss after epoch 111: 564448.0 / cumulative loss: 133566504.0 Loss after epoch 112: 557848.0 / cumulative loss: 134124352.0 Loss after epoch 113: 93376.0 / cumulative loss: 134217728.0 Loss after epoch 114: 0.0 / cumulative loss: 134217728.0 Loss after epoch 115: 0.0 / cumulative loss: 134217728.0 And it stays at `134217728.0` thereafter. The value `134217728.0` is of course exactly `128*1024*1024`, which does not seem like a coincidence. #### Steps to reproduce My code is as follows: class MyLossCalculator(CallbackAny2Vec): def __init__(self): self.epoch = 1 self.losses = [] self.cumu_losses = [] def on_epoch_end(self, model): cumu_loss = model.get_latest_training_loss() loss = cumu_loss if self.epoch <= 1 else cumu_loss - self.cumu_losses[-1] print(f"Loss after epoch {self.epoch}: {loss} / cumulative loss: {cumu_loss}") self.epoch += 1 self.losses.append(loss) self.cumu_losses.append(cumu_loss) def train_and_check(my_sentences, my_epochs, my_workers=8): print(f"Building vocab...") my_model: Word2Vec = Word2Vec(sg=1, compute_loss=True, workers=my_workers) my_model.build_vocab(my_sentences) print(f"Vocab done. Training model for {my_epochs} epochs, with {my_workers} workers...") loss_calc = MyLossCalculator() trained_word_count, raw_word_count = my_model.train(my_sentences, total_examples=my_model.corpus_count, compute_loss=True, epochs=my_epochs, callbacks=[loss_calc]) loss = loss_calc.losses[-1] print(trained_word_count, raw_word_count, loss) loss_df = pd.DataFrame({"training loss": loss_calc.losses}) loss_df.plot(color="blue") # print(f"Calculating accuracy...") # acc, details = my_model.wv.evaluate_word_analogies(questions_file, case_insensitive=True) # print(acc) return loss_calc, my_model The data is a news article corpus in Finnish; I'm not at liberty to share all of it (and anyway it's a bit big), but it looks like one would expect: [7]: df.head(2) [7]: [Row(file_and_id='data_in_json/2018/04/0001.json.gz%%3-10169118', index_in_file='853', headline='Parainen pyristelee pois lastensuojelun kriisistä: irtisanoutuneiden tilalle houkutellaan uusia sosiaalityöntekijöitä paremmilla työeduilla', publication_date='2018-04-20 11:59:35+03:00', publication_year='2018', publication_month='04', sentence='hän tiesi minkälaiseen tilanteeseen tulee', lemmatised_sentence='hän tietää minkälainen tilanne tulla', source='yle', rnd=8.436637410902392e-08), Row(file_and_id='data_in_xml/arkistosiirto2018.zip%%arkistosiirto2018/102054668.xml', index_in_file=None, headline='*** Tiedote/SDP: Medialle tiedoksi: SDP:n puheenjohtaja Antti Rinteen puhe puoluevaltuuston kokouksessa ***', publication_date='2018-04-21T12:51:44', publication_year='2018', publication_month='04', sentence='me haluamme jättää hallitukselle välikysymyksen siitä miksi nuorten ihmisten tulevaisuuden uskoa halutaan horjuttaa miksi epävarmuutta ja näköalattomuutta sekä pelkoa tulevaisuuden suhteen halutaan lisätä', lemmatised_sentence='me haluta jättää hallitus välikysymys se miksi nuori ihminen tulevaisuus usko haluta horjuttaa miksi epävarmuus ja näköalattomuus sekä pelko tulevaisuus suhteen haluta lisätä', source='stt', rnd=8.547760445010155e-07)] sentences = list(map(lambda r: r["lemmatised_sentence"].split(" "), df.select("lemmatised_sentence").collect())) [18]: sentences[0] [18]: ['hän', 'tietää', 'minkälainen', 'tilanne', 'tulla'] #### Versions The output of: ```python import platform; print(platform.platform()) import sys; print("Python", sys.version) import numpy; print("NumPy", numpy.__version__) import scipy; print("SciPy", scipy.__version__) import gensim; print("gensim", gensim.__version__) from gensim.models import word2vec;print("FAST_VERSION", word2vec.FAST_VERSION) ``` is: Windows-10-10.0.18362-SP0 Python 3.7.3 | packaged by conda-forge | (default, Jul 1 2019, 22:01:29) [MSC v.1900 64 bit (AMD64)] NumPy 1.17.3 SciPy 1.3.1 gensim 3.8.1 FAST_VERSION 1 Finally, I'm not the only one who has encountered this issue. I found the following related links: https://groups.google.com/forum/#!topic/gensim/IH5-nWoR_ZI https://stackoverflow.com/questions/59823688/gensim-word2vec-model-loss-becomes-0-after-few-epochs I'm not sure if this is only a display issue and the training continues normally even after the cumulative loss reaches its "maximum", or if the training in fact stops at that point. The trained word vectors seem reasonably ok, judging by `my_model.wv.evaluate_word_analogies()`, though they do need more training than this.
open
2020-01-26T13:28:16Z
2023-09-28T19:12:03Z
https://github.com/piskvorky/gensim/issues/2735
[ "bug", "difficulty medium" ]
tsaastam
24
albumentations-team/albumentations
machine-learning
2,104
[Add transform] Add RandomPlasmaContrast
https://kornia.readthedocs.io/en/latest/augmentation.module.html#kornia.augmentation.RandomPlasmaContrast
closed
2024-11-08T15:56:44Z
2024-11-18T03:57:20Z
https://github.com/albumentations-team/albumentations/issues/2104
[ "enhancement" ]
ternaus
1
babysor/MockingBird
pytorch
837
新版本模型训练问题与解决方法
新版本主目录里没有synthesizer_train.py,查找了一下发现...\control\cli里面有synthesizer_train.py文件,但是运行称缺一个models的依赖,pip安装models提示需要base依赖,遂安装base,提示gbk编码器错误,按照网上教程为base的setup.py第21行指定编码器为utf-8依旧报错,遂放弃使用命令行进行训练,改用web中自带的模型训练功能进行训练,正常,然后关闭,再启动命令行训练,即可正常使用。
open
2023-02-24T03:36:31Z
2023-05-16T03:08:38Z
https://github.com/babysor/MockingBird/issues/837
[]
YIZXIY
1
PrefectHQ/prefect
automation
17,127
Improve UX for flow with parameterised schedules
### Describe the current behavior Prefect 3.2 added the option to specify parameters in the schedule but it's hard to see in the UI which schedule a flow run belongs to. See: ![Image](https://github.com/user-attachments/assets/005d4767-54c9-4856-8e68-7585e2846f40) Currently, the only way to see on the flows page what schedule triggered the flow run is to click on the parameters. ### Describe the proposed behavior As a first improvement, I would propose adding the schedule slug with the flow run. This way the user can directly identify which schedule triggered a certain flow run. Further improvements could be to make it possible to filter based on the schedule slug to easily find flow runs with specific parameters. ### Example Use _No response_ ### Additional context The reason for asking for this feature is that these parameterized schedules can be used for a ML pipeline with different parameters/experiments. Being able to immediately spot which experiment we are looking at will speed up the user experience. Before parameterized schedules existed, I used a single flow function that had many deployments depending on the parameters. While this is impractical because it resulted in many deployments, at least Prefect makes it easy to search based on the deployment name (and tags). Ideally, schedules would be upgraded to a similar status with search functionality and filtering in the flow run page.
open
2025-02-13T09:10:38Z
2025-02-19T17:18:10Z
https://github.com/PrefectHQ/prefect/issues/17127
[ "enhancement" ]
mvdb-enspi
1
lepture/authlib
flask
176
Decoding a JWS token with ES256 doesn't work
**Describe the bug** Hi! I'm trying to implement an OAuth2 server with authlib and I found a bug in the library with ES256. I was trying to decode a JWS token signed with ES256 but the library has raised an exception. The generation of a JWT with ES256 works, but not the decoding. Here it's the small piece of code that raises an exception : ``` def gen_refresh_token(self, client, grant_type, user, scope): jws = JsonWebSignature(algorithms=JWS_ALGORITHMS) header = {'alg': 'ES256'} date = datetime.utcnow() payload = { 'client_id': client.get_client_id(), 'iat': int(date.timestamp()), 'user_id': user["id"], 'scope': scope, 'exp': 604800 } try: key = open("my_ec_key.pem", 'r').read() s = jws.serialize_compact(header, json.dumps(payload), key) except Exception as e: logger.exception('JWS exception', e) return s.decode("utf-8") ... class RefreshTokenGrant(grants.RefreshTokenGrant): INCLUDE_NEW_REFRESH_TOKEN = True def authenticate_refresh_token(self, refresh_token): jws = JsonWebSignature(algorithms=JWS_ALGORITHMS) try: key = open("my_ec_pub.pem", 'r').read() jws_obj = jws.deserialize_compact(refresh_token, key) ... except Exception as e: logger.exception('JWS exception', e) ``` And the traceback : ``` Traceback (most recent call last): File "/mnt/d/documents/exo1/src/exo1/rest/flask/oauth2.py", line 64, in authenticate_refresh_token jws_obj = jws.deserialize_compact(refresh_token, key) File "/home/yohann/.local/lib/python3.6/site-packages/authlib/jose/rfc7515/jws.py", line 115, in deserialize_compact self._algorithms, jws_header, payload, key) File "/home/yohann/.local/lib/python3.6/site-packages/authlib/jose/util.py", line 14, in prepare_algorithm_key key = algorithm.prepare_public_key(key) File "/home/yohann/.local/lib/python3.6/site-packages/authlib/jose/rfc7518/_backends/_key_cryptography.py", line 42, in prepare_public_key if key.startswith(b'ecdsa-sha2-'): TypeError: startswith first arg must be str or a tuple of str, not bytes ``` **To Reproduce** I put a small example of the code above. **Expected behavior** I should get the content of the token. **Environment:** - OS: Windows Subsystem for Linux - Python Version: 3.6.8 - Authlib Version: 0.13.0 **Additional context** I found a patch for my use case. You have to edit the file in : authlib/jose/rfc7518/_backends/_key_cryptography.py In the class ECKey, the method prepare_public_key doesn't convert the key to bytes. I did the following patch : ``` def prepare_public_key(self, key): if isinstance(key, EllipticCurvePublicKey): return key key = to_bytes(key) if key.startswith(b'ecdsa-sha2-'): return load_ssh_public_key(key, backend=default_backend()) else: return load_pem_public_key(key, backend=default_backend()) ``` Add any other context about the problem here.
closed
2019-12-30T09:39:09Z
2020-02-11T11:00:10Z
https://github.com/lepture/authlib/issues/176
[ "bug" ]
fenix01
5
mitmproxy/pdoc
api
296
Error when parsing reStructuredText when its located in between sections
#### Problem Description Error when parsing reStructuredText when its located in between sections as it detects it as if it is part of the section #### Code: ```py def testing(): """This is a test docstring. Parameters ---------- testing_param: str Some parameter. .. warning:: Warning for this parameter. .. note:: This is really important, but it will be rendered incorrectly. """ ``` #### Render: *Generated with `pdoc <file> -d numpy`* ![image](https://user-images.githubusercontent.com/29100934/130358641-ea0b96e1-65fa-48da-b284-b2cfbead2a97.png) #### System Information ``` pdoc: 7.4.0 Python: 3.8.10 Platform: Linux-5.13.1-051301-generic-x86_64-with-glibc2.29 ```
closed
2021-08-21T13:26:52Z
2021-08-22T18:06:35Z
https://github.com/mitmproxy/pdoc/issues/296
[ "wontfix" ]
davfsa
3
Josh-XT/AGiXT
automation
1,380
Add Commands to Convert Markdown to docx, xlsx, and pdf
We need to add commands to AGiXT that allow users to convert markdown files into docx, xlsx, and pdf formats. This feature will be useful for generating reports, documentation, and other formatted outputs directly from markdown content. The commands should handle the conversion process efficiently and support customization options for formatting and styling.
closed
2025-01-20T18:05:10Z
2025-03-16T21:43:55Z
https://github.com/Josh-XT/AGiXT/issues/1380
[ "20K Bounty" ]
Josh-XT
1
sinaptik-ai/pandas-ai
data-science
1,360
Unable to analyze the DataFrame when it contains data in list format.
### System Info OS version: MacOS Sonoma Python version: 3.12.5 The current version of `pandasai` being used: 2.2.14 ### 🐛 Describe the bug I tried using pandasai `Agent` to analyze my data in DataFrame format, but I found that if the DataFrame contains data in list format, the analysis fails, and there are no error logs in `pandasai.log`. The following is a simple example code: ```python import pandas as pd data = { 'Employee_ID': [101, 102, 103, 104], 'Employee_Name': ['Alice', 'Bob', 'Charlie', 'Diana'], 'Projects': [['Project A', 'Project B'], ['Project C'], ['Project D', 'Project E', 'Project F'], ['Project G']], 'Salary': [70000, 80000, 75000, 90000] } df = pd.DataFrame(data) agent = Agent( dfs=df, config=Config(llm=OpenAI(api_token=os.getenv("OAI_API_KEY"), model="gpt-4o")) ) print(agent.chat('Tell me the average salary of the employees')) ``` Here is the output: ```python "Unfortunately, I was not able to get your answers, because of the following error:\n\nunhashable type: 'list'\n" ```
closed
2024-09-08T06:30:09Z
2024-12-15T16:08:08Z
https://github.com/sinaptik-ai/pandas-ai/issues/1360
[ "bug" ]
ReeveWu
1
postmanlabs/httpbin
api
334
Random sub-domains no longer open the website
Up until yesterday, any random sub-domain would open the website, i.e. https://dsaf.httpbin.org/ would open the same page as https://httpbin.org/ (not redirected), just like https://eu.httpbin.org/ still does now. Since yesterday, it gives a "This site can’t be reached" error (DNS_PROBE_FINISHED_NXDOMAIN) for a random sub-domain.
closed
2017-03-17T09:03:39Z
2018-04-26T17:51:12Z
https://github.com/postmanlabs/httpbin/issues/334
[]
macroshadow
2
rougier/numpy-100
numpy
163
An alternative solution for Q.82
> 82. Compute a matrix rank (★★★) > hint: np.linalg.svd > > \# Author: Stefan van der Walt > > Z = np.random.uniform(0,1,(10,10)) > U, S, V = np.linalg.svd(Z) # Singular Value Decomposition > rank = np.sum(S > 1e-10) > print(rank) > `numpy.linalg.matrix_rank` [Doc](https://numpy.org/doc/stable/reference/generated/numpy.linalg.matrix_rank.html?highlight=rank#numpy.linalg.matrix_rank) provides an alternative way to compute matrix rank. The alternative solution will be: ```python3 from numpy.linalg import matrix_rank Z = np.random.uniform(0,1,(10,10)) print(matrix_rank(Z)) ```
open
2021-12-11T13:50:23Z
2021-12-15T18:31:13Z
https://github.com/rougier/numpy-100/issues/163
[]
iamyifan
3
pallets-eco/flask-sqlalchemy
flask
739
create_engine() missing 1 required positional argument: 'engine_opts
I'm trying to use `db.create_engine` to connect to a second database in a view. My app was working yesterday, but after I tried it on a new machine with a fresh virtualenv, it's no longer working. I think this is due to the changes in #684. ```python x = 'postgres://*****' engine = db.create_engine(x) ``` ```pytb create_engine() missing 1 required positional argument: 'engine_opts' ```
closed
2019-05-19T19:26:53Z
2020-12-05T20:21:51Z
https://github.com/pallets-eco/flask-sqlalchemy/issues/739
[]
jjRick
5
healthchecks/healthchecks
django
310
Bug: Deprecated function call breaking `manage.py collectstatic --noinput`
It seems like django deprecated `admin_static` which is causing the following exception when running `manage.py collectstatic --noinput` command. Is there a way to resolve this exception?? [Some posts suggest](https://stackoverflow.com/questions/59148185/django-error-cannot-import-name-removedindjango30warning) uninstalling and reinstalling django which I tried with no luck. ``` Traceback (most recent call last): File "/usr/lib/python3.7/site-packages/django/template/utils.py", line 66, in __getitem__ return self._engines[alias] KeyError: 'django' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.7/site-packages/django/template/backends/django.py", line 121, in get_package_libraries module = import_module(entry[1]) File "/usr/lib/python3.7/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1006, in _gcd_import File "<frozen importlib._bootstrap>", line 983, in _find_and_load File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 677, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 728, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/usr/lib/python3.7/site-packages/django/contrib/admin/templatetags/admin_static.py", line 5, in <module> from django.utils.deprecation import RemovedInDjango30Warning ImportError: cannot import name 'RemovedInDjango30Warning' from 'django.utils.deprecation' (/usr/lib/python3.7/site-packages/django/utils/deprecation.py) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "manage.py", line 10, in <module> execute_from_command_line(sys.argv) File "/usr/lib/python3.7/site-packages/django/core/management/__init__.py", line 401, in execute_from_command_line utility.execute() File "/usr/lib/python3.7/site-packages/django/core/management/__init__.py", line 395, in execute self.fetch_command(subcommand).run_from_argv(self.argv) File "/usr/lib/python3.7/site-packages/django/core/management/base.py", line 328, in run_from_argv self.execute(*args, **cmd_options) File "/usr/lib/python3.7/site-packages/django/core/management/base.py", line 369, in execute output = self.handle(*args, **options) File "/usr/lib/python3.7/site-packages/compressor/management/commands/compress.py", line 277, in handle self.handle_inner(**options) File "/usr/lib/python3.7/site-packages/compressor/management/commands/compress.py", line 300, in handle_inner offline_manifest, block_count, results = self.compress(engine, extensions, verbosity, follow_links, log) File "/usr/lib/python3.7/site-packages/compressor/management/commands/compress.py", line 100, in compress if not self.get_loaders(): File "/usr/lib/python3.7/site-packages/compressor/management/commands/compress.py", line 50, in get_loaders for e in engines.all(): File "/usr/lib/python3.7/site-packages/django/template/utils.py", line 90, in all return [self[alias] for alias in self] File "/usr/lib/python3.7/site-packages/django/template/utils.py", line 90, in <listcomp> return [self[alias] for alias in self] File "/usr/lib/python3.7/site-packages/django/template/utils.py", line 81, in __getitem__ engine = engine_cls(params) File "/usr/lib/python3.7/site-packages/django/template/backends/django.py", line 25, in __init__ options['libraries'] = self.get_templatetag_libraries(libraries) File "/usr/lib/python3.7/site-packages/django/template/backends/django.py", line 43, in get_templatetag_libraries libraries = get_installed_libraries() File "/usr/lib/python3.7/site-packages/django/template/backends/django.py", line 108, in get_installed_libraries for name in get_package_libraries(pkg): File "/usr/lib/python3.7/site-packages/django/template/backends/django.py", line 125, in get_package_libraries "trying to load '%s': %s" % (entry[1], e) django.template.library.InvalidTemplateLibrary: Invalid template library specified. ImportError raised when trying to load 'django.contrib.admin.templatetags.admin_static': cannot import name 'RemovedInDjango30Warning' from 'django.utils.deprecation' (/usr/lib/python3.7/site-packages/django/utils/deprecation.py) ``` Reference: https://django.readthedocs.io/en/2.2.x/releases/2.1.html#id2
closed
2019-12-03T16:41:29Z
2019-12-05T15:55:29Z
https://github.com/healthchecks/healthchecks/issues/310
[]
gganeshan
13
deezer/spleeter
tensorflow
564
How get output as bytes or assign to variable from Python API
<!-- Please respect the title [Discussion] tag. --> How to get the output as bytes or assign to a variable from Python API
open
2021-01-22T08:35:56Z
2022-02-15T04:28:31Z
https://github.com/deezer/spleeter/issues/564
[ "question" ]
pranavansp
1
seleniumbase/SeleniumBase
pytest
3,381
uc_gui_click_captcha fails to click cloudflare checkbox
Hello, I encountered the same problem as a guy in this issue https://github.com/seleniumbase/SeleniumBase/issues/3345 , I tried to update to the latest version, unfortunately it did not help me, my project is quite large, unfortunately, in what format would it be more convenient for you if I provided it to you? I also attach a video with the problem, in some areas of the site, if it was used with. sb.cdp.minimize() sb.cdp.medimize() click occurs at the bottom edge of the checkbox is literally pixel by pixel, suitable for clicking, but not in all cases.I have attached a case in which the click does not occur to this message. I apologize for any grammatical errors in this message, English is not my native language, and also thank you in advance for your answer ![2024-12-30 03-34-57](https://github.com/user-attachments/assets/d1abc74b-9c78-4bfb-af73-04dec1344056)
closed
2025-01-01T17:01:38Z
2025-01-01T18:35:17Z
https://github.com/seleniumbase/SeleniumBase/issues/3381
[ "invalid usage", "can't reproduce", "UC Mode / CDP Mode" ]
mkuchuman
6
eriklindernoren/ML-From-Scratch
deep-learning
111
No module named 'mlfromscratch.utils.loss_functions'
Traceback (most recent call last): File "C:\G\ML-From-Scratch\mlfromscratch\examples\gradient_boosting_regressor.py", line 9, in <module> from mlfromscratch.utils.loss_functions import SquareLoss ModuleNotFoundError: No module named 'mlfromscratch.utils.loss_functions'
open
2024-11-20T08:51:03Z
2024-11-20T08:51:03Z
https://github.com/eriklindernoren/ML-From-Scratch/issues/111
[]
LeiYangGH
0
PokeAPI/pokeapi
graphql
693
Honey tree encounters
In Generation IV, diamond pearl and platinum, you can encounter certain Pokemon only through honey trees: https://bulbapedia.bulbagarden.net/wiki/Honey_Tree The API currently does not include these encounters, so burmy and wormadam have no encounters in the game they were introduced in. Munchlax is also listed as impossible to encounter. My current project uses this API to calculate which games are required for which Pokemon. If this feature were not added, then I would have to manually override the API's calls for these Pokemon.
open
2022-02-19T19:37:14Z
2022-02-19T19:37:14Z
https://github.com/PokeAPI/pokeapi/issues/693
[]
Hyreon
0
3b1b/manim
python
1,683
DEBUG The error could be: `Undefined control sequence.` tex_file_writing.py:94
### Describe the error <!-- A clear and concise description of what you want to make. --> Unable to render `LaTex` file. And prior to this issue, the problem was [this](https://github.com/3b1b/manim/issues/1661), but I manged to installed the `texlive-full` on my ubuntu 20.04. and still unable to render tex ### Code and Error **Code**: [My code was copied from here](https://docs.manim.org.cn/getting_started/example_scenes.html#updatersexample) **Error**: ``` INFO If you want to create a local configuration file, you can config.py:232 create a file named `custom_config.yml`, or run `manimgl --config` Writing "\lim_{n \to \infty} \left\lfloor \sqrt{\frac{1}{n !} \mathrm{e}^{n} a_{n} + b_{n}^{p}} \otimes \sqrt[n]{\sum_{m = 0}^{n^{2}} \tilde{c}_{m \cdot n}^{b_{n}^{p} \cos \left( \the ERROR LaTeX Error! Not a worry, it happens to the best tex_file_writing.py:90 of us. DEBUG The error could be: `Undefined control sequence.` tex_file_writing.py:94 ``` The .tex file was generated by manimlib, and here is the log ``` This is TeX, Version 3.14159265 (TeX Live 2019/Debian) (preloaded format=tex) (./Tex/e4d7da810f12ab27.tex ! Undefined control sequence. l.1 \documentclass [preview]{standalone} ? ! Undefined control sequence. l.3 \usepackage [english]{babel} ? ! Undefined control sequence. l.4 \usepackage [utf8]{inputenc} ? ! Undefined control sequence. l.5 \usepackage [T1]{fontenc} ``` ### Environment **OS System**: Ubuntu 20.04 **manim version**: master <!-- make sure you are using the latest version of master branch --> **python version**: ``` pdfTeX 3.14159265-2.6-1.40.20 (TeX Live 2019/Debian) kpathsea version 6.3.1 Copyright 2019 Han The Thanh (pdfTeX) et al. There is NO warranty. Redistribution of this software is covered by the terms of both the pdfTeX copyright and the Lesser GNU General Public License. For more information about these matters, see the file named COPYING and the pdfTeX source. Primary author of pdfTeX: Han The Thanh (pdfTeX) et al. Compiled with libpng 1.6.37; using libpng 1.6.37 Compiled with zlib 1.2.11; using zlib 1.2.11 Compiled with xpdf version 4.01 ```
closed
2021-11-17T16:05:43Z
2021-11-18T17:50:00Z
https://github.com/3b1b/manim/issues/1683
[]
gxshao
2
Skyvern-AI/skyvern
automation
1,753
Can I change AI Mode to others,such deepseek R1
Find only can configure open ai api key, could we change to other LLM model, such as Deepseek R1?
closed
2025-02-10T06:31:10Z
2025-02-10T12:17:11Z
https://github.com/Skyvern-AI/skyvern/issues/1753
[]
zhenzhen-wang
1
NVIDIA/pix2pixHD
computer-vision
95
Strange patterns in the output image
In some of my output images, there are strange patterns occurred like the pics below: ![image](https://user-images.githubusercontent.com/7723243/51155055-2c1eba00-18b1-11e9-95ac-033966b6b37f.png) ![image](https://user-images.githubusercontent.com/7723243/51155091-4fe20000-18b1-11e9-814d-6e2df82cc7c1.png) ![image](https://user-images.githubusercontent.com/7723243/51155129-6b4d0b00-18b1-11e9-91b1-1f204e277de8.png) I can't figure out why those things came out upon my output images. Do anybody know the reason? In addition, I think my loss curves are also strange: ![image](https://user-images.githubusercontent.com/7723243/51156277-c254df00-18b5-11e9-8667-732b947969de.png) ![image](https://user-images.githubusercontent.com/7723243/51156290-ca148380-18b5-11e9-9bbc-f109c3849bb0.png) ![image](https://user-images.githubusercontent.com/7723243/51156308-d567af00-18b5-11e9-9344-b7150b7b9157.png)
open
2019-01-15T03:08:16Z
2021-01-07T08:36:05Z
https://github.com/NVIDIA/pix2pixHD/issues/95
[]
cfanyyx
6
1313e/CMasher
matplotlib
26
Instructions for R/IDL/Matlab/DS9/etc. users
Would be good to have instructions on how `R`/`IDL`/`Matlab`/etc users might be able to use the `cmasher` colourmaps. Similarly, for certain application, this might be a good idea as well. Perhaps in the README or in the online documentation. - [X] R (3ce67f9) - [ ] IDL - [ ] MATLAB - [ ] DS9
open
2020-10-20T04:45:33Z
2023-04-12T23:36:30Z
https://github.com/1313e/CMasher/issues/26
[ "documentation", "help wanted" ]
manodeep
6
flasgger/flasgger
rest-api
257
Unable to load inline defintions.
`assets with pagination. --- tags: - asset parameters: - name: authorization in: header description: Please add the authentication token to here. required: true type: string - name: search in: query description: This value will be searched in the product number, product description, and barcode fields. type: string - name: barcode in: query description: Assets that have this barcode value will be returned. type: string - name: external_resource_url in: query description: Assets that have this external resource url will be returned. type: string - name: product_group in: query description: Assets that are included this product group will be returned. type: string - name: division in: query description: Assets that are located this division will be returned. type: string - name: status_code in: query description: Assets that are in this status code will be returned. type: string - name: page in: query description: The page number that you want to recieve. type: integer - name: per_page in: query description: The resource count that you want to see per page. type: integer responses: 200: description: If the request is valid, the response will have a status of '200' and a body that includes the list of assets. schema: type: object properties: assets: type: array items: $ref: '#/definitions/Asset' pages: type: object $ref: '#/definitions/Page' examples: responses: { "assets": [ { "_link": { "fileAttachments": "http://localhost:8000/assets/MR/file_attachments/", "groupRates": "http://localhost:8000/assets/MR/group_rates/", "inspectionForms": "http://localhost:8000/assets/MR/inspection_form/", "locations": "http://localhost:8000/assets/MR/locations/", "meterHistories": "http://localhost:8000/assets/MR/meter_histories/", "productRates": "http://localhost:8000/assets/MR/product_rates/", "tags": "http://localhost:8000/assets/MR/tags/" }, "averageCostEach": 0, "barcodes": [ "BWC90001" ], "bulkItem": "Y", "cycleBill": "Y", "dateOfManufacture": "2017-07-07", "description": "MISC RENTAL PRODUCT", "fuelTypes": [], "inventory": "Y", "listPrice": 0, "make": "", "markupPercentage": 0, "currentMeter": 0, "meterType": "N", "modelNumber": "", "multipleTags": "Y", "notes": "", "productClass": "32", "productGroup": "3201", "productNumber": "MR", "quantityOnHand": 9999, "ratebook": "Y", "replacementCost": 0, "safetyNotes": [], "showOnWebsite": "", "sslNumber": "", "statusCode": "", "vendorDescription": "MISC RENTAL", "vendorNumber": "", "vendorProductNumber": "MR" }, { "_link": { "fileAttachments": "http://localhost:8000/assets/T4/file_attachments/", "groupRates": "http://localhost:8000/assets/T4/group_rates/", "inspectionForms": "http://localhost:8000/assets/T4/inspection_form/", "locations": "http://localhost:8000/assets/T4/locations/", "meterHistories": "http://localhost:8000/assets/T4/meter_histories/", "productRates": "http://localhost:8000/assets/T4/product_rates/", "tags": "http://localhost:8000/assets/T4/tags/" }, "averageCostEach": 50, "barcodes": [], "bulkItem": "Y", "cycleBill": "Y", "dateOfManufacture": "2018-07-07", "description": "OXYGEN TANK", "fuelTypes": [], "inventory": "Y", "listPrice": 61.5, "make": "", "markupPercentage": 23, "currentMeter": 0, "meterType": "N", "modelNumber": "", "multipleTags": "Y", "notes": "", "productClass": "60", "productGroup": "6001", "productNumber": "T4", "quantityOnHand": 11999995, "ratebook": "Y", "replacementCost": 0, "safetyNotes": [], "showOnWebsite": "", "sslNumber": "", "statusCode": "", "vendorDescription": "OXYGEN TANK", "vendorNumber": "8", "vendorProductNumber": "T4" } ], "pages": { "firstUrl": "http://127.0.0.1:8000/assets/?page=1&per_page=2", "lastUrl": "http://127.0.0.1:8000/assets/?page=391&per_page=2", "nextUrl": "http://127.0.0.1:8000/assets/?page=2&per_page=2", "page": 1, "pages": 391, "perPage": 2, "prevUrl": null, "total": 782 } } definitions: Asset: type: object required: - description - productNumber - vendorNumber - productGroup - productClass - statusCode - meterType properties: averageCostEach: type: number readonly: true barcodes: type: array items: type: string bulkItem: type: string enum: ["Y", "N"] maxLength: 1 cycleBill: type: string enum: ["Y", "N"] maxLength: 1 dateOfManufacture: type: string format: date description: type: string maxLength: 15 fuelTypes: type: array items: $ref: '#/definitions/AssetFuelType' inventory: type: string enum: ["Y", "N"] maxLength: 1 listPrice: type: number format: float make: type: string maxLength: 20 markupPercentage: type: number meterType: type: string enum: ["M", "B", "N"] maxLength: 1 currentMeter: type: number description: If the meter type is 'M' or 'B', this value should be prompted. initialMeter: type: number description: If the meter type is 'M' or 'B', this value should be prompted. currentMeterDate: type: string format: date description: If the meter type is 'M' or 'B', this value should be prompted. initialMeterDate: type: string format: date description: If the meter type is 'M' or 'B', this value should be prompted. meterDigits: type: number description: If the meter type is 'M', this value should be prompted. modelNumber: type: string maxLength: 15 multipleTags: type: string enum: ["Y", "N"] maxLength: 1 notes: type: string maxLength: 50 productClass: type: string maxLength: 5 productGroup: type: string maxLength: 5 productNumber: type: string maxLength: 12 quantityOnHand: type: number readonly: true ratebook: type: string enum: ["Y", "N"] maxLength: 1 replacementCost: type: number safetyNotes: type: array items: type: string showOnWebsite: type: string enum: ["Y", "N"] maxLength: 1 sslNumber: type: string maxLength: 9 statusCode: type: string maxLength: 2 vendorDescription: type: string maxLength: 30 vendorNumber: type: string maxLength: 6 vendorProductNumber: type: string maxLength: 20 externalResourceUrl: type: string maxLength: 500 Page: type: object properties: firstUrl: type: string lastUrl: type: string nextUrl: type: string page: type: number pages: type: number perPage: type: number prevUrl: type: string total: type: number AssetFuelType: type: object required: - fuelType - capacity properties: fuelType: type: string maxLength: 3 capacity: type: number I am using a customize validation function and in Flasgger validate function. `validate( request_data, 'Asset', 'swagger/assets_get.yml', validation_error_handler=validation_error_handler, validation_function=validate_response_handler, )` `def validate_response_handler(data, schema): try: jsonschema.validate(data, schema, format_checker=jsonschema.FormatChecker()) except Exception as e: raise e` as I debugged and checked schema's definitions property is blank and it has properties of Asset definition and it is now throwing an exception of RefResolution error for AssetFuelType. and how I changed the code `def validate_response_handler(data, schema): try: schema['definitions'] = { "AssetFuelType": { "type": "object", "required": [ "fuelType", "capacity" ], "properties": { "fuelType": { "type": "string", "maxLength": 3 }, "capacity": { "type": "number" } } } } jsonschema.validate(data, schema, format_checker=jsonschema.FormatChecker()) except Exception as e: raise e` and it works, so Flassger validate has any problem to parse the Yaml file or I am doing something wrong.
open
2018-11-07T06:40:18Z
2018-11-07T06:44:43Z
https://github.com/flasgger/flasgger/issues/257
[]
ghost
0
erdewit/ib_insync
asyncio
665
cancelMktDepth causes errors
I'm getting errors calling cancelMktDepth. Looks like it clears domBids and domAsks right away [1], but IB server keeps sending the level2 updates for a while, and when the decoder decodes them it tries to access domBids and domAsks, which have been cleared, causing an error. Below is a small example demonstrating the issue. It looks like there is a second issue as well with "can't find the subscribed market depth..", BTW my version.py shows __version_info__ = (0, 9, 86), but for some reason the line number in wrapper.py in the error message I get is line 921, but in the source here it looks like line 975 (?). [1] https://github.com/erdewit/ib_insync/blob/d31241f2fcb16f5a61dc075d6f458721cb95eebd/ib_insync/ib.py#L1352 [2] https://github.com/erdewit/ib_insync/blob/d31241f2fcb16f5a61dc075d6f458721cb95eebd/ib_insync/wrapper.py#L975 -Neal --- code demonstrating the issue: ``` import ib_insync as ibs ib = ibs.IB() ib.connect() ib.sleep(1) contract = ibs.Stock("SPY", "SMART", "USD") ib.qualifyContracts(contract) ib.sleep(1) ticker = ib.reqMktDepth(contract, numRows=50, isSmartDepth=True) ib.sleep(2) ib.cancelMktDepth(contract) ib.sleep(2) # now we get the following errors: # # Error 310, reqId 15049: Can't find the subscribed market depth with tickerId:15049 # Error for updateMktDepthL2: # Traceback (most recent call last): # File "/Users/neal/Desktop/IB/curses_app/ib_insync/decoder.py", line 187, in handler # method(*args) # File "/Users/neal/Desktop/IB/curses_app/ib_insync/wrapper.py", line 921, in updateMktDepthL2 # dom[position] = DOMLevel(price, size, marketMaker) # ~~~^^^^^^^^^^ # IndexError: list assignment index out of range ib.disconnect() ```
closed
2023-11-29T22:20:36Z
2023-12-01T14:23:07Z
https://github.com/erdewit/ib_insync/issues/665
[]
nealeyoung
4
jupyterhub/repo2docker
jupyter
971
Detect default branch instead of assuming master (at least for GitHub repos)
<!-- Thank you for contributing. These HTML commments will not render in the issue, but you can delete them once you've read them if you prefer! --> ### Proposed change <!-- Use this section to describe the feature you'd like to be added. --> GitHub (at least) now supports setting your default branch to something other than `master`, the recommended option being `main`. https://www.zdnet.com/article/github-to-replace-master-with-main-starting-next-month/ This can be confusing for new users of mybinder.org, e.g. https://discourse.jupyter.org/t/cant-solve-could-not-resolve-ref-for-gh-fredericfoulonlycee-math-spe-python-master/6301 We should change repo2docker's behaviour to search for/fetch the default branch name of the git repository, rather than assuming/hard code it to being `master`. I don't know if r2d assumes everything is a git repository or if this is something we'll specifically have to build into the GitHub repo handling section. ### Alternative options <!-- Use this section to describe alternative options and why you've decided on the proposed feature above. --> We continue to wait until this is more common before carrying out this work. Perhaps adding an information box to the front page of mybinder.org so folks are aware. ### Who would use this feature? <!-- Describe the audience for this feature. This information will affect who chooses to work on the feature with you. --> Everybody who has created a new GitHub repository since 1st October 2020 or have manually renamed their default branches. ### How much effort will adding it take? <!-- Try to estimate how much work adding this feature will require. This information will affect who chooses to work on the feature with you. --> I imagine there's a git command we can run to fetch the default branch name and then parse that into our standard branch handling code for non-`master` branches. ### Who can do this work? <!-- What skills are needed? Who can be recruited to add this feature? This information will affect who chooses to work on the feature with you. --> Useful skills: - Knowledge of the r2d codebase: where things are being parsed to, etc - Some git fu to detect the default branch name
closed
2020-10-16T10:43:14Z
2020-10-22T08:44:31Z
https://github.com/jupyterhub/repo2docker/issues/971
[ "needs: discussion" ]
sgibson91
1
iterative/dvc
data-science
9,899
exp show: unexpected error - 'checkpoint_tip' when using -A option
# Bug Report <!-- ## Issue name Issue names must follow the pattern `command: description` where the command is the dvc command that you are trying to run. The description should describe the consequence of the bug. Example: `repro: doesn't detect input changes` --> ## Description <!-- A clear and concise description of what the bug is. --> Command `dvc exp show -A` raises the following error without any output: ``` ERROR: unexpected error - 'checkpoint_tip' ``` ### Reproduce <!-- Step list of how to reproduce the bug --> <!-- Example: 1. dvc init 2. Copy dataset.zip to the directory 3. dvc add dataset.zip 4. dvc run -d dataset.zip -o model ./train.sh 5. modify dataset.zip 6. dvc repro --> Not sure what is causing this issue. Could it be related to breaking changes moving from DVC 2 to 3? ### Expected <!-- A clear and concise description of what you expect to happen. --> The command should show me an overview of all experiments I ran in the DVC project folder of the Git repository. ### Environment information <!-- This is required to ensure that we can reproduce the bug. --> **Output of `dvc doctor`:** ```console $ dvc doctor DVC version: 3.17.0 (conda) --------------------------- Platform: Python 3.10.6 on Linux-3.10.0-1127.8.2.el7.x86_64-x86_64-with-glibc2.17 Subprojects: dvc_data = 2.15.4 dvc_objects = 1.0.1 dvc_render = 0.5.3 dvc_task = 0.3.0 scmrepo = 1.3.1 Supports: http (aiohttp = 3.8.5, aiohttp-retry = 2.8.3), https (aiohttp = 3.8.5, aiohttp-retry = 2.8.3), s3 (s3fs = 2023.6.0, boto3 = 1.26.76) Config: Global: /home/aschuh/.config/dvc System: /etc/xdg/dvc Cache types: hardlink, symlink Cache directory: xfs on /dev/sda1 Caches: local Remotes: s3, s3 Workspace directory: xfs on /dev/sda1 Repo: dvc (subdir), git Repo.site_cache_dir: /var/tmp/dvc/repo/8d2f5d68bb223da9776a9d6301681efd ``` **Additional Information (if any):** <!-- Please check https://github.com/iterative/dvc/wiki/Debugging-DVC on ways to gather more information regarding the issue. If applicable, please also provide a `--verbose` output of the command, eg: `dvc add --verbose`. If the issue is regarding the performance, please attach the profiling information and the benchmark comparisons. --> `dvc exp list -A` works and does not cause this issue.
closed
2023-08-31T22:45:42Z
2023-09-06T05:58:54Z
https://github.com/iterative/dvc/issues/9899
[ "bug", "A: experiments" ]
aschuh-hf
6
krish-adi/barfi
streamlit
32
I want to improve barfi but i need help
Hi krish ! I'm huge fan of barfi and i want to build a projet with it but for this i need new functionnalities that i want to add but i don't really know javascript i'm profficient in python and i'm already building a script to create easily schemas with code. I need help on how to open and modify the graphic part to change the buttons and make them "streatlit activable" and also a way to group different nodes. Can you help me if you have a minute ? Thank you man ! I really love your work !
closed
2024-05-13T19:32:38Z
2025-01-16T16:08:27Z
https://github.com/krish-adi/barfi/issues/32
[ "help wanted" ]
Silectio
4
SYSTRAN/faster-whisper
deep-learning
206
requests.exceptions.ConnectionError
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/guillaumekln/faster-whisper-large-v2/revision/main (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f0e915670a0>: Failed to establish a new connection: [Errno 101] Network is unreachable'))
closed
2023-05-07T13:33:27Z
2023-05-08T07:54:16Z
https://github.com/SYSTRAN/faster-whisper/issues/206
[]
wwfcnu
4
vimalloc/flask-jwt-extended
flask
505
Direct call to decorator (jwt_required)
jwt_required(func(arg)) Previously, in version 3.x, it was called and used as above. A TypeError occurred in version 4.4.4. -> Traceback (most recent call last): File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/flask/app.py", line 2525, in wsgi_app response = self.full_dispatch_request() File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/flask/app.py", line 1823, in full_dispatch_request return self.finalize_request(rv) File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/flask/app.py", line 1842, in finalize_request response = self.make_response(rv) File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/flask/app.py", line 2162, in make_response raise TypeError( File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/flask/app.py", line 2158, in make_response rv = self.response_class.force_type( File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/werkzeug/wrappers/response.py", line 268, in force_type response = Response(*run_wsgi_app(response, environ)) File "/Users/user/PycharmProjects/flaskTestProject/venv/lib/python3.8/site-packages/werkzeug/test.py", line 1242, in run_wsgi_app app_rv = app(environ, start_response) TypeError: wrapper() takes 1 positional argument but 2 were given How should I call it in version 4.4.4? https://github.com/vimalloc/flask-jwt-extended/blob/88a628ec88eb3c0e766300613a1367ac3eb3f34f/flask_jwt_extended/view_decorators.py#L112-L158
closed
2022-12-21T02:30:47Z
2022-12-23T01:59:57Z
https://github.com/vimalloc/flask-jwt-extended/issues/505
[]
0coolcard0
2
remsky/Kokoro-FastAPI
fastapi
38
Bake Models into Docker images
Baked models in to improve stability, deployment
closed
2025-01-13T06:21:44Z
2025-01-13T06:21:45Z
https://github.com/remsky/Kokoro-FastAPI/issues/38
[]
remsky
0
rthalley/dnspython
asyncio
446
crash in dnssec with gmp-6.2.0 during initialization
This one liner: - - - import dns.rdtypes.ANY.DNSKEY - - - gives: ``` (dskm_p37) [root@hermes /usr/local/src]# python test.py Fatal Python error: Illegal instruction Current thread 0x0000000800a24000 (most recent call first): File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/Crypto/Math/_IntegerGMP.py", line 163 in __init__ File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/Crypto/PublicKey/ECC.py", line 123 in init_p256 File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/Crypto/PublicKey/ECC.py", line 138 in <module> File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap_external>", line 728 in exec_module File "<frozen importlib._bootstrap>", line 677 in _load_unlocked File "<frozen importlib._bootstrap>", line 967 in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 983 in _find_and_load File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/Crypto/Signature/DSS.py", line 42 in <module> File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap_external>", line 728 in exec_module File "<frozen importlib._bootstrap>", line 677 in _load_unlocked File "<frozen importlib._bootstrap>", line 967 in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 983 in _find_and_load File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap>", line 1035 in _handle_fromlist File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/dns/dnssec.py", line 484 in <module> File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap_external>", line 728 in exec_module File "<frozen importlib._bootstrap>", line 677 in _load_unlocked File "<frozen importlib._bootstrap>", line 967 in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 983 in _find_and_load File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/dns/rdtypes/dnskeybase.py", line 22 in <module> File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap_external>", line 728 in exec_module File "<frozen importlib._bootstrap>", line 677 in _load_unlocked File "<frozen importlib._bootstrap>", line 967 in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 983 in _find_and_load File "/usr/local/py_venv/dskm_p37/lib/python3.7/site-packages/dns/rdtypes/ANY/DNSKEY.py", line 18 in <module> File "<frozen importlib._bootstrap>", line 219 in _call_with_frames_removed File "<frozen importlib._bootstrap_external>", line 728 in exec_module File "<frozen importlib._bootstrap>", line 677 in _load_unlocked File "<frozen importlib._bootstrap>", line 967 in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 983 in _find_and_load File "test.py", line 1 in <module> Illegal instruction (core dumped) ``` This does not happen with this one liner: ``` import dns.rdtypes.ANY ``` or with gmp-6.1.2_1 Environment: ``` FreeBSD 12.1-RELEASE-p3 GENERIC amd64 OpenSSL 1.1.1d-freebsd python37-3.7.6 virtualenv-16.7.5 Tool for creating isolated Python environments virtualenv-clone-0.4.0_1 Python virtualenv cloning script virtualenvwrapper-4.8.4 Enhancements to virtualenv Package Version Location ------------ ------- ------------------- dnspython 1.16.0 DSKM 0.9.1 /usr/local/src/DSKM ecdsa 0.15 pip 20.0.2 pycryptodome 3.9.7 script 1.7.2 setuptools 46.0.0 six 1.14.0 wheel 0.34.2 ```
closed
2020-04-03T14:13:25Z
2020-05-04T12:11:13Z
https://github.com/rthalley/dnspython/issues/446
[]
mc3
4
psf/requests
python
6,051
HTTP/3 / QUIC?
Hi, is it planned to wrap around QUIC in the future?
closed
2022-01-28T01:17:34Z
2022-04-28T05:17:28Z
https://github.com/psf/requests/issues/6051
[]
ksaadDE
1
onnx/onnx
deep-learning
6,365
codeformatter / linter for yaml files?
# Ask a Question ### Question Do we have a codeformatter / linter for yaml files?
open
2024-09-14T16:20:23Z
2024-09-16T16:29:41Z
https://github.com/onnx/onnx/issues/6365
[ "question" ]
andife
4
tqdm/tqdm
jupyter
968
"Notebook validation failed" for "tqdm.notebook.tqdm" output
When using `from tqdm.notebook import tqdm` in jupyter, the jupyter server complains `Notebook validation failed` on output cells containing a progressbar. * tqdm.__version__: 4.45.0 * sys.__version__ : 3.7.6 (default, Jan 28 2020, 14:26:04) [GCC 5.4.0 20160609] * sys.platform: linux (particularly Ubuntu 16.04) * jupyter.__version__: 1.0.0 The error appears after running, and trying to save the notebook: from tqdm.notebook import tqdm for _ in tqdm(range(10)): pass The error disappears after clearing the output of the cell, and saving again. The error is a pop-up from Jupyter after saving the notebook: **Notebook validation failed** The save operation succeeded, but the notebook does not appear to be valid. The validation error was: Notebook validation failed: {'version_major': 2, 'version_minor': 0, 'model_id': '0531a121a60a46b3a9815504e28d415c'} is not valid under any of the given schemas: { "version_major": 2, "version_minor": 0, "model_id": "0531a121a60a46b3a9815504e28d415c" } This is the json-content of the cell: { "cell_type": "code", "execution_count": 14, "metadata": {}, "outputs": [ { "data": { "application/vnd.jupyter.widget-view+json": { "model_id": "51bd2bf126484d088ff2c7547803ad6b", "version_major": 2, "version_minor": 0 }, "text/plain": [ "HBox(children=(FloatProgress(value=0.0, max=10.0), HTML(value='')))" ] }, "metadata": {}, "output_type": "display_data" }, { "name": "stdout", "output_type": "stream", "text": [ "\n" ] } ], "source": [ "from tqdm.notebook import tqdm\n", "\n", "for _ in tqdm(range(10)):\n", " pass" ] }, PS: thanks for this wonderful library.
open
2020-05-12T11:36:15Z
2020-05-12T12:50:23Z
https://github.com/tqdm/tqdm/issues/968
[ "invalid ⛔", "need-feedback 📢", "p2-bug-warning ⚠", "submodule-notebook 📓" ]
prhbrt
2
ultralytics/yolov5
deep-learning
13,098
low precision
### Search before asking - [X] I have searched the YOLOv5 [issues](https://github.com/ultralytics/yolov5/issues) and [discussions](https://github.com/ultralytics/yolov5/discussions) and found no similar questions. ### Question i am using yolov5 on mydataset which is panoramic dental xrays. the goal is to detect the infected tooth and produce 3 labels for it: quadrant number, tooth number and the disease. a tooth my have more then 1 disease. so as a work around to achieve multi label object detection i split each label for a single tooth into separate line which looks like this: 000013 0.6431219110378913 0.45246010638297873 0.0671334431630972 0.2293882978723404 009 0.6431219110378913 0.45246010638297873 0.0671334431630972 0.2293882978723404 5 0.6431219110378913 0.45246010638297873 0.0671334431630972 0.2293882978723404 those are the quadrant number, tooth number and the disease for 1 tooth. i have a total of 704 images for training and validation. 70% - 30% all image sizes are not the same size and they are large. the highest precision i achieved when training yolov5m was 0.6. how can i achieve a 0.95 precision ? and what could be the issues that caused such low precision and how can i see the accuracy ? I aslo tried yolov5l and there was no difference. I trained for 300 epochs ### Additional _No response_
closed
2024-06-17T19:29:02Z
2024-10-20T19:48:04Z
https://github.com/ultralytics/yolov5/issues/13098
[ "question" ]
mamdouhhz
1
statsmodels/statsmodels
data-science
8,982
Allow `sm.OLS` to work with `Float64` values
When working with `Float64` values form `pandas`, `sm.OLS` raises an error `ValueError: Pandas data cast to numpy dtype of object. Check input data with np.asarray(data).` Example ``` impor pandas as pd import statsmodels.api as sm x = pd.Series([1,2,3,4]).astype('Float64') y = pd.Series([1,2,3,4]).astype('Float64') sm.OLS(y,x).fit() ``` It is easy to convert the `Float64` into a `float64` value, but it would be nice if `sm.OLS` can handle that.
open
2023-08-22T15:15:32Z
2023-08-22T15:37:26Z
https://github.com/statsmodels/statsmodels/issues/8982
[]
marcdelabarrera
1
Ehco1996/django-sspanel
django
548
给clash pro添加tun模式配置(主要针对Windows clash用户)
## Feature Request **Is your feature request related to a problem? Please describe:** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> NO **Describe the feature you'd like:** <!-- A clear and concise description of what you want to happen. --> 给clash pro添加tun模式配置(主要针对Windows clash用户) **Describe alternatives you've considered:** <!-- A clear and concise description of any alternative solutions or features you've considered. --> ``` yaml dns: enable: true enhanced-mode: redir-host listen: 0.0.0.0:53 ....... tun: enable: true stack: gvisor # 使用 system 需要 Clash Premium 2021.05.08 及更高版本 dns-hijack: - 198.18.0.2:53 macOS-auto-route: true macOS-auto-detect-interface: true # 自动检测出口网卡 ``` 同时Windows用户还需要根据[文档](https://docs.cfw.lbyczf.com/contents/tun.html)下载一个[Wintun](https://www.wintun.net/),然后把`Wintun`放到`Home Directory`目录下,安装`Service Mode`开启 **PR** 稍后补充PR,在本地做些测试,目前已配通,~~暂不知添加`tun`配置后是否影响macOS的`ClashX`~~,已检测不影响
closed
2021-07-01T07:38:08Z
2021-07-02T01:57:02Z
https://github.com/Ehco1996/django-sspanel/issues/548
[]
Gkirito
3