repo
stringlengths
5
51
instance_id
stringlengths
11
56
base_commit
stringlengths
40
40
fixed_commit
stringclasses
20 values
patch
stringlengths
400
56.6k
test_patch
stringlengths
0
895k
problem_statement
stringlengths
27
55.6k
hints_text
stringlengths
0
72k
created_at
int64
1,447B
1,739B
labels
listlengths
0
7
category
stringclasses
4 values
edit_functions
listlengths
1
10
added_functions
listlengths
0
19
edit_functions_length
int64
1
10
__index_level_0__
int64
1
659
ansys/pymapdl
ansys__pymapdl-3705
7e86b2a110e37383797654ee4e338d53c00df296
null
diff --git a/doc/changelog.d/3705.added.md b/doc/changelog.d/3705.added.md new file mode 100644 index 0000000000..9bc4540620 --- /dev/null +++ b/doc/changelog.d/3705.added.md @@ -0,0 +1,1 @@ +feat: speed up `requires_package` using caching \ No newline at end of file diff --git a/src/ansys/mapdl/core/misc.py b/src/ansy...
diff --git a/tests/common.py b/tests/common.py index 142e0a681a..9a7581f843 100644 --- a/tests/common.py +++ b/tests/common.py @@ -26,6 +26,7 @@ import subprocess import time from typing import Dict, List +from warnings import warn import psutil diff --git a/tests/test_mapdl.py b/tests/test_mapdl.py index 7d902...
`requires_package` is very slow It seems that `requires_package` is very slow: ![image](https://github.com/user-attachments/assets/186c6084-3765-4349-8205-81bdddae9c8e) _Originally posted by @germa89 in https://github.com/ansys/pymapdl/issues/3703#issuecomment-2615615323_
1,737,985,959,000
[ "new feature", "enhancement" ]
Performance Issue
[ "src/ansys/mapdl/core/misc.py:requires_package" ]
[ "src/ansys/mapdl/core/misc.py:is_package_installed_cached" ]
1
138
CrossGL/crosstl
CrossGL__crosstl-229
0ddc76be73115c830de6f7e64c975b78443251db
null
diff --git a/crosstl/translator/lexer.py b/crosstl/translator/lexer.py index e6540499..ac89f534 100644 --- a/crosstl/translator/lexer.py +++ b/crosstl/translator/lexer.py @@ -1,81 +1,82 @@ import re +from collections import OrderedDict -TOKENS = [ - ("COMMENT_SINGLE", r"//.*"), - ("COMMENT_MULTI", r"/\*[\s\S]*...
diff --git a/tests/test_translator/test_lexer.py b/tests/test_translator/test_lexer.py index abff042f..35c778a1 100644 --- a/tests/test_translator/test_lexer.py +++ b/tests/test_translator/test_lexer.py @@ -1,6 +1,6 @@ -from crosstl.translator.lexer import Lexer import pytest from typing import List +from crosstl.tra...
Optimize Lexer Performance with Token Caching and Pattern Combining ## Description The lexer can be optimized to improve performance and reduce memory usage. This task focuses on implementing token caching, combining similar regex patterns, and caching compiled patterns. ### Proposed Changes 1. **Token Cachi...
@Swish78 can you assign me this task I think I can complete this @NripeshN Hii I would like to work on this issue To assign issues to yourself please read this documentation: https://github.com/CrossGL/crosstl/blob/main/CONTRIBUTING.md#-assigning-or-creating-issues @CrossGL-issue-bot assign me @CrossGL-issue-bot ass...
1,735,129,841,000
[]
Performance Issue
[ "crosstl/translator/lexer.py:Lexer.tokenize", "crosstl/translator/lexer.py:Lexer.__init__" ]
[ "crosstl/translator/lexer.py:Lexer._compile_patterns", "crosstl/translator/lexer.py:Lexer._get_cached_token" ]
2
139
kiasambrook/tfl-live-tracker
kiasambrook__tfl-live-tracker-7
b38f2a397075ec42d8f2f14eaffcdc5661a6a20b
null
diff --git a/app/api/tfl_api.py b/app/api/tfl_api.py index 232b626..b688263 100644 --- a/app/api/tfl_api.py +++ b/app/api/tfl_api.py @@ -1,4 +1,5 @@ import requests +import requests_cache class TfLAPI: def __init__(self, api_key): @@ -7,9 +8,16 @@ def __init__(self, api_key): def fetch_all_stop_points(s...
Cache Station List ### Summary Currently the station list is refreshed everytime the program is launched, it would probably be best to cache this to improve performance of app and prevent being rate limited by API. ### Tasks - Cache the list of stations - Data stored should include ID, name, and location
1,734,612,485,000
[]
Performance Issue
[ "app/api/tfl_api.py:TfLAPI.fetch_all_stop_points", "app/ui/main_window.py:MainWindow.build" ]
[]
2
140
ckan/ckan
ckan__ckan-8226
15f32e57e229db6436d10adddf73bcff84bdd453
null
diff --git a/changes/6146.feature b/changes/6146.feature new file mode 100644 index 00000000000..e40b88d474f --- /dev/null +++ b/changes/6146.feature @@ -0,0 +1,1 @@ +Render snippets faster through better use of existing jinja2 tags. Use ``{% snippet 'path/to/snippet.html', arg1=test %}`` instead of ``{{ h.snippet('pat...
snippet performance issue **CKAN version** all **Describe the bug** Heavy use of template `{% snippet %}` slows page rendering times compared to alternatives like macros or `{% include %}` **Steps to reproduce** Compare the rendering time of a page using snippets with one that uses macros or `{% include %}` t...
@smotornyuk mentioned that current jinja2 now has an `{% include … without context %}` that we could have our snippet tag emit instead of calling render and inserting the contents. Should be an easy win.
1,715,738,892,000
[ "Performance", "Backport dev-v2.11" ]
Performance Issue
[ "ckan/config/middleware/flask_app.py:_ungettext_alias", "ckan/config/middleware/flask_app.py:make_flask_stack", "ckan/config/middleware/flask_app.py:helper_functions", "ckan/config/middleware/flask_app.py:c_object", "ckan/lib/helpers.py:snippet", "ckan/lib/jinja_extensions.py:CkanFileSystemLoader.get_sour...
[ "ckan/lib/jinja_extensions.py:SnippetExtension.parse" ]
7
142
fulcrumgenomics/prymer
fulcrumgenomics__prymer-99
f62c4b52b6753ed070cea2269f2fffa36591de69
null
diff --git a/prymer/offtarget/offtarget_detector.py b/prymer/offtarget/offtarget_detector.py index d1c7868..71d3201 100644 --- a/prymer/offtarget/offtarget_detector.py +++ b/prymer/offtarget/offtarget_detector.py @@ -75,6 +75,7 @@ """ # noqa: E501 import itertools +from collections import defaultdict from context...
diff --git a/tests/offtarget/test_offtarget.py b/tests/offtarget/test_offtarget.py index f8f5f8c..a2f3a46 100644 --- a/tests/offtarget/test_offtarget.py +++ b/tests/offtarget/test_offtarget.py @@ -11,6 +11,7 @@ from prymer.offtarget.bwa import BWA_EXECUTABLE_NAME from prymer.offtarget.bwa import BwaHit from prymer.o...
Speed up from O(n^2) iteration over all left/right primer hits in `OffTargetDetector._to_amplicons` `OffTargetDetector._to_amplicons` uses `itertools.product` over the left and right primer hits and evaluates each of them in a loop to identify valid left/right hit combinations for the pair. Suggestion: 1. Get hit...
I think this sounds like a great idea. I would make a couple of suggestions also: 1. You could _probably_ do the splitting into left/right/+/- and by refname all in one pass ... if you created a simple dataclass that held four lists (one for each strand and orientation). You could then have a `dict[refname, new_cl...
1,733,420,754,000
[]
Performance Issue
[ "prymer/offtarget/offtarget_detector.py:OffTargetDetector._build_off_target_result", "prymer/offtarget/offtarget_detector.py:OffTargetDetector._to_amplicons" ]
[]
2
143
nautobot/nautobot
nautobot__nautobot-6364
c86be9c46f5f0ae8f9786fe059175d81c0463e89
null
diff --git a/changes/6297.fixed b/changes/6297.fixed index f889dcbefda..e6e279aba67 100644 --- a/changes/6297.fixed +++ b/changes/6297.fixed @@ -1,2 +1,4 @@ Fixed paginator widget to display the current selected `per_page` value even if it's not one of the `PER_PAGE_DEFAULTS` options. Added pagination of related-obje...
diff --git a/nautobot/extras/tests/test_dynamicgroups.py b/nautobot/extras/tests/test_dynamicgroups.py index ccf7993186b..c2cd6974f77 100644 --- a/nautobot/extras/tests/test_dynamicgroups.py +++ b/nautobot/extras/tests/test_dynamicgroups.py @@ -304,6 +304,10 @@ def test_static_member_operations(self): sg.add_m...
Resources that are "too large" cause the nautobot process to crash ### Environment * Nautobot version (Docker tag too if applicable): 2.3.4 (Docker tag 'latest') * Python version: 3.10.15 * Database platform, version: PostgreSQL 15.1 * Middleware(s): None ### Steps to Reproduce 1. Create 200k devices 2. Create...
Thanks for the report and the investigation. I suspect that in the DynamicGroup case it's the update to the members cache that's causing an out-of-memory condition, specifically `_set_members()`: ```py to_remove = self.members.exclude(pk__in=value.values_list("pk", flat=True)) self._remo...
1,728,660,712,000
[]
Performance Issue
[ "nautobot/core/signals.py:invalidate_max_depth_cache", "nautobot/extras/models/groups.py:DynamicGroup._set_members", "nautobot/extras/models/groups.py:DynamicGroup.add_members", "nautobot/extras/models/groups.py:DynamicGroup._add_members", "nautobot/extras/models/groups.py:DynamicGroup.remove_members", "n...
[ "nautobot/core/models/tree_queries.py:TreeModel.__init_subclass__" ]
8
147
intelowlproject/GreedyBear
intelowlproject__GreedyBear-397
a7912a75e28a43e1d5c10751c111948a96faaf5b
null
diff --git a/api/serializers.py b/api/serializers.py index 937ce0f..4d2d676 100644 --- a/api/serializers.py +++ b/api/serializers.py @@ -47,14 +47,10 @@ def validate(self, data): return data -def feed_type_validation(feed_type): - feed_choices = ["log4j", "cowrie", "all"] - generalHoneypots = General...
diff --git a/tests/test_serializers.py b/tests/test_serializers.py index b8e2dbc..c538a1b 100644 --- a/tests/test_serializers.py +++ b/tests/test_serializers.py @@ -31,13 +31,20 @@ def test_valid_fields(self): for element in valid_data_choices: data_ = {"feed_type": element[0], "attack_type": el...
Feed API slow Querying the API, e.g. `/api/feeds/all/all/recent.json`, takes a while. Requesting the feed on my GreedyBear instancees with `curl -so /dev/null -w '%{time_total}\n' http://HOSTNAME/api/feeds/all/all/recent.json` takes about 6 seconds. This also produces significant CPU load on the server. However the Hon...
1,733,767,958,000
[]
Performance Issue
[ "api/serializers.py:feed_type_validation", "api/serializers.py:FeedsSerializer.validate_feed_type", "api/serializers.py:FeedsResponseSerializer.validate_feed_type", "api/views.py:feeds", "api/views.py:feeds_pagination", "api/views.py:get_queryset", "api/views.py:feeds_response" ]
[]
7
148
UXARRAY/uxarray
UXARRAY__uxarray-1144
81ac5bd50ed2a2faca24008dc261029be30c37c4
null
diff --git a/uxarray/formatting_html.py b/uxarray/formatting_html.py index 8f15978b5..733232f5f 100644 --- a/uxarray/formatting_html.py +++ b/uxarray/formatting_html.py @@ -26,7 +26,7 @@ def _grid_sections(grid, max_items_collapse=15): spherical_coordinates = list( [coord for coord in ugrid.SPHERICAL_COOR...
Slowdown in `__repr__` due to `n_nodes_per_face` construction triggering When printing a `Grid`, there is an unnecessary trigger to the `n_nodes_per_face` variable, which has a noticeable impact on performance for larger grids.
1,738,359,935,000
[]
Performance Issue
[ "uxarray/formatting_html.py:_grid_sections", "uxarray/grid/grid.py:Grid.__repr__", "uxarray/grid/grid.py:Grid.face_node_connectivity" ]
[]
3
151
UXARRAY/uxarray
UXARRAY__uxarray-1151
9926057173e143e8170e3337dd1b7d39a4d1a961
null
diff --git a/uxarray/grid/slice.py b/uxarray/grid/slice.py index 94e8e0eb8..8cce19d15 100644 --- a/uxarray/grid/slice.py +++ b/uxarray/grid/slice.py @@ -111,18 +111,23 @@ def _slice_face_indices( node_indices = np.unique(grid.face_node_connectivity.values[face_indices].ravel()) node_indices = node_indices[nod...
diff --git a/test/test_subset.py b/test/test_subset.py index 252719500..71be6dff5 100644 --- a/test/test_subset.py +++ b/test/test_subset.py @@ -24,14 +24,20 @@ def test_grid_face_isel(): for grid_path in GRID_PATHS: grid = ux.open_grid(grid_path) + grid_contains_edge_node_conn = "edge_node_conne...
`edge_node_connectivity` unnecessarily constructed when slicing When doing `Grid.isel()` with `n_face` or `n_node`, the `edge_node_connectivity` is constructed when the grid does not currently have any edges defined, which is not necessary and impacts the execution time. See also #1138
1,738,693,587,000
[]
Performance Issue
[ "uxarray/grid/slice.py:_slice_face_indices", "uxarray/subset/dataarray_accessor.py:DataArraySubsetAccessor.bounding_circle", "uxarray/subset/dataarray_accessor.py:DataArraySubsetAccessor.nearest_neighbor", "uxarray/subset/grid_accessor.py:GridSubsetAccessor.bounding_circle", "uxarray/subset/grid_accessor.py...
[]
5
152
nautobot/nautobot
nautobot__nautobot-6837
9302a6072ce3c5e97ca9a673fb7b96fc7eaccb29
null
diff --git a/changes/6767.added b/changes/6767.added new file mode 100644 index 0000000000..f9de871e1a --- /dev/null +++ b/changes/6767.added @@ -0,0 +1,1 @@ +Added cacheable `CustomField.choices` property for retrieving the list of permissible values for a select/multiselect Custom Field. diff --git a/changes/6767.fix...
diff --git a/nautobot/extras/tests/test_customfields.py b/nautobot/extras/tests/test_customfields.py index aaa3dcc4ed..ef2d6a1a69 100644 --- a/nautobot/extras/tests/test_customfields.py +++ b/nautobot/extras/tests/test_customfields.py @@ -53,12 +53,12 @@ def test_immutable_fields(self): instance.refresh_from...
Hundreds of SQL queries made when fetching device details page with ~ 20 custom fields <!-- NOTE: IF YOUR ISSUE DOES NOT FOLLOW THIS TEMPLATE, IT WILL BE CLOSED. This form is only for reporting reproducible bugs. If you need assistance with Nautobot installation, or if you have a general question, plea...
Hi @progala, thank you for submitting the issue. Do you have a high number of device components attached to the device as well (ConsoleServerPort, ConsolePort, Interface, VRFs, etc.) I want to make sure whether querying those device components is causing the slowdown instead of the custom fields. It would be really hel...
1,738,259,134,000
[]
Performance Issue
[ "nautobot/core/views/utils.py:get_csv_form_fields_from_serializer_class", "nautobot/extras/models/customfields.py:CustomFieldManager.get_for_model", "nautobot/extras/models/customfields.py:CustomField.to_form_field", "nautobot/extras/models/customfields.py:CustomField.validate" ]
[ "nautobot/extras/models/customfields.py:CustomField.choices_cache_key", "nautobot/extras/models/customfields.py:CustomField.choices", "nautobot/extras/signals.py:invalidate_choices_cache" ]
4
153
spacetelescope/stcal
spacetelescope__stcal-337
9bd365de4afd442f016103e3db3b8944d22cf769
null
diff --git a/changes/337.general.rst b/changes/337.general.rst new file mode 100644 index 00000000..f73f7e3a --- /dev/null +++ b/changes/337.general.rst @@ -0,0 +1,1 @@ +Performance improvements for jump step targeting both runtime and memory consumption. Results are mostly identical, but there are some differences in ...
diff --git a/tests/test_jump.py b/tests/test_jump.py index c620b374..734b4075 100644 --- a/tests/test_jump.py +++ b/tests/test_jump.py @@ -10,8 +10,7 @@ flag_large_events, point_inside_ellipse, find_first_good_group, - detect_jumps_data, - find_last_grp + detect_jumps_data ) DQFLAGS = { @@ -...
Jump performance Resolves [JP-3697](https://jira.stsci.edu/browse/JP-3697) Closes #337 A number of changes to `twopoint_difference.py` and `jump.py` reduce memory usage by a factor of about 2 and improve runtime by a factor of anywhere from 3 to 20. Tested output files (noted below) are unchanged, except for MIR...
1,738,693,487,000
[ "testing", "jump" ]
Performance Issue
[ "src/stcal/jump/jump.py:extend_saturation", "src/stcal/jump/jump.py:extend_ellipses", "src/stcal/jump/jump.py:find_last_grp", "src/stcal/jump/jump.py:find_faint_extended", "src/stcal/jump/jump.py:get_bigcontours", "src/stcal/jump/jump.py:diff_meddiff_int", "src/stcal/jump/jump.py:diff_meddiff_grp", "s...
[ "src/stcal/jump/jump.py:ellipse_subim", "src/stcal/jump/jump.py:convolve_fast", "src/stcal/jump/twopoint_difference.py:propagate_flags" ]
8
154
CWorthy-ocean/roms-tools
CWorthy-ocean__roms-tools-227
67df5fc400b1daeb903ba9b2efe11899d8b55dfa
null
diff --git a/docs/releases.md b/docs/releases.md index 587315aa..1d5430af 100644 --- a/docs/releases.md +++ b/docs/releases.md @@ -10,10 +10,15 @@ ### Internal Changes +* Parallelize computation of radiation correction, leading to a hugely improved memory footprint for surface forcing generation ([#227](https://gi...
diff --git a/roms_tools/tests/test_setup/test_utils.py b/roms_tools/tests/test_setup/test_utils.py index de4faad6..1012b839 100644 --- a/roms_tools/tests/test_setup/test_utils.py +++ b/roms_tools/tests/test_setup/test_utils.py @@ -15,6 +15,7 @@ def test_interpolate_from_climatology(use_dask): climatology = ERA5C...
Radiation correction is not done in parallel (even when using Dask) ## Issue If the user chooses to apply a radiation correction for the surface forcing while using Dask via ``` SurfaceForcing( grid=grid, ..., correct_radiation=True, use_dask=True ) ``` this radiation correction is not done in parallel...
1,738,006,985,000
[]
Performance Issue
[ "roms_tools/setup/datasets.py:_select_relevant_times", "roms_tools/setup/surface_forcing.py:SurfaceForcing._set_variable_info", "roms_tools/setup/surface_forcing.py:SurfaceForcing._apply_correction", "roms_tools/setup/surface_forcing.py:SurfaceForcing._validate", "roms_tools/setup/utils.py:interpolate_from_...
[]
6
155
twisted/klein
twisted__klein-773
04db6d26fd6f8dc749868268518957c8238c09d6
null
diff --git a/src/klein/_resource.py b/src/klein/_resource.py index b12f711c..4107357b 100644 --- a/src/klein/_resource.py +++ b/src/klein/_resource.py @@ -91,12 +91,11 @@ def extractURLparts(request: IRequest) -> Tuple[str, str, int, str, str]: server_port = request.getHost().port else: server_po...
Klein adds significant performance overhead over a twisted.web server A minimal hello world benchmark, just routing "/" and returning a string, is half the speed of the equivalent minimal twisted.web server. I will start investigating where the performance overhead is, and hopefully find some places to optimize.
1,726,598,482,000
[]
Performance Issue
[ "src/klein/_resource.py:extractURLparts", "src/klein/_resource.py:KleinResource.render" ]
[]
2
156
traceloop/openllmetry
traceloop__openllmetry-2577
847be92eb31f6815848e70a1d2cafb77959fe21e
null
diff --git a/packages/opentelemetry-instrumentation-groq/opentelemetry/instrumentation/groq/utils.py b/packages/opentelemetry-instrumentation-groq/opentelemetry/instrumentation/groq/utils.py index e283307b9..f8d750d4c 100644 --- a/packages/opentelemetry-instrumentation-groq/opentelemetry/instrumentation/groq/utils.py +...
🐛 Bug Report: Performance issues with opentelemetry-instrumentation-openai ### Which component is this bug for? Anthropic Instrumentation ### 📜 Description Hi! First, thank you for opentelemetry-instrumentation-openai. It makes my life easier :-) But, it's computationally extensive, significantly slowing the serv...
Thanks @Nagasaki45 for reporting! We'll investigate what's causing it. Since you're using streaming I have an assumption it's related to the token count enrichment which may affect latency (but some users find useful!). Can you try disabling it and see the effect? (Setting `enrich_token_usage` to false in the initializ...
1,738,018,114,000
[ "python", "size:S", "lgtm" ]
Performance Issue
[ "packages/opentelemetry-instrumentation-groq/opentelemetry/instrumentation/groq/utils.py:model_as_dict", "packages/opentelemetry-instrumentation-openai/opentelemetry/instrumentation/openai/shared/__init__.py:model_as_dict", "packages/opentelemetry-instrumentation-openai/opentelemetry/instrumentation/openai/util...
[]
3
157
hyeneung/tech-blog-hub-site
hyeneung__tech-blog-hub-site-49
940338fead4a77132a4d78b42232265301a20b83
null
diff --git a/serverless/apis/recommend/lambda_function.py b/serverless/apis/recommend/lambda_function.py index 3e72617..f61f099 100644 --- a/serverless/apis/recommend/lambda_function.py +++ b/serverless/apis/recommend/lambda_function.py @@ -1,32 +1,24 @@ import json import os from typing import Dict, Any, Tuple -fro...
[Feature] Migrate from OpenSearch to Elasticsearch to improve Search Performance ## Work Details - Current performance with AWS OpenSearch Free Tier shows query times averaging 1.5 seconds. - Plan to migrate to a self-hosted Elasticsearch instance on EC2(free tier) using Docker. - Expected outcome: Reduce query time...
1,736,058,281,000
[ ":star2: feature" ]
Performance Issue
[ "serverless/apis/recommend/opensearch/article_analyze.py:get_db_dataframe", "serverless/apis/recommend/opensearch/article_analyze.py:get_contents_base_recommendations", "serverless/apis/recommend/recommend.py:get_recommend_articles_by_url" ]
[]
3
161
aio-libs/aiohttp
aio-libs__aiohttp-9692
dd0b6e37339bd91b2ba666208d0c92f314252f07
null
diff --git a/CHANGES/9692.breaking.rst b/CHANGES/9692.breaking.rst new file mode 100644 index 00000000000..e0fdae11416 --- /dev/null +++ b/CHANGES/9692.breaking.rst @@ -0,0 +1,1 @@ +Changed ``ClientRequest.request_info`` to be a `NamedTuple` to improve client performance -- by :user:`bdraco`. diff --git a/aiohttp/clien...
RequestInfo is documented to be a namedtuple but its actually a dataclass or attrs https://docs.aiohttp.org/en/stable/client_reference.html#aiohttp.ClientResponse.request_info It would be about 2x as fast to create if it were a namedtuple though, and although dataclasses are faster to access its rarely accessed so n...
<img width="1500" alt="Screenshot 2024-11-05 at 8 42 07 PM" src="https://github.com/user-attachments/assets/92add3df-591f-4bfd-a138-db724e9414b6"> Production shows quite a bit of time making this one I really doubt if we have the real bottleneck here. Attrs/dataclasses have better human-readable interface than named...
1,730,950,174,000
[ "bot:chronographer:provided", "backport-3.11" ]
Performance Issue
[ "aiohttp/client_reqrep.py:ClientRequest.request_info" ]
[]
1
163
home-assistant/core
home-assistant__core-136739
a8c382566cae06c43a33c4d3d44c9bc92ef7b4d8
null
diff --git a/homeassistant/components/fritzbox/climate.py b/homeassistant/components/fritzbox/climate.py index d5a81fdef1a3d3..87a87ac691f58b 100644 --- a/homeassistant/components/fritzbox/climate.py +++ b/homeassistant/components/fritzbox/climate.py @@ -141,7 +141,7 @@ async def async_set_temperature(self, **kwargs: A...
diff --git a/tests/components/fritzbox/test_climate.py b/tests/components/fritzbox/test_climate.py index 29f5742216fb84..c7896920ce9364 100644 --- a/tests/components/fritzbox/test_climate.py +++ b/tests/components/fritzbox/test_climate.py @@ -273,20 +273,20 @@ async def test_update_error(hass: HomeAssistant, fritz: Moc...
Fritz!box sockets switch very slowly ### The problem If I want to switch on a Fritz!Box socket via Home Assistant, it takes a very long time until it is switched on and the entity is updated. ### What version of Home Assistant Core has the issue? core-2024.2.5 ### What was the last working version of Home Assistant...
Hey there @mib1185, @flabbamann, mind taking a look at this issue as it has been labeled with an integration (`fritzbox`) you are listed as a [code owner](https://github.com/home-assistant/core/blob/dev/CODEOWNERS#L446) for? Thanks! <details> <summary>Code owner commands</summary> Code owners of `fritzbox` can trig...
1,738,085,460,000
[ "cla-signed", "integration: fritzbox", "small-pr", "has-tests", "by-code-owner", "bugfix", "Quality Scale: No score" ]
Performance Issue
[ "homeassistant/components/fritzbox/climate.py:FritzboxThermostat.async_set_temperature", "homeassistant/components/fritzbox/cover.py:FritzboxCover.async_open_cover", "homeassistant/components/fritzbox/cover.py:FritzboxCover.async_close_cover", "homeassistant/components/fritzbox/cover.py:FritzboxCover.async_se...
[]
9
164
vllm-project/vllm
vllm-project__vllm-9338
8e3e7f271326e8cdb32c8f9581b2f98013a567c7
null
diff --git a/benchmarks/benchmark_serving.py b/benchmarks/benchmark_serving.py index 68f1e221c4bfb..0d205014b15bf 100644 --- a/benchmarks/benchmark_serving.py +++ b/benchmarks/benchmark_serving.py @@ -53,6 +53,8 @@ except ImportError: from argparse import ArgumentParser as FlexibleArgumentParser +MILLISECONDS_T...
[RFC]: Add Goodput Metric to Benchmark Serving ### Motivation. Currently, all metrics vLLM has are more from the perspectives of GenAI Service Providers. In order to provide a measurement from the perspectives of GenAI Service Users, we, from [Hao AI Lab](https://hao-ai-lab.github.io/home/#:~:text=Welcome%20to%2...
Hey @Imss27 ! Thanks for making this RFC! I've read the doc and overall it makes sense to me, so please feel free to implement the changes and ping me when the PR is ready for review! One suggestion I'd make is that I see you're using `inter_token_latency:30` in the example, but you probably want to use `TPOT` in...
1,728,901,192,000
[ "ready" ]
Feature Request
[ "benchmarks/benchmark_serving.py:calculate_metrics", "benchmarks/benchmark_serving.py:benchmark", "benchmarks/benchmark_serving.py:main" ]
[ "benchmarks/benchmark_serving.py:check_goodput_args", "benchmarks/benchmark_serving.py:parse_goodput" ]
3
166
celery/django-celery-beat
celery__django-celery-beat-835
17d87f4951d42498c9ca9aaa23ecbd95f2a7dbc3
null
diff --git a/django_celery_beat/schedulers.py b/django_celery_beat/schedulers.py index ce46b661..f2ea4bc9 100644 --- a/django_celery_beat/schedulers.py +++ b/django_celery_beat/schedulers.py @@ -11,14 +11,16 @@ from django.conf import settings from django.core.exceptions import ObjectDoesNotExist from django.db impo...
diff --git a/t/unit/test_schedulers.py b/t/unit/test_schedulers.py index 544ac4bd..ef081e77 100644 --- a/t/unit/test_schedulers.py +++ b/t/unit/test_schedulers.py @@ -456,22 +456,67 @@ def setup_scheduler(self, app): self.m4.save() self.m4.refresh_from_db() - dt_aware = make_aware(datetime(da...
Scheduler is slow when dealing with lots of tasks ### Summary: When there are a few thousand tasks in the config, beat becomes completely unreliable and unstable. Reconfiguration, syncing takes a lot of time, and with frequent schedule changes, hours can pass by without sending any task to the queue. * Celery Ver...
I am open to review and accept performance related improvement
1,735,849,295,000
[]
Performance Issue
[ "django_celery_beat/schedulers.py:DatabaseScheduler.all_as_schedule" ]
[ "django_celery_beat/schedulers.py:DatabaseScheduler.get_excluded_hours_for_crontab_tasks" ]
1
167
Standard-Labs/real-intent
Standard-Labs__real-intent-27
ed2a2cf98cbce1f571212aadd0cebf07f0bb5bd9
null
diff --git a/bigdbm/validate/email.py b/bigdbm/validate/email.py index 1afb4a7..417e45a 100644 --- a/bigdbm/validate/email.py +++ b/bigdbm/validate/email.py @@ -1,6 +1,8 @@ """Validate emails using MillionVerifier.""" import requests +from concurrent.futures import ThreadPoolExecutor + from bigdbm.schemas import M...
Email validation may be slow Have a hunch that email validation with MillionVerifier is very slow. Use threads to concurrently process emails and phones (check rate limits on Numverify and MillionVerifier).
1,723,076,595,000
[]
Performance Issue
[ "bigdbm/validate/email.py:EmailValidator.__init__", "bigdbm/validate/email.py:EmailValidator.validate", "bigdbm/validate/phone.py:PhoneValidator.__init__", "bigdbm/validate/phone.py:PhoneValidator.validate" ]
[]
4
169
Standard-Labs/real-intent
Standard-Labs__real-intent-102
708f4a641280c351e6778b0172e4b8aa15ec9a49
null
diff --git a/real_intent/deliver/followupboss/vanilla.py b/real_intent/deliver/followupboss/vanilla.py index 2db5ad1..6795ebb 100644 --- a/real_intent/deliver/followupboss/vanilla.py +++ b/real_intent/deliver/followupboss/vanilla.py @@ -3,6 +3,7 @@ from enum import StrEnum import base64 +from concurrent.futures imp...
Improve integration efficiency Multithreading
1,727,040,247,000
[]
Performance Issue
[ "real_intent/deliver/followupboss/vanilla.py:FollowUpBossDeliverer._deliver", "real_intent/deliver/kvcore/__init__.py:KVCoreDeliverer._deliver" ]
[]
2
170
vllm-project/vllm
vllm-project__vllm-7874
1248e8506a4d98b4f15cbfe729cf2af42fb4223a
null
diff --git a/vllm/core/scheduler.py b/vllm/core/scheduler.py index 4c2f715820317..81c78bda3b505 100644 --- a/vllm/core/scheduler.py +++ b/vllm/core/scheduler.py @@ -1027,16 +1027,21 @@ def _schedule_chunked_prefill(self) -> SchedulerOutputs: # Update waiting requests. self.waiting.extendleft(running...
diff --git a/tests/basic_correctness/test_chunked_prefill.py b/tests/basic_correctness/test_chunked_prefill.py index fc6f829c37b06..a63ac380e8598 100644 --- a/tests/basic_correctness/test_chunked_prefill.py +++ b/tests/basic_correctness/test_chunked_prefill.py @@ -116,6 +116,9 @@ def test_models_with_fp8_kv_cache( ...
[Performance]: vllm 0.5.4 with enable_chunked_prefill =True, throughput is slightly lower than 0.5.3~0.5.0. ### Your current environment <details> <summary>environment</summary> # Hardware & Nvidia driver & OS GPU models and configuration: GPU 0: NVIDIA GeForce RTX 4090 Nvidia driver version: 550.107.02...
update I managed to compile vllm 0.5.4 with torch == 2.3.1 xformers == 0.0.27 vllm-flash-attn == 2.5.9.post1 (from vllm 0.5.3 requirements ), the bug is still there. ![chunked_prefill2](https://github.com/user-attachments/assets/fbfca114-ae8b-4e13-ae3c-3afe9d3730d7) The bug only affects chunked_prefill, ever...
1,724,682,556,000
[ "ready" ]
Performance Issue
[ "vllm/core/scheduler.py:Scheduler._schedule_chunked_prefill" ]
[]
1
171
NCSU-High-Powered-Rocketry-Club/AirbrakesV2
NCSU-High-Powered-Rocketry-Club__AirbrakesV2-151
2d85a6c6bf946636c5e6b4aa74388b9d376289f3
null
diff --git a/airbrakes/airbrakes.py b/airbrakes/airbrakes.py index a4519053..f52d7200 100644 --- a/airbrakes/airbrakes.py +++ b/airbrakes/airbrakes.py @@ -1,7 +1,6 @@ """Module which provides a high level interface to the air brakes system on the rocket.""" import time -from collections import deque from typing im...
diff --git a/tests/auxil/utils.py b/tests/auxil/utils.py index 3db14c63..24b1d76f 100644 --- a/tests/auxil/utils.py +++ b/tests/auxil/utils.py @@ -3,6 +3,7 @@ from airbrakes.telemetry.packets.apogee_predictor_data_packet import ApogeePredictorDataPacket from airbrakes.telemetry.packets.context_data_packet import Cont...
Refactor logging to be more clear and performant After the changes in #146 in the logger to get #147 resolved, our logging process has slightly slowed down due to use of `msgpec.structs.asdict()` instead of `msgspec.to_builtins()`. We can still keep up with the IMU process, but the margin has reduced. I propose the f...
1,738,633,910,000
[ "enhancement" ]
Performance Issue
[ "airbrakes/airbrakes.py:AirbrakesContext.__init__", "airbrakes/airbrakes.py:AirbrakesContext.update", "airbrakes/hardware/base_imu.py:BaseIMU.get_imu_data_packets", "airbrakes/telemetry/logger.py:Logger.__init__", "airbrakes/telemetry/logger.py:Logger._prepare_log_dict", "airbrakes/telemetry/logger.py:Log...
[ "airbrakes/telemetry/logger.py:Logger._prepare_logger_packets" ]
9
172
scikit-learn/scikit-learn
scikit-learn__scikit-learn-8478
312f64053d7249a326a19a07fa635ef5b5c6ed99
null
diff --git a/doc/modules/classes.rst b/doc/modules/classes.rst index 2e7dcba82e846..243c63ab0c7e2 100644 --- a/doc/modules/classes.rst +++ b/doc/modules/classes.rst @@ -646,6 +646,7 @@ Kernels: :template: class.rst impute.SimpleImputer + impute.MICEImputer .. _kernel_approximation_ref: diff --git a/doc...
diff --git a/sklearn/tests/test_impute.py b/sklearn/tests/test_impute.py index f2bf5912e2213..954a016a835bb 100644 --- a/sklearn/tests/test_impute.py +++ b/sklearn/tests/test_impute.py @@ -1,14 +1,19 @@ +from __future__ import division + +import pytest import numpy as np from scipy import sparse -from sklearn.uti...
MICE imputer Proceeding from https://github.com/scikit-learn/scikit-learn/pull/4844#issuecomment-253043238 is a suggestion that MICE imputation be included in scikit-learn. @sergeyf has implemented it [here](https://github.com/hammerlab/fancyimpute/blob/master/fancyimpute/mice.py). Here we will discuss issues relate...
@sergeyf [wrote](https://github.com/scikit-learn/scikit-learn/pull/4844#issuecomment-259055975): > @amueller and @jnothman do either of you have a sense of what the right structure for MICE would be? It's pretty involved, so I don't want to just add a new option to `Imputer`. What about adding a new `MICE` class to `i...
1,488,322,557,000
[]
Feature Request
[ "sklearn/utils/estimator_checks.py:_yield_non_meta_checks" ]
[ "examples/plot_missing_values.py:get_results", "sklearn/impute.py:MICEImputer.__init__", "sklearn/impute.py:MICEImputer._impute_one_feature", "sklearn/impute.py:MICEImputer._get_neighbor_feat_idx", "sklearn/impute.py:MICEImputer._get_ordered_idx", "sklearn/impute.py:MICEImputer._get_abs_corr_mat", "skle...
1
177
cupy/cupy
cupy__cupy-3730
4c81aebbec5d481ddee8685d982b24372e5f272b
null
diff --git a/cupy/cuda/cufft.pyx b/cupy/cuda/cufft.pyx index 8df8ba92019..166ea4993de 100644 --- a/cupy/cuda/cufft.pyx +++ b/cupy/cuda/cufft.pyx @@ -279,6 +279,17 @@ class Plan1d(object): else: self._multi_gpu_get_plan( plan, nx, fft_type, batch, devices, out) + ...
diff --git a/tests/cupy_tests/fft_tests/test_cache.py b/tests/cupy_tests/fft_tests/test_cache.py new file mode 100644 index 00000000000..fc882f882d2 --- /dev/null +++ b/tests/cupy_tests/fft_tests/test_cache.py @@ -0,0 +1,482 @@ +import contextlib +import io +import queue +import threading +import unittest + +import pyt...
Reuse cufft plan objects Reusing a plan object have a significant impact in the complete fft performance. I measured a ~500us time drop. We should try to create a `PlanAllocator` or a pool of plans to reuse objects according to the requested size. (Plans allocate gpu memory so we have to be careful when reusing t...
Previous discussion was in #1669. @grlee77 mentioned he has a cache prototype. For the purpose of #3587 to speed up correlate/convolve, we should first focus on memoizing `Plan1d`, as it has less parameters to be used as the cache key. PyTorch refs: - usage: https://github.com/pytorch/pytorch/blob/master/docs/source/n...
1,596,694,236,000
[ "cat:enhancement", "to-be-backported" ]
Performance Issue
[ "cupy/fft/config.py:set_cufft_gpus", "cupy/fft/fft.py:_exec_fft", "cupy/fft/fft.py:_get_cufft_plan_nd", "cupyx/scipy/fftpack/_fft.py:get_fft_plan" ]
[]
4
178
scikit-learn/scikit-learn
scikit-learn__scikit-learn-17443
c3db2cdbf5244229af2c5eb54f9216a69f77a146
da36f72cc2960d47792157597d67a488d35a05f3
diff --git a/doc/modules/calibration.rst b/doc/modules/calibration.rst index d0a9737dac612..1fcd1d501d100 100644 --- a/doc/modules/calibration.rst +++ b/doc/modules/calibration.rst @@ -30,11 +30,25 @@ approximately 80% actually belong to the positive class. Calibration curves ------------------ -The following plot ...
diff --git a/sklearn/tests/test_calibration.py b/sklearn/tests/test_calibration.py index 4fe08c27fb19e..8decff0cc96d5 100644 --- a/sklearn/tests/test_calibration.py +++ b/sklearn/tests/test_calibration.py @@ -18,7 +18,7 @@ ) from sklearn.utils.extmath import softmax from sklearn.exceptions import NotFittedError -fro...
Add calibration curve to plotting module I think it would be nice to have something like ```python def plot_calibration_curve(y_true, prob_y_pred, n_bins=5, ax=None): prop_true, prop_pred = calibration_curve(y_true, prob_y_pred, n_bins=n_bins) if ax is None: ax = plt.gca() ax.plot([0, 1], [0...
should plot functions return the artist? or the axes? hm... Maybe adding a histogram as in http://scikit-learn.org/stable/auto_examples/calibration/plot_compare_calibration.html would also be nice ```python def plot_calibration_curve(y_true, y_prob, n_bins=5, ax=None, hist=True, normalize=False): prob_true, prob_...
1,591,270,294,000
[ "module:metrics" ]
Feature Request
[ "examples/calibration/plot_calibration_curve.py:plot_calibration_curve", "sklearn/metrics/_plot/base.py:_get_response" ]
[ "examples/calibration/plot_calibration_curve.py:NaivelyCalibratedLinearSVC.fit", "examples/calibration/plot_calibration_curve.py:NaivelyCalibratedLinearSVC.predict_proba", "examples/calibration/plot_compare_calibration.py:NaivelyCalibratedLinearSVC.fit", "examples/calibration/plot_compare_calibration.py:Naive...
2
179
Qiskit/qiskit
Qiskit__qiskit-13141
90e92a46643c72a21c5852299243213907453c21
null
diff --git a/crates/accelerate/src/euler_one_qubit_decomposer.rs b/crates/accelerate/src/euler_one_qubit_decomposer.rs index 98333cad39d2..e6ca094186ff 100644 --- a/crates/accelerate/src/euler_one_qubit_decomposer.rs +++ b/crates/accelerate/src/euler_one_qubit_decomposer.rs @@ -579,7 +579,7 @@ pub fn generate_circuit( ...
diff --git a/test/python/transpiler/test_unitary_synthesis.py b/test/python/transpiler/test_unitary_synthesis.py index 4abf6511d8d2..aaad7b71279b 100644 --- a/test/python/transpiler/test_unitary_synthesis.py +++ b/test/python/transpiler/test_unitary_synthesis.py @@ -18,6 +18,7 @@ import unittest import numpy as np ...
Port `UnitarySynthesis` to Rust This issue tracks porting the `UnitarySynthesis` pass to rust as part of the #12208 epic. This pass in particular will still always require a fairly large python component, as the unitary synthesis plugin interface necessitates a Python execution mode. But when the a plugin is not specif...
1,726,135,509,000
[ "performance", "Changelog: None", "Rust", "mod: transpiler" ]
Feature Request
[ "qiskit/transpiler/passes/synthesis/unitary_synthesis.py:UnitarySynthesis.run" ]
[]
1
180
dask/dask
dask__dask-8040
4a59a6827578fdf8e105d1fb7e9d50d428c9d5fa
null
diff --git a/dask/local.py b/dask/local.py index 9f035ef277a..664ae93aba2 100644 --- a/dask/local.py +++ b/dask/local.py @@ -181,7 +181,7 @@ def start_state_from_dask(dsk, cache=None, sortkey=None): waiting_data = dict((k, v.copy()) for k, v in dependents.items() if v) ready_set = set([k for k, v in waiting...
diff --git a/dask/tests/test_local.py b/dask/tests/test_local.py index 8830a216e27..a45aab7b991 100644 --- a/dask/tests/test_local.py +++ b/dask/tests/test_local.py @@ -1,3 +1,5 @@ +import pytest + import dask from dask.local import finish_task, get_sync, sortkey, start_state_from_dask from dask.order import order @...
Regression: poor memory management in threads and sync schedulers Reopen after triage from https://github.com/pydata/xarray/issues/5165 ```python import dask.array as da d = da.random.random((15000, 250000), chunks=(1, -1)).sum() d.compute(optimize_graph=False) ``` Peak RAM usage: Before #6322 (2021.3.1): 200 ...
Also cc @mrocklin @jrbourbeau . IMHO this should be treated as very high priority as it is something that is very likely to hit in the face a newbie user when he plays around with dask.array or dask.dataframe for the very first time. Hi, I'm also experiencing increased memory usage since ad0e5d140dfc934d94583f8ce50d...
1,628,956,822,000
[]
Performance Issue
[ "dask/local.py:start_state_from_dask", "dask/local.py:finish_task", "dask/local.py:get_async" ]
[]
3
181
django/django
django__django-18158
8c118c0e00846091c261b97dbed9a5b89ceb79bf
null
diff --git a/django/core/management/commands/shell.py b/django/core/management/commands/shell.py index f55b346406aa..1619561fea2e 100644 --- a/django/core/management/commands/shell.py +++ b/django/core/management/commands/shell.py @@ -2,7 +2,9 @@ import select import sys import traceback +from collections import def...
diff --git a/tests/shell/models.py b/tests/shell/models.py new file mode 100644 index 000000000000..85b40bf2058e --- /dev/null +++ b/tests/shell/models.py @@ -0,0 +1,9 @@ +from django.db import models + + +class Marker(models.Model): + pass + + +class Phone(models.Model): + name = models.CharField(max_length=50) ...
Auto-importing models feature for shell-command Description This would be an update of the existing Django shell that auto-imports models for you from your app/project. Also, the goal would be to allow the user to subclass this shell to customize its behavior and import extra things. wiki ​proposal Auto-importing mo...
['In dfac15d5: Fixed #35517, Refs #35515 -- Improved test coverage of shell command.', 1719481381.0] ['Discussion to decide on whether to have additional imports (and which ones): \u200bhttps://forum.djangoproject.com/t/default-automatic-imports-in-the-shell/33708', 1723261632.0] ['Based off the discussion, it looks li...
1,715,436,304,000
[]
Feature Request
[ "django/core/management/commands/shell.py:Command.add_arguments", "django/core/management/commands/shell.py:Command.ipython", "django/core/management/commands/shell.py:Command.bpython", "django/core/management/commands/shell.py:Command.python", "django/core/management/commands/shell.py:Command.handle" ]
[ "django/core/management/commands/shell.py:Command.get_and_report_namespace", "django/core/management/commands/shell.py:Command.get_namespace" ]
5
182
Happy-Algorithms-League/hal-cgp
Happy-Algorithms-League__hal-cgp-180
5421d9cdf0812ab3098d54c201ee115fa3129bce
null
diff --git a/cgp/genome.py b/cgp/genome.py index 3c18a87f..4e8f035d 100644 --- a/cgp/genome.py +++ b/cgp/genome.py @@ -107,7 +107,7 @@ def dna(self) -> List[int]: def dna(self, value: List[int]) -> None: self._validate_dna(value) self._dna = value - self._initialize_unkown_parameters() + ...
diff --git a/test/test_ea_mu_plus_lambda.py b/test/test_ea_mu_plus_lambda.py index cf0e1b2b..8f0521ea 100644 --- a/test/test_ea_mu_plus_lambda.py +++ b/test/test_ea_mu_plus_lambda.py @@ -196,6 +196,8 @@ def objective(individual): individual.fitness = float(individual.idx) return individual + popu...
Issue warning when `mutation_rate` is high and `mutate` potentially takes long #157 made sure that exactly the correct number of mutations occurs in a genome by checking that the mutated genome differs from the original one in `n_mutations` position. when using a high mutation rate and small number of primitives, this ...
btw, an alternative solution to this problem would be to have a maximum number iterations in the while loop (https://github.com/Happy-Algorithms-League/hal-cgp/blob/master/cgp/genome.py#L377) and return even if the desired number of differences was not reached while issuing a warning to the user What about keeping trac...
1,594,391,036,000
[]
Performance Issue
[ "cgp/genome.py:Genome.dna", "cgp/genome.py:Genome.mutate", "cgp/genome.py:Genome._mutate_output_region", "cgp/genome.py:Genome._mutate_hidden_region", "cgp/population.py:Population.__init__" ]
[ "cgp/genome.py:Genome._is_hidden_input_gene", "cgp/genome.py:Genome._select_gene_indices_for_mutation", "cgp/genome.py:Genome._determine_alternative_permissible_values", "cgp/genome.py:Genome._determine_alternative_permissible_values_hidden_gene", "cgp/genome.py:Genome._determine_alternative_permissible_val...
5
183
sgkit-dev/sgkit
sgkit-dev__sgkit-447
9150392b3b38f575d12b5f555877fc059ee44591
null
diff --git a/sgkit/distance/api.py b/sgkit/distance/api.py index 70141b598..7df70eb16 100644 --- a/sgkit/distance/api.py +++ b/sgkit/distance/api.py @@ -1,14 +1,20 @@ +import typing + import dask.array as da import numpy as np +from typing_extensions import Literal from sgkit.distance import metrics from sgkit.ty...
diff --git a/sgkit/tests/test_distance.py b/sgkit/tests/test_distance.py index 3432b6421..f5bd44861 100644 --- a/sgkit/tests/test_distance.py +++ b/sgkit/tests/test_distance.py @@ -10,7 +10,7 @@ squareform, ) -from sgkit.distance.api import pairwise_distance +from sgkit.distance.api import MetricTypes, pairwise...
Pairwise distance scalability Raising this issue to revisit the scalability of our pairwise distance calculation and whether it's worth returning to a map-reduce style implementation that would allow chunking along both dimensions. In the work that @aktech is doing on early scalability demonstrations (#345) there ar...
> Secondly, do we ever really need to run pairwise distance on arrays that are large in the variants dimension? I.e., do we care about scaling this up to large numbers of variants? xref #306 (comment) I think you mean in the *samples* dimension? (If so, I agree - ~10K is as many samples as we could hope to support ...
1,611,845,371,000
[]
Performance Issue
[ "sgkit/distance/api.py:pairwise_distance", "sgkit/distance/metrics.py:correlation", "sgkit/distance/metrics.py:euclidean" ]
[ "sgkit/distance/metrics.py:euclidean_map", "sgkit/distance/metrics.py:euclidean_reduce", "sgkit/distance/metrics.py:correlation_map", "sgkit/distance/metrics.py:correlation_reduce" ]
3
184
numpy/numpy
numpy__numpy-17394
43683b3256a86659f230dcadbcde1f8020398bfa
43f80863ba2994a186293d86d516204ef38b0043
diff --git a/doc/release/upcoming_changes/17394.new_function.rst b/doc/release/upcoming_changes/17394.new_function.rst new file mode 100644 index 000000000000..50c9d1db310d --- /dev/null +++ b/doc/release/upcoming_changes/17394.new_function.rst @@ -0,0 +1,5 @@ +``sliding_window_view`` provides a sliding window view for...
diff --git a/numpy/lib/tests/test_stride_tricks.py b/numpy/lib/tests/test_stride_tricks.py index 10d7a19abec0..efec5d24dad4 100644 --- a/numpy/lib/tests/test_stride_tricks.py +++ b/numpy/lib/tests/test_stride_tricks.py @@ -6,8 +6,10 @@ ) from numpy.lib.stride_tricks import ( as_strided, broadcast_arrays, _br...
Suggestion: Sliding Window Function Using `np.lib.stride_tricks.as_stride` one can very efficiently create a sliding window that segments an array as a preprocessing step for vectorized applications. For example a moving average of a window length `3`, stepsize `1`: ``` a = numpy.arange(10) a_strided = numpy.lib.strid...
1,601,380,131,000
[ "01 - Enhancement", "62 - Python API" ]
Feature Request
[ "numpy/lib/stride_tricks.py:as_strided" ]
[ "numpy/lib/stride_tricks.py:_sliding_window_view_dispatcher", "numpy/lib/stride_tricks.py:sliding_window_view" ]
1
185
Ouranosinc/xclim
Ouranosinc__xclim-477
cc17fde69b757443602872d1156caf00b34ba161
null
diff --git a/HISTORY.rst b/HISTORY.rst index 1d6a6ebc5..0bd3055e1 100644 --- a/HISTORY.rst +++ b/HISTORY.rst @@ -4,6 +4,7 @@ History 0.18.x ------ +* Optimization options for `xclim.sdba` : different grouping for the normalization steps of DQM and save training or fitting datasets to temporary files. * `xclim.sdba...
diff --git a/tests/test_sdba/test_adjustment.py b/tests/test_sdba/test_adjustment.py index c897ea3ca..532f8a67f 100644 --- a/tests/test_sdba/test_adjustment.py +++ b/tests/test_sdba/test_adjustment.py @@ -190,7 +190,7 @@ def test_mon_U(self, mon_series, series, mon_triangular, kind, name, spatial_dim if spatia...
sdba - improve the speed of DetrendedQuantileMapping ### Description Doing a few preliminary tests **on a single pixel and without dask**, I obtained the following times : DetrendedQuantileMapping, groups='month', 150 years of daily data : ~15 seconds DetrendedQuantileMapping, groups='day', window=31, 150 years o...
To clear out my thoughts: - The normalization process, as all non-grouping grouped operations, has additional overhead because it needs to resort the data along the time axis. And to rechunk if dask is used. - `interp_on_quantiles` currently calls griddata through a _vectorized_ `xr.apply_ufunc`, which is far from ...
1,592,246,670,000
[]
Performance Issue
[ "xclim/sdba/adjustment.py:DetrendedQuantileMapping._train", "xclim/sdba/adjustment.py:DetrendedQuantileMapping._adjust", "xclim/sdba/base.py:Grouper.get_index", "xclim/sdba/detrending.py:BaseDetrend.fit", "xclim/sdba/detrending.py:BaseDetrend.get_trend", "xclim/sdba/detrending.py:BaseDetrend._set_fitds", ...
[ "xclim/sdba/adjustment.py:DetrendedQuantileMapping.__init__", "xclim/sdba/detrending.py:BaseDetrend._set_ds" ]
9
186
scikit-learn/scikit-learn
scikit-learn__scikit-learn-10280
1557cb8a1910bce6dc33cd5cd3a08c380bdec566
null
diff --git a/doc/modules/classes.rst b/doc/modules/classes.rst index 68494051041be..99731c7fda599 100644 --- a/doc/modules/classes.rst +++ b/doc/modules/classes.rst @@ -955,6 +955,7 @@ See the :ref:`metrics` section of the user guide for further details. metrics.pairwise_distances metrics.pairwise_distances_arg...
diff --git a/sklearn/metrics/tests/test_pairwise.py b/sklearn/metrics/tests/test_pairwise.py index 799b3e4fe9bf7..0ef089c7a3619 100644 --- a/sklearn/metrics/tests/test_pairwise.py +++ b/sklearn/metrics/tests/test_pairwise.py @@ -1,11 +1,15 @@ +from types import GeneratorType + import numpy as np from numpy import lin...
[MRG] ENH: Added block_size parameter for lesser memory consumption #### Reference Issue Fixes #7287 #### What does this implement/fix? Explain your changes. This intends to add a function `pairwise_distances_blockwise` with an additional block_size parameter to avoid storing all the O(n^2) pairs' distances. Thi...
1,512,901,651,000
[]
Feature Request
[ "sklearn/_config.py:set_config", "sklearn/_config.py:config_context", "sklearn/metrics/pairwise.py:pairwise_distances_argmin_min", "sklearn/metrics/pairwise.py:pairwise_distances_argmin", "sklearn/metrics/pairwise.py:cosine_similarity", "sklearn/metrics/pairwise.py:pairwise_distances" ]
[ "sklearn/metrics/pairwise.py:_argmin_min_reduce", "sklearn/metrics/pairwise.py:_check_chunk_size", "sklearn/metrics/pairwise.py:pairwise_distances_chunked", "sklearn/utils/__init__.py:get_chunk_n_rows" ]
6
187
scikit-learn/scikit-learn
scikit-learn__scikit-learn-10471
52aaf8269235d4965022b8ec970243bdcb59c9a7
null
diff --git a/doc/whats_new/v0.20.rst b/doc/whats_new/v0.20.rst index a7ae9f7415243..b4a8046d02956 100644 --- a/doc/whats_new/v0.20.rst +++ b/doc/whats_new/v0.20.rst @@ -125,6 +125,13 @@ Classifiers and regressors only require X to be an object with finite length or shape. :issue:`9832` by :user:`Vrishank Bhardwaj...
diff --git a/sklearn/cluster/tests/test_k_means.py b/sklearn/cluster/tests/test_k_means.py index f0f2b56bd591c..de8772d761e22 100644 --- a/sklearn/cluster/tests/test_k_means.py +++ b/sklearn/cluster/tests/test_k_means.py @@ -169,7 +169,8 @@ def _check_fitted_model(km): assert_greater(km.inertia_, 0.0) # che...
KMeans optimisation for array C/F contiguity (was: Make sure that the output of PCA.fit_transform is C contiguous) otherwise, I would rather use: ``` Python pca.fit(X) X_new = pca.transform(X) ``` Because of FORTRAN data for inner product is very slow. (for example, in the KMeans) KMeans optimisation for array C/F c...
Can you give an example? Maybe we should then rather change the behavior in KMeans. I wouldn't change the output format, since we don't know what the user wants to do next. They should be the same that the output and input format of PCA.fit_transform, isn't it? Because PCA.transform is such. In _k_means._assign_label...
1,515,928,846,000
[]
Performance Issue
[ "sklearn/cluster/k_means_.py:k_means", "sklearn/cluster/k_means_.py:_kmeans_single_elkan", "sklearn/cluster/k_means_.py:KMeans._check_fit_data", "sklearn/cluster/k_means_.py:KMeans.fit", "sklearn/cluster/k_means_.py:KMeans.fit_transform", "sklearn/cluster/k_means_.py:MiniBatchKMeans.fit", "sklearn/clust...
[]
7
188
pandas-dev/pandas
pandas-dev__pandas-22762
c8ce3d01e9ffafc24c6f9dd568cd9eb7e42c610c
null
diff --git a/doc/source/whatsnew/v0.24.0.txt b/doc/source/whatsnew/v0.24.0.txt index 3d82dd042da20..29b766e616b3b 100644 --- a/doc/source/whatsnew/v0.24.0.txt +++ b/doc/source/whatsnew/v0.24.0.txt @@ -48,7 +48,7 @@ Pandas has gained the ability to hold integer dtypes with missing values. This l Here is an example of t...
diff --git a/pandas/conftest.py b/pandas/conftest.py index 621de3ffd4b12..e84657a79b51a 100644 --- a/pandas/conftest.py +++ b/pandas/conftest.py @@ -131,6 +131,30 @@ def all_arithmetic_operators(request): return request.param +_all_numeric_reductions = ['sum', 'max', 'min', + 'mean', ...
ENH: add reduce op (and groupby reduce) support to EA after #21160 Reductions for ExtensionArray Creeping up to this in https://github.com/pandas-dev/pandas/pull/22345 A few questions 1. What should this look like for EA authors? What helpers can / should we provide? 2. How does this affect users? Specifically,...
1,537,363,868,000
[ "Numeric Operations", "ExtensionArray" ]
Feature Request
[ "pandas/core/arrays/categorical.py:Categorical._reduce", "pandas/core/series.py:Series._reduce" ]
[ "pandas/core/arrays/base.py:ExtensionArray._reduce", "pandas/core/arrays/integer.py:IntegerArray._reduce" ]
2
189
tarantool/ansible-cartridge
tarantool__ansible-cartridge-172
087f6126c1cccc83e67c36ceab6868b43251e04b
null
diff --git a/CHANGELOG.md b/CHANGELOG.md index 978b7e7a..a6396c0b 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -26,6 +26,9 @@ to use the newest tag with new release to expel. It can be called second time to set up failover priority for replicasets where new instances were joined. As a result, `replicaset_he...
diff --git a/unit/test_tags.py b/unit/test_tags.py index 01e1f68e..74281892 100644 --- a/unit/test_tags.py +++ b/unit/test_tags.py @@ -48,6 +48,7 @@ def test_without_tags(self): 'Set remote_user for delegated tasks', 'Validate config', 'Set instance facts', + 'Get one i...
[2pt] Too long "Install package" task I have 100 Tarantool instances in a cluster. When I deploy them via ansible-cartridge, it takes a very long time to complete the "Install package" task, since it is performed for each instance. Moreover, this task is performed even despite the fact that it is written in the logs th...
1,612,778,877,000
[]
Performance Issue
[ "filter_plugins/filter_hosts.py:get_machine_identifier", "filter_plugins/filter_hosts.py:get_one_not_expelled_instance_for_machine" ]
[ "filter_plugins/filter_hosts.py:get_machine_hostname" ]
2
190
scikit-learn/scikit-learn
scikit-learn__scikit-learn-23197
652a6278c45f8fedfe55581503941ffc8efffaca
null
diff --git a/doc/whats_new/v1.2.rst b/doc/whats_new/v1.2.rst index 832a4e4389b19..22aecf64a85b8 100644 --- a/doc/whats_new/v1.2.rst +++ b/doc/whats_new/v1.2.rst @@ -44,6 +44,11 @@ Changes impacting all modules NumPy's SIMD optimized primitives. :pr:`23446` by :user:`Meekail Zain <micky774>` +- |Enhancement| Fin...
Optimizing assert_all_finite check Have done some work to come up with a Cython implementation of [`_assert_all_finite`]( https://github.com/scikit-learn/scikit-learn/blob/6d15840432c08b2c628babe62d7bef6d6a01fcf6/sklearn/utils/validation.py#L40 ) as the current check is rather slow. Have come up with [this implementati...
Because we are have check by default, and it can be relatively slow (https://github.com/scikit-learn/scikit-learn/pull/11487#issuecomment-405034867) in some cases, it would be definitely nice to make it faster. Ideally, it would be great to have such an improvement in numpy (with possible temporary backport in scik...
1,650,671,054,000
[ "module:utils", "cython" ]
Performance Issue
[ "sklearn/utils/setup.py:configuration", "sklearn/utils/validation.py:_assert_all_finite" ]
[]
2
191
numpy/numpy
numpy__numpy-8206
6feb18ce79bacdac09945cf2ca0ddf85f8294298
null
diff --git a/doc/release/1.16.0-notes.rst b/doc/release/1.16.0-notes.rst index 2b84bb90a311..6a73ed3548da 100644 --- a/doc/release/1.16.0-notes.rst +++ b/doc/release/1.16.0-notes.rst @@ -107,6 +107,12 @@ Previously, a ``LinAlgError`` would be raised when an empty matrix/empty matrices (with zero rows and/or columns) i...
diff --git a/numpy/lib/tests/test_function_base.py b/numpy/lib/tests/test_function_base.py index d5faed6aea49..40cca1dbb4bb 100644 --- a/numpy/lib/tests/test_function_base.py +++ b/numpy/lib/tests/test_function_base.py @@ -734,6 +734,58 @@ def test_subclass(self): assert_array_equal(out3.mask, [[], [], [], [],...
Add "insert_zero" option for np.diff I often find myself wanting the difference of an array along some axis, where the 0th element should just be the first entry of the array along that axis, so that the shape of the output matches that of the input. This is a pretty common requirement for people working with finite ...
The inverse of cumsum problem would also be solved by adding `include_identity` to cumsum (#6044), which is something I've felt a stronger need for. I prefer the ediff1d arguments to_begin and to_end, allowing for non zero and/or multiple elements to be inserted on either side. Maybe the best solution is to add the p...
1,477,311,409,000
[ "01 - Enhancement", "component: numpy.lib", "56 - Needs Release Note." ]
Feature Request
[ "numpy/lib/function_base.py:diff" ]
[]
1
192
scikit-learn/scikit-learn
scikit-learn__scikit-learn-14800
acb8ac5145cfd88fdbd2d381b34883b2c212c8c5
1e08459c742537517bb8fc4c9d4621b9d6d227bf
diff --git a/doc/whats_new/v0.24.rst b/doc/whats_new/v0.24.rst index a647de6dac628..ba137fb76698f 100644 --- a/doc/whats_new/v0.24.rst +++ b/doc/whats_new/v0.24.rst @@ -224,6 +224,11 @@ Changelog :meth:`tree.DecisionTreeRegressor.fit`, and has not effect. :pr:`17614` by :user:`Juan Carlos Alfaro Jiménez <alfaro96...
diff --git a/sklearn/datasets/tests/data/openml/1/api-v1-json-data-1.json.gz b/sklearn/datasets/tests/data/openml/1/api-v1-json-data-1.json.gz index f75912bf2def7..ba544db491637 100644 Binary files a/sklearn/datasets/tests/data/openml/1/api-v1-json-data-1.json.gz and b/sklearn/datasets/tests/data/openml/1/api-v1-json-d...
Validate MD5 checksum of downloaded ARFF in fetch_openml `fetch_openml` downloads the `data_set_description` from openml.org which includes an `md5_checksum` field. This should correspond to the MD5 checksum of the downloaded ARFF data. We should (optionally?) use `hashlib` to verify that the MD5 matches.
I'll give this a go. thanks. you may need to hack the md5s currently in sklearn/datasets/tests/data/openml to match the truncated ARFF data there Picking this to work on while in NYC WiMLDS Sprint Aug 24, 2019 I think this check should be optional. What do other core developers think? Forcing the verification ensures ...
1,566,681,065,000
[ "module:datasets" ]
Feature Request
[ "sklearn/datasets/_openml.py:_load_arff_response", "sklearn/datasets/_openml.py:_download_data_to_bunch", "sklearn/datasets/_openml.py:fetch_openml" ]
[]
3
193
pandas-dev/pandas
pandas-dev__pandas-18330
d421a09e382109c1bbe064107c4024b065839de2
null
diff --git a/doc/source/reshaping.rst b/doc/source/reshaping.rst index 1209c4a8d6be8..1b81d83bb76c7 100644 --- a/doc/source/reshaping.rst +++ b/doc/source/reshaping.rst @@ -240,7 +240,7 @@ values will be set to ``NaN``. df3 df3.unstack() -.. versionadded: 0.18.0 +.. versionadded:: 0.18.0 Alternatively, uns...
diff --git a/pandas/tests/reshape/test_reshape.py b/pandas/tests/reshape/test_reshape.py index 2722c3e92d85a..5d4aa048ae303 100644 --- a/pandas/tests/reshape/test_reshape.py +++ b/pandas/tests/reshape/test_reshape.py @@ -218,35 +218,51 @@ def test_multiindex(self): class TestGetDummies(object): - sparse = False...
ENH: allow get_dummies to accept dtype argument - [x] closes #18330 (there's no issue for this one) - [x] tests added / passed - [x] passes `git diff upstream/master -u -- "*.py" | flake8 --diff` - [x] whatsnew entry Update in version 0.19.0 made `get_dummies` return uint8 values instead of floats (#8725). While ...
1,510,867,741,000
[ "Reshaping", "Dtype Conversions" ]
Feature Request
[ "pandas/core/generic.py:NDFrame._set_axis_name", "pandas/core/reshape/reshape.py:get_dummies", "pandas/core/reshape/reshape.py:_get_dummies_1d" ]
[]
3
194
scikit-learn/scikit-learn
scikit-learn__scikit-learn-11179
a1fabce6531bc75964588ef9f7f95527cae899bc
null
diff --git a/doc/modules/classes.rst b/doc/modules/classes.rst index fb050fd87e88c..473ea1c6a3539 100644 --- a/doc/modules/classes.rst +++ b/doc/modules/classes.rst @@ -846,6 +846,7 @@ details. metrics.jaccard_similarity_score metrics.log_loss metrics.matthews_corrcoef + metrics.multilabel_confusion_matri...
diff --git a/sklearn/metrics/tests/test_classification.py b/sklearn/metrics/tests/test_classification.py index 8e18af7128350..3152521f23b77 100644 --- a/sklearn/metrics/tests/test_classification.py +++ b/sklearn/metrics/tests/test_classification.py @@ -46,6 +46,7 @@ from sklearn.metrics import recall_score from sklea...
[WIP] ENH Multilabel confusion matrix This PR considers a helper for multilabel/set-wise evaluation metrics such as precision, recall, fbeta, jaccard (#10083), fall-out, miss rate and specificity (#5516). It also incorporates suggestions from #8126 regarding efficiency of multilabel true positives calculation (but does...
What does confusion matrix mean in a multilabel context? On 20 July 2014 20:34, Arnaud Joly notifications@github.com wrote: > Currently the confusion_matrix support binary and multi-class > classification, but not multi-label data yet. > > — > Reply to this email directly or view it on GitHub > https://github.com/...
1,527,811,029,000
[]
Feature Request
[ "sklearn/metrics/classification.py:f1_score", "sklearn/metrics/classification.py:fbeta_score", "sklearn/metrics/classification.py:precision_recall_fscore_support", "sklearn/metrics/classification.py:precision_score", "sklearn/metrics/classification.py:recall_score", "sklearn/metrics/classification.py:clas...
[ "sklearn/metrics/classification.py:multilabel_confusion_matrix" ]
6
195
pandas-dev/pandas
pandas-dev__pandas-19074
24d95098789c8685e7c0ce6a57982d0372ffc3e0
null
diff --git a/doc/source/whatsnew/v0.23.0.txt b/doc/source/whatsnew/v0.23.0.txt index 246eab386b2ab..efb4707649f08 100644 --- a/doc/source/whatsnew/v0.23.0.txt +++ b/doc/source/whatsnew/v0.23.0.txt @@ -380,6 +380,7 @@ Performance Improvements - Improved performance of ``DatetimeIndex`` and ``Series`` arithmetic operati...
diff --git a/pandas/tests/indexes/test_multi.py b/pandas/tests/indexes/test_multi.py index 9664d73651185..aedc957ec67da 100644 --- a/pandas/tests/indexes/test_multi.py +++ b/pandas/tests/indexes/test_multi.py @@ -1258,6 +1258,17 @@ def test_get_loc_level(self): assert result == expected assert new_ind...
New engine for MultiIndex? Currently, ``MultiIndex.get_loc()`` and ``MultiIndex.get_indexer()`` both rely on an ``_engine`` which is either a ``MultiIndexObjectEngine`` or a ``MultiIndexHashEngine``: but both of these are thin layers over the flat ``ObjectEngine``. This means that the actual structure of labels and lev...
See also previous discussion in #1752. Aha, and #16324, which was a response to #16319, which was a reaction to #15245, which was a response to #13904, which was a follow-up to the issue @chris-b1 mentioned. And while the ``MultiIndexHashEngine`` is not what I described above (it's probably better), indeed only now ...
1,515,062,485,000
[ "Indexing", "MultiIndex" ]
Feature Request
[ "pandas/core/indexes/multi.py:MultiIndex._engine", "pandas/core/indexes/multi.py:MultiIndex.get_indexer", "pandas/core/indexes/multi.py:MultiIndex.get_loc", "pandas/core/indexes/multi.py:MultiIndex.get_loc_level" ]
[ "pandas/core/indexes/multi.py:MultiIndexUIntEngine._codes_to_ints", "pandas/core/indexes/multi.py:MultiIndexPyIntEngine._codes_to_ints" ]
4
196
rapidsai/dask-cuda
rapidsai__dask-cuda-98
f4ffc5782330ac1d83d196cba78142ffd761fe1c
null
diff --git a/dask_cuda/device_host_file.py b/dask_cuda/device_host_file.py index 02c826f7..ac0eff00 100644 --- a/dask_cuda/device_host_file.py +++ b/dask_cuda/device_host_file.py @@ -1,28 +1,94 @@ from zict import Buffer, File, Func from zict.common import ZictBase -from distributed.protocol import deserialize_bytes,...
diff --git a/dask_cuda/tests/test_device_host_file.py b/dask_cuda/tests/test_device_host_file.py index d53cbc3c..4562a00e 100644 --- a/dask_cuda/tests/test_device_host_file.py +++ b/dask_cuda/tests/test_device_host_file.py @@ -1,10 +1,11 @@ import numpy as np import cupy -from dask_cuda.device_host_file import Device...
Spill quickly from device to host I'm curious how fast we are at moving data back and forth between device and host memory. I suspect that currently we end up serializing with standard Dask serialization, which may not be aware of device memory. We should take a look at using the `"cuda"` serialization family within ...
Yes, I haven't done any profiling, and I'm sure this is something that could be improved. Thanks for pointing me to the CUDA serialization, I wasn't aware of that specialization. If the is not of utmost urgency, I can take a look at this next week, this week I'm mostly busy with EuroPython presentation and at the co...
1,564,067,150,000
[]
Performance Issue
[ "dask_cuda/device_host_file.py:_serialize_if_device", "dask_cuda/device_host_file.py:_deserialize_if_device", "dask_cuda/device_host_file.py:DeviceHostFile.__init__", "dask_cuda/device_host_file.py:DeviceHostFile.__getitem__", "dask_cuda/device_host_file.py:DeviceHostFile.__setitem__", "dask_cuda/device_h...
[ "dask_cuda/device_host_file.py:register_numba", "dask_cuda/device_host_file.py:DeviceSerialized.__init__", "dask_cuda/device_host_file.py:DeviceSerialized.__sizeof__", "dask_cuda/device_host_file.py:_", "dask_cuda/device_host_file.py:device_to_host", "dask_cuda/device_host_file.py:host_to_device" ]
6
197
django/django
django__django-6478
84c1826ded17b2d74f66717fb745fc36e37949fd
null
diff --git a/django/db/backends/oracle/compiler.py b/django/db/backends/oracle/compiler.py index 3ae567669f1b..9aa4acc0fe57 100644 --- a/django/db/backends/oracle/compiler.py +++ b/django/db/backends/oracle/compiler.py @@ -31,10 +31,17 @@ def as_sql(self, with_limits=True, with_col_aliases=False): high_whe...
diff --git a/tests/expressions/tests.py b/tests/expressions/tests.py index 18d003d57d74..8399e3c0a92c 100644 --- a/tests/expressions/tests.py +++ b/tests/expressions/tests.py @@ -12,8 +12,8 @@ Avg, Count, Max, Min, StdDev, Sum, Variance, ) from django.db.models.expressions import ( - Case, Col, ExpressionWrap...
Allow using a subquery in QuerySet.filter() Description (last modified by MikiSoft) The following function is used for filtering by generic relation (and also by one column in the model where it is) which isn't natively supported by Django. APP_LABEL = os.path.basename(os.path.dirname(__file__)) def generic_rel_fi...
1,461,135,635,000
[]
Feature Request
[ "django/db/backends/oracle/compiler.py:SQLCompiler.as_sql" ]
[ "django/db/models/expressions.py:ResolvedOuterRef.as_sql", "django/db/models/expressions.py:ResolvedOuterRef._prepare", "django/db/models/expressions.py:OuterRef.resolve_expression", "django/db/models/expressions.py:OuterRef._prepare", "django/db/models/expressions.py:Subquery.__init__", "django/db/models...
1
198
oppia/oppia
oppia__oppia-14784
16da6ea6778a7b54b601eb6b6a73abff28855766
null
diff --git a/core/domain/opportunity_services.py b/core/domain/opportunity_services.py index 71a8c378ca4c..5a0be0a9da4c 100644 --- a/core/domain/opportunity_services.py +++ b/core/domain/opportunity_services.py @@ -18,6 +18,7 @@ from __future__ import annotations +import collections import logging from core.co...
diff --git a/core/domain/opportunity_services_test.py b/core/domain/opportunity_services_test.py index e09248ae769b..680026817561 100644 --- a/core/domain/opportunity_services_test.py +++ b/core/domain/opportunity_services_test.py @@ -191,35 +191,33 @@ def test_get_translation_opportunities_with_translations_in_review(...
Slow load times for translation Opportunities <!-- - Thanks for taking the time to report a bug in the Oppia project. - Before filing a new issue, please do a quick search to check that it hasn't - already been filed on the [issue tracker](https://github.com/oppia/oppia/issues)._ --> **Describe the bug**...
Seeing the same issue for displaying reviewable suggestions, which is probably due to fetching all available reviewable suggestions up to the query limit of 1000 in core/storage/suggestion/gae_models.py. @sagangwee Hmm, It's strange when i tested this change I had about 200 suggestions, and it seemed to display pretty...
1,643,236,262,000
[ "PR: LGTM" ]
Performance Issue
[ "core/domain/opportunity_services.py:get_exp_opportunity_summary_with_in_review_translations_from_model", "core/domain/opportunity_services.py:get_translation_opportunities" ]
[ "core/domain/opportunity_services.py:_build_exp_id_to_translation_suggestion_in_review_count", "core/domain/suggestion_services.py:get_translation_suggestions_in_review_by_exp_ids", "core/storage/suggestion/gae_models.py:GeneralSuggestionModel.get_in_review_translation_suggestions_by_exp_ids" ]
2
199
modin-project/modin
modin-project__modin-3438
a43d158ea71e332380f94dad2ed02f1ded870382
null
diff --git a/docs/troubleshooting.rst b/docs/troubleshooting.rst index d0cd7d6f7e2..432cb095537 100644 --- a/docs/troubleshooting.rst +++ b/docs/troubleshooting.rst @@ -72,5 +72,127 @@ Restart your interpreter or notebook kernel. Avoid starting many Modin notebooks or interpreters in quick succession. Wait 2-3 secon...
Warn user about heterogeneous data presence during `read_csv` call and propose workaround Subpart of https://github.com/modin-project/modin/issues/3278.
1,631,877,510,000
[ "Ready for review" ]
Feature Request
[ "modin/backends/pandas/parsers.py:PandasParser.get_dtypes" ]
[]
1
200
freedomofpress/securedrop-client
freedomofpress__securedrop-client-944
bd94b7c2d0e6a4da01504b532e91cbb07716b333
null
diff --git a/securedrop_client/gui/widgets.py b/securedrop_client/gui/widgets.py index 21209e303..ccbe4d443 100644 --- a/securedrop_client/gui/widgets.py +++ b/securedrop_client/gui/widgets.py @@ -709,10 +709,15 @@ def show_sources(self, sources: List[Source]): self.empty_conversation_view.show_no_sources_...
diff --git a/tests/functional/test_download_file.py b/tests/functional/test_download_file.py index cba5409fe..9a5e02d98 100644 --- a/tests/functional/test_download_file.py +++ b/tests/functional/test_download_file.py @@ -28,7 +28,7 @@ def check_for_sources(): qtbot.waitUntil(check_for_sources, timeout=10000) ...
Client starts with blacked out / blocked UI with many sources Steps to reproduce: 1. Create a server with 1000 source (`$ NUM_SOURCES=1000 make dev`) 2. Start the client and let it download the sources. 3. Stop the client. 4. Restart the client. At step 4 the drawing of the UI appears to be blocked (I speculat...
1,584,454,550,000
[ "release blocker" ]
Performance Issue
[ "securedrop_client/gui/widgets.py:MainView.show_sources", "securedrop_client/logic.py:Controller.update_sources", "securedrop_client/storage.py:get_local_sources" ]
[ "securedrop_client/gui/widgets.py:SourceList.initial_update", "securedrop_client/gui/widgets.py:SourceList.add_source" ]
3
201
modin-project/modin
modin-project__modin-4391
4ec7f6347903f9133c65ebc5b6e0e15553b98577
null
diff --git a/docs/release_notes/release_notes-0.16.0.rst b/docs/release_notes/release_notes-0.16.0.rst index 1d1adaf0105..372c9c89cad 100644 --- a/docs/release_notes/release_notes-0.16.0.rst +++ b/docs/release_notes/release_notes-0.16.0.rst @@ -7,8 +7,9 @@ Key Features and Updates * Stability and Bugfixes * FIX-#...
diff --git a/modin/pandas/test/dataframe/test_binary.py b/modin/pandas/test/dataframe/test_binary.py index 6018bf6da7b..c338bdbd030 100644 --- a/modin/pandas/test/dataframe/test_binary.py +++ b/modin/pandas/test/dataframe/test_binary.py @@ -249,3 +249,40 @@ def test_duplicate_indexes(): modin_df2, pandas_df2 = cre...
Binary operations in Modin is slower than in pandas in case operands are Modin structures <details><summary>Benchmark code</summary> ```python import time import numpy as np import modin.pandas as pd # import pandas as pd import modin.config as cfg cfg.BenchmarkMode.put(True) # cfg.NPartitions.put(1) ...
Connected with https://github.com/modin-project/modin/issues/3100
1,650,024,288,000
[]
Performance Issue
[ "modin/core/dataframe/pandas/dataframe/dataframe.py:PandasDataframe.binary_op", "modin/core/dataframe/pandas/partitioning/partition_manager.py:PandasDataframePartitionManager.binary_operation", "modin/core/execution/ray/implementations/pandas_on_ray/partitioning/partition_manager.py:PandasOnRayDataframePartitio...
[]
3
202
pandas-dev/pandas
pandas-dev__pandas-20796
fc7bc3f74d1f5702bf66c519ad190126538a8f5b
null
diff --git a/asv_bench/benchmarks/frame_methods.py b/asv_bench/benchmarks/frame_methods.py index 3c0dd646aa502..ba2e63c20d3f8 100644 --- a/asv_bench/benchmarks/frame_methods.py +++ b/asv_bench/benchmarks/frame_methods.py @@ -103,6 +103,7 @@ def setup(self): self.df2 = DataFrame(np.random.randn(N * 50, 10)) ...
Make itertuples really an iterator/generator in implementation, not just return type `itertuples` is not really an iterator/generator and constructs a copy of whole DataFrame in memory. Ideally it would return just an iterator and construct row by row as it is being iterated over.
looks like a generator to me ``` In [1]: df = pd.DataFrame({'A': range(3), 'B': list('ABC')}) In [2]: df.itertuples() Out[2]: <map at 0x10922c080> In [3]: list(df.itertuples()) Out[3]: [Pandas(Index=0, A=0, B='A'), Pandas(Index=1, A=1, B='B'), Pandas(Index=2, A=2, B='C')] In [5]: i = df.itertuples(...
1,524,495,150,000
[ "Performance" ]
Performance Issue
[ "asv_bench/benchmarks/frame_methods.py:Iteration.setup", "asv_bench/benchmarks/frame_methods.py:Iteration.time_itertuples", "pandas/core/base.py:IndexOpsMixin.__iter__", "pandas/core/frame.py:DataFrame.itertuples" ]
[ "asv_bench/benchmarks/frame_methods.py:Iteration.time_itertuples_start", "asv_bench/benchmarks/frame_methods.py:Iteration.time_itertuples_read_first", "asv_bench/benchmarks/frame_methods.py:Iteration.time_itertuples_to_list", "asv_bench/benchmarks/frame_methods.py:Iteration.mem_itertuples_start", "asv_bench...
4
203
azavea/raster-vision
azavea__raster-vision-676
b5321235b9ebde6c9f8cc1afb84555c56209b2ec
null
diff --git a/rastervision/data/vector_source/geojson_vector_source.py b/rastervision/data/vector_source/geojson_vector_source.py index 117e1703a..5c0c2ed1f 100644 --- a/rastervision/data/vector_source/geojson_vector_source.py +++ b/rastervision/data/vector_source/geojson_vector_source.py @@ -16,4 +16,5 @@ def __init__(...
diff --git a/tests/data-files/vector_tiles/lv-mbtiles.json b/tests/data-files/vector_tiles/lv-mbtiles.json index 8aaed5571..48c9c2562 100644 --- a/tests/data-files/vector_tiles/lv-mbtiles.json +++ b/tests/data-files/vector_tiles/lv-mbtiles.json @@ -1,1 +1,1 @@ -{"type": "FeatureCollection", "features": [{"type": "Featu...
Improve vector tile performance There are a few ways to improve vector tile performance if needed: - [x] Download vector tile files to cache - [x] Filter features before doing merge and crop step - [x] In-memory cache (ideally LRU)
For each tile we run `tippecanoe-decode`. Unfortunately, running it on the USA QA tiles which is 5G takes two minutes which is too slow for doing on each scene. For example, this took 2 mins and produced a 12MB GeoJSON file. `tippecanoe-decode -f -c /opt/data/united_states_of_america.mbtiles 12 737 1606 > usa12.js...
1,548,971,768,000
[]
Performance Issue
[ "rastervision/data/vector_source/geojson_vector_source.py:GeoJSONVectorSource._get_geojson", "rastervision/data/vector_source/vector_source.py:VectorSource.get_geojson", "rastervision/data/vector_source/vector_tile_vector_source.py:process_features", "rastervision/data/vector_source/vector_tile_vector_source....
[ "rastervision/data/vector_source/vector_tile_vector_source.py:merge_geojson", "rastervision/data/vector_source/vector_tile_vector_source.py:get_tile_features" ]
5
204
dask/dask
dask__dask-7183
adb58fa7949eb673d49c45a7215e19deaaa34217
null
diff --git a/dask/array/routines.py b/dask/array/routines.py index b250e3ff44a..29542f431ce 100644 --- a/dask/array/routines.py +++ b/dask/array/routines.py @@ -620,7 +620,10 @@ def gradient(f, *varargs, **kwargs): return results -def _bincount_sum(bincounts, dtype=int): +def _bincount_agg(bincounts, dtype, **...
diff --git a/dask/array/tests/test_routines.py b/dask/array/tests/test_routines.py index 91139b83aab..3cfc023ef7b 100644 --- a/dask/array/tests/test_routines.py +++ b/dask/array/tests/test_routines.py @@ -535,10 +535,16 @@ def test_bincount(): assert da.bincount(d, minlength=6).name == da.bincount(d, minlength=6)....
Use tree reduction in `da.bincount` Currently `da.bincount` does a [reduction over `np.bincount`s from each chunk]( https://github.com/dask/dask/blob/2f308e72a46de818055819108429e8f7ce7ba03b/dask/array/routines.py#L549 ). It would be nice to change this to use [a tree reduction]( https://github.com/dask/dask/blob/maste...
This could be a fun task if it is interesting to you, @madsbk. Would provide an opportunity to learn about how reductions work in Dask and how to construct task graphs. How important is it to do a tree reduction here? I understand the value of tree reductions generally, I'm mostly just curious about when this becomes ...
1,612,557,269,000
[]
Feature Request
[ "dask/array/routines.py:_bincount_sum", "dask/array/routines.py:bincount" ]
[ "dask/array/routines.py:_bincount_agg" ]
2
205
zarr-developers/zarr-python
zarr-developers__zarr-python-738
7c31f041ca8baab860564d9c6332c52f19777e80
null
diff --git a/docs/release.rst b/docs/release.rst index 0810946ee6..e9c592a860 100644 --- a/docs/release.rst +++ b/docs/release.rst @@ -12,6 +12,9 @@ Enhancements * array indexing with [] (getitem and setitem) now supports fancy indexing. By :user:`Juan Nunez-Iglesias <jni>`; :issue:`725`. +* write_empty_chunks=Fa...
diff --git a/zarr/tests/test_core.py b/zarr/tests/test_core.py index be2feffe8a..4544a6cae9 100644 --- a/zarr/tests/test_core.py +++ b/zarr/tests/test_core.py @@ -86,9 +86,10 @@ def create_array(self, read_only=False, **kwargs): kwargs.setdefault('compressor', Zlib(level=1)) cache_metadata = kwargs.po...
Optimize storing fill_value to Array Would be nice if when storing a scalar or an array that contains only the `fill_value` of the `Array`, the chunk is merely deleted (if it exists) instead of writing out a chunk that contains only the `fill_value` for all elements. The `Array` should behave identically in either case...
1,620,766,610,000
[]
Performance Issue
[ "zarr/core.py:Array.__init__", "zarr/core.py:Array._set_basic_selection_zd", "zarr/core.py:Array._chunk_setitems", "zarr/core.py:Array._chunk_setitem_nosync", "zarr/core.py:Array._process_for_setitem", "zarr/core.py:Array.__getstate__", "zarr/creation.py:create", "zarr/creation.py:open_array" ]
[ "zarr/core.py:Array.write_empty_chunks", "zarr/core.py:Array._chunk_delitems", "zarr/core.py:Array._chunk_delitem", "zarr/storage.py:FSStore.delitems", "zarr/util.py:flatten", "zarr/util.py:all_equal" ]
8
206
paperless-ngx/paperless-ngx
paperless-ngx__paperless-ngx-1227
d8e3d91a79f0373f4b68464af0487679fdad6c8c
null
diff --git a/docs/configuration.rst b/docs/configuration.rst index a5db55927d9..c4203472c77 100644 --- a/docs/configuration.rst +++ b/docs/configuration.rst @@ -536,6 +536,8 @@ PAPERLESS_TASK_WORKERS=<num> maintain the automatic matching algorithm, check emails, consume documents, etc. This variable specifies...
diff --git a/src/documents/tests/test_settings.py b/src/documents/tests/test_settings.py deleted file mode 100644 index 9b8edab2790..00000000000 --- a/src/documents/tests/test_settings.py +++ /dev/null @@ -1,35 +0,0 @@ -import logging -from unittest import mock - -from django.test import TestCase -from paperless.settin...
Reduce PAPERLESS_TASK_WORKERS and PAPERLESS_WEBSERVER_WORKERS default values ### Description # Problem with memory and CPU usage In connection with #1084, reducing the `PAPERLESS_TASK_WORKERS` and `PAPERLESS_WEBSERVER_WORKERS` default values to `1` helps not only with reducing the memory footprint (from approx. 4...
Obviously appreciate all your thoughts, why is this a separate issue from #1084? @shamoon , this helps both with idle CPU consumption, and memory reduction as well; however, the most drastic effect is on memory. Sometimes people are funny about changing defaults, so I prefer to have a separate thread/bug/conversation ...
1,657,573,686,000
[ "documentation", "non-trivial", "backend" ]
Performance Issue
[ "src/paperless/settings.py:default_task_workers" ]
[]
1
207
scikit-learn/scikit-learn
scikit-learn__scikit-learn-13290
3f25ea034cca6fa719dec024e0cbba608bce07e6
null
diff --git a/sklearn/preprocessing/data.py b/sklearn/preprocessing/data.py index 823eedc8b7dd9..de33e9a1441a0 100644 --- a/sklearn/preprocessing/data.py +++ b/sklearn/preprocessing/data.py @@ -1526,10 +1526,10 @@ def transform(self, X): elif sparse.isspmatrix_csc(X) and self.degree < 4: return sel...
Faster PolynomialFeatures The current implementation of PolynomialFeatures could be faster. The gain is significant for matrices with less than 10.000 rows, twice faster for less than 100 rows. This is for dense matrices, not measured for sparse. #### Description PolynomialFeatures independently computes every feat...
Looks interesting. Could you please open a pull request and join some benchmark results in the description / discussion? Hi ! I am in the sprint and I am taking this up. I implemented "interaction_only" and moved the code here of my experiments https://github.com/sdpython/mlinsights/blob/master/src/mlinsights/mlmodel/_...
1,551,201,047,000
[ "module:preprocessing" ]
Performance Issue
[ "sklearn/preprocessing/data.py:PolynomialFeatures.transform" ]
[]
1
208
TransformerLensOrg/TransformerLens
TransformerLensOrg__TransformerLens-333
3cd943628b5c415585c8ef100f65989f6adc7f75
null
diff --git a/transformer_lens/HookedEncoder.py b/transformer_lens/HookedEncoder.py index e470731d3..254a72a10 100644 --- a/transformer_lens/HookedEncoder.py +++ b/transformer_lens/HookedEncoder.py @@ -229,7 +229,7 @@ def from_pretrained( official_model_name, cfg, hf_model, **from_pretrained_kwargs ...
diff --git a/tests/acceptance/test_hooked_transformer.py b/tests/acceptance/test_hooked_transformer.py index 85d4002b5..f708b0301 100644 --- a/tests/acceptance/test_hooked_transformer.py +++ b/tests/acceptance/test_hooked_transformer.py @@ -6,6 +6,7 @@ from transformers import AutoConfig from transformer_lens impor...
[Bug Report] `refactor_factored_attn_matrices` uses lots of GPU memory **Describe the bug** When loading models with `refactor_factored_attn_matrices=True` and `device="cuda"`, they take up a lot of space on GPU. Thanks to @callummcdougall for noticing this. See example code at bottom for snippets and statistics. *...
Is `refactor_factored_attn_matrices` really the issue here? As far as I can tell, using `from_pretrained` to load models to the GPU directly always takes up much more GPU memory than first loading them to the CPU and then moving them to the GPU. When I use your snippet, setting `refactor_factored_attn_matrices=True` on...
1,688,268,522,000
[]
Performance Issue
[ "transformer_lens/HookedEncoder.py:HookedEncoder.from_pretrained", "transformer_lens/HookedTransformer.py:HookedTransformer.__init__", "transformer_lens/HookedTransformer.py:HookedTransformer.move_model_modules_to_device", "transformer_lens/HookedTransformer.py:HookedTransformer.from_pretrained", "transform...
[]
7
209
NatLibFi/Annif
NatLibFi__Annif-610
3fd22027dc4d7fa40b2bd9a19f44822fca7f125f
null
diff --git a/annif/project.py b/annif/project.py index 13f470fcb..05fa7353f 100644 --- a/annif/project.py +++ b/annif/project.py @@ -9,7 +9,6 @@ import annif.corpus import annif.suggestion import annif.backend -import annif.vocab from annif.datadir import DatadirMixin from annif.exception import AnnifException, Co...
diff --git a/tests/test_vocab.py b/tests/test_vocab.py index 68f0cb983..02225ce16 100644 --- a/tests/test_vocab.py +++ b/tests/test_vocab.py @@ -19,9 +19,9 @@ def load_dummy_vocab(tmpdir): return vocab -def test_get_vocab_invalid(): +def test_get_vocab_invalid(registry): with pytest.raises(ValueError) as ...
Share vocabulary objects between projects Currently, each AnnifProject creates its own vocabulary (AnnifVocabulary object, which wraps a SubjectIndex) even though in reality many projects may use the same vocabulary - which is now even more likely since vocabularies are multilingual (#600). This leads to unnecessary us...
1,660,740,945,000
[]
Performance Issue
[ "annif/project.py:AnnifProject.vocab", "annif/registry.py:AnnifRegistry.__init__", "annif/registry.py:AnnifRegistry._create_projects", "annif/vocab.py:get_vocab" ]
[ "annif/registry.py:AnnifRegistry.get_vocab" ]
4
210
alexa-pi/AlexaPi
alexa-pi__AlexaPi-188
81d04a41212a5b6040a402c6f84907a16f5b186b
null
diff --git a/CHANGELOG.md b/CHANGELOG.md index 80fd89a6..7410c952 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -7,6 +7,10 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/). ### Changed - On Debian-based systems, the `python-pip` package gets uninstalled and `pip` is installed via `easy_ins...
Streaming audio to and from AVS While waiting for Echo push notifications for a year now, I did a small extension of https://github.com/sammachin/AlexaPi to essentially enable push by submitting pre-recorded questions to AVS when triggered by an event. It worked pretty well, but the problem I faced is that response tim...
Sounds interesting. We're not there yet to do these _future_ things - we need to refactor the AVS code first, but we ain't that far either. Let's keep it here for when we get there. Also, I'm not sure that streaming **from** AVS is gonna have significant performance impact, because Alexa's answers are usually very sm...
1,488,647,139,000
[ "enhancement", "audio" ]
Performance Issue
[ "src/alexapi/capture.py:Capture.silence_listener", "src/main.py:alexa_speech_recognizer", "src/main.py:process_response", "src/main.py:trigger_process" ]
[ "src/main.py:alexa_speech_recognizer_generate_data" ]
4
211
feast-dev/feast
feast-dev__feast-2371
45db6dcdf064688c55a7a465aa6717887e6cae34
null
diff --git a/sdk/python/feast/infra/online_stores/dynamodb.py b/sdk/python/feast/infra/online_stores/dynamodb.py index 86a96239bb..c161a5b955 100644 --- a/sdk/python/feast/infra/online_stores/dynamodb.py +++ b/sdk/python/feast/infra/online_stores/dynamodb.py @@ -11,6 +11,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY ...
diff --git a/sdk/python/tests/unit/online_store/test_dynamodb_online_store.py b/sdk/python/tests/unit/online_store/test_dynamodb_online_store.py new file mode 100644 index 0000000000..0f42230ef5 --- /dev/null +++ b/sdk/python/tests/unit/online_store/test_dynamodb_online_store.py @@ -0,0 +1,57 @@ +from dataclasses impor...
DynamoDB batch retrieval does slow sequential gets (online_read) The current DynamoDB implementation does sequential gets (https://github.com/feast-dev/feast/blob/master/sdk/python/feast/infra/online_stores/dynamodb.py#L163) ## Possible Solution A better approach is to do some multi-get operation or at least run th...
cc @vlin-lgtm @adchia, May I know what is needed to be done. I would like to contribute to this issue. > @adchia, May I know what is needed to be done. I would like to contribute to this issue. It appears we have the list of primary keys (entity_ids), so we can switch to [BatchGetItem](https://docs.aws.amazon.com/a...
1,646,533,215,000
[ "size/L", "lgtm", "approved", "ok-to-test" ]
Performance Issue
[ "sdk/python/feast/infra/online_stores/dynamodb.py:DynamoDBOnlineStore.online_read" ]
[ "sdk/python/feast/infra/online_stores/dynamodb.py:DynamoDBOnlineStore._sort_dynamodb_response" ]
1
212
huggingface/transformers
huggingface__transformers-29900
31c575bcf13c2b85b65d652dd1b5b401f99be999
null
diff --git a/src/transformers/models/whisper/feature_extraction_whisper.py b/src/transformers/models/whisper/feature_extraction_whisper.py index 42104c329346..508e85b91ffd 100644 --- a/src/transformers/models/whisper/feature_extraction_whisper.py +++ b/src/transformers/models/whisper/feature_extraction_whisper.py @@ -9...
diff --git a/tests/models/whisper/test_feature_extraction_whisper.py b/tests/models/whisper/test_feature_extraction_whisper.py index 77c7a9be3de6..8b1e25927e33 100644 --- a/tests/models/whisper/test_feature_extraction_whisper.py +++ b/tests/models/whisper/test_feature_extraction_whisper.py @@ -24,7 +24,7 @@ from datas...
[Whisper] Computing features on CPU slows down WhisperFeatureExtractor ### Feature request In the WhisperFeatureExtractor class, to reduce the time it takes to compute audio features, an option should be added to compute the features on GPU in batch mode. ### Motivation Currently, the Whisper Feature Extractor...
1,711,536,441,000
[]
Feature Request
[ "src/transformers/models/whisper/feature_extraction_whisper.py:WhisperFeatureExtractor._np_extract_fbank_features", "src/transformers/models/whisper/feature_extraction_whisper.py:WhisperFeatureExtractor._torch_extract_fbank_features", "src/transformers/models/whisper/feature_extraction_whisper.py:WhisperFeature...
[]
3
214
cisagov/manage.get.gov
cisagov__manage.get.gov-1669
e84941dc935588212146e7f32f88236728d24e2c
null
diff --git a/src/registrar/management/commands/utility/extra_transition_domain_helper.py b/src/registrar/management/commands/utility/extra_transition_domain_helper.py index 755c9b98a..c082552eb 100644 --- a/src/registrar/management/commands/utility/extra_transition_domain_helper.py +++ b/src/registrar/management/comman...
Exporting any domain report (full metadata, current-full, -federal) takes a long time, 401s, and only generates on the second attempt ### Current Behavior See title. ### Expected Behavior Generating one of these reports shouldn't take multiple minutes, throw an error, or require a retry. ### Steps to Reprod...
1,705,961,430,000
[]
Performance Issue
[ "src/registrar/management/commands/utility/extra_transition_domain_helper.py:LoadExtraTransitionDomain.update_transition_domain_models", "src/registrar/models/domain.py:Domain.get_security_email", "src/registrar/utility/csv_export.py:get_domain_infos", "src/registrar/utility/csv_export.py:write_row", "src/r...
[ "src/registrar/utility/csv_export.py:parse_row" ]
5
215
Agenta-AI/agenta
Agenta-AI__agenta-1639
bd8167ce73d0ca31e1d6397906d4ecb63d6252c3
null
diff --git a/agenta-backend/agenta_backend/models/converters.py b/agenta-backend/agenta_backend/models/converters.py index 541462c3b0..de7c6dabfb 100644 --- a/agenta-backend/agenta_backend/models/converters.py +++ b/agenta-backend/agenta_backend/models/converters.py @@ -359,6 +359,7 @@ async def environment_db_to_outpu...
[AGE-164] [Improvement] Speed up the endpoint /get_config The call to retrieve the configuration takes 1.20 seconds, indicating that 'get_config' is highly unoptimised. Here's the current process: * It fetches the base_id. * It checks permissions. * It lists the environment for the app_id, which is obtained from the ...
1,715,513,255,000
[ "Backend", "improvement", "size:M", "lgtm" ]
Performance Issue
[ "agenta-backend/agenta_backend/models/converters.py:environment_db_to_output", "agenta-backend/agenta_backend/routers/configs_router.py:get_config", "agenta-backend/agenta_backend/services/db_manager.py:fetch_base_by_id", "agenta-backend/agenta_backend/services/db_manager.py:list_variants_for_base", "agenta...
[]
5
216
matrix-org/synapse
matrix-org__synapse-8744
03e392f787df3fb8d23aca0964695a5caa075940
null
diff --git a/changelog.d/8744.bugfix b/changelog.d/8744.bugfix new file mode 100644 index 000000000000..f8f9630bd6af --- /dev/null +++ b/changelog.d/8744.bugfix @@ -0,0 +1,1 @@ +Fix a bug where appservices may be sent an excessive amount of read receipts and presence. Broke in v1.22.0. diff --git a/synapse/handlers/ap...
When enabling EDU support for appservices, synapse will start from stream position 0 when fetching presence/receipts Which is very expensive if you were say, matrix.org. If the appservice hasn't got a stream position in the database, we should go from a reasonable point in time.
1,605,111,155,000
[]
Performance Issue
[ "synapse/handlers/appservice.py:ApplicationServicesHandler._notify_interested_services_ephemeral", "synapse/storage/databases/main/receipts.py:ReceiptsWorkerStore.get_linearized_receipts_for_all_rooms" ]
[]
2
217
NVIDIA/NeMo
NVIDIA__NeMo-9100
e316b0e50432ea27fe86eb008524845b9cd4631f
null
diff --git a/nemo/collections/asr/parts/submodules/ctc_decoding.py b/nemo/collections/asr/parts/submodules/ctc_decoding.py index 67559eccf6e2..70d63c0f8c6f 100644 --- a/nemo/collections/asr/parts/submodules/ctc_decoding.py +++ b/nemo/collections/asr/parts/submodules/ctc_decoding.py @@ -213,20 +213,20 @@ def __init__(se...
diff --git a/tests/collections/asr/decoding/test_ctc_decoding.py b/tests/collections/asr/decoding/test_ctc_decoding.py index a3a5689062bf..8eceb822fd38 100644 --- a/tests/collections/asr/decoding/test_ctc_decoding.py +++ b/tests/collections/asr/decoding/test_ctc_decoding.py @@ -26,6 +26,7 @@ CTCDecoding, CTCD...
Consider refactoring CTC greedy decoding **Is your feature request related to a problem? Please describe.** As part of my work on #8673 , I have been profiling to make sure we have low latency and high throughput. I have observed that this line can sometimes take https://github.com/NVIDIA/NeMo/blob/57444ae9b63afc541...
@artbataev this may be related to your observation of TDT being faster than CTC in some cases. If I recall correctly, max does work on GPU so I don't know why I forced it onto CPU. Maybe the slice? Either way, if it fixes the slowdown let's move the .CPU() to after max calculation. It might have been that the logpro...
1,714,689,922,000
[ "ASR", "Run CICD" ]
Performance Issue
[ "nemo/collections/asr/parts/submodules/ctc_decoding.py:AbstractCTCDecoding.__init__" ]
[ "nemo/collections/asr/parts/submodules/ctc_greedy_decoding.py:GreedyBatchedCTCInfer.input_types", "nemo/collections/asr/parts/submodules/ctc_greedy_decoding.py:GreedyBatchedCTCInfer.output_types", "nemo/collections/asr/parts/submodules/ctc_greedy_decoding.py:GreedyBatchedCTCInfer.__init__", "nemo/collections/...
1
218
streamlit/streamlit
streamlit__streamlit-9754
8cfd31614446fda361aab1d007c97acb5ffcc12a
null
diff --git a/lib/streamlit/web/server/server.py b/lib/streamlit/web/server/server.py index 027db4ab064a..5b3d71f4571f 100644 --- a/lib/streamlit/web/server/server.py +++ b/lib/streamlit/web/server/server.py @@ -369,7 +369,12 @@ def _create_app(self) -> tornado.web.Application: routes.extend( ...
diff --git a/lib/tests/streamlit/web/server/routes_test.py b/lib/tests/streamlit/web/server/routes_test.py index a4f7a62781fe..b47c678c16a7 100644 --- a/lib/tests/streamlit/web/server/routes_test.py +++ b/lib/tests/streamlit/web/server/routes_test.py @@ -196,7 +196,7 @@ def get_app(self): return tornado.web.Ap...
Undesired host redirect with trailing slash and host path ### Checklist - [X] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues. - [X] I added a very descriptive title to this issue. - [X] I have provided sufficient information below to help reproduce this issue. ...
**If this issue affects you, please react with a 👍 (thumbs up emoji) to the initial post.** Your feedback helps us prioritize which bugs to investigate and address first. ![Visits](https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2Fstreamlit%2Fstreamlit%2Fissues%2F9690&title=visits&e...
1,730,317,819,000
[ "security-assessment-completed", "change:bugfix", "impact:users" ]
Security Vulnerability
[ "lib/streamlit/web/server/server.py:Server._create_app" ]
[]
1
219
streamlit/streamlit
streamlit__streamlit-9472
4c45934065ea2c0acea32fdb3b4d51fc4fb00620
null
diff --git a/lib/streamlit/runtime/state/session_state.py b/lib/streamlit/runtime/state/session_state.py index 5481cd2be27a..b9f016e19ad7 100644 --- a/lib/streamlit/runtime/state/session_state.py +++ b/lib/streamlit/runtime/state/session_state.py @@ -280,6 +280,56 @@ def _missing_key_error_message(key: str) -> str: ...
diff --git a/lib/tests/streamlit/runtime/state/session_state_test.py b/lib/tests/streamlit/runtime/state/session_state_test.py index a9408a209a9b..6eac40a58a64 100644 --- a/lib/tests/streamlit/runtime/state/session_state_test.py +++ b/lib/tests/streamlit/runtime/state/session_state_test.py @@ -38,6 +38,7 @@ from strea...
Exponential slowdown in Streamlit as the number of widgets increases (starting in v1.34) ### Checklist - [X] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues. - [X] I added a very descriptive title to this issue. - [X] I have provided sufficient information be...
**If this issue affects you, please react with a 👍 (thumbs up emoji) to the initial post.** Your feedback helps us prioritize which bugs to investigate and address first. ![Visits](https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2Fstreamlit%2Fstreamlit%2Fissues%2F9415&title=visits&e...
1,726,478,147,000
[ "security-assessment-completed", "impact:internal", "change:refactor" ]
Performance Issue
[ "lib/streamlit/runtime/state/session_state.py:SessionState.clear", "lib/streamlit/runtime/state/session_state.py:SessionState.filtered_state", "lib/streamlit/runtime/state/session_state.py:SessionState._reverse_key_wid_map", "lib/streamlit/runtime/state/session_state.py:SessionState.__getitem__", "lib/strea...
[ "lib/streamlit/runtime/state/session_state.py:KeyIdMapper.__contains__", "lib/streamlit/runtime/state/session_state.py:KeyIdMapper.__setitem__", "lib/streamlit/runtime/state/session_state.py:KeyIdMapper.__delitem__", "lib/streamlit/runtime/state/session_state.py:KeyIdMapper.id_key_mapping", "lib/streamlit/r...
8
220
streamlit/streamlit
streamlit__streamlit-8741
ad3dc4e19d261fc66560a25bab92a7a560068338
null
diff --git a/lib/streamlit/runtime/app_session.py b/lib/streamlit/runtime/app_session.py index dc9a13c17511..f68b216929db 100644 --- a/lib/streamlit/runtime/app_session.py +++ b/lib/streamlit/runtime/app_session.py @@ -144,7 +144,8 @@ def __init__( self._stop_config_listener: Callable[[], bool] | None = None ...
diff --git a/lib/tests/streamlit/runtime/app_session_test.py b/lib/tests/streamlit/runtime/app_session_test.py index ba999fc4d967..a94a00d68468 100644 --- a/lib/tests/streamlit/runtime/app_session_test.py +++ b/lib/tests/streamlit/runtime/app_session_test.py @@ -125,13 +125,13 @@ def test_shutdown(self, patched_disconn...
Streamlit unnecessarily runs the file watcher against every path in imported modules, even when watching is disabled ### Checklist - [X] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues. - [X] I added a very descriptive title to this issue. - [X] I have provided s...
**If this issue affects you, please react with a 👍 (thumbs up emoji) to the initial post.** Your feedback helps us prioritize which bugs to investigate and address first. ![Visits](https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2Fstreamlit%2Fstreamlit%2Fissues%2F8738&title=visits&e...
1,716,415,210,000
[ "security-assessment-completed", "change:bugfix", "impact:users" ]
Performance Issue
[ "lib/streamlit/runtime/app_session.py:AppSession.__init__" ]
[]
1
221
mesonbuild/meson
mesonbuild__meson-14013
0b41b364be716064285928c00330d1fc6b07cd88
null
diff --git a/mesonbuild/coredata.py b/mesonbuild/coredata.py index 98656b27d9b0..c3e435271952 100644 --- a/mesonbuild/coredata.py +++ b/mesonbuild/coredata.py @@ -658,9 +658,20 @@ def set_options(self, opts_to_set: T.Dict[OptionKey, T.Any], subproject: str = ' elif k.machine != MachineChoice.BUILD and not ...
diff --git a/test cases/unit/123 pkgsubproj/meson.build b/test cases/unit/123 pkgsubproj/meson.build new file mode 100644 index 000000000000..b4cf89fa0b6d --- /dev/null +++ b/test cases/unit/123 pkgsubproj/meson.build @@ -0,0 +1,3 @@ +project('pkg_opt_test') + +subproject('sub') diff --git a/test cases/unit/123 pkgsubp...
pkgconfig.relocatable in subproject default_options fails in parent project setup **Describe the bug** When setting up a parent project that includes a subproject with `pkgconfig.relocatable=true` in its `default_options`, meson setup fails in the parent project with the error: ``` subprojects/mysubprj/meson.build:1...
Can confirm when using [librsvg](https://gitlab.gnome.org/gnome/librsvg) as a wrap from within the GStreamer repository.
1,734,348,577,000
[]
Bug Report
[ "mesonbuild/coredata.py:CoreData.set_options" ]
[]
1
222
scikit-learn/scikit-learn
scikit-learn__scikit-learn-30327
90a51c12d6159f72ab8990ea1a3ff835861ec81d
46a7c9a5e4fe88dfdfd371bf36477f03498a3390
diff --git a/sklearn/base.py b/sklearn/base.py index d646f8d3e56bf..2c82cf05a6c5a 100644 --- a/sklearn/base.py +++ b/sklearn/base.py @@ -389,6 +389,33 @@ def __setstate__(self, state): except AttributeError: self.__dict__.update(state) + # TODO(1.7): Remove this method + def _more_tags(sel...
diff --git a/sklearn/utils/tests/test_tags.py b/sklearn/utils/tests/test_tags.py index 413fbc6bbd3de..86e4e2d7c431e 100644 --- a/sklearn/utils/tests/test_tags.py +++ b/sklearn/utils/tests/test_tags.py @@ -7,7 +7,16 @@ RegressorMixin, TransformerMixin, ) -from sklearn.utils import Tags, get_tags +from sklearn...
MAINT conversion old->new/new->old tags Towards #30298 This PR provides a way to convert old tags into new tags and new tags into old tags. We should make sure that: - An estimator implementing `_get_tags` or `_more_tags` get a warning but that the check_estimator should be working reasonably - `_safe_tags` sh...
## ✔️ Linting Passed All linting checks passed. Your pull request is in excellent shape! ☀️ <sub> _Generated for commit: [b41d1a6](https://github.com/scikit-learn/scikit-learn/pull/30302/commits/b41d1a6bafa25db5a7dc882ac21a6bdc131d744d). Link to the linter CI: [here](https://github.com/scikit-learn/scikit-learn/action...
1,732,218,511,000
[ "module:utils", "To backport", "No Changelog Needed" ]
Feature Request
[ "sklearn/utils/_tags.py:get_tags" ]
[ "sklearn/base.py:BaseEstimator._more_tags", "sklearn/base.py:BaseEstimator._get_tags", "sklearn/utils/_tags.py:_find_tags_provider", "sklearn/utils/_tags.py:_safe_tags", "sklearn/utils/_tags.py:_to_new_tags", "sklearn/utils/_tags.py:_to_old_tags" ]
1
223
scikit-learn/scikit-learn
scikit-learn__scikit-learn-30128
6bf2061f76ba0977c1f7ffb9ddc48db794e5c7ec
33f08f1c5202cc0bdb6cdc09a92442f977fd2bb3
diff --git a/sklearn/metrics/_regression.py b/sklearn/metrics/_regression.py index 62251f9b96188..c5ebe67e34a2e 100644 --- a/sklearn/metrics/_regression.py +++ b/sklearn/metrics/_regression.py @@ -58,11 +58,16 @@ def _check_reg_targets(y_true, y_pred, multioutput, dtype="numeric", xp=None): """Check that y_true a...
diff --git a/sklearn/metrics/tests/test_common.py b/sklearn/metrics/tests/test_common.py index e6abc8c433013..30722c5eab52f 100644 --- a/sklearn/metrics/tests/test_common.py +++ b/sklearn/metrics/tests/test_common.py @@ -583,8 +583,8 @@ def _require_positive_targets(y1, y2): def _require_log1p_targets(y1, y2): ""...
Reduce redundancy in floating type checks for Array API support ### Describe the workflow you want to enable While working on #29978, we noticed that the following procedure is repeated across most regression metrics in `_regression.py` for the Array API: ```python xp, _ = get_namespace(y_true, y_pred, sample_...
That sounds good to me. Hi @virchan , @adrinjalali Will it be okay if I try to work on this? > Hi @virchan , @adrinjalali > > Will it be okay if I try to work on this? Hello @sqali, This issue is currently affecting another PR I'm working on, so I feel obligated to resolve it myself. I plan to start worki...
1,729,555,735,000
[ "module:metrics", "No Changelog Needed", "Array API" ]
Feature Request
[ "sklearn/metrics/_regression.py:_check_reg_targets", "sklearn/metrics/_regression.py:mean_absolute_error", "sklearn/metrics/_regression.py:mean_absolute_percentage_error", "sklearn/metrics/_regression.py:mean_squared_error", "sklearn/metrics/_regression.py:mean_squared_log_error", "sklearn/metrics/_regres...
[ "sklearn/metrics/_regression.py:_check_reg_targets_with_floating_dtype" ]
9
224
scikit-learn/scikit-learn
scikit-learn__scikit-learn-28678
2cbf3345bce979b261f7007616cc3465f33a51d3
9870b52838091c697aadb6531e77a941a63aa725
diff --git a/sklearn/metrics/_classification.py b/sklearn/metrics/_classification.py index 0175966a21b74..b69855633a730 100644 --- a/sklearn/metrics/_classification.py +++ b/sklearn/metrics/_classification.py @@ -1727,14 +1727,18 @@ def precision_recall_fscore_support( "assigned" 0 samples. For multilabel targ...
diff --git a/sklearn/tests/test_docstring_parameters.py b/sklearn/tests/test_docstring_parameters.py index 3af463b783bc3..ffa4fe63967f9 100644 --- a/sklearn/tests/test_docstring_parameters.py +++ b/sklearn/tests/test_docstring_parameters.py @@ -12,6 +12,7 @@ import pytest import sklearn +from sklearn import metrics...
[WIP] : Added assert_consistent_docs() and related tests Fixes #9388 Added a function to check for consistency between docstring of objects. In this approach there is a python dictionary for each of Parameters, Attributes and Returns. Indexed by the name of the parameter/attribute/return with the value being a...
Do we expect this to be reasonable for everything, or only for parts of the library? something like ``random_state`` might or might not be good to document consistently - saying what the randomness is might be good. Also, we have parameters like ``alpha`` that can mean basically anything Or is the idea to call this wit...
1,711,066,814,000
[ "module:metrics", "module:utils", "No Changelog Needed", "Waiting for Second Reviewer" ]
Feature Request
[ "sklearn/metrics/_classification.py:precision_recall_fscore_support" ]
[]
1
225
PrefectHQ/prefect
PrefectHQ__prefect-16369
3abe9d067b0502e15d8963c8612edc10e221254e
null
diff --git a/src/prefect/tasks.py b/src/prefect/tasks.py index d4e7d2fbce52..0d6610009847 100644 --- a/src/prefect/tasks.py +++ b/src/prefect/tasks.py @@ -474,6 +474,10 @@ def __init__( if callable(retry_delay_seconds): self.retry_delay_seconds = retry_delay_seconds(retries) + elif not is...
diff --git a/tests/test_tasks.py b/tests/test_tasks.py index 88cc1c4d086a..06d41abd6423 100644 --- a/tests/test_tasks.py +++ b/tests/test_tasks.py @@ -4172,6 +4172,21 @@ def my_flow(): class TestTaskConstructorValidation: + async def test_task_cannot_configure_poorly_typed_retry_delay(self): + with pytes...
Missconfigured retry_delay_seconds task argument leads to confusing error message ### Bug summary When the task arg `retry_delay_seconds` is not configured with the correct type, the error message presented is misleading: ```python from datetime import timedelta from prefect import get_run_logger, task, flow @...
1,734,045,413,000
[ "bug" ]
Bug Report
[ "src/prefect/tasks.py:Task.__init__" ]
[]
1
226
PrefectHQ/prefect
PrefectHQ__prefect-16328
b60fc0c3288befd1a052a3743a315374da766ecb
null
diff --git a/src/prefect/task_engine.py b/src/prefect/task_engine.py index a07f2108c95a..d6b834d6bca6 100644 --- a/src/prefect/task_engine.py +++ b/src/prefect/task_engine.py @@ -600,6 +600,7 @@ def handle_timeout(self, exc: TimeoutError) -> None: message=message, name="TimedOut", ...
diff --git a/tests/test_task_engine.py b/tests/test_task_engine.py index 185045fb9455..c4f1847b160c 100644 --- a/tests/test_task_engine.py +++ b/tests/test_task_engine.py @@ -1511,6 +1511,58 @@ async def foo(): assert run.end_time == failed.timestamp assert run.total_run_time == failed.timestamp - run...
`TimedOut` task has no duration or end time ### Bug summary Using the following sample flow to produce a `TimedOut` task and flow, the flow run receives a duration and end time but the task run does not. ```python from prefect import flow, task import requests @task(timeout_seconds=5) def timeout_task(): ...
1,733,861,525,000
[ "bug" ]
Bug Report
[ "src/prefect/task_engine.py:SyncTaskRunEngine.handle_timeout", "src/prefect/task_engine.py:AsyncTaskRunEngine.handle_timeout" ]
[]
2
227
PrefectHQ/prefect
PrefectHQ__prefect-16145
b52c2c9a5d3d3d3e37e09efb73a4764857aa863e
null
diff --git a/src/prefect/task_engine.py b/src/prefect/task_engine.py index 43238d1cbc6c..5accda613f34 100644 --- a/src/prefect/task_engine.py +++ b/src/prefect/task_engine.py @@ -427,6 +427,11 @@ def set_state(self, state: State, force: bool = False) -> State: self.task_run.state = new_state = state + ...
diff --git a/tests/test_task_engine.py b/tests/test_task_engine.py index c6702f6a11de..185045fb9455 100644 --- a/tests/test_task_engine.py +++ b/tests/test_task_engine.py @@ -7,7 +7,7 @@ from pathlib import Path from typing import List, Optional from unittest import mock -from unittest.mock import AsyncMock, MagicMo...
Async tasks stays in Pending state until Completed on Windows Client ### Bug summary This issue is directly related to #15607 . After few hours of trial and error I managed to bring more light to it, at least i hope so. **My findings indicate that Running state had the same timestamp as Pending state.** Could...
1,732,802,886,000
[ "bug" ]
Bug Report
[ "src/prefect/task_engine.py:SyncTaskRunEngine.set_state", "src/prefect/task_engine.py:AsyncTaskRunEngine.set_state" ]
[]
2
228
PrefectHQ/prefect
PrefectHQ__prefect-16117
fb919c67427d230ecea03a5273dfdd6de59bd697
null
diff --git a/src/prefect/flows.py b/src/prefect/flows.py index 27dc6e7c4a8a..cc4f781b6499 100644 --- a/src/prefect/flows.py +++ b/src/prefect/flows.py @@ -10,6 +10,7 @@ import importlib.util import inspect import os +import sys import tempfile import warnings from functools import partial, update_wrapper @@ -137,...
diff --git a/tests/test_flows.py b/tests/test_flows.py index 47bfea76ec4a..097018f97563 100644 --- a/tests/test_flows.py +++ b/tests/test_flows.py @@ -1517,6 +1517,23 @@ def my_flow(secret: SecretStr): "secret": SecretStr("my secret") } + @pytest.mark.skipif( + sys.version_info < (3, 9...
flow.validate_parameters regression ### Bug summary Certain flow deployments that worked fine before cannot be executed anymore after updating from Prefect 2.20.9 to any newer Prefect version. This is due to a change in the parameter validation. MRE: ```python from prefect import flow @flow def my_flow(par...
hi @j-tr - thanks for the issue! can you share what version of `pydantic` you are using? I cannot seem to directly reproduce this ```python In [1]: from prefect import flow ...: ...: @flow ...: def my_flow(param: dict[str, str]): ...: pass ...: ...: flow_run_params = { ...: "par...
1,732,631,512,000
[ "bug", "fix", "2.x" ]
Bug Report
[ "src/prefect/flows.py:Flow.validate_parameters", "src/prefect/flows.py:select_flow" ]
[]
2
229
PrefectHQ/prefect
PrefectHQ__prefect-15909
b26d891d78122801a70827ab46d2fd73dd57ea3d
null
diff --git a/src/integrations/prefect-dask/prefect_dask/task_runners.py b/src/integrations/prefect-dask/prefect_dask/task_runners.py index f4edbab25295..60066faaf870 100644 --- a/src/integrations/prefect-dask/prefect_dask/task_runners.py +++ b/src/integrations/prefect-dask/prefect_dask/task_runners.py @@ -319,7 +319,7 ...
diff --git a/src/integrations/prefect-dask/tests/test_task_runners.py b/src/integrations/prefect-dask/tests/test_task_runners.py index 748de299c65e..d1bc5c8bec0b 100644 --- a/src/integrations/prefect-dask/tests/test_task_runners.py +++ b/src/integrations/prefect-dask/tests/test_task_runners.py @@ -452,3 +452,49 @@ def ...
DaskTaskRunner can not work together with `wait_for` ### Bug summary When using the `DaskTaskRunner` using wait_for is broken. Using the `ThreadPoolTaskRunner` the same flow works just fine. ```py from prefect import flow, task from prefect.task_runners import ThreadPoolTaskRunner from prefect_dask.task_runner...
hi @EraYaN - thank you for the issue! will take a look at this
1,730,486,057,000
[ "bug", "integrations", "fix" ]
Bug Report
[ "src/integrations/prefect-dask/prefect_dask/task_runners.py:DaskTaskRunner.submit" ]
[]
1
230
PrefectHQ/prefect
PrefectHQ__prefect-15468
8ebcefd79839168433fa2080c101221131050e05
null
diff --git a/src/prefect/cli/deploy.py b/src/prefect/cli/deploy.py index cf15c23ab6ee..cfa321482e55 100644 --- a/src/prefect/cli/deploy.py +++ b/src/prefect/cli/deploy.py @@ -41,6 +41,7 @@ ) from prefect.cli.root import app, is_interactive from prefect.client.schemas.actions import DeploymentScheduleCreate +from pre...
diff --git a/tests/cli/test_deploy.py b/tests/cli/test_deploy.py index 661494a4ce4a..1f455b62306f 100644 --- a/tests/cli/test_deploy.py +++ b/tests/cli/test_deploy.py @@ -276,7 +276,7 @@ async def test_project_deploy(self, project_dir, prefect_client: PrefectClient): await run_sync_in_worker_thread( ...
Add ability to configure `ConcurrenyOptions` to the CLI ### Describe the current behavior Currently users can set the the concurrency_limit for a deployment when creating it with the CLI by specifying the limit with and int as so: `deploy "my-deployment -cl 1". Upcoming work #15291, adds options for handling con...
1,727,107,212,000
[ "enhancement" ]
Feature Request
[ "src/prefect/cli/deploy.py:deploy" ]
[]
1
231
PrefectHQ/prefect
PrefectHQ__prefect-14693
80181d73383659c27f4e16122328f6c383e852fe
null
diff --git a/src/integrations/prefect-docker/prefect_docker/worker.py b/src/integrations/prefect-docker/prefect_docker/worker.py index bd85f4525325..9ebba4986f16 100644 --- a/src/integrations/prefect-docker/prefect_docker/worker.py +++ b/src/integrations/prefect-docker/prefect_docker/worker.py @@ -101,6 +101,7 @@ class...
diff --git a/src/integrations/prefect-docker/tests/test_worker.py b/src/integrations/prefect-docker/tests/test_worker.py index 1fa081ccd3b4..55130027f4af 100644 --- a/src/integrations/prefect-docker/tests/test_worker.py +++ b/src/integrations/prefect-docker/tests/test_worker.py @@ -850,6 +850,37 @@ async def test_task_...
Configure `ipc_mode` or `shm_size` on container creation For some flows that utilize `multiprocessing` for ML workloads, I need to run containers with increased shared memory available under `/dev/shm` (`--shm_size`) or use the host's namespace (--ipc_mode='host'). Looking through the code, I believe this would expand...
Thanks for the issue @geoffrp! I think this is a great opportunity to expose more configurability for the Docker worker so that we don't need to update the code each time a new parameter is needed. I would tend to agree, lots of arguments to play with for custom containers. Are you suggesting removing `DockerWorkerJo...
1,721,662,099,000
[ "enhancement", "integrations", "2.x" ]
Feature Request
[ "src/integrations/prefect-docker/prefect_docker/worker.py:DockerWorker._build_container_settings" ]
[]
1
232
PrefectHQ/prefect
PrefectHQ__prefect-14608
ad8e126b6a07fd6f8e08c9d50c22dabc247f765b
null
diff --git a/src/prefect/futures.py b/src/prefect/futures.py index d0df04ea9a5a..7379aa8c8b06 100644 --- a/src/prefect/futures.py +++ b/src/prefect/futures.py @@ -1,4 +1,5 @@ import abc +import collections import concurrent.futures import inspect import uuid @@ -256,13 +257,7 @@ def wait(self, timeout: Optional[flo...
diff --git a/tests/test_futures.py b/tests/test_futures.py index e009e0fa1716..43fcd2dc2460 100644 --- a/tests/test_futures.py +++ b/tests/test_futures.py @@ -16,6 +16,7 @@ PrefectFutureList, PrefectWrappedFuture, resolve_futures_to_states, + wait, ) from prefect.states import Completed, Failed fro...
Add `wait` utility for `PrefectFuture`s Add a `wait` utility that mirrors `concurrent.futures.wait`. Should accept an iterable of futures to wait for completion and a timeout. It should return a tuple of futures that have completed and those that have not yet completed.
1,721,054,806,000
[ "feature", "3.x" ]
Feature Request
[ "src/prefect/futures.py:PrefectFutureList.wait" ]
[]
1
233
PrefectHQ/prefect
PrefectHQ__prefect-14259
af200f20b9ba72609f404b5eb92da3ed30682695
null
diff --git a/src/integrations/prefect-sqlalchemy/prefect_sqlalchemy/database.py b/src/integrations/prefect-sqlalchemy/prefect_sqlalchemy/database.py index bdda680e3772..f98b109fb371 100644 --- a/src/integrations/prefect-sqlalchemy/prefect_sqlalchemy/database.py +++ b/src/integrations/prefect-sqlalchemy/prefect_sqlalche...
diff --git a/src/integrations/prefect-sqlalchemy/tests/test_database.py b/src/integrations/prefect-sqlalchemy/tests/test_database.py index bd17537ea2c2..a2f0327ace48 100644 --- a/src/integrations/prefect-sqlalchemy/tests/test_database.py +++ b/src/integrations/prefect-sqlalchemy/tests/test_database.py @@ -12,6 +12,7 @@...
return `CursorResult` from prefect_sqlalchemy execute/execute_many methods ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar request and didn't find it. - [X] I searched the Prefect documentation for this feature. ### Prefect Version 3.x ### Describe t...
1,719,173,597,000
[]
Feature Request
[ "src/integrations/prefect-sqlalchemy/prefect_sqlalchemy/database.py:SqlAlchemyConnector.execute", "src/integrations/prefect-sqlalchemy/prefect_sqlalchemy/database.py:SqlAlchemyConnector.execute_many" ]
[]
2
234
PrefectHQ/prefect
PrefectHQ__prefect-13756
38b424138539decd1ce396d0d73c6b73451e547b
null
diff --git a/src/prefect/tasks.py b/src/prefect/tasks.py index ec2dcd497d0b..ece4963f9dc0 100644 --- a/src/prefect/tasks.py +++ b/src/prefect/tasks.py @@ -849,7 +849,11 @@ def submit( flow_run_context = FlowRunContext.get() if not flow_run_context: - raise ValueError("Task.submit() must b...
diff --git a/tests/test_background_tasks.py b/tests/test_background_tasks.py index 40a9e34c3015..9dd3a70047d8 100644 --- a/tests/test_background_tasks.py +++ b/tests/test_background_tasks.py @@ -333,7 +333,7 @@ async def test_call_with_return_state(self, async_foo_task): class TestMap: async def test_map(self, ...
Add a `deferred` parameter to `Task.map` to control whether to defer or submit tasks Currently, `Task.map` always submits tasks within a flow or defers tasks outside of a flow. But now that we support using `delay` and `apply_async` within flows correctly, a flow might want to defer tasks with `map` and may not be usin...
1,717,441,194,000
[ "enhancement" ]
Feature Request
[ "src/prefect/tasks.py:Task.submit", "src/prefect/tasks.py:Task.map" ]
[]
2
235
PrefectHQ/prefect
PrefectHQ__prefect-13367
a0630c10580b78e6d7d41fe1b469cb5cb1a0bca2
null
diff --git a/src/prefect/cli/deployment.py b/src/prefect/cli/deployment.py index 398fd457653a..401c3dc80d70 100644 --- a/src/prefect/cli/deployment.py +++ b/src/prefect/cli/deployment.py @@ -22,7 +22,7 @@ from prefect.cli._types import PrefectTyper from prefect.cli._utilities import exit_with_error, exit_with_success...
diff --git a/tests/cli/deployment/test_deployment_cli.py b/tests/cli/deployment/test_deployment_cli.py index 57c19166a434..7b79206102b8 100644 --- a/tests/cli/deployment/test_deployment_cli.py +++ b/tests/cli/deployment/test_deployment_cli.py @@ -5,12 +5,10 @@ from prefect import flow from prefect.client.orchestrat...
Remove `prefect deployment apply` CLI from the `main` branch Remove the `prefect.cli.deployment.apply` function and any references to it. Any private code that was used solely for that function should also be deleted. This CLI command will be included in 2.x releases, but will not be included in the 3.0 release.
1,715,782,782,000
[]
Feature Request
[ "src/prefect/cli/deployment.py:apply" ]
[]
1
236
PrefectHQ/prefect
PrefectHQ__prefect-13366
77d752294bd90a3b8d1be1600d16b9a2509eebe2
null
diff --git a/src/prefect/cli/deployment.py b/src/prefect/cli/deployment.py index 129b25aa3a1c..398fd457653a 100644 --- a/src/prefect/cli/deployment.py +++ b/src/prefect/cli/deployment.py @@ -6,7 +6,7 @@ import sys import textwrap import warnings -from datetime import datetime, timedelta, timezone +from datetime impo...
diff --git a/tests/cli/deployment/test_deployment_build.py b/tests/cli/deployment/test_deployment_build.py deleted file mode 100644 index 0ef5a829d1e7..000000000000 --- a/tests/cli/deployment/test_deployment_build.py +++ /dev/null @@ -1,1311 +0,0 @@ -from datetime import timedelta -from pathlib import Path -from textwr...
Remove `prefect deployment build` CLI from the `main` branch Remove the `prefect.cli.deployment.build` function and any references to it. Any private code that was used solely for that function should also be deleted. This CLI command will be included in 2.x releases, but will not be included in the 3.0 release.
1,715,781,881,000
[]
Feature Request
[ "src/prefect/cli/deployment.py:build" ]
[]
1
237
PrefectHQ/prefect
PrefectHQ__prefect-13259
dbf132b346df079c4565abbe2cb31f61117b1b13
null
diff --git a/src/prefect/client/orchestration.py b/src/prefect/client/orchestration.py index 2e4334ba4bfe..53aa8bc5bf9a 100644 --- a/src/prefect/client/orchestration.py +++ b/src/prefect/client/orchestration.py @@ -1958,7 +1958,7 @@ async def update_deployment_schedule( kwargs = {} if active is not No...
diff --git a/tests/client/test_prefect_client.py b/tests/client/test_prefect_client.py index 4d13d6de0e12..203ec7a19216 100644 --- a/tests/client/test_prefect_client.py +++ b/tests/client/test_prefect_client.py @@ -2416,7 +2416,12 @@ async def test_read_deployment_schedules_success(self, prefect_client, deploymen ...
`client/orchestration.py/update_deployment_schedule` should be update with `active` and `schedule` at once ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar request and didn't find it. - [X] I searched the Prefect documentation for this feature. ##...
1,715,015,278,000
[]
Feature Request
[ "src/prefect/client/orchestration.py:PrefectClient.update_deployment_schedule" ]
[]
1
238
PrefectHQ/prefect
PrefectHQ__prefect-12410
b2f209a88888f56406c2806f45dc901b12fe3fa3
null
diff --git a/src/prefect/cli/deploy.py b/src/prefect/cli/deploy.py index 9e33c86cc053..28e4dc4cecb9 100644 --- a/src/prefect/cli/deploy.py +++ b/src/prefect/cli/deploy.py @@ -147,6 +147,12 @@ async def deploy( None, "-v", "--variable", + help=("DEPRECATED: Please use --jv/--job-variabl...
diff --git a/tests/cli/test_deploy.py b/tests/cli/test_deploy.py index 1b5e65724c08..363cb6a90b6e 100644 --- a/tests/cli/test_deploy.py +++ b/tests/cli/test_deploy.py @@ -277,7 +277,7 @@ async def test_project_deploy( invoke_and_assert, command=( "deploy ./flows/hello.py:my_fl...
CLI: Update `prefect deploy` options to support a `--jv/job-variable` option ### First check - [X] I am a contributor to the Prefect codebase ### Description In https://github.com/PrefectHQ/prefect/pull/12195/ we introduced a new option to `prefect deployment run` called `--jv/--job-variable` for providing job vari...
1,711,384,116,000
[ "enhancement" ]
Feature Request
[ "src/prefect/cli/deploy.py:deploy" ]
[]
1
239
PrefectHQ/prefect
PrefectHQ__prefect-12045
5e58d7bd06a6466ec2e4994716562c925ea2990e
null
diff --git a/src/prefect/utilities/dockerutils.py b/src/prefect/utilities/dockerutils.py index 3c05cb7fa2f9..adbe302066de 100644 --- a/src/prefect/utilities/dockerutils.py +++ b/src/prefect/utilities/dockerutils.py @@ -104,13 +104,18 @@ def silence_docker_warnings() -> Generator[None, None, None]: @contextmanager def...
diff --git a/tests/utilities/test_importtools.py b/tests/utilities/test_importtools.py index bb9f10f77957..7284fd1fafce 100644 --- a/tests/utilities/test_importtools.py +++ b/tests/utilities/test_importtools.py @@ -3,6 +3,7 @@ import sys from pathlib import Path from types import ModuleType +from unittest.mock impor...
Improve error message when Docker is not running and Prefect tries to build a Docker image ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar request and didn't find it. - [X] I searched the Prefect documentation for this feature. ### Prefect Version 2.x...
@serinamarie can you assign it to me? would love to take it as a first issue. @discdiver can you assign this issue to me? Assigned. Thank you @eladm26! You can check out the contributor docs[ here](https://docs.prefect.io/latest/contributing/overview/). Please let us know if you encounter any issues or anything in th...
1,708,543,037,000
[]
Feature Request
[ "src/prefect/utilities/dockerutils.py:docker_client" ]
[]
1
240
Delgan/loguru
Delgan__loguru-1239
65fe4a8db9a8f297ae3648f51b5e3050b30945a9
null
diff --git a/loguru/_better_exceptions.py b/loguru/_better_exceptions.py index e9ee1124..17d36d61 100644 --- a/loguru/_better_exceptions.py +++ b/loguru/_better_exceptions.py @@ -534,13 +534,39 @@ def _format_exception( yield from self._indent("-" * 35, group_nesting + 1, prefix="+-") def _forma...
diff --git a/tests/exceptions/output/others/one_liner_recursion.txt b/tests/exceptions/output/others/one_liner_recursion.txt new file mode 100644 index 00000000..635d3a23 --- /dev/null +++ b/tests/exceptions/output/others/one_liner_recursion.txt @@ -0,0 +1,79 @@ + +Traceback (most recent call last): + File "tests/exce...
Add unit tests for traceback formatting of RecursionError For example, the following snippet: ```python from loguru import logger def rec(): rec() try: rec() except Exception: logger.exception("Oups") ``` Should produce the following logs: ```txt 2024-02-17 12:06:00.577 | ERROR | __...
1,732,453,430,000
[]
Feature Request
[ "loguru/_better_exceptions.py:ExceptionFormatter._format_list" ]
[]
1
241
Lightning-AI/pytorch-lightning
Lightning-AI__pytorch-lightning-20403
20d19d2f5728f7049272f2db77a9748ff4cf5ccd
null
diff --git a/src/lightning/pytorch/trainer/connectors/checkpoint_connector.py b/src/lightning/pytorch/trainer/connectors/checkpoint_connector.py index c73ceb32ec77f..a41f87d418ebe 100644 --- a/src/lightning/pytorch/trainer/connectors/checkpoint_connector.py +++ b/src/lightning/pytorch/trainer/connectors/checkpoint_conn...
diff --git a/tests/tests_pytorch/trainer/connectors/test_checkpoint_connector.py b/tests/tests_pytorch/trainer/connectors/test_checkpoint_connector.py index 39911f9eddc7f..d29e2285e983c 100644 --- a/tests/tests_pytorch/trainer/connectors/test_checkpoint_connector.py +++ b/tests/tests_pytorch/trainer/connectors/test_che...
Support restoring callbacks' status when predicting ### Description & Motivation I implemented EMA using Callback. However, when predicting, the callback's weight isn't loaded back ### Pitch remove the condition at: https://github.com/Lightning-AI/pytorch-lightning/blob/f5d82504da252a255df268bd2bb99bf3f2886d27/src/...
1,730,960,186,000
[ "waiting on author", "callback", "pl" ]
Feature Request
[ "src/lightning/pytorch/trainer/connectors/checkpoint_connector.py:_CheckpointConnector._restore_modules_and_callbacks" ]
[]
1
242
dask/dask
dask__dask-11609
076b1f4d59283709127e5f7fe3b0c045127b7a92
null
diff --git a/dask/dataframe/io/io.py b/dask/dataframe/io/io.py index bfd330b5dcc..cf6657a6c67 100644 --- a/dask/dataframe/io/io.py +++ b/dask/dataframe/io/io.py @@ -33,11 +33,12 @@ insert_meta_param_description, is_series_like, make_meta, + pyarrow_strings_enabled, ) from dask.delayed import Delayed...
diff --git a/dask/dataframe/tests/test_dataframe.py b/dask/dataframe/tests/test_dataframe.py index b08718b8e54..bc5808aa020 100644 --- a/dask/dataframe/tests/test_dataframe.py +++ b/dask/dataframe/tests/test_dataframe.py @@ -24,7 +24,7 @@ import dask.dataframe.groupby from dask import delayed from dask._compatibilit...
Xarray to_dask_dataframe consume more memory on V2024.12.0 compared to V2024.10.0 <!-- Please include a self-contained copy-pastable example that generates the issue if possible. Please be concise with code posted. See guidelines below on how to provide a good bug report: - Craft Minimal Bug Reports http://matthe...
I can confirm the regression using the default threaded executor. It looks like this is not happening when using a distributed LocalCluster so it's probably some logic in `local.py::get_async` that isn't releasing data the way it should be (I'm just guessing, haven't debugged this yet) Yeah, like @fjetter said, I can r...
1,734,466,175,000
[]
Performance Issue
[ "dask/dataframe/io/io.py:from_dask_array" ]
[]
1
243
dask/dask
dask__dask-11605
44c75d08f6b693c53315c7926b5681af33696f5d
null
diff --git a/dask/array/overlap.py b/dask/array/overlap.py index 035882ecb2b..4d52db81033 100644 --- a/dask/array/overlap.py +++ b/dask/array/overlap.py @@ -368,6 +368,12 @@ def ensure_minimum_chunksize(size, chunks): return tuple(output) +def _get_overlap_rechunked_chunks(x, depth2): + depths = [max(d) if ...
diff --git a/dask/array/tests/test_overlap.py b/dask/array/tests/test_overlap.py index 4b4a3515102..c3d6019c367 100644 --- a/dask/array/tests/test_overlap.py +++ b/dask/array/tests/test_overlap.py @@ -567,6 +567,18 @@ def func(x): assert y.shape == (3,) +def test_map_overlap_trim_false_chunking(): + a =...
map_overlap fails if any chunk is smaler than overlap <!-- Please include a self-contained copy-pastable example that generates the issue if possible. Please be concise with code posted. See guidelines below on how to provide a good bug report: - Craft Minimal Bug Reports http://matthewrocklin.com/blog/work/2018/...
1,734,427,653,000
[]
Bug Report
[ "dask/array/overlap.py:overlap", "dask/array/overlap.py:map_overlap" ]
[ "dask/array/overlap.py:_get_overlap_rechunked_chunks" ]
2
244
dask/dask
dask__dask-11541
5b115c4360fec6a4aa6e0edf8ad1d89a87c986dd
null
diff --git a/dask/array/reshape.py b/dask/array/reshape.py index 1cc79090a64..b824110c7d5 100644 --- a/dask/array/reshape.py +++ b/dask/array/reshape.py @@ -473,6 +473,9 @@ def reshape_blockwise( _sanity_checks(x, shape) + if len(shape) == x.ndim and shape == x.shape: + return Array(x.dask, x.name, x...
diff --git a/dask/array/tests/test_reshape.py b/dask/array/tests/test_reshape.py index 14fb84ba919..1473d97d120 100644 --- a/dask/array/tests/test_reshape.py +++ b/dask/array/tests/test_reshape.py @@ -411,12 +411,11 @@ def test_reshape_blockwise_raises_for_expansion(): reshape_blockwise(arr, (3, 2, 9)) -@p...
bug in dask.array.reshape_blockwise **Minimal Complete Verifiable Example**: ```python import dask.array reshape_blockwise(dask.array.ones((3,)), shape=(3,), chunks=((3,),)) ``` ``` ValueError: Setting chunks is not allowed when reducing the number of dimensions. ``` cc @phofl
This is expected, we only need chunks if you increase the number of dimensions. We can calculate the chunks ourselves if we reduce the dimensions and I'd rather not validate user input. Do you have a use-case where you need this for a case like this? Yes, https://github.com/pydata/xarray/pull/9800/files, but I worke...
1,732,093,944,000
[]
Bug Report
[ "dask/array/reshape.py:reshape_blockwise" ]
[]
1
245
dask/dask
dask__dask-11479
b5ed6c5529c347a210cda856f6cf9f1609203e2a
null
diff --git a/dask/array/overlap.py b/dask/array/overlap.py index 265a50b141c..43c4ba7a252 100644 --- a/dask/array/overlap.py +++ b/dask/array/overlap.py @@ -1,13 +1,16 @@ from __future__ import annotations import warnings +from functools import reduce from numbers import Integral, Number +from operator import mul ...
diff --git a/dask/array/tests/test_overlap.py b/dask/array/tests/test_overlap.py index beff0650e4c..2c1b2a7ee8c 100644 --- a/dask/array/tests/test_overlap.py +++ b/dask/array/tests/test_overlap.py @@ -831,6 +831,63 @@ def test_sliding_window_errors(window_shape, axis): sliding_window_view(arr, window_shape, ax...
xarray.rolling().construct() and sliding_window_view chunking behaviour Using ``.rolling().construct()`` will inject new dimensions into an array with ``sliding_window_view`` under the hood. NumPy is clever about this by arranging things to be a view of the underlying array and thus not increasing the memory footprint....
> My suggestions would be to split up the other chunks in sliding_window_view to keep chunk sizes the same after construct, this would slow us down for cases where we indeed don't make a copy, but I think this is sufficiently rare in the general construct case. :+1: > Construct is also used internally for reducti...
1,730,403,606,000
[]
Performance Issue
[ "dask/array/overlap.py:sliding_window_view" ]
[]
1
246
dask/dask
dask__dask-11450
017adfb1794cd3debef0b7a4b0e8b6a587da8531
null
diff --git a/dask/array/slicing.py b/dask/array/slicing.py index c18e529392c..ec12479cf99 100644 --- a/dask/array/slicing.py +++ b/dask/array/slicing.py @@ -195,7 +195,9 @@ def slice_with_newaxes(out_name, in_name, blockdims, index): where_none[i] -= n # Pass down and do work - dsk, blockdims2 = ...
diff --git a/dask/array/tests/test_slicing.py b/dask/array/tests/test_slicing.py index a5ea273707f..2d1ab3006c8 100644 --- a/dask/array/tests/test_slicing.py +++ b/dask/array/tests/test_slicing.py @@ -1036,3 +1036,13 @@ def test_take_sorted_indexer(): ) }, } == dict(result.dask) + + +def test...
Slicing an Array with full slices along every dimension should just map the keys ``` ds = xr.open_zarr( "gs://weatherbench2/datasets/era5/1959-2023_01_10-wb13-6h-1440x721.zarr", #consolidated=True ) ds = ds[variables].sel(time=slice("2020-01-01", "2022-12-31")) ``` The dataset has one point per chunk a...
1,729,499,729,000
[]
Performance Issue
[ "dask/array/slicing.py:slice_with_newaxes", "dask/array/slicing.py:slice_wrap_lists", "dask/array/slicing.py:slice_slices_and_integers" ]
[]
3
247
dask/dask
dask__dask-11434
091f4e9bd226836faeaf30bdc3b5182ba14d22a6
null
diff --git a/dask/array/core.py b/dask/array/core.py index 7d4f023fa61..faedbb6ef3c 100644 --- a/dask/array/core.py +++ b/dask/array/core.py @@ -4795,7 +4795,7 @@ def broadcast_shapes(*shapes): if np.isnan(sizes).any(): dim = np.nan else: - dim = 0 if 0 in sizes else np.max(siz...
diff --git a/dask/array/tests/test_array_core.py b/dask/array/tests/test_array_core.py index 7d3e98558c2..121f8a69896 100644 --- a/dask/array/tests/test_array_core.py +++ b/dask/array/tests/test_array_core.py @@ -10,6 +10,7 @@ np = pytest.importorskip("numpy") +import itertools import math import operator impor...
`array.broadcast_shapes` to return a tuple of `int`, not a tuple of NumPy scalars <!-- Please do a quick search of existing issues to make sure that this has not been asked before. --> Before NumPy v2, the `repr` of NumPy scalars returned a plain number. This changed in [NEP 51](https://numpy.org/neps/nep-0051-scala...
I don't think that we want to deviate from numpy here > I don't think that we want to deviate from numpy here Thanks! I realise I should have compared the behaviour before posting, sorry. Anyway, I can now confirm that Dask is deviating from NumPy: ```python from dask import array as da import numpy as np ...
1,729,089,240,000
[]
Feature Request
[ "dask/array/core.py:broadcast_shapes" ]
[]
1
248
dask/dask
dask__dask-11393
e71d4361a57e7990e4a325021e6382877ac9ab49
null
diff --git a/dask/dataframe/utils.py b/dask/dataframe/utils.py index d748db315d9..456843efbff 100644 --- a/dask/dataframe/utils.py +++ b/dask/dataframe/utils.py @@ -434,10 +434,14 @@ def check_matching_columns(meta, actual): if extra or missing: extra_info = f" Extra: {extra}\n Missing: {missi...
diff --git a/dask/dataframe/tests/test_utils_dataframe.py b/dask/dataframe/tests/test_utils_dataframe.py index 40dc3bbce7a..6b0af9935fd 100644 --- a/dask/dataframe/tests/test_utils_dataframe.py +++ b/dask/dataframe/tests/test_utils_dataframe.py @@ -463,7 +463,12 @@ def test_check_matching_columns_raises_appropriate_err...
"Order of columns does not match" error should give an extra info about expected order It is not clear what order we should provide when we see the error "Order of columns does not match": ```python import pandas as pd import dask.dataframe as dd data = { 'id': [1, 1, 1, 2, 2, 2], 'date': pd.to_datetime...
100% agree, are you interested in submitting a PR? Making the error message more helpful would be great generally
1,726,588,598,000
[]
Feature Request
[ "dask/dataframe/utils.py:check_matching_columns" ]
[]
1
249
dask/dask
dask__dask-11354
31d57d34c1c72c551200355a738bf6cf1b1aed78
null
diff --git a/dask/array/core.py b/dask/array/core.py index 76ab569f4ee..d7e80038c97 100644 --- a/dask/array/core.py +++ b/dask/array/core.py @@ -22,7 +22,7 @@ import numpy as np from numpy.typing import ArrayLike -from tlz import accumulate, concat, first, frequencies, groupby, partition +from tlz import accumulate...
diff --git a/dask/array/tests/test_rechunk.py b/dask/array/tests/test_rechunk.py index 02eea953f08..26a7e405e4f 100644 --- a/dask/array/tests/test_rechunk.py +++ b/dask/array/tests/test_rechunk.py @@ -825,7 +825,18 @@ def test_rechunk_avoid_needless_chunking(): (100, 50, 10, (10,) * 10), (100, 100, 10...
Improve how normalize_chunks selects chunk sizes if auto is given Specifying auto for a dimension when new chunks are selected like in rechunking allows us to choose a chunk size that we think is a good fit and efficient. This is for now an issue to collect thoughts from others and feedback, I am pretty sure that I...
Hi, Sounds like good idea to me. I presume such a suggested change would only apply if the `previous_chunks` keyword to `normalize_chunks` was used - is that the idea? I.e. `normalize_chunks(..., previous_chunks=None)` would still give the current "strict" auto behaviour. David Yes, that is my current idea, b...
1,724,936,697,000
[]
Feature Request
[ "dask/array/core.py:normalize_chunks", "dask/array/core.py:_compute_multiplier", "dask/array/core.py:auto_chunks" ]
[ "dask/array/core.py:_convert_int_chunk_to_tuple", "dask/array/core.py:_split_up_single_chunk" ]
3
250
dask/dask
dask__dask-11303
67b2852d5f48d46527e3ac9c78122b483aef4057
null
diff --git a/dask/order.py b/dask/order.py index 4de05c2b49b..62fb678b2fb 100644 --- a/dask/order.py +++ b/dask/order.py @@ -208,6 +208,9 @@ def order( _sort_keys_cache: dict[Key, tuple[int, int, int, int, str]] = {} + leafs_connected_to_loaded_roots: set[Key] = set() + processed_roots = set() + def...
diff --git a/dask/tests/test_order.py b/dask/tests/test_order.py index 0a456cbcccd..bea8caf7383 100644 --- a/dask/tests/test_order.py +++ b/dask/tests/test_order.py @@ -256,7 +256,7 @@ def test_prefer_deep(abcde): | | d a - Prefer longer chains first so we should start with c + Prefer longer chain...
order: dask order returns suboptimal ordering for xr.rechunk().groupby().reduce("cohorts") @fjetter and I looked into a problematic pattern for xarray this morning and identified that dask order is missbehaving here. dask order is currently returning a suboptimal execution order for the pattern: ``` xarr =. ... ...
1,723,472,698,000
[]
Performance Issue
[ "dask/order.py:order" ]
[]
1
251
dask/dask
dask__dask-11183
1755a718125ef2ea18e9c079ecf1509b7b6be730
null
diff --git a/dask/dataframe/__init__.py b/dask/dataframe/__init__.py index 029eef5c54a..1a3e128f6ca 100644 --- a/dask/dataframe/__init__.py +++ b/dask/dataframe/__init__.py @@ -5,19 +5,37 @@ from packaging.version import Version +# The "dataframe.query-planning" config can only be processed once +DASK_EXPR_ENABLED...
diff --git a/dask/dataframe/tests/test_dataframe.py b/dask/dataframe/tests/test_dataframe.py index 639e0e0c5ea..900a2068835 100644 --- a/dask/dataframe/tests/test_dataframe.py +++ b/dask/dataframe/tests/test_dataframe.py @@ -6476,3 +6476,13 @@ def test_enforce_runtime_divisions(): RuntimeError, match="`enf...
[FEA] Add official mechanism to check if query-planning is enabled in ``dask.dataframe`` The value of the `"dataframe.query-planning"` config currently dictates whether or not `dask.dataframe` will use dask-expr under the hood. However, after `dask.dataframe` is first imported, the value of the `"dataframe.query-planni...
This kind of problem occurs every now and then for various kinds of configs. Typically we are setting some global variable in this case. >This kind of problem occurs every now and then for various kinds of configs. Typically we are setting some global variable in this case. Thanks Florian! Just to clarify: This mean...
1,718,640,869,000
[ "dataframe" ]
Feature Request
[ "dask/dataframe/__init__.py:_dask_expr_enabled" ]
[]
1
252
dask/dask
dask__dask-10963
fa2d072c4b494b95a4352b17217fde4a301ed76b
null
diff --git a/dask/bag/core.py b/dask/bag/core.py index e3a6f675def..31f640b5841 100644 --- a/dask/bag/core.py +++ b/dask/bag/core.py @@ -1621,7 +1621,15 @@ def to_dataframe(self, meta=None, columns=None, optimize_graph=True): dsk = dfs.dask divisions = [None] * (self.npartitions + 1) - re...
diff --git a/dask/bag/tests/test_bag.py b/dask/bag/tests/test_bag.py index c27f8bed2fb..d00addb5dc9 100644 --- a/dask/bag/tests/test_bag.py +++ b/dask/bag/tests/test_bag.py @@ -739,9 +739,7 @@ def test_from_long_sequence(): def test_from_empty_sequence(): - dd = pytest.importorskip("dask.dataframe") - if dd....
Support bag.to_dataframe when query planning is enabled As far as I can tell, dask bags can only be converted to dask dataframes when query planning is disabled. It would be great to support both query planning, and the flexibility of turning bags into dataframes. Currently calling to_dataframe with query planning enab...
1,709,076,359,000
[]
Feature Request
[ "dask/bag/core.py:Bag.to_dataframe" ]
[]
1
253
dask/dask
dask__dask-10796
3e9ae492934e133547e5eaec5020186fad39898a
null
diff --git a/dask/dataframe/groupby.py b/dask/dataframe/groupby.py index 8acdeb251c4..8e4cf321af5 100644 --- a/dask/dataframe/groupby.py +++ b/dask/dataframe/groupby.py @@ -48,7 +48,14 @@ ) from dask.highlevelgraph import HighLevelGraph from dask.typing import no_default -from dask.utils import M, _deprecated, deriv...
diff --git a/dask/dataframe/tests/test_groupby.py b/dask/dataframe/tests/test_groupby.py index 54dcb59ba92..51873785a3c 100644 --- a/dask/dataframe/tests/test_groupby.py +++ b/dask/dataframe/tests/test_groupby.py @@ -1761,13 +1761,6 @@ def test_cumulative(func, key, sel): g, dg = (d.groupby(key)[sel] for d in (df,...
Deprecate axis in GroupBy cumulative transformers Deprecate axis in - cumsum - cumprod - cumcount and please skip tests that test this behaviour for dask-expr
1,705,313,831,000
[]
Feature Request
[ "dask/dataframe/groupby.py:_GroupBy.cumcount" ]
[]
1
254